Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

winperfcounters receiver reports no metrics if any scraper fails #16712

Closed
chrislbs opened this issue Dec 5, 2022 · 4 comments
Closed

winperfcounters receiver reports no metrics if any scraper fails #16712

chrislbs opened this issue Dec 5, 2022 · 4 comments

Comments

@chrislbs
Copy link
Contributor

chrislbs commented Dec 5, 2022

Component(s)

receiver/windowsperfcounters

What happened?

Description

In the windows perf counter receiver, if any instance or object fails to receive, all metrics are thrown away rather than just the scrapers that failed to successfully read.

Steps to Reproduce

Create a windowsperfcounter receiver for an instance that exists, and another that doesn't.

Expected Result

Metrics will be received and exported for the perf counter instance that exists.

Actual Result

Metrics are not received or exported for the perf counter instance that exists.

Collector version

v0.66.0

Environment information

No response

OpenTelemetry Collector configuration

receivers:
  windowsperfcounters:
    metrics:
      dotnet.gen0.heap.size:
        description: dotnet gen 0 heap size
        unit: By
        gauge:
    collection_interval: 30s
    perfcounters:
    - object: .NET CLR Memory
      instance:
      - "notexist.exe"
      - "exists.exe"
      counters:
        - name: "Gen 0 heap size"
          metric: dotnet.gen0.heap.size

service:
  pipelines:
    metrics:
      receivers: [windowsperfcounters]

Log output

No response

Additional context

No response

@chrislbs chrislbs added bug Something isn't working needs triage New item requiring triage labels Dec 5, 2022
@chrislbs
Copy link
Contributor Author

chrislbs commented Dec 5, 2022

/label receiver/windowsperfcountersreceiver

@github-actions
Copy link
Contributor

github-actions bot commented Feb 6, 2023

This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping @open-telemetry/collector-contrib-triagers. If this issue is still relevant, please ping the code owners or leave a comment explaining why it is still relevant. Otherwise, please close it.

Pinging code owners:

See Adding Labels via Comments if you do not have permissions to add labels yourself.

@github-actions github-actions bot added the Stale label Feb 6, 2023
@fatsheep9146 fatsheep9146 removed the Stale label Feb 6, 2023
@atoulme atoulme removed the needs triage New item requiring triage label Mar 7, 2023
@github-actions
Copy link
Contributor

github-actions bot commented Jul 4, 2023

This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping @open-telemetry/collector-contrib-triagers. If this issue is still relevant, please ping the code owners or leave a comment explaining why it is still relevant. Otherwise, please close it.

Pinging code owners:

See Adding Labels via Comments if you do not have permissions to add labels yourself.

@github-actions github-actions bot added the Stale label Jul 4, 2023
@github-actions
Copy link
Contributor

github-actions bot commented Sep 3, 2023

This issue has been closed as inactive because it has been stale for 120 days with no activity.

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Sep 3, 2023
mx-psi pushed a commit that referenced this issue Apr 26, 2024
…32370)

Returning partial errors for failures during scraping to prevent
throwing out all successfully retrieved metrics.

**Description:** Fixing a bug where a single instance of a windows
performance object prevents metrics for all other valid instances from
being published.

**Link to tracking Issue:**
[16712](#16712)

**Testing:** I added additional tests that verifies if some scrapers
fail to retrieve metrics, the ones that don't fail continue to publish
their metrics.

**Documentation:**
rimitchell pushed a commit to rimitchell/opentelemetry-collector-contrib that referenced this issue May 8, 2024
…pen-telemetry#32370)

Returning partial errors for failures during scraping to prevent
throwing out all successfully retrieved metrics.

**Description:** Fixing a bug where a single instance of a windows
performance object prevents metrics for all other valid instances from
being published.

**Link to tracking Issue:**
[16712](open-telemetry#16712)

**Testing:** I added additional tests that verifies if some scrapers
fail to retrieve metrics, the ones that don't fail continue to publish
their metrics.

**Documentation:**
jlg-io pushed a commit to jlg-io/opentelemetry-collector-contrib that referenced this issue May 14, 2024
…pen-telemetry#32370)

Returning partial errors for failures during scraping to prevent
throwing out all successfully retrieved metrics.

**Description:** Fixing a bug where a single instance of a windows
performance object prevents metrics for all other valid instances from
being published.

**Link to tracking Issue:**
[16712](open-telemetry#16712)

**Testing:** I added additional tests that verifies if some scrapers
fail to retrieve metrics, the ones that don't fail continue to publish
their metrics.

**Documentation:**
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants