-
Notifications
You must be signed in to change notification settings - Fork 2.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[exporter/datadog] Trace stats are severely underreporting hit counts #31713
Comments
Pinging code owners:
See Adding Labels via Comments if you do not have permissions to add labels yourself. |
I thinks this is probably related to #31219 (and a bug). If you're not using the connector at all and just the exporter, the DD exporter is sending |
This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping Pinging code owners:
See Adding Labels via Comments if you do not have permissions to add labels yourself. |
@sirianni are you still facing the issue? can you please us know if you were able to try the latest version of connector. |
This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping Pinging code owners:
See Adding Labels via Comments if you do not have permissions to add labels yourself. |
I’m facing a similar issue in an application, where there’s a discrepancy between the counters of metrics generated from span tags and the metrics automatically generated by the otelhttp library. For example, the API receives 1,000 hits within a certain period, and during execution, it consistently makes calls to another API. When making these calls, we generate a client-type span, which leads to Datadog creating a client APM metric. This allows us to filter and monitor the application's dependencies. However, the value of this metric significantly differs from the metrics generated by otelhttp, causing inconsistencies in the data we're monitoring. I’m not sure if this is related to the warning in the Datadog documentation that mentions metrics from traces can indeed have inconsistencies. |
This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping Pinging code owners:
See Adding Labels via Comments if you do not have permissions to add labels yourself. |
This issue has been closed as inactive because it has been stale for 120 days with no activity. |
Component(s)
exporter/datadog
What happened?
Description
The
trace.*
stats calculated within thedatadogexporter
are severely underreporting hit counts in many cases.We confirmed this by comparing the span counts
trace.*
metricsStrangely, this discrepancy is not observed for a few HTTP routes (e.g.
/metrics/logical_clusters
in the above screenshot).We do not have sampling enabled anywhere in our OTel Collector configuration.
Collector version
v0.93.0
Environment information
Environment
Kubernetes
OpenTelemetry Collector configuration
Log output
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: