CONTROL COST WHILE MAINTAINING OPTIMAL TRACE FIDELITY
A different approach to tracing: Fine-grained controls to dynamically manage tracing volumes matched with unique tracing insights contextually linked with metrics, traces, and logs for faster troubleshooting and incident remediation.
Faster Triage
Auto generated trace insights and contextual linkage between metrics, logs, and events makes trace data more useful.
Automated critical path analysis highlights segments of each span that impact the total duration of a trace to quickly identify latency issues.
Control Costs
Dynamically “turn up” head sampling rates during an incident and back down when the issue has been identified.
Set sampling rates based on how critical the service is to the business.
Wider Adoption
Designed to provide value for both the casual and power user, eliminating the need to rely on a couple of “heroes” in the organization.
Powerful filters make it easy for developers to get value when troubleshooting without being a daily user of the solution.
No Vendor Lock-In
100% compatible with OpenTelemetry, Jaeger, or Zipkin.
Send metric, log and trace data via the native OpenTelemetry Collector or the Chronosphere Collector.
DYNAMIC HEAD AND TAIL BASED SAMPLING
You can’t egress, store, and query every trace you generate — it is too expensive. But how do you make sure you don’t miss important traces?
With Chronosphere, you can use Datasets to group traces based on value.
Set a global low head sampling rate to reduce costs. When an incident occurs for your high-value services you can dynamically turn up the sampling rate for just those traces. When the system is healthy, turn the rate back down again - no redeploy required.
Use dynamic head sampling in conjunction with tail sampling to derive the most value from your trace data when needed to control costs without sacrificing visibility.
Get a real-time view of incoming traces grouped by tags and their relative frequency. Use this information to help understand how often your applications emit traces and troubleshoot spikes in ingest rates.
Alert and/or create dashboards by generating a new metric data point based on an entire trace or part of a trace. Trace metrics can be generated in seconds without re-instrumenting your application.
Use them to capture errors on specific high-cardinality dimensions that might get buried or rolled-away by traditional Prometheus metrics.
Flexible search, powerful filters, and a full topology map helps you quickly identify, triage, and understand the root cause of problems.
You can view all trace data that's relevant to a particular issue, and compare that trace against a previous time period to better understand where errors are occurring