In many organizations, the logging architecture grows organically. Teams add tooling to solve immediate needs: a new platform leads to new collection agents to maintain; growing log volumes result in a pipeline to buffer data streams, and so on. Over time you end up in operational chaos and fragmented configurations.
A common “current state” pattern might look like this:
- Multiple types of agents in parallel (some vendor-specific, some open source) that align to different backends. The agents can be deployed inconsistently across clusters/hosts, increasing configuration drift and upgrade toil.
- Multiple backends for different consumers (for example, observability, security, business intelligence). This forces teams to maintain different schemas, routing rules, and access controls.
- Kafka (or similar queues) used as a “pipeline” for buffering and fan-out – often because teams already run event-driven infrastructure and want to extend it to telemetry.
- Format sprawl, combined with weak governance. Logs arrive in many formats with different default message types. Without governance, the resulting data becomes harder to use and more expensive to store. Pipelines/queues that can ingest from many agents and formats can unintentionally amplify lack of standardization unless governance is enforced.
This “sprawling but functional” setup is usually the moment teams start looking for a target state that centralizes processing/routing and reduces agent sprawl. This target state – an integrated logging architecture – becomes a control system to help you keep costs down and remove complexity.
In the sections that follow, we’ll break down a practical, vendor-neutral approach built on three layers – log agents, a telemetry pipeline, and a log management/analysis platform – and show how they work together.
The target state: what an integrated logging architecture means
“Integrated” doesn’t mean “one vendor.” It means shared governance (schema, tags, redaction rules, retention goals) and multiple control points across the log data path.
A practical reference model has three layers:
- Collection agents to capture logs reliably at the edge.
- A telemetry pipeline to reduce, transform, enrich, redact, and route logs before they hit destinations.
- A log platform to search, alert, investigate, and manage lifecycle. (Bonus points if this platform can inform upstream control.)
This structure replaces all-or-nothing choices (“ingest everything” vs. “drop blindly”) with deliberate optimization.
Vendor-neutral reference architecture: responsibilities by layer (and why each exists)
Layer 1: Log collection agents (edge collection and resilience)
Agents run where logs are produced (nodes, VMs, clusters). They handle discovery, rotation, lightweight parsing/metadata, and secure forwarding. The key best practice is standardizing on a single set of agents so you don’t multiply configuration and rollout work.
Ideally, your team should rely on an open source log collection agent. Doing so will help you mitigate vendor lock-in in the long-term.
Layer 2: Telemetry pipeline (pre-egress control and routing)
The telemetry pipeline is the policy/control point for logs in motion. Centralize processing and routing rules here instead of distributing them across hundreds of agent configs.
Typical pipeline responsibilities include:
- Reducing data volumes by removing noise
- Transforming and standardizing schemas
- Enriching data with context that’s helpful to end-users downstream
- Protecting/removing sensitive information from fields
- Routing subsets of data to the right destinations (analytics/SIEM/archive)
This breaks tight coupling between producers and backends and reduces egress/ingest waste by shaping data upstream.
Layer 3: Log management & analysis platform (investigation, lifecycle, and feedback)
Layer 3 is where teams interact with logs: search, dashboards, alerts, and investigations – plus lifecycle controls (hot vs. warm vs. archive).
A capability to prioritize is the ability to analyze how your team uses its data. If you know what’s queried vs. ignored, you can optimize safely across all layers:
- Tune noisy sources
- Apply confident pipeline filters/routing
- Right-size indexing and retention based on real access patterns
Together, you get a feedback system: upstream controls reduce waste early, downstream usage insight identifies what’s valuable, and governance keeps teams aligned. The payoff is more predictable spend, faster incident response, less operational toil, and easier backend changes or migrations.
Practical patterns you can apply
With an integrated logging architecture in place, you can implement several use cases that add value to your observability and security practices. These patterns assume a three-layer model (agent → pipeline → log platform) and are designed to be vendor-neutral. The common thread: treat logs as data you can shape and route intentionally – based on use case, cost, and how teams actually consume the data.
Pattern 1: Multi-destination routing without multi-agent chaos
Instead of deploying/maintaining separate agents per backend, use a telemetry pipeline as the fan-out point. This keeps collection stable while letting you route subsets to the right tools (observability, SIEM, archive). The pipeline can also “shape” the same dataset differently per destination when schema/format needs vary. 
Pattern 2: Route to the most cost-effective destination by dataset
Not all logs deserve hot, searchable storage. Route high-value operational logs to your primary log platform, route security-relevant streams to SIEM, and send everything (or curated subsets) to low-cost object storage for retention and replay.
This pattern aligns cost with intent: “hot” where it’s actively used, “cheap” where it’s primarily retained.
Pattern 3: Normalize once, query everywhere
Different sources emit different formats; inconsistent fields turn search and correlation into an art project. A pipeline can standardize formats across sources so downstream teams can query consistently, without requiring every dev team to rewrite their logging output. 
Pattern 4: Security/compliance as upstream controls
Make redaction/masking a first-class pipeline policy so sensitive fields are removed before data leaves your environment. This eliminates the possibility of PII ending up in your SaaS log management tooling, protecting you from compliance or security risks.
Pattern 5: Migration-safe logging (and faster platform evaluations)
Use the pipeline to stream the same data to multiple backends during migrations or evaluations – without reinstalling new agents everywhere. This de-risks cutovers and can shorten evaluation timelines by enabling side-by-side comparisons.  
Build control points, then iterate with data
If logging feels expensive and chaotic, the fix usually isn’t a bigger backend—it’s a better architecture. Standardize collection, add a telemetry pipeline as the pre-egress control point, and choose a log platform that combines investigation workflows with usage insight. Start with normalization and routing, then use real usage data to decide what stays hot, what gets archived, and what you can confidently drop.
Manning book: Logging Best Practices
Transform logging from a reactive debugging tool into a proactive competitive advantage.