
Security Data Pipelines: What Analysts, Customers, and Prospects Are Saying About Security Data Needs
Over the past year, a clear pattern has emerged from conversations with analysts, customers, and prospective buyers: security teams are rethinking how security data moves through their environments.
What used to be treated as a subcomponent of observability tooling is rapidly evolving into a distinct architectural layer. Increasingly, analysts and practitioners are referring to this layer as Security Data Pipelines (sometimes called security telemetry pipelines).
This shift reflects a deeper realization: modern security operations are fundamentally a data problem, not just a tooling problem.
Below are several key trends shaping this emerging category and what they mean for security architects and SOC leaders.
The Emergence of Security Data Pipelines
Until recently, telemetry pipelines were often seen as part of the broader observability stack alongside metrics and traces. But security teams are increasingly recognizing that security telemetry has very different requirements.
Security data volumes are exploding due to:
- Endpoint telemetry
- Identity logs
- Cloud infrastructure logs
- Network monitoring
- SaaS audit data
- Threat intelligence enrichment
As the amount of data grows, SIEM costs rise with it. Organizations frequently discover that poor data management—not their detection platform—is what threatens their security budget.
Without effective pipeline controls, organizations face:
- runaway ingestion costs
- noisy and low-quality telemetry
- poor detection fidelity
- delayed investigations
Security leaders are increasingly realizing that data quality and data control must be handled upstream of the SIEM. We explore this in more detail in this article on how high-quality security data dramatically improves detection and cost efficiency.
This is one of the reasons analysts are beginning to treat Security Data Pipelines as a new product category—a dedicated control plane for security data.
Data Quality Is Becoming a Strategic Security Capability
Many organizations initially approach pipeline management as a cost-reduction initiative. But in practice, the real strategic value is data quality.
High-quality data enables:
- more reliable detections
- better enrichment and correlation
- faster investigations
- improved AI model performance
- lower false positives
Poor-quality data has the opposite effect: it increases SIEM costs while simultaneously degrading security outcomes.
Security teams are therefore focusing on pipeline capabilities like:
- classification
- enrichment
- normalization
- schema standardization
- deduplication
- filtering
These capabilities transform raw telemetry into structured, meaningful security signals before they reach downstream tools. For a deeper dive into why pipeline data quality is becoming a core security capability, see this post on why data quality makes security pipelines strategic.
Reducing Complexity to Enable Architectural Flexibility
While cost reduction is a major driver, the most urgent challenge for security architects is infrastructure complexity.
Large organizations rarely operate a single security analytics platform. Instead, they often run combinations of platforms such as:
- Splunk
- Microsoft Sentinel
- Elastic
At the same time, organizations frequently introduce additional tools such as:
- MDR providers
- UEBA systems
- data lakes
- AI-driven detection platforms
This creates a brittle architecture where changing one downstream tool requires rebuilding large parts of the data collection layer.
Security architects are therefore looking for future-proof collection architectures that:
- centralize ingestion
- classify data in the pipeline
- normalize data once
- make it easy to route the right data to each destination (for example, low-value log messages to cold storage, security-relevant data to the SIEM)
- enable easy tool replacement
This architectural pattern is often referred to as pipeline unification, where a single pipeline layer manages telemetry for both observability and security workloads. You can explore this architecture further in the overview of unified telemetry pipelines.
A real-world example of this approach can be seen in this case study of a government organization performing a Google SecOps migration using a unified pipeline architecture.
The Connection to Detection Engineering
Another emerging trend is the tight coupling between telemetry pipelines and detection engineering.
Modern detection engineering teams rely on structured, well-understood data to build reliable detection logic. Without high-quality telemetry, even the most sophisticated detection rules fail.
As one detection engineer put it:
If you want to do detection engineering, you have to have the right data—and understand what it means.
Highly advanced security organizations are therefore splitting responsibilities across specialized teams:
- Detection Engineering: Detection logic, rule development
- Security Data Engineering: Telemetry pipelines, normalization, enrichment
This division allows detection engineers to focus on threat logic, while pipeline engineers ensure the right data reaches the right tools. For example, one large healthcare organization dramatically improved observability performance by implementing pipeline-based data filtering and routing.
Processing on the Edge: Shifting Left in the Pipeline
Another important trend is the shift-left of data processing and detection capabilities.
Traditionally, nearly all detection logic ran inside the SIEM. But SIEM platforms are expensive environments for heavy data processing.
Security teams increasingly want to perform streaming analysis directly in the telemetry pipeline. By moving classification, filtering, and normalization, enrichment and labeling into the pipeline, you’re paving the way to also incorporate:
- Sigma rule matching
- threat intelligence matching
- pattern and string detection
By performing these operations before data reaches the SIEM, organizations can:
- drop noisy or low-value data early
- reduce ingestion costs
- accelerate detection workflows
- improve SIEM performance
This architecture effectively turns the telemetry pipeline into a real-time security data processing layer. This shift is discussed in more detail in this article about why SOCs are rethinking the role of the security data pipeline.
Optimizing SIEM Performance Through Smarter Data Design
Another challenge security teams encounter is poorly optimized SIEM indexing strategies.
Platforms like Splunk rely heavily on indexed fields for fast searching and correlation. But poorly designed indexing strategies can dramatically increase costs.
Security telemetry pipelines can help solve this by:
- pre-normalizing fields
- standardizing key attributes
- controlling indexed field creation
- reducing unnecessary field cardinality
When pipelines prepare and structure data before ingestion, SIEMs become faster, cheaper, and easier to operate. For example, this guide explains how organizations can optimize Splunk indexed fields using upstream pipeline processing.
The Rise of Autonomous Data Layer
As data volumes continue to grow, many organizations are moving toward automated governance of security data flows. Rather than manually tuning ingestion rules across dozens of tools, security teams want centralized control over:
- data routing
- enrichment policies
- filtering logic
- cost controls
- compliance requirements
This concept is often described as an autonomous data layer, where the pipeline automatically performs:
- data curation
- data volume and cost reduction,
- tiered storage, and also provides
- pipeline visibility and health monitoring
You can explore this concept further in the discussion of the autonomous data layer for controlling security data cost and cyber risk.
Enabling AI Security Agents
Perhaps the fastest-growing topic is the role of AI in security operations. Organizations are experimenting with:
- AI SOC assistants
- automated investigation agents
- threat hunting copilots
- LLM-based detection systems
But these tools require large volumes of clean, structured data. Raw telemetry is typically inconsistent across sources, making it difficult for AI systems to interpret reliably. This is where security data pipelines become critical.
By normalizing data into standard schemas (such as OCSF), the pipeline layer can feed clean, structured telemetry directly into AI-driven security tools.
This approach also supports decoupled security architectures, where AI systems, data lakes, and SIEMs all consume the same normalized telemetry. This architecture is explored further in this article on the rise of decoupled SIEM architectures.
Security Data Pipeline Management Is Becoming a Core Discipline
Taken together, these trends point to the emergence of a new operational discipline: security data pipeline management.
Security teams increasingly need capabilities to:
- collect telemetry at scale
- normalize and enrich data
- reduce noise and cost
- route data across multiple tools
- enable detection engineering
- prepare data for AI systems
In other words, the pipeline layer is becoming the control plane for security data. You can learn more about this emerging operational model in this overview of security data pipeline management.
The Bottom Line
Security telemetry pipelines are evolving from a behind-the-scenes infrastructure component into a strategic layer of the modern security architecture.
Organizations that treat security data as a first-class asset—and build dedicated pipelines to manage it—gain several advantages:
In a world where security effectiveness increasingly depends on data, the telemetry pipeline is quickly becoming one of the most important layers in the entire security stack.
Follow Our Progress!
We are excited to be realizing our vision above with a full Axoflow product suite.
Sign Me UpFighting data Loss?

Book a free 30-min consultation with syslog-ng creator Balázs Scheidler
