From SIEM cost control to AI-ready data, security telemetry pipelines are emerging as a critical layer for managing security data at scale.

Security Data Pipelines: What Analysts, Customers, and Prospects Are Saying About Security Data Needs

Over the past year, a clear pattern has emerged from conversations with analysts, customers, and prospective buyers: security teams are rethinking how security data moves through their environments.

What used to be treated as a subcomponent of observability tooling is rapidly evolving into a distinct architectural layer. Increasingly, analysts and practitioners are referring to this layer as Security Data Pipelines (sometimes called security telemetry pipelines).

This shift reflects a deeper realization: modern security operations are fundamentally a data problem, not just a tooling problem.

Below are several key trends shaping this emerging category and what they mean for security architects and SOC leaders.

The Emergence of Security Data Pipelines

Until recently, telemetry pipelines were often seen as part of the broader observability stack alongside metrics and traces. But security teams are increasingly recognizing that security telemetry has very different requirements.

Security data volumes are exploding due to:

  • Endpoint telemetry
  • Identity logs
  • Cloud infrastructure logs
  • Network monitoring
  • SaaS audit data
  • Threat intelligence enrichment

As the amount of data grows, SIEM costs rise with it. Organizations frequently discover that poor data management—not their detection platform—is what threatens their security budget.

Without effective pipeline controls, organizations face:

  • runaway ingestion costs
  • noisy and low-quality telemetry
  • poor detection fidelity
  • delayed investigations

Security leaders are increasingly realizing that data quality and data control must be handled upstream of the SIEM. We explore this in more detail in this article on how high-quality security data dramatically improves detection and cost efficiency.

This is one of the reasons analysts are beginning to treat Security Data Pipelines as a new product category—a dedicated control plane for security data.

Data Quality Is Becoming a Strategic Security Capability

Many organizations initially approach pipeline management as a cost-reduction initiative. But in practice, the real strategic value is data quality.

High-quality data enables:

  • more reliable detections
  • better enrichment and correlation
  • faster investigations
  • improved AI model performance
  • lower false positives

Poor-quality data has the opposite effect: it increases SIEM costs while simultaneously degrading security outcomes.

Security teams are therefore focusing on pipeline capabilities like:

  • classification
  • enrichment
  • normalization
  • schema standardization
  • deduplication
  • filtering

These capabilities transform raw telemetry into structured, meaningful security signals before they reach downstream tools. For a deeper dive into why pipeline data quality is becoming a core security capability, see this post on why data quality makes security pipelines strategic.

Reducing Complexity to Enable Architectural Flexibility

While cost reduction is a major driver, the most urgent challenge for security architects is infrastructure complexity.

Large organizations rarely operate a single security analytics platform. Instead, they often run combinations of platforms such as:

At the same time, organizations frequently introduce additional tools such as:

  • MDR providers
  • UEBA systems
  • data lakes
  • AI-driven detection platforms

This creates a brittle architecture where changing one downstream tool requires rebuilding large parts of the data collection layer.

Security architects are therefore looking for future-proof collection architectures that:

  • centralize ingestion
  • classify data in the pipeline
  • normalize data once
  • make it easy to route the right data to each destination (for example, low-value log messages to cold storage, security-relevant data to the SIEM)
  • enable easy tool replacement

This architectural pattern is often referred to as pipeline unification, where a single pipeline layer manages telemetry for both observability and security workloads. You can explore this architecture further in the overview of unified telemetry pipelines.

A real-world example of this approach can be seen in this case study of a government organization performing a Google SecOps migration using a unified pipeline architecture.

The Connection to Detection Engineering

Another emerging trend is the tight coupling between telemetry pipelines and detection engineering.

Modern detection engineering teams rely on structured, well-understood data to build reliable detection logic. Without high-quality telemetry, even the most sophisticated detection rules fail.

As one detection engineer put it:

If you want to do detection engineering, you have to have the right data—and understand what it means.

Highly advanced security organizations are therefore splitting responsibilities across specialized teams:

  • Detection Engineering: Detection logic, rule development
  • Security Data Engineering: Telemetry pipelines, normalization, enrichment

This division allows detection engineers to focus on threat logic, while pipeline engineers ensure the right data reaches the right tools. For example, one large healthcare organization dramatically improved observability performance by implementing pipeline-based data filtering and routing.

Processing on the Edge: Shifting Left in the Pipeline

Another important trend is the shift-left of data processing and detection capabilities.

Traditionally, nearly all detection logic ran inside the SIEM. But SIEM platforms are expensive environments for heavy data processing.

Security teams increasingly want to perform streaming analysis directly in the telemetry pipeline. By moving classification, filtering, and normalization, enrichment and labeling into the pipeline, you’re paving the way to also incorporate:

  • Sigma rule matching
  • threat intelligence matching
  • pattern and string detection

By performing these operations before data reaches the SIEM, organizations can:

  • drop noisy or low-value data early
  • reduce ingestion costs
  • accelerate detection workflows
  • improve SIEM performance

This architecture effectively turns the telemetry pipeline into a real-time security data processing layer. This shift is discussed in more detail in this article about why SOCs are rethinking the role of the security data pipeline.

Optimizing SIEM Performance Through Smarter Data Design

Another challenge security teams encounter is poorly optimized SIEM indexing strategies.

Platforms like Splunk rely heavily on indexed fields for fast searching and correlation. But poorly designed indexing strategies can dramatically increase costs.

Security telemetry pipelines can help solve this by:

  • pre-normalizing fields
  • standardizing key attributes
  • controlling indexed field creation
  • reducing unnecessary field cardinality

When pipelines prepare and structure data before ingestion, SIEMs become faster, cheaper, and easier to operate. For example, this guide explains how organizations can optimize Splunk indexed fields using upstream pipeline processing.

The Rise of Autonomous Data Layer

As data volumes continue to grow, many organizations are moving toward automated governance of security data flows. Rather than manually tuning ingestion rules across dozens of tools, security teams want centralized control over:

  • data routing
  • enrichment policies
  • filtering logic
  • cost controls
  • compliance requirements

This concept is often described as an autonomous data layer, where the pipeline automatically performs:

  • data curation
  • data volume and cost reduction,
  • tiered storage, and also provides
  • pipeline visibility and health monitoring

You can explore this concept further in the discussion of the autonomous data layer for controlling security data cost and cyber risk.

Enabling AI Security Agents

Perhaps the fastest-growing topic is the role of AI in security operations. Organizations are experimenting with:

  • AI SOC assistants
  • automated investigation agents
  • threat hunting copilots
  • LLM-based detection systems

But these tools require large volumes of clean, structured data. Raw telemetry is typically inconsistent across sources, making it difficult for AI systems to interpret reliably. This is where security data pipelines become critical.

By normalizing data into standard schemas (such as OCSF), the pipeline layer can feed clean, structured telemetry directly into AI-driven security tools.

This approach also supports decoupled security architectures, where AI systems, data lakes, and SIEMs all consume the same normalized telemetry. This architecture is explored further in this article on the rise of decoupled SIEM architectures.

Security Data Pipeline Management Is Becoming a Core Discipline

Taken together, these trends point to the emergence of a new operational discipline: security data pipeline management.

Security teams increasingly need capabilities to:

  • collect telemetry at scale
  • normalize and enrich data
  • reduce noise and cost
  • route data across multiple tools
  • enable detection engineering
  • prepare data for AI systems

In other words, the pipeline layer is becoming the control plane for security data. You can learn more about this emerging operational model in this overview of security data pipeline management.

The Bottom Line

Security telemetry pipelines are evolving from a behind-the-scenes infrastructure component into a strategic layer of the modern security architecture.

Organizations that treat security data as a first-class asset—and build dedicated pipelines to manage it—gain several advantages:

In a world where security effectiveness increasingly depends on data, the telemetry pipeline is quickly becoming one of the most important layers in the entire security stack.

Follow Our Progress!

We are excited to be realizing our vision above with a full Axoflow product suite.

Sign Me Up
This button is added to each code block on the live site, then its parent is removed from here.

Fighting data Loss?

Balázs Scheidler

Book a free 30-min consultation with syslog-ng creator Balázs Scheidler

Recent Posts

The Stack We Built One Problem at a Time
The End of the Monolithic SIEM: Why Decoupled Security Architectures Are Growing In Popularity
How’s that AI copilot working out for you?