Use Case High-Cardinality Metrics at Scale

Store Billions of Unique Metric Streams Without Performance Penalties or Cost Explosions

Modern cloud-native architectures demand high-cardinality observability but traditional time series databases force you to choose between complete visibility and sustainable costs. Apica Forge eliminates this tradeoff. And as AI agents generate per-token costs, latency histograms, and tool invocation rates at machine speed, Forge is the metrics backbone your agentic infrastructure depends on.
Billions
Unique metric streams supported
Low-Latency
Consistently fast query performance regardless of unique metric stream count
4,000
Character tag name support — label everything that matters, freely
Zero
Cardinality-driven pricing penalties. Store every metric stream Kubernetes and your AI agents generate
Common scenarios we solve
Prometheus hits cardinality limits under Kubernetes load — monitoring blackout during peak traffic
BEFORE
Apica Forge handles billions of unique metric streams without limits, sampling, or performance loss
AFTER
Engineers self-censor meaningful tags to avoid cardinality cost spikes — blind spots by design
BEFORE
4,000-character tag support, zero cardinality pricing — label everything that matters, freely
AFTER
Querying high-cardinality dimensions takes minutes — root cause analysis is painfully slow
BEFORE
Consistent millisecond query performance regardless of unique metric stream count or data age
AFTER
AI agents generating per-token costs, latency histograms, and tool call rates — no TSDB built to handle this cardinality at machine speed
BEFORE
Apica Forge stores AI-specific metric signals at full resolution — the metrics foundation agentic infrastructure depends on
AFTER
The Problem

The High-Cardinality Crisis in Cloud-Native Environments — And AI Is About to Make It Exponentially Worse

Cloud-native and microservices architectures have fundamentally changed the metrics landscape. Kubernetes deployments with hundreds of ephemeral pods, containerized applications with dynamic service meshes, and auto-scaling infrastructure generate explosive cardinality growth — millions of unique time series metrics with dozens of tags and dimensions per metric. Traditional time series databases weren’t designed for this reality.

AI agents compound the cardinality problem in ways that general-purpose time series databases (TSDBs) are fundamentally unable to address. Every AI agent in production generates per-token costs, per-request latency histograms, tool invocation rates, and multi-agent call chain metrics, continuously, at machine speed, at cardinality levels that break platforms built for human-scale infrastructure monitoring. The organizations deploying AI agents today need a metrics foundation engineered for this reality. Apica Forge is purpose-built for exactly this.

  • Cardinality management challenges

    Competing platforms often tie cardinality limits to pricing tiers, forcing trade-offs between monitoring granularity and cost.

  • Query performance degradation

    Traditional TSDBs slow to a crawl when querying across high-cardinality dimensions — minutes to return results that should take milliseconds.

  • Rapidly escalating cost scaling

    Per-metric or per-custom-metric pricing models penalize cloud-native architectures, with costs growing unsustainably as Kubernetes environments scale — a pattern Apica customers consistently report.

  • Premature aggregation

    Teams forced to aggregate data at collection time, destroying granularity and creating blind spots during incident investigation.

  • Tag anxiety

    Engineers self-censor meaningful labels to stay within cardinality budgets, reducing observability value.

  • AI metrics blindness

    General-purpose TSDBs can't handle the cardinality and frequency of AI agent metric signals, per-token costs, latency distributions, tool call rates, without degradation or cost explosion.value.

Organizations face an impossible choice: reduce metric granularity and accept degraded troubleshooting capabilities — or maintain full-context observability at unsustainable cost with crippled query performance.
The cardinality tax you're paying
Rapidly escalating
Observability costs on per-metric pricing platforms as Kubernetes and AI workloads multiply unique metric combinations, based on Apica customer data
Minutes
Query response time on traditional TSDBs against high-cardinality dimensions — should be milliseconds
Zero visibility
Into cardinality budget consumption until monitoring fails. Traditional TSDBs provide no warning before hitting limits
AI-native signals
Per-token costs, per-agent latency histograms, tool invocation rates, multi-agent call chain metrics. Cardinality traditional TSDBs can't handle at machine speed
Our Solution

Apica Forge: Built for Infinite Cardinality — Including Your AI Agents

Apica Forge is purpose-built for the cardinality realities of modern cloud-native infrastructure and the AI workloads running on top of it. No artificial limits. No per-metric pricing. No aggregation trade-offs. Store billions of unique metric streams, including the AI-specific signals that break general-purpose TSDBs, with consistently fast query performance regardless of cardinality.

Before Apica
  • Cardinality limits: Artificial caps force trade-offs between monitoring completeness and cost
  • Query performance degradation: Traditional TSDBs slow dramatically as unique metric combinations grow
  • Exponential cost scaling: Per-metric pricing penalizes cloud-native architectures that generate high cardinality
  • Forced aggregation: Destroying granularity at collection time creates permanent blind spots
  • Tag anxiety: Engineers self-censoring meaningful labels to avoid cardinality cost spikes
  • AI metrics gap: No TSDB built to handle per-token costs, agent latency histograms, and tool invocation rates at the cardinality and frequency AI agents generate in production
With Apica
  • No cardinality limits: Store billions of unique metric streams without performance penalties or cost explosions
  • Consistent query performance: Millisecond response times regardless of unique metric stream count
  • Architecture-aware pricing: No per-metric charges that penalize Kubernetes and microservices scale — pricing is structured to scale with actual usage; contact the Apica team for details on what's right for your environment
  • Full granularity preserved: Collect at full resolution, aggregate on query — never at collection time
  • 4,000-character tag support: Use meaningful, descriptive labels without artificial constraints
  • AI-native metrics foundation: Store per-token costs, latency histograms, tool invocation rates, and multi-agent call chain metrics at full resolution — the cardinality-scale metrics backbone that agentic infrastructure demands

The Apica advantage: We eliminate the cardinality tax, store every metric stream at full resolution without compromising query performance or breaking your budget. And we're the metrics foundation your AI agents can actually depend on.

How It Works

Infinite Cardinality, Consistent Performance

Apica Forge's architecture was designed from the ground up for the high-cardinality realities of cloud-native observability, delivering millisecond query performance regardless of how many unique metric streams you generate. And because Forge was explicitly built for AI-native metric signals, it extends the same unlimited cardinality and millisecond performance to the agentic workloads now entering production.

Apica Forge Time Series Engine

  • Purpose-built TSDB handling billions of unique time series without performance degradation
  • Metric names and tags supporting up to 4,000 characters — no artificial truncation
  • No per-metric or per-custom-metric pricing constraints: Pricing scales with actual usage; contact the Apica team for the model that fits your environment
  • Consistent millisecond query performance at any cardinality scale

Kubernetes-Scale Cardinality

  • Handle millions of unique metric combinations from ephemeral pod environments
  • Every pod, container, deployment, namespace, and node fully monitored without cardinality limits
  • Service mesh metrics at full granularity across all interactions
  • Auto-scaling doesn't create cardinality explosions or cost spikes

Full-Resolution Collection

  • Collect metrics at full resolution — never pre-aggregate or downsample at collection time
  • Apply aggregation only when querying — preserve all granularity for incident investigation
  • Dynamic rollup policies applied retrospectively without data loss
  • Compare current behavior against any historical point at full resolution

Query Performance at Scale

  • Millisecond query response across billions of unique metric streams
  • Efficient storage architecture that doesn't degrade as cardinality grows
  • Multi-dimensional queries across any tag combination without performance penalties
  • Historical data accessible at consistently fast performance. No archival delay in standard deployments

AI-Native Metrics and SLO Signals (Powered by Apica Forge)

Apica Forge was built to be the metrics backbone for agentic infrastructure, storing and querying the signals that AI agents generate at machine speed:

  • Per-token costs: Track LLM API costs at per-request granularity across all models and features. No aggregation that hides cost spikes
  • Per-agent latency histograms: Store full latency distributions for every AI agent interaction, calculate accurate P95/P99 without raw sample retention
  • Tool invocation rates: Monitor every tool call, retrieval operation, and API interaction at full cardinality: Per agent, per model, per feature
  • Multi-agent call chain metrics: Track the metric signals that emerge from complex agentic workflows, cardinality that breaks general-purpose TSDBs
  • SLO signals at agentic scale: Define and track Service Level Objectives against agentic reliability contracts, powered by Forge's high-cardinality metrics engine
  • Continuous operations under failure: Multi-node replication keeps Forge ingesting and serving queries even during node failures. No gaps in agent telemetry, no alerting blind spots
The Result

No Cardinality Limits. No Performance Penalties. No Surprises.

Billions
Unique metric streams stored without performance penalties or cost explosions
Low-Latency
Consistently fast query performance regardless of cardinality
Significant
Reduction in metrics storage costs through histogram-native storage, which stores statistical distributions rather than every raw sample, eliminating the storage burden without sacrificing percentile accuracy
Zero
Engineering blind spots from tag self-censorship — label everything your infrastructure and AI agents generate
Customer Results

Results based on Apica customer deployments. Individual results may vary based on environment complexity and implementation scope.

Cloud-Native SaaS: Platform Engineering

Challenge

Kubernetes deployment generating 500M+ unique metric streams. Prometheus hit cardinality limits and failed under load. Datadog costs grew rapidly, driven by per-metric cardinality pricing.

Solution

Apica Forge as the high-cardinality metrics backend, replacing Prometheus for Kubernetes-scale time series storage.

Results
  • 100% cardinality coverage — every metric stream captured with no limits or sampling
  • Query response time improved from minutes to under 50ms at full scale
  • 65% reduction in metrics storage costs vs. per-metric pricing platform
  • Engineering team removed tag self-censorship — full context for every incident
Customer Results

Results based on Apica customer deployments. Individual results may vary based on environment complexity and implementation scope.

Financial Services: Observability Platform Team

Challenge

Multi-region trading platform with 2B+ unique metric streams across microservices. Legacy TSDB degraded to 5-minute query times at scale.

Solution

Apica Forge replacing legacy TSDB for high-cardinality financial metrics with consistent query performance.

Results
  • 2B+ metric streams stored with consistent sub-100ms query performance
  • Zero metric loss during peak trading cardinality events
  • 78% reduction in TSDB operational overhead
  • Root cause analysis time reduced from hours to minutes with full-granularity access
Why Apica

Built for Cardinality at Enterprise Scale — Including the AI Workloads Generating It

Apica Forge wasn't retrofitted for high cardinality — it was architected from scratch to handle the metric reality of modern cloud-native infrastructure. That same unlimited cardinality and millisecond performance extends naturally to the AI agents now entering production environments, handling the per-token costs, latency histograms, and tool invocation rates that general-purpose TSDBs can't sustain at scale.

Architecture-Native Cardinality

Technical Foundation

Not a traditional TSDB with cardinality bolted on — Apica Forge's storage architecture is designed for billions of unique time series. No performance degradation, no artificial limits, no scaling surprises.

Pricing That Scales with Value

Economic Model

No per-metric or per-custom-metric charges that penalize cloud-native architectures. Predictable costs that don't explode as Kubernetes generates more unique metric combinations. Pricing is structured to scale with actual usage — contact the Apica team to discuss the right model for your environment.

Full Granularity Always

Data Integrity

Never pre-aggregate or downsample at collection time. Collect at full resolution, apply aggregation on query. Full context for incident investigation at any historical point.

4,000-Character Tag Support

Label Freedom

Use descriptive, meaningful tag names without artificial constraints. Stop self-censoring important context to stay within cardinality budgets — label everything that matters.

The Metrics Backbone for Agentic AI

Forge Product Principle

Apica Forge was built to be where metrics become the foundation for agentic intelligence. Per-token costs, per-agent latency histograms, tool invocation rates, multi-agent call chain metrics — stored at full resolution, queryable in milliseconds, at the cardinality levels that production AI agents generate. The new SLO dashboard in Ascent 2.16 uses Forge's high-cardinality metrics engine to track reliability contracts for AI-integrated services — not just traditional uptime metrics. Forge is the metrics layer that scales from Kubernetes to agentic AI without architectural compromise..