Skip to content

Stabilarity Hub

Menu
  • Home
  • Research
    • Healthcare & Life Sciences
      • Medical ML Diagnosis
    • Enterprise & Economics
      • AI Economics
      • Cost-Effective AI
      • Spec-Driven AI
    • Geopolitics & Strategy
      • Anticipatory Intelligence
      • Future of AI
      • Geopolitical Risk Intelligence
    • AI & Future Signals
      • Capability–Adoption Gap
      • AI Observability
      • AI Intelligence Architecture
    • Data Science & Methods
      • HPF-P Framework
      • Intellectual Data Analysis
    • Publications
      • External Publications
    • Robotics & Engineering
      • Open Humanoid
    • Benchmarks & Measurement
      • Universal Intelligence Benchmark
      • Shadow Economy Dynamics
  • Tools
    • Healthcare & Life Sciences
      • ScanLab
      • AI Data Readiness Assessment
    • Enterprise Strategy
      • AI Use Case Classifier
      • ROI Calculator
      • Risk Calculator
    • Portfolio & Analytics
      • HPF Portfolio Optimizer
      • Adoption Gap Monitor
      • Data Mining Method Selector
    • Geopolitics & Prediction
      • War Prediction Model
      • Ukraine Crisis Prediction
      • Gap Analyzer
    • Technical & Observability
      • OTel AI Inspector
    • Robotics & Engineering
      • Humanoid Simulation
    • Benchmarks
      • UIB Benchmark Tool
  • API Gateway
  • About
  • Contact
  • Join Community
  • Terms of Service
  • Geopolitical Stability Dashboard
Menu

Observability for AI Systems: Why OpenTelemetry Is Not Enough and What the Community Needs

Posted on March 4, 2026March 6, 2026 by
AI Observability & MonitoringTechnical Research · Article 1 of 1
By Oleh Ivchenko

πŸ”§ Try it: OTel AI Inspector β€” paste your OTel trace JSON and get your L1-L4 coverage score instantly. Free, client-side, no data sent to server.

Observability for AI Systems: Why OpenTelemetry Is Not Enough and What the Community Needs

πŸ“š Academic Citation: Ivchenko, Oleh (2026). Observability for AI Systems: Why OpenTelemetry Is Not Enough and What the Community Needs. Research article: Observability for AI Systems: Why OpenTelemetry Is Not Enough and What the Community Needs. ONPU. DOI: 10.5281/zenodo.18864333

Oleh Ivchenko Odessa National Polytechnic University March 2026


Abstract

Modern AI systems deployed in production remain fundamentally opaque to the engineers who operate them. While OpenTelemetry has emerged as the de facto standard for distributed systems observability, its extension to AI and large language model (LLM) workloads exposes critical gaps: latency traces do not capture hallucination rates, infrastructure metrics do not surface semantic drift, and no vendor-agnostic standard exists for quality observability. This paper surveys the current landscape of AI observability tooling, identifies structural gaps in the existing standards, and proposes a four-layer taxonomy β€” from infrastructure signals to business impact β€” that could form the foundation for a unified community standard. We establish a research agenda for the AI Observability & Monitoring series.


1. Introduction: The Black Box in Production

The deployment of machine learning systems into production environments has outpaced our ability to understand their behavior at runtime. For two decades, the software engineering community built robust observability practices around distributed systems: the “three pillars” of logs, metrics, and traces [1], formalized by the OpenTelemetry project [2] and operationalized through platforms like Prometheus [3], Grafana [4], Datadog [5], and Dynatrace [6]. These tools answer the question: Is the system behaving as its code specifies?

For AI systems, this question is necessary but insufficient. A language model serving responses with 200ms P99 latency and zero HTTP 5xx errors may simultaneously be hallucinating factual claims at a 12% rate, producing outputs that have drifted from its fine-tuning distribution, or consuming token budgets at a rate that will bankrupt the product in three months. None of these failures are visible in a conventional APM dashboard.

This is the foundational problem of AI observability: the contract of an AI system is semantic, not syntactic. Its correctness cannot be verified by inspecting code paths alone β€” it requires reasoning about the relationship between inputs, outputs, and intended behavior in a way that classical observability infrastructure was never designed to support.

The stakes are not theoretical. As organizations move AI systems from pilot to production β€” deploying LLM agents, retrieval-augmented generation (RAG) pipelines, and multi-model orchestration at scale β€” the absence of principled observability creates compounding operational debt. Engineers lack the instrumentation to detect regressions, attribute costs, debug agent behavior, or demonstrate compliance.

This paper surveys what the community has built, what is missing, and what a comprehensive standard would require.


2. What Exists Today: The Current Landscape

2.1 OpenTelemetry and the GenAI Semantic Conventions

OpenTelemetry (OTel) has achieved remarkable adoption as a vendor-neutral observability framework, providing SDKs, APIs, and a Collector architecture for traces, metrics, and logs across distributed systems [2]. In 2024, the OpenTelemetry community established the GenAI Special Interest Group (SIG), which produced a set of semantic conventions specifically for AI workloads [7].

The GenAI semantic conventions define span attributes for LLM calls, including:

  • gen_ai.system β€” the AI system being used (e.g., openai, anthropic)
  • gen_ai.request.model β€” the model identifier
  • genai.request.maxtokens β€” requested token limit
  • genai.response.finishreason β€” why generation stopped
  • genai.usage.prompttokens and genai.usage.completiontokens β€” token counts

The opentelemetry-instrumentation-openai package [8] provides automatic instrumentation for OpenAI API calls, generating spans that conform to these conventions. This represents genuine progress: engineers deploying OpenAI-backed systems can now see LLM calls as first-class spans in their distributed traces.

However, the GenAI semantic conventions as of their current specification address only the infrastructure surface of AI calls β€” what was requested and what was returned, measured in time and tokens. They do not address output quality, semantic correctness, or behavioral drift.

2.2 OpenLLMetry: OTel-Native LLM Observability

Traceloop’s OpenLLMetry [9] extends OpenTelemetry specifically for LLM applications. It provides instrumentation for major LLM providers and frameworks (LangChain, LlamaIndex, Haystack, Semantic Kernel) and generates OTel-compatible spans with LLM-specific attributes. OpenLLMetry’s core contribution is breadth: it makes OTel instrumentation available across the fragmented ecosystem of LLM libraries without requiring per-library instrumentation code.

OpenLLMetry also introduces the concept of workflow spans that can group related LLM calls into logical tasks β€” a partial solution to the multi-hop agent tracing problem. However, it remains a library-level instrumentation layer and inherits the same semantic limitations as the OTel GenAI conventions.

2.3 Application-Layer Observability Platforms

A cluster of purpose-built platforms has emerged to address what OTel does not:

PlatformPrimary FocusOTel-NativeSelf-HostableQuality Signals
LangSmith (LangChain)LLM debugging & evaluationPartialNoYes (via evaluators)
LangfuseLLM observability & tracingYes (OTLP)YesYes (scores API)
HeliconeCost & usage analyticsNoProxy-basedLimited
Phoenix (Arize AI)ML + LLM observabilityYesYesYes (evaluations)

LangSmith [10] offers deep integration with LangChain applications, providing trace visualization, prompt management, and an evaluation framework. Its evaluation system allows users to define quality metrics and run them against traced outputs β€” the closest existing approach to semantic observability. However, it is tightly coupled to the LangChain ecosystem and requires sending data to LangChain’s cloud infrastructure.

Langfuse [11] is an open-source alternative with strong OTLP support, a scores API for attaching quality signals to traces, and self-hosting capability. Its architecture is more OTel-aligned than LangSmith’s, making it a viable bridge platform for teams already invested in OTel infrastructure.

Phoenix by Arize AI [12] brings the company’s background in ML monitoring to LLM observability, offering embedding visualizations, retrieval analysis for RAG pipelines, and hallucination detection via LLM-as-judge evaluation. Its ML heritage gives it stronger drift detection capabilities than application-focused platforms.

Helicone [13] operates as a proxy layer, intercepting LLM API calls to capture usage and cost data without SDK integration. Its simplicity is its strength, but the proxy architecture limits it to request/response capture and forecloses deeper span instrumentation.

2.4 Experiment Tracking vs. Production Monitoring

MLflow [14] and Weights & Biases [15] represent the experiment tracking paradigm: tooling designed for the model development phase, tracking hyperparameters, metrics, and artifacts across training runs. Both have extended their platforms toward deployment and monitoring, but the conceptual gap remains. Experiment tracking optimizes for reproducibility (can I recreate this result?), while production monitoring optimizes for reliability (is this system behaving correctly right now?). The data models, storage requirements, and operational patterns are fundamentally different, and the conflation of these concerns has impeded the development of purpose-built production monitoring standards.

2.5 Prometheus and Grafana Patterns for ML

The Prometheus + Grafana stack [3][4] is widely used for ML infrastructure monitoring: GPU utilization, memory consumption, inference throughput, and queue depth. Custom metrics can be exported via the Prometheus client libraries, and teams have developed patterns for tracking model-level statistics β€” prediction distributions, output value histograms, and request rates by input feature category.

This approach provides genuine value for detecting infrastructure-level anomalies and broad behavioral shifts. However, it requires significant custom engineering effort, lacks standardized metric naming conventions for AI workloads, and provides no native support for semantic quality signals.


3. The Structural Gaps

The landscape survey reveals four categories of gaps that no existing solution adequately addresses:

3.1 Model Quality Signals Are Not First-Class Observables

Existing observability infrastructure treats model outputs as opaque byte strings. Token counts are measured; semantic content is not. There is no standard mechanism for attaching quality signals β€” factuality scores, semantic similarity to reference outputs, coherence ratings, or refusal rates β€” to trace spans in a way that propagates through the observability pipeline to dashboards, alerts, and SLO calculations.

Some platforms provide post-hoc evaluation frameworks (LangSmith evaluators, Phoenix evaluations), but these operate outside the trace collection path and cannot power real-time alerting or SLO compliance calculations.

3.2 Multi-Hop Agent Tracing Is Unsolved

LLM agents that orchestrate tool calls, sub-agent invocations, and multi-step reasoning chains present a tracing challenge that OTel’s span model handles poorly. A ReAct-pattern agent [16] that executes five tool calls across three external APIs before producing a final response generates a trace tree where:

  1. The causally significant spans are at leaf level (tool outputs that shaped the final response)
  2. The overall quality of the response cannot be attributed to any individual span
  3. The “reasoning” between tool calls is invisible to the tracer

The OTel span model supports parent-child relationships but provides no semantic primitives for representing reasoning chains, attention to intermediate results, or counterfactual attribution.

3.3 Drift Detection Is External and Ad Hoc

Data drift and model drift β€” the degradation of model performance as the statistical distribution of inputs diverges from training data β€” are recognized problems with mature detection algorithms [17]. However, drift detection is universally implemented as a separate batch pipeline, operating on collected data with significant latency, rather than as a first-class observable integrated into the real-time telemetry pipeline.

The consequence is that drift is detected reactively: after user-facing quality has degraded, after business metrics have moved, and often only after manual investigation. No existing observability standard treats drift as a signal that should appear in the same pipeline as latency or error rate.

3.4 Cost Attribution Is Coarse

LLM inference costs are substantial and variable. Token-level cost attribution β€” mapping cost to specific users, features, request types, or business units β€” requires custom engineering on top of existing telemetry. The GenAI semantic conventions capture token counts, but the translation from token counts to costs (which varies by model, provider, and time) is not standardized, and cost signals are not integrated into standard observability dashboards or alerting frameworks.


4. A Proposed Taxonomy: Four Layers of AI Observability

We propose a four-layer taxonomy that structures AI observability concerns from infrastructure to business impact:

graph TD
    L4["L4: Business Impact\nTask completion rate\nUser satisfaction proxy\nRevenue attribution\nSLO compliance"]
    L3["L3: Semantic Quality\nFactuality scores\nRelevance and coherence\nHallucination rate\nRefusal rate"]
    L2["L2: Model Behavior\nLatency (TTFT, TBT)\nToken usage and costs\nError rates by type\nFinish reasons"]
    L1["L1: Infrastructure\nCPU / GPU / Memory\nNetwork and queue depth\nService health\nClassic OTel spans"]

    L4 --> L3
    L3 --> L2
    L2 --> L1

L1: Infrastructure covers the traditional observability domain: compute resource utilization, network performance, service availability, and the OTel spans that already instrument HTTP calls and database queries. This layer is mature; the community has solved it.

L2: Model Behavior covers the signals specific to model inference: time to first token (TTFT), token generation throughput, token budget consumption, finish reason distributions, and error categorization (rate limit, content filter, context length exceeded, etc.). The GenAI semantic conventions address this layer incompletely; it is achievable with existing OTel primitives.

L3: Semantic Quality is the critical unsolved layer. It encompasses signals that require reasoning about the content of model outputs: factuality relative to source documents, semantic similarity to expected outputs, coherence and fluency, refusal and safety filter activation rates, and hallucination detection. These signals require either model-based evaluation (LLM-as-judge) or embedding-based similarity computation β€” neither of which fits naturally into the synchronous, low-overhead span collection path.

L4: Business Impact connects model behavior to user and business outcomes: task completion rates, user satisfaction proxies (thumbs up/down, session continuation, return rate), revenue attribution for AI-assisted transactions, and SLO compliance expressed in business terms. This layer requires joining telemetry data with product analytics, which existing observability infrastructure does not natively support.

4.1 Tool Coverage by Layer

graph LR
    subgraph "L1: Infrastructure"
        OTel["OTel + Prometheus"]
        Grafana["Grafana"]
    end
    subgraph "L2: Model Behavior"
        OpenLLMetry["OpenLLMetry"]
        Helicone["Helicone"]
        Langfuse2["Langfuse"]
    end
    subgraph "L3: Semantic Quality"
        LangSmith["LangSmith"]
        Phoenix["Phoenix (Arize)"]
        Langfuse3["Langfuse (scores)"]
    end
    subgraph "L4: Business Impact"
        Custom["Custom Engineering Required"]
    end

5. Agent Trace Flow: The Multi-Hop Problem

A representative agent trace demonstrates the instrumentation challenges at the architectural level:

sequenceDiagram
    participant User
    participant Agent as LLM Agent
    participant OTel as OTel Collector
    participant Tool1 as Search API
    participant LLM as LLM Provider

    User->>Agent: Query
    Agent->>OTel: span: agent.run [start]
    
    Agent->>LLM: plan(query)
    LLM-->>Agent: tool call plan
    Agent->>OTel: span: gen_ai.completion [tokens=340]
    
    Agent->>Tool1: search(query)
    Tool1-->>Agent: results
    Agent->>OTel: span: tool.search [latency=820ms]
    
    Agent->>LLM: reason(query, results)
    LLM-->>Agent: final_response
    Agent->>OTel: span: gen_ai.completion [tokens=580]
    
    Agent->>User: response
    Agent->>OTel: span: agent.run [end, total_tokens=920]

    Note over OTel: Missing: hallucination rate, semantic quality, drift signal

The diagram highlights the instrumentation boundary: OTel can capture every span, latency, and token count. It cannot capture whether the final response is factually correct, whether the search results were relevant, or whether the model’s reasoning chain is coherent. These gaps are not gaps in OTel’s implementation β€” they are gaps in its conceptual model.


6. What the Community Needs: A Research Agenda

Addressing the structural gaps requires contributions at multiple levels:

6.1 Semantic Conventions for Quality Signals

The OpenTelemetry GenAI SIG should extend its semantic conventions to include quality signal attributes at L3. This requires defining:

  • A standard schema for attaching evaluation scores to spans (by score type, evaluator identity, and confidence)
  • Conventions for representing retrieval quality in RAG pipelines (hit rate, relevance score, faithfulness)
  • Standard span events for semantic quality checkpoints

6.2 Asynchronous Evaluation Integration

Real-time semantic evaluation (LLM-as-judge at inference time) adds latency and cost that may be prohibitive. The community needs standardized patterns for asynchronous evaluation: collecting raw inputs and outputs at inference time (L2), running quality evaluation as a background process, and propagating results back to the trace with a causal link to the original span.

6.3 Drift as a First-Class Signal

Drift detection should be defined as an observable with a standard metric type and alert semantics. Proposed conventions:

  • ai.drift.input_distribution β€” divergence from training input distribution
  • ai.drift.output_distribution β€” shift in output value distribution
  • ai.drift.embedding_distance β€” cosine distance from reference embedding centroid

These metrics should be exportable via standard Prometheus exposition format and queryable via PromQL.

6.4 Cost Attribution at the Trace Level

A standard for cost attribution would define how token counts (captured at L2) are translated to monetary costs and associated with business-level dimensions (user, feature, organization, deployment). This requires:

  • A provider-maintained registry of token costs per model
  • A standard span attribute for computed cost (genai.usage.costusd)
  • Cost rollup semantics for agent traces with multiple LLM calls

6.5 Community Reference Implementation

The most impactful near-term contribution would be an open-source reference implementation demonstrating all four layers in a single instrumented application: a Python-based LLM agent with complete L1-L4 observability, exportable to any OTel-compatible backend. This reference implementation would demonstrate the async evaluation pattern, provide example Grafana dashboards covering all four layers, include drift detection with Prometheus metrics, and enable the community to experiment with the proposed conventions before standardization.


7. Conclusion

OpenTelemetry represents an extraordinary achievement in standardizing observability infrastructure. Its extension to AI workloads through the GenAI semantic conventions is a meaningful step, but it addresses only the first two layers of a four-layer challenge. The community has produced valuable application-layer tooling β€” Langfuse, Phoenix, LangSmith β€” but these tools remain siloed, vendor-specific, or ecosystem-constrained.

The path to principled AI observability runs through the OpenTelemetry community: extending semantic conventions to L3 quality signals, standardizing asynchronous evaluation patterns, and treating drift detection as a first-class observable rather than a batch analytics afterthought.

This paper establishes the taxonomy and identifies the research agenda. Subsequent articles in this series will address each gap in depth β€” from practical instrumentation patterns to the statistical foundations of semantic drift detection β€” with the goal of producing community-ready specification proposals.

The black box in production is a solvable problem. Solving it requires the same community consensus-building that produced OpenTelemetry itself.


References

[1] Majors, C., Fong-Jones, L., & Miranda, G. (2022). Observability Engineering. O’Reilly Media.

[2] OpenTelemetry Authors. (2024). OpenTelemetry Specification. https://opentelemetry.io/docs/specs/otel/

[3] Prometheus Authors. (2024). Prometheus: Monitoring System & Time Series Database. https://prometheus.io/

[4] Grafana Labs. (2024). Grafana: The open observability platform. https://grafana.com/

[5] Datadog Inc. (2024). AI Observability. https://www.datadoghq.com/product/ai-observability/

[6] Dynatrace. (2024). AI-powered observability. https://www.dynatrace.com/

[7] OpenTelemetry GenAI SIG. (2024). Semantic Conventions for Generative AI Systems. https://opentelemetry.io/docs/specs/semconv/gen-ai/

[8] OpenTelemetry Authors. (2024). opentelemetry-instrumentation-openai. https://github.com/open-telemetry/opentelemetry-python-contrib

[9] Traceloop. (2024). OpenLLMetry: Open-source observability for your LLM application. https://github.com/traceloop/openllmetry

[10] LangChain. (2024). LangSmith. https://smith.langchain.com/

[11] Langfuse. (2024). Open Source LLM Engineering Platform. https://langfuse.com/

[12] Arize AI. (2024). Phoenix: AI Observability & Evaluation. https://phoenix.arize.com/

[13] Helicone. (2024). Open-source LLM Observability. https://www.helicone.ai/

[14] Zaharia, M., et al. (2018). Accelerating the Machine Learning Lifecycle with MLflow. IEEE Data Eng. Bull., 41(4), 39-45.

[15] Weights & Biases. (2024). Weights & Biases: The AI Developer Platform. https://wandb.ai/

[16] Yao, S., et al. (2022). ReAct: Synergizing Reasoning and Acting in Language Models. arXiv:2210.03629.

[17] Gama, J., Zliobaite, I., Bifet, A., Pechenizkiy, M., & Bouchachia, A. (2014). A survey on concept drift adaptation. ACM Computing Surveys, 46(4), 1-37.

[18] Fowler, M. (2019). Observability and Monitoring. https://martinfowler.com/articles/domain-oriented-observability.html

[19] DORA Research Program. (2023). Accelerate State of DevOps Report. Google Cloud.

[20] Liang, P., et al. (2022). Holistic Evaluation of Language Models. arXiv:2211.09110.

[21] Zhao, W. X., et al. (2023). A Survey of Large Language Models. arXiv preprint. arXiv:2302.14016

[22] Liang, P., et al. (2022). Holistic Evaluation of Language Models (HELM). Stanford CRFM. arXiv:2211.09110

[23] Shen, T., et al. (2023). Large Language Models Aligned to Human Preferences through RLHF. arXiv:2307.01003

← Previous
Start of series
Next →
Next article coming soon
All AI Observability & Monitoring articles (1)1 / 1
Version History Β· 3 revisions
+
RevDateStatusActionBySize
v1Mar 4, 2026DRAFTInitial draft
First version created
(w) Author21,809 (+21809)
v2Mar 4, 2026PUBLISHEDPublished
Article published to research hub
(w) Author21,809 (~0)
v3Mar 6, 2026CURRENTContent update
Section additions or elaboration
(w) Author22,141 (+332)

Versioning is automatic. Each revision reflects editorial updates, reference validation, or formatting changes.

Recent Posts

  • Container Orchestration for AI β€” Kubernetes Cost Optimization
  • The Computer & Math 33%: Why the Most AI-Capable Occupation Group Still Automates Only a Third of Its Tasks
  • Frontier AI Consolidation Economics: Why the Big Get Bigger
  • Silicon War Economics: The Cost Structure of Chip Nationalism
  • Enterprise AI Agents as the New Insider Threat: A Cost-Effectiveness Analysis of Autonomous Risk

Recent Comments

  1. Oleh on Google Antigravity: Redefining AI-Assisted Software Development

Archives

  • March 2026
  • February 2026

Categories

  • ai
  • AI Economics
  • AI Observability & Monitoring
  • AI Portfolio Optimisation
  • Ancient IT History
  • Anticipatory Intelligence
  • Capability-Adoption Gap
  • Cost-Effective Enterprise AI
  • Future of AI
  • Geopolitical Risk Intelligence
  • hackathon
  • healthcare
  • HPF-P Framework
  • innovation
  • Intellectual Data Analysis
  • medai
  • Medical ML Diagnosis
  • Open Humanoid
  • Research
  • Shadow Economy Dynamics
  • Spec-Driven AI Development
  • Technology
  • Uncategorized
  • Universal Intelligence Benchmark
  • War Prediction

About

Stabilarity Research Hub is dedicated to advancing the frontiers of AI, from Medical ML to Anticipatory Intelligence. Our mission is to build robust and efficient AI systems for a safer future.

Language

  • Medical ML Diagnosis
  • AI Economics
  • Cost-Effective AI
  • Anticipatory Intelligence
  • Data Mining
  • πŸ”‘ API for Researchers

Connect

Facebook Group: Join

Telegram: @Y0man

Email: contact@stabilarity.com

© 2026 Stabilarity Research Hub

© 2026 Stabilarity Hub | Powered by Superbs Personal Blog theme
Stabilarity Research Hub

Open research platform for AI, machine learning, and enterprise technology. All articles are preprints with DOI registration via Zenodo.

185+
Articles
8
Series
DOI
Archived

Research Series

  • Medical ML Diagnosis
  • Anticipatory Intelligence
  • Intellectual Data Analysis
  • AI Economics
  • Cost-Effective AI
  • Spec-Driven AI

Community

  • Join Community
  • MedAI Hack
  • Zenodo Archive
  • Contact Us

Legal

  • Terms of Service
  • About Us
  • Contact
Operated by
Stabilarity OÜ
Registry: 17150040
Estonian Business Register β†’
© 2026 Stabilarity OÜ. Content licensed under CC BY 4.0
Terms About Contact
Language: πŸ‡¬πŸ‡§ EN πŸ‡ΊπŸ‡¦ UK πŸ‡©πŸ‡ͺ DE πŸ‡΅πŸ‡± PL πŸ‡«πŸ‡· FR
Display Settings
Theme
Light
Dark
Auto
Width
Default
Column
Wide
Text 100%

We use cookies to enhance your experience and analyze site traffic. By clicking "Accept All", you consent to our use of cookies. Read our Terms of Service for more information.