Capability · Framework — observability

Datadog LLM Observability

Datadog LLM Observability sits inside the Datadog APM platform. It captures every LLM call as a span, visualises token and cost usage, runs out-of-the-box quality evaluations (toxicity, PII leakage, hallucination), and joins LLM traces to the rest of your distributed trace so you can debug end-to-end. The `ddtrace` SDK auto-instruments OpenAI, Anthropic, LangChain, LlamaIndex, Bedrock, and more.

Framework facts

Category
observability
Language
Python / Node.js / Java / Go
License
Commercial SaaS

Install

pip install ddtrace
export DD_API_KEY=...
export DD_LLMOBS_ENABLED=1
export DD_LLMOBS_ML_APP=my-llm-app

Quickstart

from ddtrace.llmobs import LLMObs
LLMObs.enable(ml_app='my-llm-app', api_key='...')

from openai import OpenAI
OpenAI().chat.completions.create(
    model='gpt-4o-mini',
    messages=[{'role':'user','content':'hi'}]
)
# traces visible in Datadog LLM Observability UI

Alternatives

  • Arize Phoenix — open-source
  • LangSmith — LangChain-native
  • New Relic AI Monitoring — direct competitor
  • Dynatrace / Splunk O11y — APM peers adding LLM features

Frequently asked questions

Does it work without LangChain?

Yes — `ddtrace` has auto-instrumentation for the major provider SDKs (OpenAI, Anthropic, Bedrock, Vertex) and a `@workflow`/`@llm` decorator for custom code.

How is pricing calculated?

Datadog charges per LLM span indexed, with a monthly free tier. Check current pricing; volumes in production can grow quickly.

Sources

  1. Datadog LLM Observability docs — accessed 2026-04-20
  2. ddtrace GitHub — accessed 2026-04-20