Capability · Framework — observability

Opik

Opik captures every LLM call as a structured trace, lets you annotate and compare them, and ships LLM-as-judge evaluation suites out of the box. It's designed for engineering teams who want LangSmith-style visibility without being locked into the LangChain ecosystem.

Framework facts

Category
observability
Language
Python / TypeScript
License
Apache-2.0
Repository
https://github.com/comet-ml/opik

Install

pip install opik
opik configure  # choose cloud or self-host

Quickstart

from opik import track
import anthropic

@track
def ask(question: str) -> str:
    client = anthropic.Anthropic()
    r = client.messages.create(
        model='claude-opus-4-7',
        max_tokens=256,
        messages=[{'role': 'user', 'content': question}],
    )
    return r.content[0].text

print(ask('What is VSET?'))

Alternatives

  • Langfuse — open-source equivalent
  • LangSmith — LangChain-native
  • Arize Phoenix — open-source
  • Helicone — proxy-based

Frequently asked questions

Opik vs Langfuse — which should I pick?

Both are strong open-source observability platforms. Opik leans into evaluation (LLM-as-judge, datasets, experiments); Langfuse leans into analytics and prompt management. Try both on a weekend.

Is Opik only for Comet users?

No. It's fully standalone and self-hostable. Comet offers a managed tier for teams that prefer SaaS.

Sources

  1. Opik — GitHub — accessed 2026-04-20
  2. Opik — docs — accessed 2026-04-20