Technology RadarTechnology Radar
This item was not updated in last three versions of the Radar. Should it have appeared in one of the more recent editions, there is a good chance it remains pertinent. However, if the item dates back further, its relevance may have diminished and our current evaluation could vary. Regrettably, our capacity to consistently revisit items from past Radar editions is limited.
Trial

Langfuse is an open-source LLM observability platform — providing tracing, monitoring, cost tracking, and evaluation for AI applications in production. It's the standard starting point for teams that need visibility into what their AI systems are actually doing.

Buy vs Build

Strong buy story: Langfuse Cloud (hosted) gets you up and running in minutes. Strong build story for data-sensitive teams: self-host via Docker with the open-source version. Both options use the same SDK.

Why Engineering Managers Care About LLM Observability

Traditional application monitoring (APM) tells you a request returned a 200 status and took 230ms. That's not enough for AI systems. An LLM can return a 200 in 230ms and still:

  • Give completely wrong information
  • Generate unsafe content
  • Waste tokens on irrelevant reasoning
  • Fail silently on an edge case it was never tested on

Langfuse adds the missing layer: it captures what the model was asked, what it answered, how much it cost, and whether the answer was actually good.

What It Tracks

  • Traces: The full input → model call → output for each request, including nested tool calls in agentic workflows
  • Token costs: Per-request and aggregated cost breakdown across models and users
  • Latency: Time to first token, total response time, queue time
  • Evaluation scores: Automated LLM-as-a-judge scoring or human annotation via the Langfuse UI
  • Sessions: Grouping related traces into a conversation or user session

Key Features

  • Works with any LLM framework (LangChain, LangGraph, CrewAI, direct API calls)
  • Python and TypeScript SDKs with framework-specific integrations
  • Self-hostable (Docker, no external dependencies beyond Postgres + Redis)
  • Prompt management: version, A/B test, and deploy prompts from the UI without a code deploy

Getting Started

pip install langfuse

from langfuse.decorators import observe

@observe()  # automatically traces this function
def generate_summary(text: str) -> str:
    response = anthropic.messages.create(
        model="claude-sonnet-4-6",
        messages=[{"role": "user", "content": f"Summarise: {text}"}]
    )
    return response.content[0].text

Key Characteristics

Property Value
License MIT (core) + commercial (enterprise features)
Self-hostable Yes (Docker + Postgres + Redis)
Cloud pricing Free up to 50k observations/month; Pro $59/month
Frameworks LangChain, LangGraph, OpenAI, Anthropic, direct, and more
Provider Langfuse GmbH
Website langfuse.com
GitHub langfuse/langfuse

Further Reading