All task areas

LLM Observability

Tracing, evals, prompt management for AI apps

3 tasks4 tools
1

Trace & debug LLM calls

langfuse
LLM Observability

Langfuse

Open-source LLM observability platform for tracing, evaluating, and debugging AI applications — self-host or use the cloud.

TypeScriptPythonOpen Source
85
Trust
Strong
langsmith
LLM Observability

LangSmith

LangChain's observability and evaluation platform — trace, debug, and evaluate LLM applications with deep LangChain ecosystem integration.

TypeScriptPythonLangChain
80
Trust
Strong
2

Evaluate LLM output quality

langfuse
LLM Observability

Langfuse

Open-source LLM observability platform for tracing, evaluating, and debugging AI applications — self-host or use the cloud.

TypeScriptPythonOpen Source
85
Trust
Strong
braintrust
LLM Observability

Braintrust

AI evaluation and observability platform focused on running structured evals, scoring LLM outputs, and prompt iteration workflows.

TypeScriptPythonEvals
70
Trust
Good
3

Track token cost & usage

helicone
LLM Observability

Helicone

Lightweight LLM observability via a proxy URL swap — get cost tracking, request logging, and caching with a one-line integration.

TypeScriptOpen SourceProxy
73
Trust
Good
langfuse
LLM Observability

Langfuse

Open-source LLM observability platform for tracing, evaluating, and debugging AI applications — self-host or use the cloud.

TypeScriptPythonOpen Source
85
Trust
Strong