
Langfuse
Collaborative platform for building, monitoring, and debugging LLM applications.
- Stars
- 20,896
- License
- —
- Last commit
- 1 hour ago
Explore leading tools in the LLM Evaluation & Observability category, including open-source options and SaaS products. Compare features, use cases, and find the best fit for your workflow.
10+ open-source projects · 3 SaaS products
These projects are active, self-hostable choices for knowledge management teams evaluating alternatives to SaaS tools.

Collaborative platform for building, monitoring, and debugging LLM applications.

Open-source platform for tracing, evaluating, and optimizing LLM applications

AI observability platform for tracing, evaluation, and prompt management

Full‑stack observability for LLM applications via OpenTelemetry

Full‑life‑cycle platform for building, testing, and monitoring AI agents
Collaborative platform for building, monitoring, and debugging LLM applications.
Agenta enables engineering and product teams to build reliable LLM applications faster through collaborative prompt management, systematic evaluation, and real‑time observability, supporting 50+ models and custom integrations.
Expect a strong TypeScript presence among maintained projects.
Understand the commercial incumbents teams migrate from and how many open-source alternatives exist for each product.
DeepEval-powered LLM evaluation platform to test, benchmark, and safeguard apps
AIOps platform for streaming anomaly detection, root cause analysis, and incident prediction
LLM/agent observability with tracing, monitoring, and alerts
Confident AI (from the creators of DeepEval) provides metrics, regression testing, tracing, and guardrails to compare prompts/models, catch regressions, and monitor LLM applications.
Frequently replaced when teams want private deployments and lower TCO.
Browse neighbouring categories in ML & AI to widen your evaluation.