
Langfuse
Collaborative platform for building, monitoring, and debugging LLM applications.
- Stars
- 18,876
- License
- Unknown
- Last commit
- 3 days ago
Explore leading tools in the LLM Evaluation & Observability category, including open-source options and SaaS products. Compare features, use cases, and find the best fit for your workflow.
10+ open-source projects · 3 SaaS products
These projects are active, self-hostable choices for knowledge management teams evaluating alternatives to SaaS tools.

Collaborative platform for building, monitoring, and debugging LLM applications.

Open-source platform for tracing, evaluating, and optimizing LLM applications

AI observability platform for tracing, evaluation, and prompt management

Full‑stack observability for LLM applications via OpenTelemetry

Full‑life‑cycle platform for building, testing, and monitoring AI agents
Collaborative platform for building, monitoring, and debugging LLM applications.
Laminar provides automatic OpenTelemetry tracing, cost and token metrics, parallel evaluation, and dataset export for LLM apps, all via a Rust backend and SDKs for Python and TypeScript.
Understand the commercial incumbents teams migrate from and how many open-source alternatives exist for each product.
DeepEval-powered LLM evaluation platform to test, benchmark, and safeguard apps
AIOps platform for streaming anomaly detection, root cause analysis, and incident prediction
LLM/agent observability with tracing, monitoring, and alerts
Confident AI (from the creators of DeepEval) provides metrics, regression testing, tracing, and guardrails to compare prompts/models, catch regressions, and monitor LLM applications.
Frequently replaced when teams want private deployments and lower TCO.
Browse neighbouring categories in ML & AI to widen your evaluation.