Monitor, evaluate & improve
your LLM apps

Langtrace is an open-source observability tool that collects and analyzes traces and metrics to help you improve your LLM apps.

Trusted and recognized by

Comcast's logo
Mayflower's logo
Elastic's logo
SearchStax's logo


Neoskop's logo
Pulse Energy's logo

Pulse Energy

Simple non-intrusive setup

Access the Langtrace SDK with 2 lines of code
from langtrace_python_sdk import langtrace


Supports popular LLMs, frameworks and vector databases

Why Langtrace?

Open-Source & Secure

Langtrace can be self-hosted and supports OpenTelemetry standard traces, which can be ingested by any observability tool of your choice, resulting in no vendor lock-in.

End-to-end Observability

Get visibility and insights into your entire ML pipeline, whether it is a RAG or a fine-tuned model with traces and logs that cut across framework, vectorDB and LLM requests.

Establish a Feedback Loop

Annotate and create golden datasets with traced LLM interactions, and use them to continuously test and enhance your AI applications. Langtrace includes built-in heuristic, statistical, and model-based evaluations to support this process.

Build and deploy with confidence



Trace requests, detect bottlenecks, and optimize performance with traces.


Annotate and Manually evaluate the LLM requests, and create golden datasets.


Run LLM based automated evaluations to track performance overtime.


Compare the performance of your prompts across different models.


Track cost and latency at project, model and user levels.

Built by a world class team of builders from

Join the Langtrace community