Find failures fast with agent tracing
Quickly debug and understand non-deterministic LLM app behavior with tracing. See what your agent is doing step by step —then fix issues to improve latency and response quality.
LangSmith Observability gives you complete visibility into agent behavior with tracing, real-time monitoring, alerting, and high-level insights into usage.




Quickly debug and understand non-deterministic LLM app behavior with tracing. See what your agent is doing step by step —then fix issues to improve latency and response quality.
Track business-critical metrics like costs, latency, and response quality with live dashboards. Get alerts when issues happen and drill into the root cause.


See clusters of similar conversations to understand what users actually want and quickly find all instances of similar problems to address systemic issues.
See exactly what's happening at every step of your agent. Steer your agent to accomplish critical tasks the way you intended.
Rapidly move through build, test, deploy, learn, repeat with workflows across the entire agent engineering lifecycle.
Ship at scale with agent infrastructure designed for long-running workloads and human oversight.
Keep your current stack. LangSmith works with your preferred open-source framework or custom code.
LangSmith works with any framework. If you’re already using LangChain or LangGraph, just set one environment variable to get started with tracing your AI application.