Ship agents with confidence.

LangSmith is a unified observability & evals platform where teams can debug, test, and monitor AI app performance — whether building with LangChain or not.

Helping top teams ship AI agents that actually work.

How does LangSmith help?

Real-time visibility

Trace every agent run to pinpoint bottlenecks and understand mistakes in development.

Agent Evals

Improve agent performance with test-driven development, guided by automated evals and datasets.

Faster prompt iteration

Non-developer teammates can refine prompts and kick off evaluations in the UI.

Production-grade monitoring

Track latency, cost, and issues with quality before your users do.

Find failures fast with agent observability.

Get started tracing your app

Quickly debug and understand non-deterministic LLM app behavior with tracing. See what your agent is doing step by step —then fix issues to improve latency and response quality.

Evaluate your agent's performance.

Evaluate your app by saving production traces to datasets —  then score performance with LLM-as-Judge evaluators. Gather human feedback from subject-matter experts to assess response relevance, correctness, harmfulness, and other criteria.

Learn how to run an eval

Iterate and collaborate on prompts.

Experiment with models and prompts in the Playground, and compare outputs across different prompt versions. Any teammate can use the Prompt Canvas UI to directly recommend and improve prompts.

Create and test a prompt

Monitor what matters to the business.

Track business-critical metrics like costs, latency, and response quality with live dashboards — then drill into the root cause when problems arise.

See how to create a custom dashboard

Learn best practices for evaluating your AI agents, from design to production.

LangSmith FAQs

Can I use LangSmith if I don’t use LangChain or LangGraph?

Yes! Many companies who don’t build with LangChain/LangGraph use LangSmith. You can log traces to LangSmith via the Python SDK, the TypeScript SDK, or the API. See here for more information.

How easy is it to start using LangSmith if I use LangChain or LangGraph?

Getting started on LangSmith requires just two environment variables in your LangChain or LangGraph code. See how to send traces from your LangGraph agent or your LangChain app.

My application isn’t written in Python or TypeScript. Will LangSmith be helpful?

Yes, you can log traces to LangSmith using a standard OpenTelemetry client to access all LangSmith features, including tracing, running evals, and prompt engineering. See the docs.

How can LangSmith help with observability and evaluation?

LangSmith traces contain the full information of all the inputs and outputs of each step of the application, giving users full visibility into their agent or LLM app behavior. LangSmith also allows users to instantly run evals to assess agent or LLM app performance — including LLM-as-Judge evaluators for auto-scoring and the ability to attach human feedback. Learn more.

I can’t have data leave my environment. Can I self-host LangSmith?

Yes, we allow customers to self-host LangSmith on our enterprise plan. We deliver the software to run on your Kubernetes cluster, and data will not leave your environment. For more information, check out our documentation.

Where is LangSmith data stored?

For Cloud SaaS, traces are stored in GCP us-central-1 or GCP europe-west4, depending on your plan. Learn more.

Will LangSmith add latency to my application?

No, LangSmith does not add any latency to your application. In the LangSmith SDK, there’s a callback handler that sends traces to a LangSmith trace collector which runs as an async, distributed process. Additionally, if LangSmith experiences an incident, your application performance will not be disrupted.

Will you train on the data that I send LangSmith?

We will not train on your data, and you own all rights to your data. See LangSmith Terms of Service for more information.

How much does LangSmith cost?

See our pricing page for more information, and find a plan that works for you.

Ready to start shipping 
reliable GenAI apps faster?

Get started with LangChain, LangSmith, and LangGraph to enhance your LLM app development, from prototype to production.