LangSmith Unveils Agent Development Lifecycle: Build, Test, Deploy, Monitor
Written byMango
Drafted with AI; edited and reviewed by a human.
![]()
TL;DR
- LangSmith is introducing a structured four-part agent development lifecycle: Build, Test, Deploy, and Monitor.
- This lifecycle aims to make agent creation systematic and repeatable, moving beyond one-off demos.
- Open-source frameworks like LangChain, LangGraph, and Deep Agents are key to the "Build" phase.
- The LangSmith Platform provides essential tools for observability, evaluation, and deployment.
The development of intelligent agents, once a realm of complex, isolated projects, is now evolving into a more systematic and repeatable practice. LangSmith has introduced a comprehensive four-part agent development lifecycle—Build, Test, Deploy, and Monitor—designed to guide teams through the entire process, from initial creation to ongoing improvement. This structured approach emphasizes shipping early, learning from real-world usage, and iterating rapidly, transforming agent experimentation into a robust system for continuous delivery and enhancement.
The core principle behind this lifecycle is the understanding that effective agent development is not about creating a single, perfect agent, but about establishing a workflow that allows for consistent progress. The phases are intentionally ordered: teams are encouraged to Test agents thoroughly before Deployment, Monitor their performance in production, and then feed those learnings back into the next Build and evaluation cycle. For individual agents, this process can remain relatively simple. However, as organizations scale to manage multiple agents, it necessitates robust infrastructure and governance to control costs, manage tool access, inspect interactions, reuse context, and determine appropriate levels of human oversight.
The "Build" phase offers a spectrum of tooling to suit different needs and expertise levels. For those who prefer a code-first approach, the LangChain ecosystem provides powerful open-source frameworks. LangChain itself is ideal for composing model calls, tools, prompts, and retrieval strategies. For more complex agentic systems requiring branching, looping, and state persistence, LangGraph offers advanced control. Meanwhile, Deep Agents is designed for longer-running tasks, providing extensive capabilities for prompts, skills, and middleware. Beyond these, no-code and low-code solutions like LangSmith Fleet also empower a broader range of users to participate in agent development, bridging the gap between domain expertise and technical implementation.
Thorough "Testing" is paramount before any agent reaches production. This doesn't imply the need for a perfect evaluation suite from day one, but rather the establishment of sufficient evaluations to catch significant errors and compare versions without blindly deploying changes. Most evaluation workflows begin with a focused dataset of representative tasks, drawn from expected use cases, manual testing, or even prior production traces. Metrics are crucial here, whether measuring direct correctness for tasks with clear ground truth or relying on criteria-based evaluation for agents with multiple valid outcomes, such as writing responses or summarizing conversations. Experiments then leverage these datasets and metrics to compare different prompts, models, or orchestration patterns, revealing whether an agent is improving or regressing over time.
Summary
- LangSmith's new agent development lifecycle features four key stages: Build, Test, Deploy, and Monitor.
- This framework aims to standardize agent creation, turning experimentation into a repeatable process for iterative improvement.
- Developers can leverage open-source tools like LangChain, LangGraph, and Deep Agents in the Build phase.
- The LangSmith Platform is essential for observability, evaluation, and seamless deployment of agents.
Source: The Agent Development Lifecycle
Read next

Hermes Agent Brings Self-Evolving AI to NVIDIA RTX PCs & Qwen 3.6 Locally
Hermes Agent introduces self-improving AI capabilities, optimized for local deployment on NVIDIA RTX PCs and workstations, and enhanced by new Qwen 3.6 open-weight LLMs.
Continue reading