Alpha Notice: These docs cover the v1-alpha release. Content is incomplete and subject to change.For the latest stable version, see the v0 LangChain Python or LangChain JavaScript docs.
When a model proposes an action that might require review — for example, writing to a file or executing SQL — the middleware can pause execution and wait for a decision. It does this by checking each tool call against a configurable policy. If intervention is needed, the middleware issues an interrupt that halts execution. The graph state is saved using LangGraph’s persistence layer, so execution can pause safely and resume later. A human response then determines what happens next: the action can be accepted as-is (
accept
), modified before running (edit
), or rejected with feedback (respond
).
Interrupt response types
The middleware defines three built-in ways a human can respond to an interrupt:Response Type | Description | Example Use Case |
---|---|---|
✅ accept | The action is approved as-is and executed without changes. | Send an email draft exactly as written |
✏️ edit | The tool call is executed with modifications. | Change the recipient before sending an email |
❌ respond | The tool call is rejected, with an explanation added to the conversation. | Reject an email draft and explain how to rewrite it |
interrupt_on
.
When multiple tool calls are paused at the same time, each action requires a separate response.
Responses must be provided in the same order as the actions appear in the interrupt request.
When editing tool arguments, make changes conservatively. Significant modifications to the original arguments may cause the model to re-evaluate its approach and potentially execute the tool multiple times or take unexpected actions.
Configuring interrupts
To use HITL, add the middleware to the agent’smiddleware
list when creating the agent.
You configure it with a mapping of tool actions to the response types that are allowed for each action. The middleware will interrupt execution when a tool call matches an action in the mapping.
You must configure a checkpointer to persist the graph state across interrupts.
In production, use a persistent checkpointer like AsyncPostgresSaver. For testing or prototyping, use InMemorySaver.When invoking the agent, pass a
config
that includes the thread ID to associate execution with a conversation thread.
See the LangGraph human-in-the-loop documentation for details.Responding to interrupts
When you invoke the agent, it runs until it either completes or an interrupt is raised. An interrupt is triggered when a tool call matches the policy you configured ininterrupt_on
. In that case, the invocation result will include an __interrupt__
field with the actions that require review. You can then present those actions to a reviewer and resume execution once responses are provided.
Response types
Use
accept
to approve the tool call as-is and execute it without changes.Execution lifecycle
The middleware defines anafter_model
hook that runs after the model generates a response but before any tool calls are executed:
- The agent invokes the model to generate a response.
- The middleware inspects the response for tool calls.
- If any calls require human input, the middleware builds a list of
HumanInterrupt
objects and calls interrupt. - The agent waits for human responses.
- Based on responses, the middleware executes approved or edited calls, synthesizes @[ToolMessage]‘s for rejected calls, and resumes execution.
UI integration
The prebuiltHumanInTheLoopMiddleware
is designed to work out-of-the-box with LangChain provided UI applications like Agent ChatUI. The middleware’s interrupt messages include all the information needed to render a review interface, including tool names, arguments, and allowed response types.