import { BaseChannel } from "../channels/base.js"; import { OverwriteValue, START } from "../constants.js"; import { LastValue } from "../channels/last_value.js"; import { InteropZodToStateDefinition } from "../graph/zod/meta.js"; import { LangGraphRunnableConfig, Runtime } from "../pregel/runnable_types.js"; import { AnnotationRoot, SingleReducer, StateDefinition } from "../graph/annotation.js"; import { CompiledStateGraph } from "../graph/state.js"; import { Messages } from "../graph/messages_reducer.js"; import { MessagesAnnotation } from "../graph/messages_annotation.js"; import { ToolNode } from "./tool_node.js"; import { All, BaseCheckpointSaver, BaseStore } from "@langchain/langgraph-checkpoint"; import { Runnable, RunnableBinding, RunnableLike, RunnableSequence, RunnableToolLike } from "@langchain/core/runnables"; import * as _langchain_core_messages0 from "@langchain/core/messages"; import { BaseMessage, BaseMessageLike, SystemMessage } from "@langchain/core/messages"; import { InteropZodObject, InteropZodType } from "@langchain/core/utils/types"; import { DynamicTool, StructuredToolInterface } from "@langchain/core/tools"; import { LanguageModelLike } from "@langchain/core/language_models/base"; //#region src/prebuilt/react_agent_executor.d.ts /** * @deprecated `AgentState` has been moved to {@link https://www.npmjs.com/package/langchain langchain} package. * Update your import to `import { AgentState } from "langchain";` */ interface AgentState = Record> { messages: BaseMessage[]; structuredResponse: StructuredResponseType; } type N = typeof START | "agent" | "tools"; type StructuredResponseSchemaOptions = { schema: InteropZodType | Record; prompt?: string; strict?: boolean; [key: string]: unknown; }; type ServerTool = Record; type ClientTool = StructuredToolInterface | DynamicTool | RunnableToolLike; type Prompt = SystemMessage | string | ((state: typeof MessagesAnnotation.State, config: LangGraphRunnableConfig) => BaseMessageLike[]) | ((state: typeof MessagesAnnotation.State, config: LangGraphRunnableConfig) => Promise) | Runnable; /** @deprecated Use Prompt instead. */ type StateModifier = Prompt; /** @deprecated Use Prompt instead. */ type MessageModifier = SystemMessage | string | ((messages: BaseMessage[]) => BaseMessage[]) | ((messages: BaseMessage[]) => Promise) | Runnable; declare const createReactAgentAnnotation: = Record>() => AnnotationRoot<{ messages: BaseChannel, _langchain_core_messages0.MessageType>[], OverwriteValue, _langchain_core_messages0.MessageType>[]> | Messages, unknown>; structuredResponse: { (annotation: SingleReducer): BaseChannel, unknown>; (): LastValue; Root: (sd: S) => AnnotationRoot; }; }>; declare const PreHookAnnotation: AnnotationRoot<{ llmInputMessages: BaseChannel, _langchain_core_messages0.MessageType>[], OverwriteValue, _langchain_core_messages0.MessageType>[]> | Messages, unknown>; }>; type PreHookAnnotation = typeof PreHookAnnotation; type AnyAnnotationRoot = AnnotationRoot; type ToAnnotationRoot = A extends AnyAnnotationRoot ? A : A extends InteropZodObject ? AnnotationRoot> : never; /** * @deprecated `CreateReactAgentParams` has been moved to {@link https://www.npmjs.com/package/langchain langchain} package. * Update your import to `import { CreateAgentParams } from "langchain";` */ type CreateReactAgentParams = Record, C extends AnyAnnotationRoot | InteropZodObject = AnyAnnotationRoot> = { /** The chat model that can utilize OpenAI-style tool calling. */llm: LanguageModelLike | ((state: ToAnnotationRoot["State"] & PreHookAnnotation["State"], runtime: Runtime["State"]>) => Promise | LanguageModelLike); /** A list of tools or a ToolNode. */ tools: ToolNode | (ServerTool | ClientTool)[]; /** * @deprecated Use prompt instead. */ messageModifier?: MessageModifier; /** * @deprecated Use prompt instead. */ stateModifier?: StateModifier; /** * An optional prompt for the LLM. This takes full graph state BEFORE the LLM is called and prepares the input to LLM. * * Can take a few different forms: * * - str: This is converted to a SystemMessage and added to the beginning of the list of messages in state["messages"]. * - SystemMessage: this is added to the beginning of the list of messages in state["messages"]. * - Function: This function should take in full graph state and the output is then passed to the language model. * - Runnable: This runnable should take in full graph state and the output is then passed to the language model. * * Note: * Prior to `v0.2.46`, the prompt was set using `stateModifier` / `messagesModifier` parameters. * This is now deprecated and will be removed in a future release. */ prompt?: Prompt; /** * Additional state schema for the agent. */ stateSchema?: A; /** * An optional schema for the context. */ contextSchema?: C; /** An optional checkpoint saver to persist the agent's state. */ checkpointSaver?: BaseCheckpointSaver | boolean; /** An optional checkpoint saver to persist the agent's state. Alias of "checkpointSaver". */ checkpointer?: BaseCheckpointSaver | boolean; /** An optional list of node names to interrupt before running. */ interruptBefore?: N[] | All; /** An optional list of node names to interrupt after running. */ interruptAfter?: N[] | All; store?: BaseStore; /** * An optional schema for the final agent output. * * If provided, output will be formatted to match the given schema and returned in the 'structuredResponse' state key. * If not provided, `structuredResponse` will not be present in the output state. * * Can be passed in as: * - Zod schema * - JSON schema * - { prompt, schema }, where schema is one of the above. * The prompt will be used together with the model that is being used to generate the structured response. * * @remarks * **Important**: `responseFormat` requires the model to support `.withStructuredOutput()`. * * **Note**: The graph will make a separate call to the LLM to generate the structured response after the agent loop is finished. * This is not the only strategy to get structured responses, see more options in [this guide](https://langchain-ai.github.io/langgraph/how-tos/react-agent-structured-output/). */ responseFormat?: InteropZodType | StructuredResponseSchemaOptions | Record; /** * An optional name for the agent. */ name?: string; /** * An optional description for the agent. * This can be used to describe the agent to the underlying supervisor LLM. */ description?: string | undefined; /** * Use to specify how to expose the agent name to the underlying supervisor LLM. - undefined: Relies on the LLM provider {@link AIMessage#name}. Currently, only OpenAI supports this. - `"inline"`: Add the agent name directly into the content field of the {@link AIMessage} using XML-style tags. Example: `"How can I help you"` -> `"agent_nameHow can I help you?"` */ includeAgentName?: "inline" | undefined; /** * An optional node to add before the `agent` node (i.e., the node that calls the LLM). * Useful for managing long message histories (e.g., message trimming, summarization, etc.). */ preModelHook?: RunnableLike["State"] & PreHookAnnotation["State"], ToAnnotationRoot["Update"] & PreHookAnnotation["Update"], LangGraphRunnableConfig>; /** * An optional node to add after the `agent` node (i.e., the node that calls the LLM). * Useful for implementing human-in-the-loop, guardrails, validation, or other post-processing. */ postModelHook?: RunnableLike["State"], ToAnnotationRoot["Update"], LangGraphRunnableConfig>; /** * Determines the version of the graph to create. * * Can be one of * - `"v1"`: The tool node processes a single message. All tool calls in the message are * executed in parallel within the tool node. * - `"v2"`: The tool node processes a single tool call. Tool calls are distributed across * multiple instances of the tool node using the Send API. * * @default `"v1"` */ version?: "v1" | "v2"; }; /** * @deprecated `createReactAgent` has been moved to {@link https://www.npmjs.com/package/langchain langchain} package. * Update your import to `import { createAgent } from "langchain";` * * Creates a StateGraph agent that relies on a chat model utilizing tool calling. * * @example * ```ts * import { ChatOpenAI } from "@langchain/openai"; * import { tool } from "@langchain/core/tools"; * import { z } from "zod"; * import { createReactAgent } from "@langchain/langgraph/prebuilt"; * * const model = new ChatOpenAI({ * model: "gpt-4o", * }); * * const getWeather = tool((input) => { * if (["sf", "san francisco"].includes(input.location.toLowerCase())) { * return "It's 60 degrees and foggy."; * } else { * return "It's 90 degrees and sunny."; * } * }, { * name: "get_weather", * description: "Call to get the current weather.", * schema: z.object({ * location: z.string().describe("Location to get the weather for."), * }) * }) * * const agent = createReactAgent({ llm: model, tools: [getWeather] }); * * const inputs = { * messages: [{ role: "user", content: "what is the weather in SF?" }], * }; * * const stream = await agent.stream(inputs, { streamMode: "values" }); * * for await (const { messages } of stream) { * console.log(messages); * } * // Returns the messages in the state at each step of execution * ``` */ declare function createReactAgent = Record, C extends AnyAnnotationRoot | InteropZodObject = AnyAnnotationRoot>(params: CreateReactAgentParams): CompiledStateGraph["State"], ToAnnotationRoot["Update"], any, typeof MessagesAnnotation.spec & ToAnnotationRoot["spec"], ReturnType>["spec"] & ToAnnotationRoot["spec"]>; //#endregion export { AgentState, CreateReactAgentParams, createReactAgent, createReactAgentAnnotation }; //# sourceMappingURL=react_agent_executor.d.ts.map