import { CreateChatCompletionRequest, ConfigurationParameters, CreateChatCompletionResponse, ChatCompletionFunctions, CreateChatCompletionRequestFunctionCall } from "openai"; import { AzureOpenAIInput, OpenAICallOptions, OpenAIChatInput } from "../types/openai-types.js"; import type { StreamingAxiosConfiguration } from "../util/axios-types.js"; import { BaseChatModel, BaseChatModelParams } from "./base.js"; import { BaseMessage, ChatGenerationChunk, ChatResult } from "../schema/index.js"; import { CallbackManagerForLLMRun } from "../callbacks/manager.js"; import { StructuredTool } from "../tools/base.js"; export { OpenAICallOptions, OpenAIChatInput, AzureOpenAIInput }; interface TokenUsage { completionTokens?: number; promptTokens?: number; totalTokens?: number; } interface OpenAILLMOutput { tokenUsage: TokenUsage; } export interface ChatOpenAICallOptions extends OpenAICallOptions { function_call?: CreateChatCompletionRequestFunctionCall; functions?: ChatCompletionFunctions[]; tools?: StructuredTool[]; promptIndex?: number; } /** * Wrapper around OpenAI large language models that use the Chat endpoint. * * To use you should have the `openai` package installed, with the * `OPENAI_API_KEY` environment variable set. * * To use with Azure you should have the `openai` package installed, with the * `AZURE_OPENAI_API_KEY`, * `AZURE_OPENAI_API_INSTANCE_NAME`, * `AZURE_OPENAI_API_DEPLOYMENT_NAME` * and `AZURE_OPENAI_API_VERSION` environment variable set. * `AZURE_OPENAI_BASE_PATH` is optional and will override `AZURE_OPENAI_API_INSTANCE_NAME` if you need to use a custom endpoint. * * @remarks * Any parameters that are valid to be passed to {@link * https://platform.openai.com/docs/api-reference/chat/create | * `openai.createChatCompletion`} can be passed through {@link modelKwargs}, even * if not explicitly available on this class. */ export declare class ChatOpenAI extends BaseChatModel implements OpenAIChatInput, AzureOpenAIInput { get callKeys(): (keyof ChatOpenAICallOptions)[]; lc_serializable: boolean; get lc_secrets(): { [key: string]: string; } | undefined; get lc_aliases(): Record; temperature: number; topP: number; frequencyPenalty: number; presencePenalty: number; n: number; logitBias?: Record; modelName: string; modelKwargs?: OpenAIChatInput["modelKwargs"]; stop?: string[]; timeout?: number; streaming: boolean; maxTokens?: number; openAIApiKey?: string; azureOpenAIApiVersion?: string; azureOpenAIApiKey?: string; azureOpenAIApiInstanceName?: string; azureOpenAIApiDeploymentName?: string; azureOpenAIBasePath?: string; private client; private clientConfig; constructor(fields?: Partial & Partial & BaseChatModelParams & { configuration?: ConfigurationParameters; }, /** @deprecated */ configuration?: ConfigurationParameters); /** * Get the parameters used to invoke the model */ invocationParams(options?: this["ParsedCallOptions"]): Omit; /** @ignore */ _identifyingParams(): { apiKey?: string | Promise | ((name: string) => string) | ((name: string) => Promise) | undefined; organization?: string | undefined; username?: string | undefined; password?: string | undefined; accessToken?: string | Promise | ((name?: string | undefined, scopes?: string[] | undefined) => string) | ((name?: string | undefined, scopes?: string[] | undefined) => Promise) | undefined; basePath?: string | undefined; baseOptions?: any; formDataCtor?: (new () => any) | undefined; function_call?: CreateChatCompletionRequestFunctionCall | undefined; stop?: import("openai").CreateChatCompletionRequestStop | undefined; stream?: boolean | null | undefined; user?: string | undefined; functions?: ChatCompletionFunctions[] | undefined; model: string; temperature?: number | null | undefined; top_p?: number | null | undefined; n?: number | null | undefined; max_tokens?: number | undefined; presence_penalty?: number | null | undefined; frequency_penalty?: number | null | undefined; logit_bias?: object | null | undefined; model_name: string; }; _streamResponseChunks(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): AsyncGenerator; startStream(request: CreateChatCompletionRequest, options?: StreamingAxiosConfiguration): { next(): Promise; [Symbol.asyncIterator](): any; }; /** * Get the identifying parameters for the model */ identifyingParams(): { apiKey?: string | Promise | ((name: string) => string) | ((name: string) => Promise) | undefined; organization?: string | undefined; username?: string | undefined; password?: string | undefined; accessToken?: string | Promise | ((name?: string | undefined, scopes?: string[] | undefined) => string) | ((name?: string | undefined, scopes?: string[] | undefined) => Promise) | undefined; basePath?: string | undefined; baseOptions?: any; formDataCtor?: (new () => any) | undefined; function_call?: CreateChatCompletionRequestFunctionCall | undefined; stop?: import("openai").CreateChatCompletionRequestStop | undefined; stream?: boolean | null | undefined; user?: string | undefined; functions?: ChatCompletionFunctions[] | undefined; model: string; temperature?: number | null | undefined; top_p?: number | null | undefined; n?: number | null | undefined; max_tokens?: number | undefined; presence_penalty?: number | null | undefined; frequency_penalty?: number | null | undefined; logit_bias?: object | null | undefined; model_name: string; }; /** @ignore */ _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise; getNumTokensFromMessages(messages: BaseMessage[]): Promise<{ totalCount: number; countPerMessage: number[]; }>; /** @ignore */ completionWithRetry(request: CreateChatCompletionRequest, options?: StreamingAxiosConfiguration): Promise; _llmType(): string; /** @ignore */ _combineLLMOutput(...llmOutputs: OpenAILLMOutput[]): OpenAILLMOutput; } export declare class PromptLayerChatOpenAI extends ChatOpenAI { promptLayerApiKey?: string; plTags?: string[]; returnPromptLayerId?: boolean; constructor(fields?: ConstructorParameters[0] & { promptLayerApiKey?: string; plTags?: string[]; returnPromptLayerId?: boolean; }); _generate(messages: BaseMessage[], options?: string[] | ChatOpenAICallOptions, runManager?: CallbackManagerForLLMRun): Promise; }