import { AzureOpenAIChatInput, OpenAICoreRequestOptions } from "../../types.js"; import { ChatOpenAIResponses, ChatOpenAIResponsesCallOptions } from "../../chat_models/responses.js"; import { AzureChatOpenAIFields } from "./common.js"; import { LangSmithParams } from "@langchain/core/language_models/chat_models"; import { Serialized } from "@langchain/core/load/serializable"; //#region src/azure/chat_models/responses.d.ts declare class AzureChatOpenAIResponses extends ChatOpenAIResponses implements Partial { azureOpenAIApiVersion?: string; azureOpenAIApiKey?: string; azureADTokenProvider?: () => Promise; azureOpenAIApiInstanceName?: string; azureOpenAIApiDeploymentName?: string; azureOpenAIBasePath?: string; azureOpenAIEndpoint?: string; _llmType(): string; get lc_aliases(): Record; get lc_secrets(): { [key: string]: string; } | undefined; get lc_serializable_keys(): string[]; getLsParams(options: this["ParsedCallOptions"]): LangSmithParams; constructor(deploymentName: string, fields?: Omit); constructor(fields?: AzureChatOpenAIFields); _getClientOptions(options: OpenAICoreRequestOptions | undefined): OpenAICoreRequestOptions; toJSON(): Serialized; } //#endregion export { AzureChatOpenAIResponses }; //# sourceMappingURL=responses.d.ts.map