import type { BaseChatModelParams, LangSmithParams } from "@langchain/core/language_models/chat_models"; import { type OpenAIClient, type ChatOpenAICallOptions, type OpenAIChatInput, type OpenAICoreRequestOptions, ChatOpenAI } from "@langchain/openai"; type NovitaUnsupportedArgs = "frequencyPenalty" | "presencePenalty" | "logitBias" | "functions"; type NovitaUnsupportedCallOptions = "functions" | "function_call"; export interface ChatNovitaCallOptions extends Omit { response_format: { type: "json_object"; schema: Record; }; } export interface ChatNovitaInput extends Omit, BaseChatModelParams { /** * Novita API key * @default process.env.NOVITA_API_KEY */ novitaApiKey?: string; /** * API key alias * @default process.env.NOVITA_API_KEY */ apiKey?: string; } /** * Novita chat model implementation */ export declare class ChatNovitaAI extends ChatOpenAI { static lc_name(): string; _llmType(): string; get lc_secrets(): { [key: string]: string; } | undefined; lc_serializable: boolean; constructor(fields?: Partial> & BaseChatModelParams & { novitaApiKey?: string; apiKey?: string; }); getLsParams(options: this["ParsedCallOptions"]): LangSmithParams; toJSON(): import("@langchain/core/load/serializable").Serialized; completionWithRetry(request: OpenAIClient.Chat.ChatCompletionCreateParamsStreaming, options?: OpenAICoreRequestOptions): Promise>; completionWithRetry(request: OpenAIClient.Chat.ChatCompletionCreateParamsNonStreaming, options?: OpenAICoreRequestOptions): Promise; } export {};