import { GenerationChunk } from "../../outputs.js"; import { CallbackManagerForLLMRun } from "../../callbacks/manager.js"; import { BaseLLMParams, LLM } from "../../language_models/llms.js"; //#region src/utils/testing/llms.d.ts declare class FakeLLM extends LLM { response?: string; thrownErrorString?: string; constructor(fields: { response?: string; thrownErrorString?: string; } & BaseLLMParams); _llmType(): string; _call(prompt: string, _options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise; } declare class FakeStreamingLLM extends LLM { sleep?: number; responses?: string[]; thrownErrorString?: string; constructor(fields: { sleep?: number; responses?: string[]; thrownErrorString?: string; } & BaseLLMParams); _llmType(): string; _call(prompt: string): Promise; _streamResponseChunks(input: string, _options?: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): AsyncGenerator; } //#endregion export { FakeLLM, FakeStreamingLLM }; //# sourceMappingURL=llms.d.ts.map