/** * @module teams-ai */ /** * Copyright (c) Microsoft Corporation. All rights reserved. * Licensed under the MIT License. */ import { TurnContext } from 'botbuilder'; import { Memory } from '../../MemoryFork'; import { PromptCompletionModel } from '../../models/PromptCompletionModel'; import { PromptResponse, PromptResponseStatus } from '../../types'; import { Message, PromptFunctions, PromptTemplate } from '../../prompts'; import { Tokenizer } from '../../tokenizers'; /** * A test model that can be used to test the prompt completion system. */ export declare class TestModel implements PromptCompletionModel { /** * * @param {PromptResponseStatus} status Optional. Status of the prompt response. Defaults to `success`. * @param {Message} response Optional. Response to the prompt. Defaults to `{ role: 'assistant', content: 'Hello World' }`. * @param {Error} error Optional. Error to return. Defaults to `undefined`. */ constructor(status?: PromptResponseStatus, response?: Message, error?: Error); /** * Status of the prompt response. */ status: PromptResponseStatus; /** * Response to the prompt. */ response: Message; /** * Error to return. */ error?: Error; /** * Completes a prompt. * @param {TurnContext} context Current turn context. * @param {Memory} memory An interface for accessing state values. * @param {PromptFunctions} functions Functions to use when rendering the prompt. * @param {Tokenizer} tokenizer Tokenizer to use when rendering the prompt. * @param {PromptTemplate} template Prompt template to complete. * @returns {Promise>} A `PromptResponse` with the status and message. */ completePrompt(context: TurnContext, memory: Memory, functions: PromptFunctions, tokenizer: Tokenizer, template: PromptTemplate): Promise>; } //# sourceMappingURL=TestModel.d.ts.map