import { LLM, BaseLLMParams } from "./base.js"; export interface HFInput { /** Model to use */ model: string; /** Sampling temperature to use */ temperature?: number; /** * Maximum number of tokens to generate in the completion. */ maxTokens?: number; /** Total probability mass of tokens to consider at each step */ topP?: number; /** Integer to define the top tokens considered within the sample operation to create new text. */ topK?: number; /** Penalizes repeated tokens according to frequency */ frequencyPenalty?: number; /** API key to use. */ apiKey?: string; } export declare class HuggingFaceInference extends LLM implements HFInput { get lc_secrets(): { [key: string]: string; } | undefined; model: string; temperature: number | undefined; maxTokens: number | undefined; topP: number | undefined; topK: number | undefined; frequencyPenalty: number | undefined; apiKey: string | undefined; constructor(fields?: Partial & BaseLLMParams); _llmType(): string; /** @ignore */ _call(prompt: string, options: this["ParsedCallOptions"]): Promise; /** @ignore */ static imports(): Promise<{ HfInference: typeof import("@huggingface/inference").HfInference; }>; }