import { AppConfig, ChatConfig } from "./config"; import { Tokenizer } from "@mlc-ai/web-tokenizers"; export declare function hasModelInCache(modelId: string, appConfig?: AppConfig): Promise; export declare function deleteModelAllInfoInCache(modelId: string, appConfig?: AppConfig): Promise; export declare function deleteModelInCache(modelId: string, appConfig?: AppConfig): Promise; export declare function deleteChatConfigInCache(modelId: string, appConfig?: AppConfig): Promise; export declare function deleteModelWasmInCache(modelId: string, appConfig?: AppConfig): Promise; /** * * @param baseUrl The link to which we can find tokenizer files, usually is a `ModelRecord.model`. * @param config A ChatConfig, usually loaded from `mlc-chat-config.json` in `baseUrl`. * @param appConfig An AppConfig, usually `webllm.prebuiltAppConfig` if not defined by user. * @param logger Logging function, console.log by default. * @returns */ export declare function asyncLoadTokenizer(baseUrl: string, config: ChatConfig, appConfig: AppConfig, logger?: (msg: string) => void): Promise; //# sourceMappingURL=cache_util.d.ts.map