// Copyright (c) Microsoft Corporation. All rights reserved. // Licensed under the MIT License. // WebNN API currently does not have a TypeScript definition file. This file is a workaround with types generated from // WebNN API specification. // https://github.com/webmachinelearning/webnn/issues/677 /// import type {Tensor} from 'onnxruntime-common'; /* eslint-disable @typescript-eslint/naming-convention */ export declare namespace JSEP { type BackendType = unknown; type AllocFunction = (size: number) => number; type FreeFunction = (size: number) => number; type UploadFunction = (dataOffset: number, gpuDataId: number, size: number) => void; type DownloadFunction = (gpuDataId: number, dataOffset: number, size: number) => Promise; type CreateKernelFunction = (name: string, kernel: number, attribute: unknown) => void; type ReleaseKernelFunction = (kernel: number) => void; type RunFunction = (kernel: number, contextDataOffset: number, sessionHandle: number, errors: Array>) => number; type CaptureBeginFunction = () => void; type CaptureEndFunction = () => void; type ReplayFunction = () => void; export interface Module extends WebGpuModule, WebNnModule { /** * Mount the external data file to an internal map, which will be used during session initialization. * * @param externalDataFilePath - specify the relative path of the external data file. * @param externalDataFileData - specify the content data. */ mountExternalData(externalDataFilePath: string, externalDataFileData: Uint8Array): void; /** * Unmount all external data files from the internal map. */ unmountExternalData(): void; /** * This is the entry of JSEP initialization. This function is called once when initializing ONNX Runtime per * backend. This function initializes Asyncify support. If name is 'webgpu', also initializes WebGPU backend and * registers a few callbacks that will be called in C++ code. */ jsepInit(name: 'webgpu', initParams: [ backend: BackendType, alloc: AllocFunction, free: FreeFunction, upload: UploadFunction, download: DownloadFunction, createKernel: CreateKernelFunction, releaseKernel: ReleaseKernelFunction, run: RunFunction, captureBegin: CaptureBeginFunction, captureEnd: CaptureEndFunction, replay: ReplayFunction ]): void; jsepInit(name: 'webnn', initParams?: never): void; } export interface WebGpuModule { /** * [exported from wasm] Specify a kernel's output when running OpKernel::Compute(). * * @param context - specify the kernel context pointer. * @param index - specify the index of the output. * @param data - specify the pointer to encoded data of type and dims. */ _JsepOutput(context: number, index: number, data: number): number; /** * [exported from wasm] Get name of an operator node. * * @param kernel - specify the kernel pointer. * @returns the pointer to a C-style UTF8 encoded string representing the node name. */ _JsepGetNodeName(kernel: number): number; /** * [exported from pre-jsep.js] Register a user GPU buffer for usage of a session's input or output. * * @param sessionId - specify the session ID. * @param index - specify an integer to represent which input/output it is registering for. For input, it is the * input_index corresponding to the session's inputNames. For output, it is the inputCount + output_index * corresponding to the session's ouputNames. * @param buffer - specify the GPU buffer to register. * @param size - specify the original data size in byte. * @returns the GPU data ID for the registered GPU buffer. */ jsepRegisterBuffer: (sessionId: number, index: number, buffer: GPUBuffer, size: number) => number; /** * [exported from pre-jsep.js] Get the GPU buffer by GPU data ID. * * @param dataId - specify the GPU data ID * @returns the GPU buffer. */ jsepGetBuffer: (dataId: number) => GPUBuffer; /** * [exported from pre-jsep.js] Create a function to be used to create a GPU Tensor. * * @param gpuBuffer - specify the GPU buffer * @param size - specify the original data size in byte. * @param type - specify the tensor type. * @returns the generated downloader function. */ jsepCreateDownloader: (gpuBuffer: GPUBuffer, size: number, type: Tensor.GpuBufferDataTypes) => () => Promise; /** * [exported from pre-jsep.js] Called when InferenceSession.run started. This function will be called before * _OrtRun[WithBinding]() is called. * @param sessionId - specify the session ID. */ jsepOnRunStart: (sessionId: number) => void; /** * [exported from pre-jsep.js] Release a session. This function will be called before _OrtReleaseSession() is * called. * @param sessionId - specify the session ID. * @returns */ jsepOnReleaseSession: (sessionId: number) => void; } export interface WebNnModule { /** * Active MLContext used to create WebNN EP. */ currentContext: MLContext; } } export interface OrtInferenceAPIs { _OrtInit(numThreads: number, loggingLevel: number): number; _OrtGetLastError(errorCodeOffset: number, errorMessageOffset: number): void; _OrtCreateSession(dataOffset: number, dataLength: number, sessionOptionsHandle: number): Promise; _OrtReleaseSession(sessionHandle: number): void; _OrtGetInputOutputCount(sessionHandle: number, inputCountOffset: number, outputCountOffset: number): number; _OrtGetInputName(sessionHandle: number, index: number): number; _OrtGetOutputName(sessionHandle: number, index: number): number; _OrtFree(stringHandle: number): void; _OrtCreateTensor( dataType: number, dataOffset: number, dataLength: number, dimsOffset: number, dimsLength: number, dataLocation: number): number; _OrtGetTensorData(tensorHandle: number, dataType: number, dataOffset: number, dimsOffset: number, dimsLength: number): number; _OrtReleaseTensor(tensorHandle: number): void; _OrtCreateBinding(sessionHandle: number): number; _OrtBindInput(bindingHandle: number, nameOffset: number, tensorHandle: number): Promise; _OrtBindOutput(bindingHandle: number, nameOffset: number, tensorHandle: number, location: number): number; _OrtClearBoundOutputs(ioBindingHandle: number): void; _OrtReleaseBinding(ioBindingHandle: number): void; _OrtRunWithBinding( sessionHandle: number, ioBindingHandle: number, outputCount: number, outputsOffset: number, runOptionsHandle: number): Promise; _OrtRun( sessionHandle: number, inputNamesOffset: number, inputsOffset: number, inputCount: number, outputNamesOffset: number, outputCount: number, outputsOffset: number, runOptionsHandle: number): Promise; _OrtCreateSessionOptions( graphOptimizationLevel: number, enableCpuMemArena: boolean, enableMemPattern: boolean, executionMode: number, enableProfiling: boolean, profileFilePrefix: number, logId: number, logSeverityLevel: number, logVerbosityLevel: number, optimizedModelFilePath: number): number; _OrtAppendExecutionProvider(sessionOptionsHandle: number, name: number): number; _OrtAddFreeDimensionOverride(sessionOptionsHandle: number, name: number, dim: number): number; _OrtAddSessionConfigEntry(sessionOptionsHandle: number, configKey: number, configValue: number): number; _OrtReleaseSessionOptions(sessionOptionsHandle: number): void; _OrtCreateRunOptions(logSeverityLevel: number, logVerbosityLevel: number, terminate: boolean, tag: number): number; _OrtAddRunConfigEntry(runOptionsHandle: number, configKey: number, configValue: number): number; _OrtReleaseRunOptions(runOptionsHandle: number): void; _OrtEndProfiling(sessionHandle: number): number; } export interface OrtTrainingAPIs { _OrtTrainingLoadCheckpoint(dataOffset: number, dataLength: number): number; _OrtTrainingReleaseCheckpoint(checkpointHandle: number): void; _OrtTrainingCreateSession( sessionOptionsHandle: number, checkpointHandle: number, trainOffset: number, trainLength: number, evalOffset: number, evalLength: number, optimizerOffset: number, optimizerLength: number): number; _OrtTrainingLazyResetGrad(trainingHandle: number): number; _OrtTrainingRunTrainStep( trainingHandle: number, inputsOffset: number, inputCount: number, outputsOffset: number, outputCount: number, runOptionsHandle: number): number; _OrtTrainingOptimizerStep(trainingHandle: number, runOptionsHandle: number): number; _OrtTrainingEvalStep( trainingHandle: number, inputsOffset: number, inputCount: number, outputsOffset: number, outputCount: number, runOptionsHandle: number): number; _OrtTrainingGetParametersSize(trainingHandle: number, paramSizeT: number, trainableOnly: boolean): number; _OrtTrainingCopyParametersToBuffer( trainingHandle: number, parametersBuffer: number, parameterCount: number, trainableOnly: boolean): number; _OrtTrainingCopyParametersFromBuffer( trainingHandle: number, parametersBuffer: number, parameterCount: number, trainableOnly: boolean): number; _OrtTrainingGetModelInputOutputCount( trainingHandle: number, inputCount: number, outputCount: number, isEvalModel: boolean): number; _OrtTrainingGetModelInputOutputName(trainingHandle: number, index: number, isInput: boolean, isEvalModel: boolean): number; _OrtTrainingReleaseSession(trainingHandle: number): void; } /** * The interface of the WebAssembly module for ONNX Runtime, compiled from C++ source code by Emscripten. */ export interface OrtWasmModule extends EmscriptenModule, OrtInferenceAPIs, Partial, Partial { // #region emscripten functions stackSave(): number; stackRestore(stack: number): void; stackAlloc(size: number): number; UTF8ToString(offset: number, maxBytesToRead?: number): string; lengthBytesUTF8(str: string): number; stringToUTF8(str: string, offset: number, maxBytes: number): void; // #endregion // #region config numThreads?: number; // #endregion }