/// import { ChannelCredentials, ChannelOptions, UntypedServiceImplementation, handleUnaryCall, Client, ClientUnaryCall, Metadata, CallOptions, ServiceError } from '@grpc/grpc-js'; import _m0 from 'protobufjs/minimal'; import { BatchInferenceTask_Status, BatchInferenceTask } from '../../../../../yandex/cloud/ai/batch_inference/v1/batch_inference_task'; export declare const protobufPackage = "yandex.cloud.ai.batch_inference.v1"; export interface BatchInferenceMetadata { taskId: string; taskStatus: BatchInferenceTask_Status; completedBatches: number; totalBatches: number; } export interface BatchInferenceResponse { task?: BatchInferenceTask; } export interface DescribeBatchInferenceRequest { /** Required task id */ taskId: string; } export interface DescribeBatchInferenceResponse { task?: BatchInferenceTask; } export interface ListBatchInferencesRequest { /** Folder ID for which the list of tasks will be provided. */ folderId: string; pageSize: number; pageToken: string; /** Batch inference status for filtering */ status: BatchInferenceTask_Status; } export interface ListBatchInferencesResponse { tasks: BatchInferenceTask[]; nextPageToken: string; } export interface CancelBatchInferenceRequest { /** Task ID that should be canceled. */ taskId: string; } export interface CancelBatchInferenceResponse { taskId: string; } export interface DeleteBatchInferenceRequest { /** Task ID that should be deleted. */ taskId: string; } export interface DeleteBatchInferenceResponse { taskId: string; } export declare const BatchInferenceMetadata: { encode(message: BatchInferenceMetadata, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): BatchInferenceMetadata; fromJSON(object: any): BatchInferenceMetadata; toJSON(message: BatchInferenceMetadata): unknown; fromPartial, never>>(object: I): BatchInferenceMetadata; }; export declare const BatchInferenceResponse: { encode(message: BatchInferenceResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): BatchInferenceResponse; fromJSON(object: any): BatchInferenceResponse; toJSON(message: BatchInferenceResponse): unknown; fromPartial, never>) | undefined; } & Record, never>) | undefined; dataLoggingEnabled?: boolean | undefined; jsonObject?: boolean | undefined; jsonSchema?: ({ schema?: { [x: string]: any; } | undefined; } & { schema?: ({ [x: string]: any; } & { [x: string]: any; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; status?: BatchInferenceTask_Status | undefined; resultDatasetId?: string | undefined; labels?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; createdBy?: string | undefined; createdAt?: Date | undefined; startedAt?: Date | undefined; finishedAt?: Date | undefined; errors?: ({ status?: { code?: number | undefined; message?: string | undefined; details?: { typeUrl?: string | undefined; value?: Buffer | undefined; }[] | undefined; } | undefined; lineErrors?: { lineNumber?: number | undefined; message?: string | undefined; }[] | undefined; batchErrors?: { batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; }[] | undefined; } & { status?: ({ code?: number | undefined; message?: string | undefined; details?: { typeUrl?: string | undefined; value?: Buffer | undefined; }[] | undefined; } & { code?: number | undefined; message?: string | undefined; details?: ({ typeUrl?: string | undefined; value?: Buffer | undefined; }[] & ({ typeUrl?: string | undefined; value?: Buffer | undefined; } & { typeUrl?: string | undefined; value?: Buffer | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; lineErrors?: ({ lineNumber?: number | undefined; message?: string | undefined; }[] & ({ lineNumber?: number | undefined; message?: string | undefined; } & { lineNumber?: number | undefined; message?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; batchErrors?: ({ batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; }[] & ({ batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; } & { batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): BatchInferenceResponse; }; export declare const DescribeBatchInferenceRequest: { encode(message: DescribeBatchInferenceRequest, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): DescribeBatchInferenceRequest; fromJSON(object: any): DescribeBatchInferenceRequest; toJSON(message: DescribeBatchInferenceRequest): unknown; fromPartial, never>>(object: I): DescribeBatchInferenceRequest; }; export declare const DescribeBatchInferenceResponse: { encode(message: DescribeBatchInferenceResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): DescribeBatchInferenceResponse; fromJSON(object: any): DescribeBatchInferenceResponse; toJSON(message: DescribeBatchInferenceResponse): unknown; fromPartial, never>) | undefined; } & Record, never>) | undefined; dataLoggingEnabled?: boolean | undefined; jsonObject?: boolean | undefined; jsonSchema?: ({ schema?: { [x: string]: any; } | undefined; } & { schema?: ({ [x: string]: any; } & { [x: string]: any; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; status?: BatchInferenceTask_Status | undefined; resultDatasetId?: string | undefined; labels?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; createdBy?: string | undefined; createdAt?: Date | undefined; startedAt?: Date | undefined; finishedAt?: Date | undefined; errors?: ({ status?: { code?: number | undefined; message?: string | undefined; details?: { typeUrl?: string | undefined; value?: Buffer | undefined; }[] | undefined; } | undefined; lineErrors?: { lineNumber?: number | undefined; message?: string | undefined; }[] | undefined; batchErrors?: { batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; }[] | undefined; } & { status?: ({ code?: number | undefined; message?: string | undefined; details?: { typeUrl?: string | undefined; value?: Buffer | undefined; }[] | undefined; } & { code?: number | undefined; message?: string | undefined; details?: ({ typeUrl?: string | undefined; value?: Buffer | undefined; }[] & ({ typeUrl?: string | undefined; value?: Buffer | undefined; } & { typeUrl?: string | undefined; value?: Buffer | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; lineErrors?: ({ lineNumber?: number | undefined; message?: string | undefined; }[] & ({ lineNumber?: number | undefined; message?: string | undefined; } & { lineNumber?: number | undefined; message?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; batchErrors?: ({ batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; }[] & ({ batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; } & { batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): DescribeBatchInferenceResponse; }; export declare const ListBatchInferencesRequest: { encode(message: ListBatchInferencesRequest, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): ListBatchInferencesRequest; fromJSON(object: any): ListBatchInferencesRequest; toJSON(message: ListBatchInferencesRequest): unknown; fromPartial, never>>(object: I): ListBatchInferencesRequest; }; export declare const ListBatchInferencesResponse: { encode(message: ListBatchInferencesResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): ListBatchInferencesResponse; fromJSON(object: any): ListBatchInferencesResponse; toJSON(message: ListBatchInferencesResponse): unknown; fromPartial, never>) | undefined; } & Record, never>) | undefined; dataLoggingEnabled?: boolean | undefined; jsonObject?: boolean | undefined; jsonSchema?: ({ schema?: { [x: string]: any; } | undefined; } & { schema?: ({ [x: string]: any; } & { [x: string]: any; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; status?: BatchInferenceTask_Status | undefined; resultDatasetId?: string | undefined; labels?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; createdBy?: string | undefined; createdAt?: Date | undefined; startedAt?: Date | undefined; finishedAt?: Date | undefined; errors?: ({ status?: { code?: number | undefined; message?: string | undefined; details?: { typeUrl?: string | undefined; value?: Buffer | undefined; }[] | undefined; } | undefined; lineErrors?: { lineNumber?: number | undefined; message?: string | undefined; }[] | undefined; batchErrors?: { batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; }[] | undefined; } & { status?: ({ code?: number | undefined; message?: string | undefined; details?: { typeUrl?: string | undefined; value?: Buffer | undefined; }[] | undefined; } & { code?: number | undefined; message?: string | undefined; details?: ({ typeUrl?: string | undefined; value?: Buffer | undefined; }[] & ({ typeUrl?: string | undefined; value?: Buffer | undefined; } & { typeUrl?: string | undefined; value?: Buffer | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; lineErrors?: ({ lineNumber?: number | undefined; message?: string | undefined; }[] & ({ lineNumber?: number | undefined; message?: string | undefined; } & { lineNumber?: number | undefined; message?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; batchErrors?: ({ batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; }[] & ({ batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; } & { batchNumber?: number | undefined; firstLine?: number | undefined; lastLine?: number | undefined; message?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>)[] & Record, never>) | undefined; nextPageToken?: string | undefined; } & Record, never>>(object: I): ListBatchInferencesResponse; }; export declare const CancelBatchInferenceRequest: { encode(message: CancelBatchInferenceRequest, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): CancelBatchInferenceRequest; fromJSON(object: any): CancelBatchInferenceRequest; toJSON(message: CancelBatchInferenceRequest): unknown; fromPartial, never>>(object: I): CancelBatchInferenceRequest; }; export declare const CancelBatchInferenceResponse: { encode(message: CancelBatchInferenceResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): CancelBatchInferenceResponse; fromJSON(object: any): CancelBatchInferenceResponse; toJSON(message: CancelBatchInferenceResponse): unknown; fromPartial, never>>(object: I): CancelBatchInferenceResponse; }; export declare const DeleteBatchInferenceRequest: { encode(message: DeleteBatchInferenceRequest, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): DeleteBatchInferenceRequest; fromJSON(object: any): DeleteBatchInferenceRequest; toJSON(message: DeleteBatchInferenceRequest): unknown; fromPartial, never>>(object: I): DeleteBatchInferenceRequest; }; export declare const DeleteBatchInferenceResponse: { encode(message: DeleteBatchInferenceResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): DeleteBatchInferenceResponse; fromJSON(object: any): DeleteBatchInferenceResponse; toJSON(message: DeleteBatchInferenceResponse): unknown; fromPartial, never>>(object: I): DeleteBatchInferenceResponse; }; export declare const BatchInferenceServiceService: { /** Describes a concrete task */ readonly describe: { readonly path: "/yandex.cloud.ai.batch_inference.v1.BatchInferenceService/Describe"; readonly requestStream: false; readonly responseStream: false; readonly requestSerialize: (value: DescribeBatchInferenceRequest) => Buffer; readonly requestDeserialize: (value: Buffer) => DescribeBatchInferenceRequest; readonly responseSerialize: (value: DescribeBatchInferenceResponse) => Buffer; readonly responseDeserialize: (value: Buffer) => DescribeBatchInferenceResponse; }; /** Lists tasks in folder */ readonly list: { readonly path: "/yandex.cloud.ai.batch_inference.v1.BatchInferenceService/List"; readonly requestStream: false; readonly responseStream: false; readonly requestSerialize: (value: ListBatchInferencesRequest) => Buffer; readonly requestDeserialize: (value: Buffer) => ListBatchInferencesRequest; readonly responseSerialize: (value: ListBatchInferencesResponse) => Buffer; readonly responseDeserialize: (value: Buffer) => ListBatchInferencesResponse; }; /** Cancels a concrete task */ readonly cancel: { readonly path: "/yandex.cloud.ai.batch_inference.v1.BatchInferenceService/Cancel"; readonly requestStream: false; readonly responseStream: false; readonly requestSerialize: (value: CancelBatchInferenceRequest) => Buffer; readonly requestDeserialize: (value: Buffer) => CancelBatchInferenceRequest; readonly responseSerialize: (value: CancelBatchInferenceResponse) => Buffer; readonly responseDeserialize: (value: Buffer) => CancelBatchInferenceResponse; }; /** Deletes a concrete task */ readonly delete: { readonly path: "/yandex.cloud.ai.batch_inference.v1.BatchInferenceService/Delete"; readonly requestStream: false; readonly responseStream: false; readonly requestSerialize: (value: DeleteBatchInferenceRequest) => Buffer; readonly requestDeserialize: (value: Buffer) => DeleteBatchInferenceRequest; readonly responseSerialize: (value: DeleteBatchInferenceResponse) => Buffer; readonly responseDeserialize: (value: Buffer) => DeleteBatchInferenceResponse; }; }; export interface BatchInferenceServiceServer extends UntypedServiceImplementation { /** Describes a concrete task */ describe: handleUnaryCall; /** Lists tasks in folder */ list: handleUnaryCall; /** Cancels a concrete task */ cancel: handleUnaryCall; /** Deletes a concrete task */ delete: handleUnaryCall; } export interface BatchInferenceServiceClient extends Client { /** Describes a concrete task */ describe(request: DescribeBatchInferenceRequest, callback: (error: ServiceError | null, response: DescribeBatchInferenceResponse) => void): ClientUnaryCall; describe(request: DescribeBatchInferenceRequest, metadata: Metadata, callback: (error: ServiceError | null, response: DescribeBatchInferenceResponse) => void): ClientUnaryCall; describe(request: DescribeBatchInferenceRequest, metadata: Metadata, options: Partial, callback: (error: ServiceError | null, response: DescribeBatchInferenceResponse) => void): ClientUnaryCall; /** Lists tasks in folder */ list(request: ListBatchInferencesRequest, callback: (error: ServiceError | null, response: ListBatchInferencesResponse) => void): ClientUnaryCall; list(request: ListBatchInferencesRequest, metadata: Metadata, callback: (error: ServiceError | null, response: ListBatchInferencesResponse) => void): ClientUnaryCall; list(request: ListBatchInferencesRequest, metadata: Metadata, options: Partial, callback: (error: ServiceError | null, response: ListBatchInferencesResponse) => void): ClientUnaryCall; /** Cancels a concrete task */ cancel(request: CancelBatchInferenceRequest, callback: (error: ServiceError | null, response: CancelBatchInferenceResponse) => void): ClientUnaryCall; cancel(request: CancelBatchInferenceRequest, metadata: Metadata, callback: (error: ServiceError | null, response: CancelBatchInferenceResponse) => void): ClientUnaryCall; cancel(request: CancelBatchInferenceRequest, metadata: Metadata, options: Partial, callback: (error: ServiceError | null, response: CancelBatchInferenceResponse) => void): ClientUnaryCall; /** Deletes a concrete task */ delete(request: DeleteBatchInferenceRequest, callback: (error: ServiceError | null, response: DeleteBatchInferenceResponse) => void): ClientUnaryCall; delete(request: DeleteBatchInferenceRequest, metadata: Metadata, callback: (error: ServiceError | null, response: DeleteBatchInferenceResponse) => void): ClientUnaryCall; delete(request: DeleteBatchInferenceRequest, metadata: Metadata, options: Partial, callback: (error: ServiceError | null, response: DeleteBatchInferenceResponse) => void): ClientUnaryCall; } export declare const BatchInferenceServiceClient: { new (address: string, credentials: ChannelCredentials, options?: Partial): BatchInferenceServiceClient; service: typeof BatchInferenceServiceService; }; type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined; export type DeepPartial = T extends Builtin ? T : T extends Array ? Array> : T extends ReadonlyArray ? ReadonlyArray> : T extends {} ? { [K in keyof T]?: DeepPartial; } : Partial; type KeysOfUnion = T extends T ? keyof T : never; export type Exact = P extends Builtin ? P : P & { [K in keyof P]: Exact; } & Record>, never>; export {};