/// import { ChannelCredentials, ChannelOptions, UntypedServiceImplementation, handleUnaryCall, handleBidiStreamingCall, Client, ClientUnaryCall, Metadata, CallOptions, ClientDuplexStream, ServiceError } from '@grpc/grpc-js'; import _m0 from 'protobufjs/minimal'; import { Duration } from '../../../../../google/protobuf/duration'; import { Operation } from '../../../../../yandex/cloud/operation/operation'; export declare const protobufPackage = "yandex.cloud.ai.stt.v2"; export interface LongRunningRecognitionRequest { config?: RecognitionConfig; audio?: RecognitionAudio; } export interface LongRunningRecognitionResponse { chunks: SpeechRecognitionResult[]; } export interface StreamingRecognitionRequest { config?: RecognitionConfig | undefined; audioContent: Buffer | undefined; } export interface StreamingRecognitionResponse { chunks: SpeechRecognitionChunk[]; } export interface RecognitionAudio { content: Buffer | undefined; uri: string | undefined; } export interface RecognitionConfig { specification?: RecognitionSpec; folderId: string; } export interface RecognitionSpec { audioEncoding: RecognitionSpec_AudioEncoding; /** 8000, 16000, 48000 only for pcm */ sampleRateHertz: number; /** code in BCP-47 */ languageCode: string; profanityFilter: boolean; model: string; /** * If set true, tentative hypotheses may be returned as they become available (final=false flag) * If false or omitted, only final=true result(s) are returned. * Makes sense only for StreamingRecognize requests. */ partialResults: boolean; singleUtterance: boolean; /** Used only for long running recognize. */ audioChannelCount: number; /** This mark allows disable normalization text */ rawResults: boolean; /** Rewrite text in literature style (default: false) */ literatureText: boolean; } export declare enum RecognitionSpec_AudioEncoding { AUDIO_ENCODING_UNSPECIFIED = 0, /** LINEAR16_PCM - 16-bit signed little-endian (Linear PCM) */ LINEAR16_PCM = 1, OGG_OPUS = 2, /** MP3 - transcription only */ MP3 = 3, UNRECOGNIZED = -1 } export declare function recognitionSpec_AudioEncodingFromJSON(object: any): RecognitionSpec_AudioEncoding; export declare function recognitionSpec_AudioEncodingToJSON(object: RecognitionSpec_AudioEncoding): string; export interface SpeechRecognitionChunk { alternatives: SpeechRecognitionAlternative[]; /** This flag shows that the received chunk contains a part of the recognized text that won't be changed. */ final: boolean; /** This flag shows that the received chunk is the end of an utterance. */ endOfUtterance: boolean; } export interface SpeechRecognitionResult { alternatives: SpeechRecognitionAlternative[]; channelTag: number; } export interface SpeechRecognitionAlternative { text: string; confidence: number; words: WordInfo[]; } export interface WordInfo { startTime?: Duration; endTime?: Duration; word: string; confidence: number; } export declare const LongRunningRecognitionRequest: { encode(message: LongRunningRecognitionRequest, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): LongRunningRecognitionRequest; fromJSON(object: any): LongRunningRecognitionRequest; toJSON(message: LongRunningRecognitionRequest): unknown; fromPartial, never>) | undefined; folderId?: string | undefined; } & Record, never>) | undefined; audio?: ({ content?: Buffer | undefined; uri?: string | undefined; } & { content?: Buffer | undefined; uri?: string | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): LongRunningRecognitionRequest; }; export declare const LongRunningRecognitionResponse: { encode(message: LongRunningRecognitionResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): LongRunningRecognitionResponse; fromJSON(object: any): LongRunningRecognitionResponse; toJSON(message: LongRunningRecognitionResponse): unknown; fromPartial, never>) | undefined; endTime?: ({ seconds?: number | undefined; nanos?: number | undefined; } & { seconds?: number | undefined; nanos?: number | undefined; } & Record, never>) | undefined; word?: string | undefined; confidence?: number | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>)[] & Record, never>) | undefined; channelTag?: number | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>>(object: I): LongRunningRecognitionResponse; }; export declare const StreamingRecognitionRequest: { encode(message: StreamingRecognitionRequest, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): StreamingRecognitionRequest; fromJSON(object: any): StreamingRecognitionRequest; toJSON(message: StreamingRecognitionRequest): unknown; fromPartial, never>) | undefined; folderId?: string | undefined; } & Record, never>) | undefined; audioContent?: Buffer | undefined; } & Record, never>>(object: I): StreamingRecognitionRequest; }; export declare const StreamingRecognitionResponse: { encode(message: StreamingRecognitionResponse, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): StreamingRecognitionResponse; fromJSON(object: any): StreamingRecognitionResponse; toJSON(message: StreamingRecognitionResponse): unknown; fromPartial, never>) | undefined; endTime?: ({ seconds?: number | undefined; nanos?: number | undefined; } & { seconds?: number | undefined; nanos?: number | undefined; } & Record, never>) | undefined; word?: string | undefined; confidence?: number | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>)[] & Record, never>) | undefined; final?: boolean | undefined; endOfUtterance?: boolean | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>>(object: I): StreamingRecognitionResponse; }; export declare const RecognitionAudio: { encode(message: RecognitionAudio, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): RecognitionAudio; fromJSON(object: any): RecognitionAudio; toJSON(message: RecognitionAudio): unknown; fromPartial, never>>(object: I): RecognitionAudio; }; export declare const RecognitionConfig: { encode(message: RecognitionConfig, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): RecognitionConfig; fromJSON(object: any): RecognitionConfig; toJSON(message: RecognitionConfig): unknown; fromPartial, never>) | undefined; folderId?: string | undefined; } & Record, never>>(object: I): RecognitionConfig; }; export declare const RecognitionSpec: { encode(message: RecognitionSpec, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): RecognitionSpec; fromJSON(object: any): RecognitionSpec; toJSON(message: RecognitionSpec): unknown; fromPartial, never>>(object: I): RecognitionSpec; }; export declare const SpeechRecognitionChunk: { encode(message: SpeechRecognitionChunk, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SpeechRecognitionChunk; fromJSON(object: any): SpeechRecognitionChunk; toJSON(message: SpeechRecognitionChunk): unknown; fromPartial, never>) | undefined; endTime?: ({ seconds?: number | undefined; nanos?: number | undefined; } & { seconds?: number | undefined; nanos?: number | undefined; } & Record, never>) | undefined; word?: string | undefined; confidence?: number | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>)[] & Record, never>) | undefined; final?: boolean | undefined; endOfUtterance?: boolean | undefined; } & Record, never>>(object: I): SpeechRecognitionChunk; }; export declare const SpeechRecognitionResult: { encode(message: SpeechRecognitionResult, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SpeechRecognitionResult; fromJSON(object: any): SpeechRecognitionResult; toJSON(message: SpeechRecognitionResult): unknown; fromPartial, never>) | undefined; endTime?: ({ seconds?: number | undefined; nanos?: number | undefined; } & { seconds?: number | undefined; nanos?: number | undefined; } & Record, never>) | undefined; word?: string | undefined; confidence?: number | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>)[] & Record, never>) | undefined; channelTag?: number | undefined; } & Record, never>>(object: I): SpeechRecognitionResult; }; export declare const SpeechRecognitionAlternative: { encode(message: SpeechRecognitionAlternative, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SpeechRecognitionAlternative; fromJSON(object: any): SpeechRecognitionAlternative; toJSON(message: SpeechRecognitionAlternative): unknown; fromPartial, never>) | undefined; endTime?: ({ seconds?: number | undefined; nanos?: number | undefined; } & { seconds?: number | undefined; nanos?: number | undefined; } & Record, never>) | undefined; word?: string | undefined; confidence?: number | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>>(object: I): SpeechRecognitionAlternative; }; export declare const WordInfo: { encode(message: WordInfo, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): WordInfo; fromJSON(object: any): WordInfo; toJSON(message: WordInfo): unknown; fromPartial, never>) | undefined; endTime?: ({ seconds?: number | undefined; nanos?: number | undefined; } & { seconds?: number | undefined; nanos?: number | undefined; } & Record, never>) | undefined; word?: string | undefined; confidence?: number | undefined; } & Record, never>>(object: I): WordInfo; }; export declare const SttServiceService: { readonly longRunningRecognize: { readonly path: "/yandex.cloud.ai.stt.v2.SttService/LongRunningRecognize"; readonly requestStream: false; readonly responseStream: false; readonly requestSerialize: (value: LongRunningRecognitionRequest) => Buffer; readonly requestDeserialize: (value: Buffer) => LongRunningRecognitionRequest; readonly responseSerialize: (value: Operation) => Buffer; readonly responseDeserialize: (value: Buffer) => Operation; }; readonly streamingRecognize: { readonly path: "/yandex.cloud.ai.stt.v2.SttService/StreamingRecognize"; readonly requestStream: true; readonly responseStream: true; readonly requestSerialize: (value: StreamingRecognitionRequest) => Buffer; readonly requestDeserialize: (value: Buffer) => StreamingRecognitionRequest; readonly responseSerialize: (value: StreamingRecognitionResponse) => Buffer; readonly responseDeserialize: (value: Buffer) => StreamingRecognitionResponse; }; }; export interface SttServiceServer extends UntypedServiceImplementation { longRunningRecognize: handleUnaryCall; streamingRecognize: handleBidiStreamingCall; } export interface SttServiceClient extends Client { longRunningRecognize(request: LongRunningRecognitionRequest, callback: (error: ServiceError | null, response: Operation) => void): ClientUnaryCall; longRunningRecognize(request: LongRunningRecognitionRequest, metadata: Metadata, callback: (error: ServiceError | null, response: Operation) => void): ClientUnaryCall; longRunningRecognize(request: LongRunningRecognitionRequest, metadata: Metadata, options: Partial, callback: (error: ServiceError | null, response: Operation) => void): ClientUnaryCall; streamingRecognize(): ClientDuplexStream; streamingRecognize(options: Partial): ClientDuplexStream; streamingRecognize(metadata: Metadata, options?: Partial): ClientDuplexStream; } export declare const SttServiceClient: { new (address: string, credentials: ChannelCredentials, options?: Partial): SttServiceClient; service: typeof SttServiceService; }; type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined; export type DeepPartial = T extends Builtin ? T : T extends Array ? Array> : T extends ReadonlyArray ? ReadonlyArray> : T extends {} ? { [K in keyof T]?: DeepPartial; } : Partial; type KeysOfUnion = T extends T ? keyof T : never; export type Exact = P extends Builtin ? P : P & { [K in keyof P]: Exact; } & Record>, never>; export {};
= P extends Builtin ? P : P & { [K in keyof P]: Exact
; } & Record>, never>; export {};