import _m0 from 'protobufjs/minimal'; export declare const protobufPackage = "yandex.cloud.dataproc.manager.v1"; /** Yandex Data Processing job. */ export interface Job { /** * Unique ID of the Yandex Data Processing job. * This ID is assigned by MDB in the process of creating Yandex Data Processing job. */ id: string; /** ID of the Yandex Data Processing cluster. */ clusterId: string; /** The time when the Yandex Data Processing job was created. */ createdAt?: Date; /** The time when the Yandex Data Processing job was started. */ startedAt?: Date; /** The time when the Yandex Data Processing job was finished. */ finishedAt?: Date; /** Name of the Yandex Data Processing job. */ name: string; /** The id of the user who created the job */ createdBy: string; /** Status. */ status: Job_Status; mapreduceJob?: MapreduceJob | undefined; sparkJob?: SparkJob | undefined; pysparkJob?: PysparkJob | undefined; hiveJob?: HiveJob | undefined; /** Attributes of YARN application. */ applicationInfo?: ApplicationInfo; } export declare enum Job_Status { STATUS_UNSPECIFIED = 0, /** PROVISIONING - Job created in metadb and is waiting agent to acquire. */ PROVISIONING = 1, /** PENDING - Job acquired by agent and is waiting for execution. */ PENDING = 2, /** RUNNING - Job is running. */ RUNNING = 3, /** ERROR - Job failed. */ ERROR = 4, /** DONE - Job finished. */ DONE = 5, /** CANCELLED - Job cancelled. */ CANCELLED = 6, /** CANCELLING - Job is waiting for cancellation. */ CANCELLING = 7, UNRECOGNIZED = -1 } export declare function job_StatusFromJSON(object: any): Job_Status; export declare function job_StatusToJSON(object: Job_Status): string; export interface ApplicationAttempt { /** ID of YARN application attempt */ id: string; /** ID of YARN Application Master container */ amContainerId: string; } export interface ApplicationInfo { /** ID of YARN application */ id: string; /** YARN application attempts */ applicationAttempts: ApplicationAttempt[]; } export interface MapreduceJob { /** Optional arguments to the driver. */ args: string[]; /** URIs of file to run. */ jarFileUris: string[]; /** URIs of files to be copied to the working directory of Yandex Data Processing drivers and distributed tasks. */ fileUris: string[]; /** URIs of archives to be extracted in the working directory of Yandex Data Processing drivers and tasks. */ archiveUris: string[]; /** A mapping of property names to values, used to configure Yandex Data Processing. */ properties: { [key: string]: string; }; /** The HCFS URI of the jar file containing the main class. */ mainJarFileUri: string | undefined; /** The name of the driver's main class. */ mainClass: string | undefined; } export interface MapreduceJob_PropertiesEntry { key: string; value: string; } export interface SparkJob { /** Optional arguments to the driver. */ args: string[]; /** Jar file URIs to add to the CLASSPATHs of the Yandex Data Processing driver and tasks. */ jarFileUris: string[]; /** URIs of files to be copied to the working directory of Yandex Data Processing drivers and distributed tasks. */ fileUris: string[]; /** URIs of archives to be extracted in the working directory of Yandex Data Processing drivers and tasks. */ archiveUris: string[]; /** A mapping of property names to values, used to configure Yandex Data Processing. */ properties: { [key: string]: string; }; /** The HCFS URI of the jar file containing the main class. */ mainJarFileUri: string; /** The name of the driver's main class. */ mainClass: string; /** List of maven coordinates of jars to include on the driver and executor classpaths. */ packages: string[]; /** List of additional remote repositories to search for the maven coordinates given with --packages. */ repositories: string[]; /** List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. */ excludePackages: string[]; } export interface SparkJob_PropertiesEntry { key: string; value: string; } export interface PysparkJob { /** Optional arguments to the driver. */ args: string[]; /** Jar file URIs to add to the CLASSPATHs of the Yandex Data Processing driver and tasks. */ jarFileUris: string[]; /** URIs of files to be copied to the working directory of Yandex Data Processing drivers and distributed tasks. */ fileUris: string[]; /** URIs of archives to be extracted in the working directory of Yandex Data Processing drivers and tasks. */ archiveUris: string[]; /** A mapping of property names to values, used to configure Yandex Data Processing. */ properties: { [key: string]: string; }; /** URI of the main Python file to use as the driver. Must be a .py file. */ mainPythonFileUri: string; /** URIs of Python files to pass to the PySpark framework. */ pythonFileUris: string[]; /** List of maven coordinates of jars to include on the driver and executor classpaths. */ packages: string[]; /** List of additional remote repositories to search for the maven coordinates given with --packages. */ repositories: string[]; /** List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. */ excludePackages: string[]; } export interface PysparkJob_PropertiesEntry { key: string; value: string; } export interface QueryList { queries: string[]; } export interface HiveJob { /** A mapping of property names to values, used to configure Hive. */ properties: { [key: string]: string; }; /** Whether to continue executing queries if a query fails. */ continueOnFailure: boolean; /** Mapping of query variable names to values. */ scriptVariables: { [key: string]: string; }; /** Jar file URIs to add to the CLASSPATHs of the Hive driver and tasks. */ jarFileUris: string[]; /** URI of the script that contains Hive queries. */ queryFileUri: string | undefined; queryList?: QueryList | undefined; } export interface HiveJob_PropertiesEntry { key: string; value: string; } export interface HiveJob_ScriptVariablesEntry { key: string; value: string; } /** Yandex Data Processing support job. */ export interface SupportJob { /** * Required. Unique ID of the Yandex Data Processing job. * This ID is assigned by MDB in the process of creating Yandex Data Processing job. */ id: string; /** Required. Unique ID of the Yandex Data Processing cluster. */ clusterId: string; /** The time when the Yandex Data Processing job was created. */ createdAt?: Date; /** The time when the Yandex Data Processing job was started. */ startedAt?: Date; /** The time when the Yandex Data Processing job was finished. */ finishedAt?: Date; /** Status. */ status: SupportJob_Status; /** Command. */ cmd: string; /** Execution timeout in seconds. */ timeout: number; /** The id of the user who created the job */ createdBy: string; } export declare enum SupportJob_Status { STATUS_UNSPECIFIED = 0, /** PROVISIONING - Job created in metadb and is waiting agent to acquire. */ PROVISIONING = 1, /** PENDING - Job acquired by agent and is waiting for execution. */ PENDING = 2, /** RUNNING - Job is running. */ RUNNING = 3, /** ERROR - Job failed. */ ERROR = 4, /** DONE - Job finished. */ DONE = 5, /** CANCELLED - Job cancelled. */ CANCELLED = 6, /** CANCELLING - Job is waiting for cancellation. */ CANCELLING = 7, UNRECOGNIZED = -1 } export declare function supportJob_StatusFromJSON(object: any): SupportJob_Status; export declare function supportJob_StatusToJSON(object: SupportJob_Status): string; export declare const Job: { encode(message: Job, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): Job; fromJSON(object: any): Job; toJSON(message: Job): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; } & Record, never>) | undefined; sparkJob?: ({ args?: string[] | undefined; jarFileUris?: string[] | undefined; fileUris?: string[] | undefined; archiveUris?: string[] | undefined; properties?: { [x: string]: string | undefined; } | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: string[] | undefined; repositories?: string[] | undefined; excludePackages?: string[] | undefined; } & { args?: (string[] & string[] & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; pysparkJob?: ({ args?: string[] | undefined; jarFileUris?: string[] | undefined; fileUris?: string[] | undefined; archiveUris?: string[] | undefined; properties?: { [x: string]: string | undefined; } | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: string[] | undefined; packages?: string[] | undefined; repositories?: string[] | undefined; excludePackages?: string[] | undefined; } & { args?: (string[] & string[] & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: (string[] & string[] & Record, never>) | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; hiveJob?: ({ properties?: { [x: string]: string | undefined; } | undefined; continueOnFailure?: boolean | undefined; scriptVariables?: { [x: string]: string | undefined; } | undefined; jarFileUris?: string[] | undefined; queryFileUri?: string | undefined; queryList?: { queries?: string[] | undefined; } | undefined; } & { properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; continueOnFailure?: boolean | undefined; scriptVariables?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; queryFileUri?: string | undefined; queryList?: ({ queries?: string[] | undefined; } & { queries?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; applicationInfo?: ({ id?: string | undefined; applicationAttempts?: { id?: string | undefined; amContainerId?: string | undefined; }[] | undefined; } & { id?: string | undefined; applicationAttempts?: ({ id?: string | undefined; amContainerId?: string | undefined; }[] & ({ id?: string | undefined; amContainerId?: string | undefined; } & { id?: string | undefined; amContainerId?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): Job; }; export declare const ApplicationAttempt: { encode(message: ApplicationAttempt, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): ApplicationAttempt; fromJSON(object: any): ApplicationAttempt; toJSON(message: ApplicationAttempt): unknown; fromPartial, never>>(object: I): ApplicationAttempt; }; export declare const ApplicationInfo: { encode(message: ApplicationInfo, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): ApplicationInfo; fromJSON(object: any): ApplicationInfo; toJSON(message: ApplicationInfo): unknown; fromPartial, never>)[] & Record, never>) | undefined; } & Record, never>>(object: I): ApplicationInfo; }; export declare const MapreduceJob: { encode(message: MapreduceJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): MapreduceJob; fromJSON(object: any): MapreduceJob; toJSON(message: MapreduceJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; } & Record, never>>(object: I): MapreduceJob; }; export declare const MapreduceJob_PropertiesEntry: { encode(message: MapreduceJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): MapreduceJob_PropertiesEntry; fromJSON(object: any): MapreduceJob_PropertiesEntry; toJSON(message: MapreduceJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): MapreduceJob_PropertiesEntry; }; export declare const SparkJob: { encode(message: SparkJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SparkJob; fromJSON(object: any): SparkJob; toJSON(message: SparkJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>>(object: I): SparkJob; }; export declare const SparkJob_PropertiesEntry: { encode(message: SparkJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SparkJob_PropertiesEntry; fromJSON(object: any): SparkJob_PropertiesEntry; toJSON(message: SparkJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): SparkJob_PropertiesEntry; }; export declare const PysparkJob: { encode(message: PysparkJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PysparkJob; fromJSON(object: any): PysparkJob; toJSON(message: PysparkJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: (string[] & string[] & Record, never>) | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>>(object: I): PysparkJob; }; export declare const PysparkJob_PropertiesEntry: { encode(message: PysparkJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PysparkJob_PropertiesEntry; fromJSON(object: any): PysparkJob_PropertiesEntry; toJSON(message: PysparkJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): PysparkJob_PropertiesEntry; }; export declare const QueryList: { encode(message: QueryList, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): QueryList; fromJSON(object: any): QueryList; toJSON(message: QueryList): unknown; fromPartial, never>) | undefined; } & Record, never>>(object: I): QueryList; }; export declare const HiveJob: { encode(message: HiveJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): HiveJob; fromJSON(object: any): HiveJob; toJSON(message: HiveJob): unknown; fromPartial, never>) | undefined; continueOnFailure?: boolean | undefined; scriptVariables?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; queryFileUri?: string | undefined; queryList?: ({ queries?: string[] | undefined; } & { queries?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): HiveJob; }; export declare const HiveJob_PropertiesEntry: { encode(message: HiveJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): HiveJob_PropertiesEntry; fromJSON(object: any): HiveJob_PropertiesEntry; toJSON(message: HiveJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): HiveJob_PropertiesEntry; }; export declare const HiveJob_ScriptVariablesEntry: { encode(message: HiveJob_ScriptVariablesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): HiveJob_ScriptVariablesEntry; fromJSON(object: any): HiveJob_ScriptVariablesEntry; toJSON(message: HiveJob_ScriptVariablesEntry): unknown; fromPartial, never>>(object: I): HiveJob_ScriptVariablesEntry; }; export declare const SupportJob: { encode(message: SupportJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SupportJob; fromJSON(object: any): SupportJob; toJSON(message: SupportJob): unknown; fromPartial, never>>(object: I): SupportJob; }; type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined; export type DeepPartial = T extends Builtin ? T : T extends Array ? Array> : T extends ReadonlyArray ? ReadonlyArray> : T extends {} ? { [K in keyof T]?: DeepPartial; } : Partial; type KeysOfUnion = T extends T ? keyof T : never; export type Exact = P extends Builtin ? P : P & { [K in keyof P]: Exact; } & Record>, never>; export {};