import _m0 from 'protobufjs/minimal'; export declare const protobufPackage = "yandex.cloud.dataproc.v1"; /** A Yandex Data Processing job. For details about the concept, see [documentation](/docs/data-proc/concepts/jobs). */ export interface Job { /** ID of the job. Generated at creation time. */ id: string; /** ID of the Yandex Data Processing cluster that the job belongs to. */ clusterId: string; /** Creation timestamp. */ createdAt?: Date; /** The time when the job was started. */ startedAt?: Date; /** The time when the job was finished. */ finishedAt?: Date; /** Name of the job, specified in the [JobService.Create] request. */ name: string; /** The id of the user who created the job */ createdBy: string; /** Job status. */ status: Job_Status; /** Specification for a MapReduce job. */ mapreduceJob?: MapreduceJob | undefined; /** Specification for a Spark job. */ sparkJob?: SparkJob | undefined; /** Specification for a PySpark job. */ pysparkJob?: PysparkJob | undefined; /** Specification for a Hive job. */ hiveJob?: HiveJob | undefined; /** Attributes of YARN application. */ applicationInfo?: ApplicationInfo; } export declare enum Job_Status { STATUS_UNSPECIFIED = 0, /** PROVISIONING - Job is logged in the database and is waiting for the agent to run it. */ PROVISIONING = 1, /** PENDING - Job is acquired by the agent and is in the queue for execution. */ PENDING = 2, /** RUNNING - Job is being run in the cluster. */ RUNNING = 3, /** ERROR - Job failed to finish the run properly. */ ERROR = 4, /** DONE - Job is finished. */ DONE = 5, /** CANCELLED - Job is cancelled. */ CANCELLED = 6, /** CANCELLING - Job is waiting for cancellation. */ CANCELLING = 7, UNRECOGNIZED = -1 } export declare function job_StatusFromJSON(object: any): Job_Status; export declare function job_StatusToJSON(object: Job_Status): string; export interface ApplicationAttempt { /** ID of YARN application attempt */ id: string; /** ID of YARN Application Master container */ amContainerId: string; } export interface ApplicationInfo { /** ID of YARN application */ id: string; /** YARN application attempts */ applicationAttempts: ApplicationAttempt[]; } export interface MapreduceJob { /** Optional arguments to pass to the driver. */ args: string[]; /** JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task. */ jarFileUris: string[]; /** * URIs of resource files to be copied to the working directory of Yandex Data Processing drivers * and distributed Hadoop tasks. */ fileUris: string[]; /** URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks. */ archiveUris: string[]; /** Property names and values, used to configure Yandex Data Processing and MapReduce. */ properties: { [key: string]: string; }; /** HCFS URI of the .jar file containing the driver class. */ mainJarFileUri: string | undefined; /** The name of the driver class. */ mainClass: string | undefined; } export interface MapreduceJob_PropertiesEntry { key: string; value: string; } export interface SparkJob { /** Optional arguments to pass to the driver. */ args: string[]; /** JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task. */ jarFileUris: string[]; /** * URIs of resource files to be copied to the working directory of Yandex Data Processing drivers * and distributed Hadoop tasks. */ fileUris: string[]; /** URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks. */ archiveUris: string[]; /** Property names and values, used to configure Yandex Data Processing and Spark. */ properties: { [key: string]: string; }; /** The HCFS URI of the JAR file containing the `main` class for the job. */ mainJarFileUri: string; /** The name of the driver class. */ mainClass: string; /** List of maven coordinates of jars to include on the driver and executor classpaths. */ packages: string[]; /** List of additional remote repositories to search for the maven coordinates given with --packages. */ repositories: string[]; /** List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. */ excludePackages: string[]; } export interface SparkJob_PropertiesEntry { key: string; value: string; } export interface PysparkJob { /** Optional arguments to pass to the driver. */ args: string[]; /** JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task. */ jarFileUris: string[]; /** * URIs of resource files to be copied to the working directory of Yandex Data Processing drivers * and distributed Hadoop tasks. */ fileUris: string[]; /** URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks. */ archiveUris: string[]; /** Property names and values, used to configure Yandex Data Processing and PySpark. */ properties: { [key: string]: string; }; /** URI of the file with the driver code. Must be a .py file. */ mainPythonFileUri: string; /** URIs of Python files to pass to the PySpark framework. */ pythonFileUris: string[]; /** List of maven coordinates of jars to include on the driver and executor classpaths. */ packages: string[]; /** List of additional remote repositories to search for the maven coordinates given with --packages. */ repositories: string[]; /** List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. */ excludePackages: string[]; } export interface PysparkJob_PropertiesEntry { key: string; value: string; } export interface QueryList { /** List of Hive queries. */ queries: string[]; } export interface HiveJob { /** Property names and values, used to configure Yandex Data Processing and Hive. */ properties: { [key: string]: string; }; /** Flag indicating whether a job should continue to run if a query fails. */ continueOnFailure: boolean; /** Query variables and their values. */ scriptVariables: { [key: string]: string; }; /** JAR file URIs to add to CLASSPATH of the Hive driver and each task. */ jarFileUris: string[]; /** URI of the script with all the necessary Hive queries. */ queryFileUri: string | undefined; /** List of Hive queries to be used in the job. */ queryList?: QueryList | undefined; } export interface HiveJob_PropertiesEntry { key: string; value: string; } export interface HiveJob_ScriptVariablesEntry { key: string; value: string; } export declare const Job: { encode(message: Job, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): Job; fromJSON(object: any): Job; toJSON(message: Job): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; } & Record, never>) | undefined; sparkJob?: ({ args?: string[] | undefined; jarFileUris?: string[] | undefined; fileUris?: string[] | undefined; archiveUris?: string[] | undefined; properties?: { [x: string]: string | undefined; } | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: string[] | undefined; repositories?: string[] | undefined; excludePackages?: string[] | undefined; } & { args?: (string[] & string[] & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; pysparkJob?: ({ args?: string[] | undefined; jarFileUris?: string[] | undefined; fileUris?: string[] | undefined; archiveUris?: string[] | undefined; properties?: { [x: string]: string | undefined; } | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: string[] | undefined; packages?: string[] | undefined; repositories?: string[] | undefined; excludePackages?: string[] | undefined; } & { args?: (string[] & string[] & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: (string[] & string[] & Record, never>) | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; hiveJob?: ({ properties?: { [x: string]: string | undefined; } | undefined; continueOnFailure?: boolean | undefined; scriptVariables?: { [x: string]: string | undefined; } | undefined; jarFileUris?: string[] | undefined; queryFileUri?: string | undefined; queryList?: { queries?: string[] | undefined; } | undefined; } & { properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; continueOnFailure?: boolean | undefined; scriptVariables?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; queryFileUri?: string | undefined; queryList?: ({ queries?: string[] | undefined; } & { queries?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; applicationInfo?: ({ id?: string | undefined; applicationAttempts?: { id?: string | undefined; amContainerId?: string | undefined; }[] | undefined; } & { id?: string | undefined; applicationAttempts?: ({ id?: string | undefined; amContainerId?: string | undefined; }[] & ({ id?: string | undefined; amContainerId?: string | undefined; } & { id?: string | undefined; amContainerId?: string | undefined; } & Record, never>)[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): Job; }; export declare const ApplicationAttempt: { encode(message: ApplicationAttempt, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): ApplicationAttempt; fromJSON(object: any): ApplicationAttempt; toJSON(message: ApplicationAttempt): unknown; fromPartial, never>>(object: I): ApplicationAttempt; }; export declare const ApplicationInfo: { encode(message: ApplicationInfo, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): ApplicationInfo; fromJSON(object: any): ApplicationInfo; toJSON(message: ApplicationInfo): unknown; fromPartial, never>)[] & Record, never>) | undefined; } & Record, never>>(object: I): ApplicationInfo; }; export declare const MapreduceJob: { encode(message: MapreduceJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): MapreduceJob; fromJSON(object: any): MapreduceJob; toJSON(message: MapreduceJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; } & Record, never>>(object: I): MapreduceJob; }; export declare const MapreduceJob_PropertiesEntry: { encode(message: MapreduceJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): MapreduceJob_PropertiesEntry; fromJSON(object: any): MapreduceJob_PropertiesEntry; toJSON(message: MapreduceJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): MapreduceJob_PropertiesEntry; }; export declare const SparkJob: { encode(message: SparkJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SparkJob; fromJSON(object: any): SparkJob; toJSON(message: SparkJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>>(object: I): SparkJob; }; export declare const SparkJob_PropertiesEntry: { encode(message: SparkJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SparkJob_PropertiesEntry; fromJSON(object: any): SparkJob_PropertiesEntry; toJSON(message: SparkJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): SparkJob_PropertiesEntry; }; export declare const PysparkJob: { encode(message: PysparkJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PysparkJob; fromJSON(object: any): PysparkJob; toJSON(message: PysparkJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: (string[] & string[] & Record, never>) | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>>(object: I): PysparkJob; }; export declare const PysparkJob_PropertiesEntry: { encode(message: PysparkJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PysparkJob_PropertiesEntry; fromJSON(object: any): PysparkJob_PropertiesEntry; toJSON(message: PysparkJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): PysparkJob_PropertiesEntry; }; export declare const QueryList: { encode(message: QueryList, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): QueryList; fromJSON(object: any): QueryList; toJSON(message: QueryList): unknown; fromPartial, never>) | undefined; } & Record, never>>(object: I): QueryList; }; export declare const HiveJob: { encode(message: HiveJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): HiveJob; fromJSON(object: any): HiveJob; toJSON(message: HiveJob): unknown; fromPartial, never>) | undefined; continueOnFailure?: boolean | undefined; scriptVariables?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; queryFileUri?: string | undefined; queryList?: ({ queries?: string[] | undefined; } & { queries?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): HiveJob; }; export declare const HiveJob_PropertiesEntry: { encode(message: HiveJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): HiveJob_PropertiesEntry; fromJSON(object: any): HiveJob_PropertiesEntry; toJSON(message: HiveJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): HiveJob_PropertiesEntry; }; export declare const HiveJob_ScriptVariablesEntry: { encode(message: HiveJob_ScriptVariablesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): HiveJob_ScriptVariablesEntry; fromJSON(object: any): HiveJob_ScriptVariablesEntry; toJSON(message: HiveJob_ScriptVariablesEntry): unknown; fromPartial, never>>(object: I): HiveJob_ScriptVariablesEntry; }; type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined; export type DeepPartial = T extends Builtin ? T : T extends Array ? Array> : T extends ReadonlyArray ? ReadonlyArray> : T extends {} ? { [K in keyof T]?: DeepPartial; } : Partial; type KeysOfUnion = T extends T ? keyof T : never; export type Exact = P extends Builtin ? P : P & { [K in keyof P]: Exact; } & Record>, never>; export {};