import _m0 from 'protobufjs/minimal'; export declare const protobufPackage = "yandex.cloud.spark.v1"; /** Spark job. */ export interface Job { /** * Required. Unique ID of the Spark job. * This ID is assigned by MDB in the process of creating Spark job. */ id: string; /** Required. Unique ID of the Spark cluster. */ clusterId: string; /** The time when the Spark job was created. */ createdAt?: Date; /** The time when the Spark job was started. */ startedAt?: Date; /** The time when the Spark job was finished. */ finishedAt?: Date; /** Name of the Spark job. */ name: string; /** The id of the user who created the job */ createdBy: string; /** Status. */ status: Job_Status; sparkJob?: SparkJob | undefined; pysparkJob?: PysparkJob | undefined; /** Spark UI Url. */ uiUrl: string; } export declare enum Job_Status { STATUS_UNSPECIFIED = 0, /** PROVISIONING - Job created and is waiting to acquire. */ PROVISIONING = 1, /** PENDING - Job acquired and is waiting for execution. */ PENDING = 2, /** RUNNING - Job is running. */ RUNNING = 3, /** ERROR - Job failed. */ ERROR = 4, /** DONE - Job finished. */ DONE = 5, /** CANCELLED - Job cancelled. */ CANCELLED = 6, /** CANCELLING - Job is waiting for cancellation. */ CANCELLING = 7, UNRECOGNIZED = -1 } export declare function job_StatusFromJSON(object: any): Job_Status; export declare function job_StatusToJSON(object: Job_Status): string; export interface SparkJob { /** Optional arguments to pass to the driver. */ args: string[]; /** Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks. */ jarFileUris: string[]; /** URIs of files to be copied to the working directory of Spark drivers and distributed tasks. */ fileUris: string[]; /** URIs of archives to be extracted in the working directory of Spark drivers and tasks. */ archiveUris: string[]; /** A mapping of property names to values, used to configure Spark. */ properties: { [key: string]: string; }; /** URI of the jar file containing the main class. */ mainJarFileUri: string; /** The name of the driver's main class. */ mainClass: string; /** List of maven coordinates of jars to include on the driver and executor classpaths. */ packages: string[]; /** List of additional remote repositories to search for the maven coordinates given with --packages. */ repositories: string[]; /** List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. */ excludePackages: string[]; } export interface SparkJob_PropertiesEntry { key: string; value: string; } export interface PysparkJob { /** Optional arguments to pass to the driver. */ args: string[]; /** Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks. */ jarFileUris: string[]; /** URIs of files to be copied to the working directory of Spark drivers and distributed tasks. */ fileUris: string[]; /** URIs of archives to be extracted in the working directory of Spark drivers and tasks. */ archiveUris: string[]; /** A mapping of property names to values, used to configure Spark. */ properties: { [key: string]: string; }; /** URI of the main Python file to use as the driver. Must be a .py file. */ mainPythonFileUri: string; /** URIs of Python files to pass to the PySpark framework. */ pythonFileUris: string[]; /** List of maven coordinates of jars to include on the driver and executor classpaths. */ packages: string[]; /** List of additional remote repositories to search for the maven coordinates given with --packages. */ repositories: string[]; /** List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. */ excludePackages: string[]; } export interface PysparkJob_PropertiesEntry { key: string; value: string; } export declare const Job: { encode(message: Job, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): Job; fromJSON(object: any): Job; toJSON(message: Job): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; pysparkJob?: ({ args?: string[] | undefined; jarFileUris?: string[] | undefined; fileUris?: string[] | undefined; archiveUris?: string[] | undefined; properties?: { [x: string]: string | undefined; } | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: string[] | undefined; packages?: string[] | undefined; repositories?: string[] | undefined; excludePackages?: string[] | undefined; } & { args?: (string[] & string[] & Record, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: (string[] & string[] & Record, never>) | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>) | undefined; uiUrl?: string | undefined; } & Record, never>>(object: I): Job; }; export declare const SparkJob: { encode(message: SparkJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SparkJob; fromJSON(object: any): SparkJob; toJSON(message: SparkJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainJarFileUri?: string | undefined; mainClass?: string | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>>(object: I): SparkJob; }; export declare const SparkJob_PropertiesEntry: { encode(message: SparkJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): SparkJob_PropertiesEntry; fromJSON(object: any): SparkJob_PropertiesEntry; toJSON(message: SparkJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): SparkJob_PropertiesEntry; }; export declare const PysparkJob: { encode(message: PysparkJob, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PysparkJob; fromJSON(object: any): PysparkJob; toJSON(message: PysparkJob): unknown; fromPartial, never>) | undefined; jarFileUris?: (string[] & string[] & Record, never>) | undefined; fileUris?: (string[] & string[] & Record, never>) | undefined; archiveUris?: (string[] & string[] & Record, never>) | undefined; properties?: ({ [x: string]: string | undefined; } & { [x: string]: string | undefined; } & Record, never>) | undefined; mainPythonFileUri?: string | undefined; pythonFileUris?: (string[] & string[] & Record, never>) | undefined; packages?: (string[] & string[] & Record, never>) | undefined; repositories?: (string[] & string[] & Record, never>) | undefined; excludePackages?: (string[] & string[] & Record, never>) | undefined; } & Record, never>>(object: I): PysparkJob; }; export declare const PysparkJob_PropertiesEntry: { encode(message: PysparkJob_PropertiesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PysparkJob_PropertiesEntry; fromJSON(object: any): PysparkJob_PropertiesEntry; toJSON(message: PysparkJob_PropertiesEntry): unknown; fromPartial, never>>(object: I): PysparkJob_PropertiesEntry; }; type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined; export type DeepPartial = T extends Builtin ? T : T extends Array ? Array> : T extends ReadonlyArray ? ReadonlyArray> : T extends {} ? { [K in keyof T]?: DeepPartial; } : Partial; type KeysOfUnion = T extends T ? keyof T : never; export type Exact = P extends Builtin ? P : P & { [K in keyof P]: Exact; } & Record>, never>; export {};