import _m0 from 'protobufjs/minimal'; export declare const protobufPackage = "yandex.cloud.mdb.greenplum.v1"; export interface PXFConfig { /** * Timeout for connection to the Apache Tomcat® server when making read requests. * * Specify values in seconds. */ connectionTimeout?: number; /** * Timeout for connection to the Apache Tomcat® server when making write requests. * * Specify the values in seconds. */ uploadTimeout?: number; /** * Maximum number of the Apache Tomcat® threads. * * To prevent situations when requests get stuck or fail due to running out of memory or malfunctioning of the Java garbage collector, specify the number of the Apache Tomcat® threads. Learn more about adjusting the number of threads in the [VMware Greenplum® Platform Extension Framework](https://docs.vmware.com/en/VMware-Greenplum-Platform-Extension-Framework/6.9/greenplum-platform-extension-framework/cfg_mem.html) documentation. */ maxThreads?: number; /** Determines whether the timeout for core streaming threads is permitted. */ poolAllowCoreThreadTimeout?: boolean; /** Number of core streaming threads per pool. */ poolCoreSize?: number; /** * Maximum number of requests you can add to a pool queue for core streaming threads. * * If `0`, no pool queue is generated. */ poolQueueCapacity?: number; /** Maximum allowed number of core streaming threads. */ poolMaxSize?: number; /** Initial size, in megabytes, of the JVM heap for the PXF daemon. */ xmx?: number; /** Maximum size, in megabytes, of the JVM heap for the PXF daemon. */ xms?: number; } export interface PXFConfigSet { effectiveConfig?: PXFConfig; /** User-defined settings. */ userConfig?: PXFConfig; /** Default configuration. */ defaultConfig?: PXFConfig; } export interface PXFDatasourceS3 { /** Public key to access S3 storage. */ accessKey: string; /** Secret key to access S3 storage. */ secretKey: string; /** * Manages a fast upload of big files to S3 storage. In case of the `false` value, the PXF generates files on disk before sending them to the S3 storage. In case of the `true` value, the PXF generates files in RAM (the PXF writes to disc only if there is not enough RAM). * * The fast upload is enabled by default. */ fastUpload?: boolean; /** S3 storage address. The default value is `storage.yandexcloud.net` used for Yandex Object Storage. */ endpoint: string; } export interface PXFDatasourceJDBC { /** * JDBC driver class in Java. The possible values are the following: * * * `com.simba.athena.jdbc.Driver` * * `com.clickhouse.jdbc.ClickHouseDriver` * * `com.ibm.as400.access.AS400JDBCDriver` * * `com.microsoft.sqlserver.jdbc.SQLServerDriver` * * `com.mysql.cj.jdbc.Driver` * * `org.postgresql.Driver` * * `oracle.jdbc.driver.OracleDriver` * * `net.snowflake.client.jdbc.SnowflakeDriver` * * `io.trino.jdbc.TrinoDriver` */ driver: string; /** * URL that the JDBC driver uses to connect to the database. Examples: * * * `jdbc:mysql://mysqlhost:3306/testdb`: Local MySQL DB. * * `jdbc:postgresql://c-.rw.mdb.yandexcloud.net:6432/db1`: Managed Service for PostgreSQL cluster. The address contains the special FQDN of the cluster's master. * * `jdbc:oracle:thin:@host.example:1521:orcl`: Oracle DB. */ url: string; /** Username of the DB owner. */ user: string; /** Password of the DB owner. */ password: string; /** * Number of rows to read in an external table, in a batch. * * The default value is `100`. */ statementBatchSize?: number; /** * Number of rows to fetch (buffer) when reading from an external table. * * The default value is `1000`. */ statementFetchSize?: number; /** * Amount of time (in seconds) the JDBC driver waits for a statement to run. This timeout applies to statements created for both read and write operations. * * The default value is `60`. */ statementQueryTimeout?: number; /** Determines whether JDBC connection pooling is used in a server configuration. By default, it is used. */ poolEnabled?: boolean; /** * Maximum number of connections to the DB backend. * * The default value is `5`. */ poolMaximumSize?: number; /** * Maximum time, in milliseconds, to wait for a connection from the pool. * * The default value is `30000`. */ poolConnectionTimeout?: number; /** * Maximum amount of time, in milliseconds, after which an inactive connection is considered idle. * * The default value is `30000`. */ poolIdleTimeout?: number; /** * Minimum number of idle connections maintained in the connection pool. * * The default value is `0`. */ poolMinimumIdle?: number; } export interface PXFDatasourceCore { /** URI whose scheme and authority determine the file system implementation. */ defaultFs: string; /** Rules for mapping Kerberos principals to operating system user accounts. */ securityAuthToLocal: string; } export interface PXFDatasourceKerberos { /** Determines whether the Kerberos authentication server is used. By default, it is not used. */ enable?: boolean; /** Host of the primary KDC server (Key Distribution Center). */ primary: string; /** Kerberos realm for a Greenplum® DB. */ realm: string; /** KDC server hosts. */ kdcServers: string[]; /** Administration server host. Usually, this is the primary Kerberos server. */ adminServer: string; /** Domain that is used for the host name extension. Applicable when Kerberos 4 service members become Kerberos 5 service members (for example, when rcmd.hostname is replaced with host/hostname.domain). */ defaultDomain: string; /** Base64 encoded contents of the keytab file. */ keytabBase64: string; } export interface PXFDatasourceHDFSDfsNamenode { rpcAddress: string; serviceRpcAddress: string; httpAddress: string; httpsAddress: string; } export interface PXFDatasourceHDFSDfs { /** * Determines whether automatic failover is enabled for the high availability of the file system. * * The automatic failover is enabled by default. */ haAutomaticFailoverEnabled?: boolean; /** * If `true`, access tokens are used as capabilities for accessing datanodes. If `false`, no access tokens are checked on accessing datanodes. * * The check of access tokens is enabled by default. */ blockAccessTokenEnabled?: boolean; /** Determines whether the datanode hostname is used when connecting to datanodes. */ useDatanodeHostname?: boolean; /** * List of HDFS service logical names. * * Specify them separated by commas. The names can be arbitrary. */ namenodes: { [key: string]: PXFDatasourceHDFSDfsNamenode; }; /** Corresponds well-known HDFS client setting "dfs.nameservices" for this datasource */ nameservices: string; } export interface PXFDatasourceHDFSDfs_NamenodesEntry { key: string; value?: PXFDatasourceHDFSDfsNamenode; } export interface PXFDatasourceHDFSYarnHaRm { resourcemanagerAddress: string; resourcemanagerSchedulerAddress: string; resourcemanagerResourceTrackerAddress: string; resourcemanagerAdminAddress: string; resourcemanagerWebappAddress: string; resourcemanagerWebappHttpsAddress: string; } export interface PXFDatasourceHDFSYarn { /** * Determines whether high availability is enabled for YARN's ResourceManager services. * * The high availability is enabled by default. */ resourcemanagerHaEnabled?: boolean; /** * Determines whether another ResourceManager should automatically become active when the active ResourceManager has failed and does not respond. * * The switch of ResourceManagers is enabled by default if the high availability is enabled. */ resourcemanagerHaAutoFailoverEnabled?: boolean; /** Determines whether the embedded ActiveStandbyElector method should be used for the election of the active ResourceManager. If the current active ResourceManager has failed and does not respond, the ActiveStandbyElector method makes another ResourceManager active which then takes over. */ resourcemanagerHaAutoFailoverEmbedded?: boolean; /** Cluster ID. Specify it, so the ResourceManager service does not become active for a different cluster. */ resourcemanagerClusterId: string; /** Highly available ResourceManager service. */ haRm: { [key: string]: PXFDatasourceHDFSYarnHaRm; }; } export interface PXFDatasourceHDFSYarn_HaRmEntry { key: string; value?: PXFDatasourceHDFSYarnHaRm; } export interface PXFDatasourceHDFS { /** Settings of the file system and security rules. */ core?: PXFDatasourceCore; /** Settings of the Kerberos network authentication protocol. */ kerberos?: PXFDatasourceKerberos; /** * Enables authentication on behalf of the Greenplum® user when connecting to the remote file storage or DBMS. * * The authentication is disabled by default. */ userImpersonation?: boolean; /** Login username for the remote file storage or DBMS if authentication on behalf of the Greenplum® user is enabled. */ username: string; /** * Maximum number of times that PXF retries a SASL connection request after a refused connection returns a `GSS initiate failed` error. * * The default value is `5`. */ saslConnectionRetries?: number; /** * ZooKeeper server hosts. * * Specify values in the `
:` format. */ zkHosts: string[]; /** Settings of the distributed file system. */ dfs?: PXFDatasourceHDFSDfs; /** Settings of the ResourceManager service that is responsible for tracking resources in a cluster and scheduling applications (e.g., MapReduce jobs). */ yarn?: PXFDatasourceHDFSYarn; } export interface PXFDatasourceHive { /** Settings of the file system and security rules. */ core?: PXFDatasourceCore; /** Settings of the Kerberos network authentication protocol. */ kerberos?: PXFDatasourceKerberos; /** * Enables authentication on behalf of the Greenplum® user when connecting to the remote file storage or DBMS. * * The authentication is disabled by default. */ userImpersonation?: boolean; /** Login username for the remote file storage or DBMS if authentication on behalf of the Greenplum® user is enabled. */ username: string; /** * Maximum number of times that PXF retries a SASL connection request after a refused connection returns a `GSS initiate failed` error. * * The default value is `5`. */ saslConnectionRetries?: number; /** * ZooKeeper server hosts. * * Specify values in the `
:` format. */ zkHosts: string[]; /** * Specifies if predicate pushdown is enabled for queries on external tables. * * The predicate pushdown is enabled by default. */ ppd?: boolean; /** List of URIs separated by commas. To request metadata, the remote DBMS connects to Metastore by one of these URIs. */ metastoreUris: string[]; /** Service principal for the Metastore Thrift server. */ metastoreKerberosPrincipal: string; /** Kerberos server principal. */ authKerberosPrincipal: string; } export interface PXFDatasource { /** Data source name. */ name: string; /** Settings of an external S3 data source. */ s3?: PXFDatasourceS3 | undefined; /** Settings of an external JDBC data source. */ jdbc?: PXFDatasourceJDBC | undefined; /** Settings of an external HDFS data source. */ hdfs?: PXFDatasourceHDFS | undefined; /** Settings of an external Hive data source. */ hive?: PXFDatasourceHive | undefined; } export declare const PXFConfig: { encode(message: PXFConfig, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFConfig; fromJSON(object: any): PXFConfig; toJSON(message: PXFConfig): unknown; fromPartial, never>>(object: I): PXFConfig; }; export declare const PXFConfigSet: { encode(message: PXFConfigSet, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFConfigSet; fromJSON(object: any): PXFConfigSet; toJSON(message: PXFConfigSet): unknown; fromPartial, never>) | undefined; userConfig?: ({ connectionTimeout?: number | undefined; uploadTimeout?: number | undefined; maxThreads?: number | undefined; poolAllowCoreThreadTimeout?: boolean | undefined; poolCoreSize?: number | undefined; poolQueueCapacity?: number | undefined; poolMaxSize?: number | undefined; xmx?: number | undefined; xms?: number | undefined; } & { connectionTimeout?: number | undefined; uploadTimeout?: number | undefined; maxThreads?: number | undefined; poolAllowCoreThreadTimeout?: boolean | undefined; poolCoreSize?: number | undefined; poolQueueCapacity?: number | undefined; poolMaxSize?: number | undefined; xmx?: number | undefined; xms?: number | undefined; } & Record, never>) | undefined; defaultConfig?: ({ connectionTimeout?: number | undefined; uploadTimeout?: number | undefined; maxThreads?: number | undefined; poolAllowCoreThreadTimeout?: boolean | undefined; poolCoreSize?: number | undefined; poolQueueCapacity?: number | undefined; poolMaxSize?: number | undefined; xmx?: number | undefined; xms?: number | undefined; } & { connectionTimeout?: number | undefined; uploadTimeout?: number | undefined; maxThreads?: number | undefined; poolAllowCoreThreadTimeout?: boolean | undefined; poolCoreSize?: number | undefined; poolQueueCapacity?: number | undefined; poolMaxSize?: number | undefined; xmx?: number | undefined; xms?: number | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): PXFConfigSet; }; export declare const PXFDatasourceS3: { encode(message: PXFDatasourceS3, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceS3; fromJSON(object: any): PXFDatasourceS3; toJSON(message: PXFDatasourceS3): unknown; fromPartial, never>>(object: I): PXFDatasourceS3; }; export declare const PXFDatasourceJDBC: { encode(message: PXFDatasourceJDBC, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceJDBC; fromJSON(object: any): PXFDatasourceJDBC; toJSON(message: PXFDatasourceJDBC): unknown; fromPartial, never>>(object: I): PXFDatasourceJDBC; }; export declare const PXFDatasourceCore: { encode(message: PXFDatasourceCore, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceCore; fromJSON(object: any): PXFDatasourceCore; toJSON(message: PXFDatasourceCore): unknown; fromPartial, never>>(object: I): PXFDatasourceCore; }; export declare const PXFDatasourceKerberos: { encode(message: PXFDatasourceKerberos, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceKerberos; fromJSON(object: any): PXFDatasourceKerberos; toJSON(message: PXFDatasourceKerberos): unknown; fromPartial, never>) | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & Record, never>>(object: I): PXFDatasourceKerberos; }; export declare const PXFDatasourceHDFSDfsNamenode: { encode(message: PXFDatasourceHDFSDfsNamenode, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFSDfsNamenode; fromJSON(object: any): PXFDatasourceHDFSDfsNamenode; toJSON(message: PXFDatasourceHDFSDfsNamenode): unknown; fromPartial, never>>(object: I): PXFDatasourceHDFSDfsNamenode; }; export declare const PXFDatasourceHDFSDfs: { encode(message: PXFDatasourceHDFSDfs, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFSDfs; fromJSON(object: any): PXFDatasourceHDFSDfs; toJSON(message: PXFDatasourceHDFSDfs): unknown; fromPartial, never>) | undefined; } & Record, never>) | undefined; nameservices?: string | undefined; } & Record, never>>(object: I): PXFDatasourceHDFSDfs; }; export declare const PXFDatasourceHDFSDfs_NamenodesEntry: { encode(message: PXFDatasourceHDFSDfs_NamenodesEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFSDfs_NamenodesEntry; fromJSON(object: any): PXFDatasourceHDFSDfs_NamenodesEntry; toJSON(message: PXFDatasourceHDFSDfs_NamenodesEntry): unknown; fromPartial, never>) | undefined; } & Record, never>>(object: I): PXFDatasourceHDFSDfs_NamenodesEntry; }; export declare const PXFDatasourceHDFSYarnHaRm: { encode(message: PXFDatasourceHDFSYarnHaRm, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFSYarnHaRm; fromJSON(object: any): PXFDatasourceHDFSYarnHaRm; toJSON(message: PXFDatasourceHDFSYarnHaRm): unknown; fromPartial, never>>(object: I): PXFDatasourceHDFSYarnHaRm; }; export declare const PXFDatasourceHDFSYarn: { encode(message: PXFDatasourceHDFSYarn, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFSYarn; fromJSON(object: any): PXFDatasourceHDFSYarn; toJSON(message: PXFDatasourceHDFSYarn): unknown; fromPartial, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): PXFDatasourceHDFSYarn; }; export declare const PXFDatasourceHDFSYarn_HaRmEntry: { encode(message: PXFDatasourceHDFSYarn_HaRmEntry, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFSYarn_HaRmEntry; fromJSON(object: any): PXFDatasourceHDFSYarn_HaRmEntry; toJSON(message: PXFDatasourceHDFSYarn_HaRmEntry): unknown; fromPartial, never>) | undefined; } & Record, never>>(object: I): PXFDatasourceHDFSYarn_HaRmEntry; }; export declare const PXFDatasourceHDFS: { encode(message: PXFDatasourceHDFS, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHDFS; fromJSON(object: any): PXFDatasourceHDFS; toJSON(message: PXFDatasourceHDFS): unknown; fromPartial, never>) | undefined; kerberos?: ({ enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: string[] | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & { enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: (string[] & string[] & Record, never>) | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & Record, never>) | undefined; userImpersonation?: boolean | undefined; username?: string | undefined; saslConnectionRetries?: number | undefined; zkHosts?: (string[] & string[] & Record, never>) | undefined; dfs?: ({ haAutomaticFailoverEnabled?: boolean | undefined; blockAccessTokenEnabled?: boolean | undefined; useDatanodeHostname?: boolean | undefined; namenodes?: { [x: string]: { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } | undefined; } | undefined; nameservices?: string | undefined; } & { haAutomaticFailoverEnabled?: boolean | undefined; blockAccessTokenEnabled?: boolean | undefined; useDatanodeHostname?: boolean | undefined; namenodes?: ({ [x: string]: { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } | undefined; } & { [x: string]: ({ rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } & { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; nameservices?: string | undefined; } & Record, never>) | undefined; yarn?: ({ resourcemanagerHaEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEmbedded?: boolean | undefined; resourcemanagerClusterId?: string | undefined; haRm?: { [x: string]: { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } | undefined; } | undefined; } & { resourcemanagerHaEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEmbedded?: boolean | undefined; resourcemanagerClusterId?: string | undefined; haRm?: ({ [x: string]: { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } | undefined; } & { [x: string]: ({ resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } & { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): PXFDatasourceHDFS; }; export declare const PXFDatasourceHive: { encode(message: PXFDatasourceHive, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasourceHive; fromJSON(object: any): PXFDatasourceHive; toJSON(message: PXFDatasourceHive): unknown; fromPartial, never>) | undefined; kerberos?: ({ enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: string[] | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & { enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: (string[] & string[] & Record, never>) | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & Record, never>) | undefined; userImpersonation?: boolean | undefined; username?: string | undefined; saslConnectionRetries?: number | undefined; zkHosts?: (string[] & string[] & Record, never>) | undefined; ppd?: boolean | undefined; metastoreUris?: (string[] & string[] & Record, never>) | undefined; metastoreKerberosPrincipal?: string | undefined; authKerberosPrincipal?: string | undefined; } & Record, never>>(object: I): PXFDatasourceHive; }; export declare const PXFDatasource: { encode(message: PXFDatasource, writer?: _m0.Writer): _m0.Writer; decode(input: _m0.Reader | Uint8Array, length?: number): PXFDatasource; fromJSON(object: any): PXFDatasource; toJSON(message: PXFDatasource): unknown; fromPartial, never>) | undefined; jdbc?: ({ driver?: string | undefined; url?: string | undefined; user?: string | undefined; password?: string | undefined; statementBatchSize?: number | undefined; statementFetchSize?: number | undefined; statementQueryTimeout?: number | undefined; poolEnabled?: boolean | undefined; poolMaximumSize?: number | undefined; poolConnectionTimeout?: number | undefined; poolIdleTimeout?: number | undefined; poolMinimumIdle?: number | undefined; } & { driver?: string | undefined; url?: string | undefined; user?: string | undefined; password?: string | undefined; statementBatchSize?: number | undefined; statementFetchSize?: number | undefined; statementQueryTimeout?: number | undefined; poolEnabled?: boolean | undefined; poolMaximumSize?: number | undefined; poolConnectionTimeout?: number | undefined; poolIdleTimeout?: number | undefined; poolMinimumIdle?: number | undefined; } & Record, never>) | undefined; hdfs?: ({ core?: { defaultFs?: string | undefined; securityAuthToLocal?: string | undefined; } | undefined; kerberos?: { enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: string[] | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } | undefined; userImpersonation?: boolean | undefined; username?: string | undefined; saslConnectionRetries?: number | undefined; zkHosts?: string[] | undefined; dfs?: { haAutomaticFailoverEnabled?: boolean | undefined; blockAccessTokenEnabled?: boolean | undefined; useDatanodeHostname?: boolean | undefined; namenodes?: { [x: string]: { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } | undefined; } | undefined; nameservices?: string | undefined; } | undefined; yarn?: { resourcemanagerHaEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEmbedded?: boolean | undefined; resourcemanagerClusterId?: string | undefined; haRm?: { [x: string]: { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } | undefined; } | undefined; } | undefined; } & { core?: ({ defaultFs?: string | undefined; securityAuthToLocal?: string | undefined; } & { defaultFs?: string | undefined; securityAuthToLocal?: string | undefined; } & Record, never>) | undefined; kerberos?: ({ enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: string[] | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & { enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: (string[] & string[] & Record, never>) | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & Record, never>) | undefined; userImpersonation?: boolean | undefined; username?: string | undefined; saslConnectionRetries?: number | undefined; zkHosts?: (string[] & string[] & Record, never>) | undefined; dfs?: ({ haAutomaticFailoverEnabled?: boolean | undefined; blockAccessTokenEnabled?: boolean | undefined; useDatanodeHostname?: boolean | undefined; namenodes?: { [x: string]: { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } | undefined; } | undefined; nameservices?: string | undefined; } & { haAutomaticFailoverEnabled?: boolean | undefined; blockAccessTokenEnabled?: boolean | undefined; useDatanodeHostname?: boolean | undefined; namenodes?: ({ [x: string]: { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } | undefined; } & { [x: string]: ({ rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } & { rpcAddress?: string | undefined; serviceRpcAddress?: string | undefined; httpAddress?: string | undefined; httpsAddress?: string | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; nameservices?: string | undefined; } & Record, never>) | undefined; yarn?: ({ resourcemanagerHaEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEmbedded?: boolean | undefined; resourcemanagerClusterId?: string | undefined; haRm?: { [x: string]: { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } | undefined; } | undefined; } & { resourcemanagerHaEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEnabled?: boolean | undefined; resourcemanagerHaAutoFailoverEmbedded?: boolean | undefined; resourcemanagerClusterId?: string | undefined; haRm?: ({ [x: string]: { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } | undefined; } & { [x: string]: ({ resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } & { resourcemanagerAddress?: string | undefined; resourcemanagerSchedulerAddress?: string | undefined; resourcemanagerResourceTrackerAddress?: string | undefined; resourcemanagerAdminAddress?: string | undefined; resourcemanagerWebappAddress?: string | undefined; resourcemanagerWebappHttpsAddress?: string | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; } & Record, never>) | undefined; hive?: ({ core?: { defaultFs?: string | undefined; securityAuthToLocal?: string | undefined; } | undefined; kerberos?: { enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: string[] | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } | undefined; userImpersonation?: boolean | undefined; username?: string | undefined; saslConnectionRetries?: number | undefined; zkHosts?: string[] | undefined; ppd?: boolean | undefined; metastoreUris?: string[] | undefined; metastoreKerberosPrincipal?: string | undefined; authKerberosPrincipal?: string | undefined; } & { core?: ({ defaultFs?: string | undefined; securityAuthToLocal?: string | undefined; } & { defaultFs?: string | undefined; securityAuthToLocal?: string | undefined; } & Record, never>) | undefined; kerberos?: ({ enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: string[] | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & { enable?: boolean | undefined; primary?: string | undefined; realm?: string | undefined; kdcServers?: (string[] & string[] & Record, never>) | undefined; adminServer?: string | undefined; defaultDomain?: string | undefined; keytabBase64?: string | undefined; } & Record, never>) | undefined; userImpersonation?: boolean | undefined; username?: string | undefined; saslConnectionRetries?: number | undefined; zkHosts?: (string[] & string[] & Record, never>) | undefined; ppd?: boolean | undefined; metastoreUris?: (string[] & string[] & Record, never>) | undefined; metastoreKerberosPrincipal?: string | undefined; authKerberosPrincipal?: string | undefined; } & Record, never>) | undefined; } & Record, never>>(object: I): PXFDatasource; }; type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined; export type DeepPartial = T extends Builtin ? T : T extends Array ? Array> : T extends ReadonlyArray ? ReadonlyArray> : T extends {} ? { [K in keyof T]?: DeepPartial; } : Partial; type KeysOfUnion = T extends T ? keyof T : never; export type Exact = P extends Builtin ? P : P & { [K in keyof P]: Exact; } & Record>, never>; export {};