Administrator
- Administrator
- All Products
Property
| Description
|
---|---|
Database
| The name of the schema in Databricks.
The name can contain only alphanumeric characters and hyphen (-).
This property is optional for SQL warehouse, all-purpose cluster, and job cluster.
If you do not specify a value, all databases available in the
workspace are listed. The value you specify overrides the schema
specified in the SQL Warehouse JDBC URL
connection property. |
JDBC Driver Class Name
| The name of the JDBC driver class.
This property is optional for SQL warehouse, all-purpose cluster, and job cluster.
Default is
com.databricks.client.jdbc.Driver
|
Staging Environment
| The staging environment where your data is temporarily stored before
processing. This property is required for SQL warehouse,
all-purpose cluster, and job cluster. Select one of the
following options as the staging environment:
Default is Volume.
You
cannot switch between clusters after you establish a connection.
Effective in the October 2024 release, personal
staging location is deprecated. Deprecated functionality is
supported, but Informatica intends to drop support in a future
release. Informatica requests that you use a volume to stage the
data. |
Volume Path
| The absolute path in Volume where you want to stage the data
temporarily. Specify the path in the following format: /Volumes/<catalog_identifier>/<schema_identifier>/<volume_identifier>/
If you do not specify a volume path, the Secure Agent creates
a managed volume in Databricks. |
Databricks Host
| The host name of the endpoint the Databricks account belongs to.
This property is required only for all-purpose cluster and job cluster. Doesn't apply to SQL warehouse.
You can get the Databricks Host from the JDBC URL. The URL is available in the Advanced Options of JDBC or ODBC in the Databricks all-purpose cluster.
The following example shows the Databricks Host in JDBC URL:
jdbc:databricks:// <Databricks Host> :443/
default;transportMode=http; ssl=1;httpPath=sql/
protocolv1/o/<Org Id>/<Cluster ID>; AuthMech=3;
UID=token; PWD=<personal-access-token> The value of PWD in Databricks Host, Organization Id, and Cluster ID is always
<personal-access-token> .
|
Cluster ID
| The ID of the cluster.
This property is required only for all-purpose cluster and job cluster. Doesn't apply to SQL warehouse.
You can get the cluster ID from the JDBC URL. The URL is available in the Advanced Options of JDBC or ODBC in the Databricks all-purpose cluster
The following example shows the Cluster ID in JDBC URL:
jdbc:databricks://<Databricks Host>:443/
default;transportMode=http; ssl=1;httpPath=sql/
protocolv1/o/<Org Id>/ <Cluster ID> ;
AuthMech=3;UID=token; PWD=<personal-access-token> |
Organization ID
| The unique organization ID for the workspace in Databricks.
This property is required only for all-purpose cluster and job cluster. Doesn't apply to SQL warehouse.
You can get the Organization ID from the JDBC URL. The URL is available in the Advanced Options of JDBC or ODBC in the Databricks all-purpose cluster
The following example shows the Organization ID in JDBC URL:
jdbc:databricks://<Databricks Host>:443/
default;transportMode=http; ssl=1;httpPath=sql/
protocolv1/o/ <Organization ID> / <Cluster
ID>;AuthMech=3;UID=token; PWD=<personal-access-token> |
Min Workers 1
| The minimum number of worker nodes to be used for the Spark job. Minimum value is 1.
This property is required only for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
|
Max Workers 1
| The maximum number of worker nodes to be used for the Spark job. If you don't want to autoscale, set Max Workers = Min Workers or don't set Max Workers.
This property is optional for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
|
DB Runtime Version 1
| The version of job cluster to spawn when you connect to job cluster to process mappings.
This property is required only for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
Select the Databricks runtime version 9.1 LTS, 13.3 LTS, or
15.4 LTS. To use version 15.4 LTS, ensure that you set the
spark.databricks.driver.dbfsLibraryInstallationAllowed
parameter to true in the Spark Configuration
connection property. |
Worker Node Type 1
| The worker node instance type that is used to run the Spark job.
This property is required only for all-purpose cluster and job cluster. Doesn't apply to SQL warehouse.
For example, the worker node type for AWS can be i3.2xlarge. The worker node type for Azure can be Standard_DS3_v2.
|
Driver Node Type 1
| The driver node instance type that is used to collect data from the Spark workers.
This property is optional for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
For example, the driver node type for AWS can be i3.2xlarge. The driver node type for Azure can be Standard_DS3_v2.
If you don't specify the driver node type, Databricks uses the value you specify in the worker node type field.
|
Instance Pool ID 1
| The instance pool ID used for the Spark cluster.
This property is optional for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
If you specify the Instance Pool ID to run mappings , the following connection properties are ignored:
|
Elastic Disk 1
| Enables the cluster to get additional disk space.
This property is optional for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
Enable this option if the Spark workers are running low on disk space.
|
Spark Configuration 1
| The Spark configuration to use in the job cluster.
This property is optional for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
The configuration must be in the following format:
"key1"="value1";"key2"="value2";...
For example,
"spark.executor.userClassPathFirst"="False"
To use Databricks runtime version 15.4, specify the following
paramter: 'spark.databricks.driver.dbfsLibraryInstallationAllowed'='true' Doesn't apply to
Data Ingestion and Replication tasks.
|
Spark Environment Variables 1
| The environment variables to export before launching the Spark driver and workers.
This property is optional for job cluster. Doesn't apply to SQL warehouse and all-purpose cluster.
The variables must be in the following format:
"key1"="value1";"key2"="value2";...
For example,
"MY_ENVIRONMENT_VARIABLE"="true"
Doesn't apply to
Data Ingestion and Replication tasks.
|
1 Doesn't apply to mappings in advanced mode.
|