Common Content for Data Engineering
- Common Content for Data Engineering 10.4.1
- All Products
Issue
| Description
|
---|---|
BDM-36415
| If a mapping uses a parameter value to specify the execution environment, and you specify "native" or "Hadoop" as arguments when running one of the following commands, the Data Integration Service ignores the command and the mapping is not updated. The issue affects the following commands:
The issue occurs because the commands do not support parameter resolution.
Workaround: Apply EBF-20109 and use an additional argument for the command, ValidationEnvironmentFilterName. For example,
ValidationEnvironmentFilterName=Hadoop
|
BDM-34065
| When you import the cluster configuration from Compute cluster configured on Cloudera and you choose to create associated connections, the creation process fails to create the Hive connection and displays the following error: [EDR_1069] Failed to resolve the option [connectString] of connection [HIVE_cco_ac] because the associated cluster configuration [CCO_AC] does not contain the property [hive.metastore.uris]. Workaround:
|
BDM-34036
| When the Spark engine runs a mapping or previews data on a Cloudera CDP cluster, the job incorrectly writes date/time data to the target when the date is earlier than 1583 and the mapping uses one of the following sources:
|
BDM-33922
| In a Kerberos-enabled domain, a mapping that runs on the Spark engine using an impersonation user fails if the following conditions are true:
Workaround: Provide the DECRYPT privilege in the encryption zone to the Active Directory user, or use a cluster staging directory that is outside of the encryption zone. For example, use the /tmp directory.
|
BDM-33915
| When the Blaze engines performs a lookup on a compressed Hive table of text format, the mapping fails with the following error:
"[CORE_3] An internal exception occurred with message: java.io.IOException: incorrect header check java.io.IOException: incorrect header check"
|
BDM-33914
| When the Blaze engine reads a modified Hive partition table of text format, the source read fails with the following error:
"[CORE_3] An internal exception occurred with message: java.lang.RuntimeException: No data found for parttion state java.lang.RuntimeException: No data found for parttion state"
|
BDM-33894
| When the Spark engine runs a mapping on a Cloudera CDP cluster with dynamic resource allocation enabled, the mapping fails with the following SPARK_1003 error:
"Failed to connect to <node running Spark shuffle service>"
For more information, see the KB article "ERROR: "[SPARK_1003] Failed to connect to <node running Spark shuffle service>" when the Spark engine runs a mapping on Cloudera CDP."
|
BDM-33822
| When the Blaze engine reads a partition table where partitions have been added using the Hive command
ALTER TABLE ADD PARTITION, the Blaze engine does not recognize the new partition data. This error condition occurs with the following file formats:
|
BDM-33585
| When a mapping that runs on the Blaze engine reads from or writes to an encryption zone in HDFS, the mapping fails.
Workaround: Configure the user that runs a mapping on Blaze to be the same as the impersonation user. To use a different user, give DECRYPT privileges to the Blaze user.
|
BDM-22677
| When you create a Cloudera CDP cluster configuration and connection during installation, the installer uses the default engine type, MRv2, rather than Tez in the Hadoop connection.
In the Administrator tool, edit the Hadoop connection
Common Properties . Set
Engine Type to Tez.
|
CORE-8958
| When you run a mapping that reads from a Hive source on a Cloudera CDP cluster, the mapping fails with the following error:
java.lang.NoClassDefFoundError: org/apache/jute/Record
For CDP Private Cloud, in the Hive connection properties, set
Metadata Connection String to the following value:
jdbc:hive2://<node running HiveServer2>:10000/default;principal=hive/_HOST@PLATFORMKRB.COM;ssl=true;
This workaround does not apply to CDP Public Cloud.
|
OCON-24969
| When you create a custom SQL query in a Read transformation to read from a Hive source that contains complex data types, you cannot select the elements of the complex data field in the Advanced Query properties. You can manually select the name of an element in the query.
|
Issue
| Description
|
---|---|
BDM-33967
| Databricks clusters do not support concurrent mapping processing. While very limited concurrency is sometimes possible, concurrency support depends on multiple factors and is difficult to predict.
|