Common Content for Data Engineering
- Common Content for Data Engineering 10.2.2 HotFix 1
- All Products
Bug
| Description
|
---|---|
OCON-17874
| When you run a Sqoop mapping with a TLS-enabled domain and
SparkSqoopDisSideInvocation property is set to true, the mapping fails because the location of the truststore system property of the Data Integration Service gets modified from
-Djavax.net.ssl.trustStore= $INFA_HOME/services/shared/security/infa_truststore.jks to
-Djavax.net.ssl.trustStore=$INFA_HOME/externaljdbcjars/../java/jre/lib/security/cacerts .
|
OCON-16535
| When you run a Sqoop mapping on the Spark engine, the MapReduce job fails because the Data Integration Service ignores the
hadoop.security.token.service.use_ip configuration property in the
core-site.xml file.
|
IN-3362
| A data object with an intelligent structure model might accept JSON input files formatted with newlines between attributes as suitable for partitioning, even though the files cannot be partitioned. This might cause data that corresponds to the model to be identified as lost data in the Spark log.
|
IDL-8652
| If you run a mass ingestion specification that contains more than 2,000 tables, a majority of the ingestion mapping jobs fail with the following error:
|
IDL-10354
| When a run instance of a mass ingestion specification completes, the Summary page in the Mass Ingestion tool continues to display the status as RUNNING. The Summary page updates the status only after you navigate to the Execution Statistics page for the run instance.
|
BDM-23906
| The infacmd ccps listClusters command fails when the domain contains a Databricks cluster that does not have any custom tags.
Workaround: Manually add a custom tag to the cluster or create the cluster through a cluster workflow.
|
BDM-23599
| Connectivity to Hive sources and targets on Amazon S3 with SSE-KMS authorization fails on a Hortonworks HDP 2.6.1 or higher version cluster with an error reading "The encryption method specified is not supported."
|
BDM-23199
| When a sorted Joiner transformation joins data from a single source, data loss can occur if the row size for key columns exceed the data cache size.
|
BDM-22862
| If the Spark engine runs a mapping that creates temporary Hive tables and the data contains string values with newline characters (\n), the newline character is treated as the end of one record. For example, if the temporary Hive table stores the string
'abc \n def' , the first record is
'abc' and the second record is
'def' .
One situation where the issue occurs is in a mapping that contains an Update Strategy transformation. When the mapping runs on the Spark engine, the Update Strategy transformation stages data in temporary Hive tables.
|
BDM-22754
| When the DTM runs out of memory and shuts down, the subsequent mappings sometimes fail with a NullPointerException during mapping initialization.
|
BDM-22736
| When you run a mapping with a large number of Joiner transformations on the Spark engine, the Data Integration Service runs out of memory.
|
BDM-21657
| Blaze mapping reads NULL data from a Parquet format Hive table.
|
BDM-21324
| Incorrect query generated for a Hive update strategy when the column name starts with a number.
|
BDM-18221
| If the tez.tez-ui.history-url.base is set to NULL or if it is not present in the cluster configuration, the Tez jobs fail with a null pointer exception.
|
BDM-18195
| When you disable yarn.timeline-service in the cluster configuration, the Tez view does not show the Hadoop job.
|
BDM-18173
| The duration taken to load the Summary Statistics and Execution Statistics in the Administrator tool for a non-Administrator user is more when compared to an Administrator.
|
BDM-17879
| Slow performance of domain when OS profiles are enabled.
|
BDM-17849
| When you use infacmd command line program to export a mapping that contains an XML payload of a complex data type, the export fails with the following error:
|
BDM-17831
| The JVM options for debugging are disabled for Data Integration Services.
|
BDM-17755
| When a decimal port on a partitioned, bucketed Hive target is disconnected, a passthrough mapping that runs on the Blaze engine fails.
|
BDM-17698
| When you use the Cloudera Connector Powered by Teradata to run a Sqoop TDCH mapping, the mapping fails with the following error:
Missing argument for option: staging-database
|
BDM-17685
| The Data Integration Service does not remove temporary empty files from the operating system profile target directory when all the following conditions are true:
|
BDM-17541
| When you run a mapping by choosing
Window > Preferences > Run configurations > Mapping , the Developer tool attempts to re-establish connections with all configured domains. Depending on the connection quality, this may result in dramatically slower performance.
|
BDM-17175
| After you import the workflow into the Model repository, you cannot create a parameter set for the SQL parameter with the empty default value.
|
BDM-16843
| After you import an SQL parameter into the Model repository, you can edit the parameter value but you cannot reset it later to the default empty or NULL value.
|
BDM-14626
| When you do not assign a connection in the physical data object, the workflow application fails to deploy.
|
BDM-14139
| In a Hortonworks HDP or an Amazon EMR environment, a mapping run on the Spark engine fails when all of the following conditions are true:
For example, when you write to an Avro target with a schema such as array(array(struct)) or array(map(string, struct)), the mapping fails.
|
BDM-13107
| When you run a mapping with Spark monitoring enabled, performance bottlenecks can appear. The performance varies based on the mapping complexity.
|