Common Content for Data Engineering
- Common Content for Data Engineering 10.2.2 Service Pack 1
- All Products
Bug
| Description
|
---|---|
OCON-9377
| When you configure Sqoop and run a Teradata Parallel Transporter mapping on a Cloudera cluster to export data of the Byte or Varbyte data type to a Teradata target, the mapping fails on the Blaze engine.
|
OCON-9376
| If you configure Sqoop to export data of the Blob or Clob data type to a Teradata target, TDCH mappings fail on the Spark engine.
|
OCON-9143
| In the read and write operations for a complex file data object, you cannot edit the precision and scale of elements within a field that is of a complex data type.
For example, if Field1 is of type array with string elements, you cannot edit the precision and scale of the string elements.
|
OCON-9005
| When you run TDCH mappings on the Hive engine to write time data to a Teradata target, the nanosecond part is truncated.
|
OCON-8850
| If you configure Sqoop to export data of the Timestamp data type from a Hive source to a Microsoft Azure SQL Data Warehouse target, the mapping fails.
|
OCON-8779
| If you configure Sqoop to export data of the Real data type to IBM DB2 z/OS targets, the mapping fails.
|
OCON-7687
| When you export data through Sqoop and the columns contain mixed case characters, the mapping fails.
|
OCON-7669
| When you configure Sqoop and OraOop, and export data to an Oracle target that contains mixed case characters in the table name, the mapping fails.
Workaround: Use the generic Oracle JDBC driver to export data.
|
OCON-7429
| When you run a Teradata Parallel Transporter mapping on a Hortonworks cluster and on the Blaze engine to write data of the Byte or Varbyte data type to a Teradata target, the data gets corrupted. This issue occurs when you use the
hdp-connector-for-teradata-1.5.1.2.5.0.0-1245-distro.tar.gz JAR.
Workaround: Use the
hdp-connector-for-teradata-1.4.1.2.3.2.0-2950-distro.tar.gz JAR.
|
OCON-730
| When you export data through Sqoop and there are primary key violations, the mapping fails and bad records are not written to the bad file. (456616)
|
OCON-7216
| If a Sqoop source or target contains a column name with double quotes, the mapping fails on the Blaze engine. However, the Blaze Job Monitor incorrectly indicates that the mapping ran successfully and that rows were written into the target.
|
OCON-7212
| If there are unconnected ports in a target, Sqoop mappings fail on the Blaze engine. This issue occurs when you run the Sqoop mapping on any cluster other than a Cloudera cluster.
Workaround: Before you run the mapping, create a table in the target database with columns corresponding to the connected ports.
|
OCON-7205
| When you run a Sqoop mapping on the Blaze engine to export data of the Numeric data type from Netezza, the scale part of the data is truncated.
|
OCON-17245
| When you run a Sqoop mapping on a Kerberos-enabled HDInsight cluster with ADLS as storage, the mapping fails on the Blaze engine.
|
OCON-17194
| Microsoft Azure Blob Storage incorrectly appears as a source option for the Lookup transformation on the Spark and Databricks Spark engines, but this option is not supported. Mapping validation fails with the following error when you select Microsoft Azure Blob Storage as a lookup source on the Spark or Databricks Spark engine:
|
OCON-16315
| The Data Integration Service fails with out of memory errors when you run a large number of concurrent mappings with data object read and write operations that project columns as complex data types.
Workaround: Perform any of the following steps:
|
IN-3362
| A data object with an intelligent structure model might accept JSON input files formatted with newlines between attributes as suitable for partitioning, even though the files cannot be partitioned. This might cause data that corresponds to the model to be identified as lost data in the Spark log.
|
BDM-4597
| A mapping with a joiner transformation that processes more than 4,294,967,294 rows in a single partition will fail.
Workaround: If possible, increase partitioning on the source.
|
BDM-23575
| The number of nodes is incorrectly displayed in the session log for the mappings that run on nodes that are labeled in a cluster that runs on Blaze and Spark engines.
|
BDM-23550
| When an Update Strategy transformation contains an insert, update, or a delete operation, and a JDBC target, incorrect number of inserted, updated, or deleted rows appear from the Spark events.
|
BDM-23392
| When a mapping that runs on Spark uses an Update Strategy transformation, table constraints might cause a BatchUpdateException and mapping failures.
Workaround: Edit the Spark.JdbcNumPartition setting in the mapping Runtime Properties to reduce the number of partitions to 1. This causes all rows to be processed as a single partition.
|
BDM-23317
| When you monitor statistics for the Data Integration Service configured with file-based queuing, the jobs incorrectly appear to run even when the job state is queued.
|
BDM-22490
| When a dynamic mapping that runs on Spark uses an Update Strategy transformation, adding a column to the Hive target table schema causes the mapping to fail.
|
BDM-22481
| When the Spark engine processes an input value of zero in a decimal port that is configured with equivalent precision and scale, the engine treats the value as data overflow and the return value is NULL.
|
BDM-22282
| The Spark engine might take up to 30 minutes to run a mapping that contains a Python transformation if you pass a large number of ports to the Python transformation.
|
BDM-22260
| Cannot get Spark monitoring statistics for a mapping run that uses any of the following connections: Google BigQuery, Google Cloud Storage, Google Cloud Spanner, and Google Analytics.
|
BDM-2222
| The Spark engine does not run the footer row command configured for a flat file target. (459942)
|
BDM-2141
| Mapping with a Hive source and target that uses an ABS function with an IIF function fails in the Hadoop environment. (424789)
|
BDM-2137
| Mapping in the Hadoop environment fails when it contains a Hive source and a filter condition that uses the default table name prefixed to the column name.
Workaround: Edit the filter condition to remove the table name prefixed to the column name and run the mapping again. (422627)
|
BDM-2136
| Mapping in the Hadoop environment fails because the Hadoop connection uses 128 characters in its name. (421834)
|
BDM-20856
| When you import a cluster workflow, the import wizard does not include the option to choose the non-native connection that was associated with the Create Cluster task.
Workaround: After you import the workflow, manually assign a Databricks or Hadoop connection to the Create Cluster task.
|
BDM-20697
| If you use the numberOfErrorRows system-defined mapping output in a mapping that runs on the Spark engine, the engine returns an incorrect value for the mapping output.
|
BDM-18140
| A mapping that reads a large number of reference tables may take longer than expected to run on the Spark engine. The issue is observed when the mapping includes transformations that collectively read 140 reference tables.
Workaround: Run the mapping on the Blaze engine.
|
BDM-17485
| Mapping or mapplet that has the same name or any non-reusable transformation with the same name as mapplet or mapping fails to import into the Model repository.
|
BDM-17174
| When memory usage reaches the maximum container size, YARN kills the container.
Memory usage on the OOP Container Manager reaches the maximum container size if the following conditions are true:
|
BDM-16521
| On Oracle and DB2, when a Lookup transformation contains a Text data type column and you import the mapping into the Developer tool, the Text datatype is mapped to the Clob datatype and the mapping fails with an error. Similarly, the Binary data type gets imported into the Model repository as a Blob data type and the mapping fails.
Workaround: Edit the column type in the Lookup transformation to run the mapping.
|