Table of Contents

Search

  1. Abstract
  2. Informatica 10.2.2 Service Pack 1
  3. Support Changes
  4. Verify System Requirements
  5. Upgrade Path
  6. Installation
  7. Post-Installation Steps
  8. 10.2.2 Service Pack 1 Fixed Limitations
  9. 10.2.2 Service Pack 1 Known Limitations
  10. 10.2.2 Fixed Limitations and Closed Enhancements
  11. 10.2.2 Known Limitations
  12. Emergency Bug Fixes Merged into 10.2.2 Service Pack 1
  13. Informatica Global Customer Support

Big Data Release Notes

Big Data Release Notes

Big Data Management Known Limitations

Big Data Management Known Limitations

The following table describes known limitations:
Bug
Description
OCON-9377
When you configure Sqoop and run a Teradata Parallel Transporter mapping on a Cloudera cluster to export data of the Byte or Varbyte data type to a Teradata target, the mapping fails on the Blaze engine.
OCON-9376
If you configure Sqoop to export data of the Blob or Clob data type to a Teradata target, TDCH mappings fail on the Spark engine.
OCON-9143
In the read and write operations for a complex file data object, you cannot edit the precision and scale of elements within a field that is of a complex data type.
For example, if Field1 is of type array with string elements, you cannot edit the precision and scale of the string elements.
OCON-9005
When you run TDCH mappings on the Hive engine to write time data to a Teradata target, the nanosecond part is truncated.
OCON-8850
If you configure Sqoop to export data of the Timestamp data type from a Hive source to a Microsoft Azure SQL Data Warehouse target, the mapping fails.
OCON-8779
If you configure Sqoop to export data of the Real data type to IBM DB2 z/OS targets, the mapping fails.
OCON-7687
When you export data through Sqoop and the columns contain mixed case characters, the mapping fails.
OCON-7669
When you configure Sqoop and OraOop, and export data to an Oracle target that contains mixed case characters in the table name, the mapping fails.
Workaround: Use the generic Oracle JDBC driver to export data.
OCON-7429
When you run a Teradata Parallel Transporter mapping on a Hortonworks cluster and on the Blaze engine to write data of the Byte or Varbyte data type to a Teradata target, the data gets corrupted. This issue occurs when you use the
hdp-connector-for-teradata-1.5.1.2.5.0.0-1245-distro.tar.gz
JAR.
Workaround: Use the
hdp-connector-for-teradata-1.4.1.2.3.2.0-2950-distro.tar.gz
JAR.
OCON-730
When you export data through Sqoop and there are primary key violations, the mapping fails and bad records are not written to the bad file. (456616)
OCON-7216
If a Sqoop source or target contains a column name with double quotes, the mapping fails on the Blaze engine. However, the Blaze Job Monitor incorrectly indicates that the mapping ran successfully and that rows were written into the target.
OCON-7212
If there are unconnected ports in a target, Sqoop mappings fail on the Blaze engine. This issue occurs when you run the Sqoop mapping on any cluster other than a Cloudera cluster.
Workaround: Before you run the mapping, create a table in the target database with columns corresponding to the connected ports.
OCON-7205
When you run a Sqoop mapping on the Blaze engine to export data of the Numeric data type from Netezza, the scale part of the data is truncated.
OCON-17245
When you run a Sqoop mapping on a Kerberos-enabled HDInsight cluster with ADLS as storage, the mapping fails on the Blaze engine.
OCON-17194
Microsoft Azure Blob Storage incorrectly appears as a source option for the Lookup transformation on the Spark and Databricks Spark engines, but this option is not supported. Mapping validation fails with the following error when you select Microsoft Azure Blob Storage as a lookup source on the Spark or Databricks Spark engine:
[Spark Validation] The Integration Service cannot use the Spark engine to push the transformation [test_input_txt_Read1] of type [LookupTx] to the Hadoop cluster. [ID: TRANSFORMATION_NOT_SUPPORTED] test_input_txt_Read1
OCON-16315
The Data Integration Service fails with out of memory errors when you run a large number of concurrent mappings with data object read and write operations that project columns as complex data types.
Workaround: Perform any of the following steps:
  • Increase the heap memory settings on the machine where the Data Integration Service runs.
  • Reduce the number of concurrent mappings that process complex data types.
  • Set the Data Integration Service custom property ExecutionOptions.CBOStatCacheSize to a reasonably small number.
IN-3362
A data object with an intelligent structure model might accept JSON input files formatted with newlines between attributes as suitable for partitioning, even though the files cannot be partitioned. This might cause data that corresponds to the model to be identified as lost data in the Spark log.
BDM-4597
A mapping with a joiner transformation that processes more than 4,294,967,294 rows in a single partition will fail.
Workaround: If possible, increase partitioning on the source.
BDM-23575
The number of nodes is incorrectly displayed in the session log for the mappings that run on nodes that are labeled in a cluster that runs on Blaze and Spark engines.
BDM-23550
When an Update Strategy transformation contains an insert, update, or a delete operation, and a JDBC target, incorrect number of inserted, updated, or deleted rows appear from the Spark events.
BDM-23392
When a mapping that runs on Spark uses an Update Strategy transformation, table constraints might cause a BatchUpdateException and mapping failures.
Workaround: Edit the Spark.JdbcNumPartition setting in the mapping Runtime Properties to reduce the number of partitions to 1. This causes all rows to be processed as a single partition.
BDM-23317
When you monitor statistics for the Data Integration Service configured with file-based queuing, the jobs incorrectly appear to run even when the job state is queued.
BDM-22490
When a dynamic mapping that runs on Spark uses an Update Strategy transformation, adding a column to the Hive target table schema causes the mapping to fail.
BDM-22481
When the Spark engine processes an input value of zero in a decimal port that is configured with equivalent precision and scale, the engine treats the value as data overflow and the return value is NULL.
BDM-22282
The Spark engine might take up to 30 minutes to run a mapping that contains a Python transformation if you pass a large number of ports to the Python transformation.
BDM-22260
Cannot get Spark monitoring statistics for a mapping run that uses any of the following connections: Google BigQuery, Google Cloud Storage, Google Cloud Spanner, and Google Analytics.
BDM-2222
The Spark engine does not run the footer row command configured for a flat file target. (459942)
BDM-2141
Mapping with a Hive source and target that uses an ABS function with an IIF function fails in the Hadoop environment. (424789)
BDM-2137
Mapping in the Hadoop environment fails when it contains a Hive source and a filter condition that uses the default table name prefixed to the column name.
Workaround: Edit the filter condition to remove the table name prefixed to the column name and run the mapping again. (422627)
BDM-2136
Mapping in the Hadoop environment fails because the Hadoop connection uses 128 characters in its name. (421834)
BDM-20856
When you import a cluster workflow, the import wizard does not include the option to choose the non-native connection that was associated with the Create Cluster task.
Workaround: After you import the workflow, manually assign a Databricks or Hadoop connection to the Create Cluster task.
BDM-20697
If you use the numberOfErrorRows system-defined mapping output in a mapping that runs on the Spark engine, the engine returns an incorrect value for the mapping output.
BDM-18140
A mapping that reads a large number of reference tables may take longer than expected to run on the Spark engine. The issue is observed when the mapping includes transformations that collectively read 140 reference tables.
Workaround: Run the mapping on the Blaze engine.
BDM-17485
Mapping or mapplet that has the same name or any non-reusable transformation with the same name as mapplet or mapping fails to import into the Model repository.
BDM-17174
When memory usage reaches the maximum container size, YARN kills the container.
Memory usage on the OOP Container Manager reaches the maximum container size if the following conditions are true:
  • Concurrent jobs run for longer than two days.
  • The Blaze engine has not reached the idle timeout limit or the sunset time.
BDM-16521
On Oracle and DB2, when a Lookup transformation contains a Text data type column and you import the mapping into the Developer tool, the Text datatype is mapped to the Clob datatype and the mapping fails with an error. Similarly, the Binary data type gets imported into the Model repository as a Blob data type and the mapping fails.
Workaround: Edit the column type in the Lookup transformation to run the mapping.

0 COMMENTS

We’d like to hear from you!