Table of Contents

Search

  1. Abstract
  2. Informatica 10.2.2 HotFix 1 Service Pack 2 Installation
  3. Emergency Bug Fixes Merged into 10.2.2 HotFix 1 Service Pack 2
  4. 10.2.2 HotFix 1 Service Pack 2 Fixed Limitations
  5. 10.2.2 HotFix 1 Service Pack 2 Known Limitations
  6. 10.2.2 HotFix 1 Service Pack 1 Fixed Limitations
  7. 10.2.2 HotFix 1 Service Pack 1 Known Limitations
  8. 10.2.2 HotFix 1 Fixed Limitations
  9. 10.2.2 HotFix 1 Known Limitations
  10. 10.2.2 Service Pack 1 Fixed Limitations
  11. 10.2.2 Service Pack 1 Known Limitations
  12. 10.2.2 Fixed Limitations and Closed Enhancements
  13. 10.2.2 Known Limitations
  14. Informatica Global Customer Support

Big Data Release Notes (10.2.2 HotFix 1 Service Pack 2)

Big Data Release Notes (10.2.2 HotFix 1 Service Pack 2)

Big Data Management Fixed Limitations (10.2.2)

Big Data Management Fixed Limitations (10.2.2)

Review the Release Notes of previous releases for information about previous fixed limitations.
The following table describes fixed limitations:
Bug
Description
OCON-17874
When you run a Sqoop mapping with a TLS-enabled domain and
SparkSqoopDisSideInvocation
property is set to true, the mapping fails because the location of the truststore system property of the Data Integration Service gets modified from
-Djavax.net.ssl.trustStore= $INFA_HOME/services/shared/security/infa_truststore.jks
to
-Djavax.net.ssl.trustStore=$INFA_HOME/externaljdbcjars/../java/jre/lib/security/cacerts
.
OCON-16535
When you run a Sqoop mapping on the Spark engine, the MapReduce job fails because the Data Integration Service ignores the
hadoop.security.token.service.use_ip
configuration property in the
core-site.xml
file.
IN-3362
A data object with an intelligent structure model might accept JSON input files formatted with newlines between attributes as suitable for partitioning, even though the files cannot be partitioned. This might cause data that corresponds to the model to be identified as lost data in the Spark log.
IDL-8652
If you run a mass ingestion specification that contains more than 2,000 tables, a majority of the ingestion mapping jobs fail with the following error:
SEVERE: UserDefinedJob execution failed
IDL-10354
When a run instance of a mass ingestion specification completes, the Summary page in the Mass Ingestion tool continues to display the status as RUNNING. The Summary page updates the status only after you navigate to the Execution Statistics page for the run instance.
BDM-23906
The infacmd ccps listClusters command fails when the domain contains a Databricks cluster that does not have any custom tags.
Workaround: Manually add a custom tag to the cluster or create the cluster through a cluster workflow.
BDM-23599
Connectivity to Hive sources and targets on Amazon S3 with SSE-KMS authorization fails on a Hortonworks HDP 2.6.1 or higher version cluster with an error reading "The encryption method specified is not supported."
BDM-23199
When a sorted Joiner transformation joins data from a single source, data loss can occur if the row size for key columns exceed the data cache size.
BDM-22862
If the Spark engine runs a mapping that creates temporary Hive tables and the data contains string values with newline characters (\n), the newline character is treated as the end of one record. For example, if the temporary Hive table stores the string
'abc \n def'
, the first record is
'abc'
and the second record is
'def'
.
One situation where the issue occurs is in a mapping that contains an Update Strategy transformation. When the mapping runs on the Spark engine, the Update Strategy transformation stages data in temporary Hive tables.
BDM-22754
When the DTM runs out of memory and shuts down, the subsequent mappings sometimes fail with a NullPointerException during mapping initialization.
BDM-22736
When you run a mapping with a large number of Joiner transformations on the Spark engine, the Data Integration Service runs out of memory.
BDM-21657
Blaze mapping reads NULL data from a Parquet format Hive table.
BDM-21324
Incorrect query generated for a Hive update strategy when the column name starts with a number.
BDM-18221
If the tez.tez-ui.history-url.base is set to NULL or if it is not present in the cluster configuration, the Tez jobs fail with a null pointer exception.
BDM-18195
When you disable yarn.timeline-service in the cluster configuration, the Tez view does not show the Hadoop job.
BDM-18173
The duration taken to load the Summary Statistics and Execution Statistics in the Administrator tool for a non-Administrator user is more when compared to an Administrator.
BDM-17879
Slow performance of domain when OS profiles are enabled.
BDM-17849
When you use infacmd command line program to export a mapping that contains an XML payload of a complex data type, the export fails with the following error:
com.sun.xml.xsom.parser.XSOMParser cannot be found by com.infa.products.sdk.adapter.parser.xml.utils_1.1.0 com/sun/xml/xsom/parser/XSOMParser [ICMD_10033] Command [ExportObjects] failed with error [[OIECMD_10006] Error occurred during export function. See previous error.].
BDM-17831
The JVM options for debugging are disabled for Data Integration Services.
BDM-17755
When a decimal port on a partitioned, bucketed Hive target is disconnected, a passthrough mapping that runs on the Blaze engine fails.
BDM-17698
When you use the Cloudera Connector Powered by Teradata to run a Sqoop TDCH mapping, the mapping fails with the following error:
Missing argument for option: staging-database
BDM-17685
The Data Integration Service does not remove temporary empty files from the operating system profile target directory when all the following conditions are true:
  • The Data Integration Service is enabled to use operating system profiles.
  • The mapping writes to an HDFS target.
  • The mapping runs on the Hive engine.
BDM-17541
When you run a mapping by choosing
Window > Preferences > Run configurations > Mapping
, the Developer tool attempts to re-establish connections with all configured domains. Depending on the connection quality, this may result in dramatically slower performance.
BDM-17175
After you import the workflow into the Model repository, you cannot create a parameter set for the SQL parameter with the empty default value.
BDM-16843
After you import an SQL parameter into the Model repository, you can edit the parameter value but you cannot reset it later to the default empty or NULL value.
BDM-14626
When you do not assign a connection in the physical data object, the workflow application fails to deploy.
BDM-14139
In a Hortonworks HDP or an Amazon EMR environment, a mapping run on the Spark engine fails when all of the following conditions are true:
  • The mapping writes to an Avro target.
  • The schema contains an array or a map that exceeds two levels of nesting.
  • A struct element exists at the third or higher level of nesting.
For example, when you write to an Avro target with a schema such as array(array(struct)) or array(map(string, struct)), the mapping fails.
BDM-13107
When you run a mapping with Spark monitoring enabled, performance bottlenecks can appear. The performance varies based on the mapping complexity.

0 COMMENTS

We’d like to hear from you!