Table of Contents

Search

  1. Abstract
  2. PowerExchange for Amazon Redshift
  3. PowerExchange for Amazon S3
  4. PowerExchange for Cassandra
  5. PowerExchange for Cassandra JDBC
  6. PowerExchange for DataSift
  7. PowerExchange for Facebook
  8. PowerExchange for Google Analytics
  9. PowerExchange for Google BigQuery
  10. PowerExchange for Google Cloud Spanner
  11. PowerExchange for Google Cloud Storage
  12. PowerExchange for Greenplum
  13. PowerExchange for HBase
  14. PowerExchange for HDFS
  15. PowerExchange for Hive
  16. PowerExchange for JD Edwards EnterpriseOne
  17. PowerExchange for JDBC V2 (10.4.0)
  18. PowerExchange for LDAP
  19. PowerExchange for LinkedIn
  20. PowerExchange for MapR-DB (10.4.0)
  21. PowerExchange for Microsoft Azure Blob Storage
  22. PowerExchange for Microsoft Azure Cosmos DB SQL API
  23. PowerExchange for Microsoft Azure Data Lake Storage Gen1
  24. PowerExchange for Microsoft Azure Data Lake Storage Gen2
  25. PowerExchange for Microsoft Azure SQL Data Warehouse
  26. PowerExchange for Microsoft Dynamics CRM
  27. PowerExchange for MongoDB
  28. PowerExchange for Netezza
  29. PowerExchange for OData
  30. PowerExchange for Salesforce
  31. PowerExchange for Salesforce Marketing Cloud
  32. PowerExchange for SAP NetWeaver
  33. PowerExchange for Snowflake (10.4.0)
  34. PowerExchange for Tableau
  35. PowerExchange for Tableau V3
  36. PowerExchange for Teradata Parallel Transporter API
  37. PowerExchange for Twitter
  38. PowerExchange for Web Content-Kapow Katalyst
  39. Informatica Global Customer Support

PowerExchange Adapters for Informatica Release Notes

PowerExchange Adapters for Informatica Release Notes

PowerExchange for HDFS Fixed Limitations (10.4.0)

PowerExchange for HDFS Fixed Limitations (10.4.0)

Review the Release Notes of previous releases for information about previous fixed limitations.
The following table describes fixed limitations:
Bug
Description
OCON-19803
When you run a mapping on the Spark engine to read data from a local complex file source, if the folders inside the source directory contains files with same names and you use a wildcard pattern to read the entire parent directory, the Data Integration Service reads data from any one of the files with the same names and there is no impact in reading from the other files.
OCON-18292
When you use the FileName port in a complex file target in Binary format and run a mapping in the native environment, the Data Integration Service writes all the files to a single folder.
OCON-18208
When you run a mapping to write data to a complex file using
Create target
option for Avro/Parquet formats with mapping flow enabled, the schema is created with primitive datatypes and the Data Integration Service skips the rows having null values.
OCON-18169
When you run a mapping on the Spark engine to read data from a complex file with ORC (Optimized Row Columnar) dayatype using the wildcard character questionmark '?' in MapR distro, the mapping fails.
BDM-14811
When you validate a mapping and select a connection parameter type that is not valid, the parameter name appears incorrectly.
This error occurs when you import a flat file from the Hadoop environment, parameterize the connection name, and change the parameter type to a type that is not valid.

0 COMMENTS

We’d like to hear from you!