Table of Contents

Search

  1. Abstract
  2. PowerExchange for Amazon Redshift
  3. PowerExchange for Amazon S3
  4. PowerExchange for Cassandra
  5. PowerExchange for Cassandra JDBC
  6. PowerExchange for DataSift
  7. PowerExchange for Facebook
  8. PowerExchange for Google Analytics
  9. PowerExchange for Google BigQuery
  10. PowerExchange for Google Cloud Spanner
  11. PowerExchange for Google Cloud Storage
  12. PowerExchange for Greenplum
  13. PowerExchange for HBase
  14. PowerExchange for HDFS (10.4.1)
  15. PowerExchange for HDFS (10.4.0)
  16. PowerExchange for Hive
  17. PowerExchange for JD Edwards EnterpriseOne
  18. PowerExchange for JDBC V2
  19. PowerExchange for LDAP
  20. PowerExchange for LinkedIn
  21. PowerExchange for MapR-DB
  22. PowerExchange for Microsoft Azure Blob Storage
  23. PowerExchange for Microsoft Azure Cosmos DB SQL API
  24. PowerExchange for Microsoft Azure Data Lake Storage Gen1
  25. PowerExchange for Microsoft Azure Data Lake Storage Gen2
  26. PowerExchange for Microsoft Azure SQL Data Warehouse
  27. PowerExchange for Microsoft Dynamics CRM
  28. PowerExchange for MongoDB
  29. PowerExchange for Netezza
  30. PowerExchange for OData
  31. PowerExchange for Salesforce
  32. PowerExchange for Salesforce Marketing Cloud
  33. PowerExchange for SAP NetWeaver
  34. PowerExchange for Snowflake
  35. PowerExchange for Tableau
  36. PowerExchange for Tableau V3
  37. PowerExchange for Teradata Parallel Transporter API
  38. PowerExchange for Twitter
  39. PowerExchange for Web Content-Kapow Katalyst
  40. Informatica Global Customer Support

PowerExchange Adapters for Informatica Release Notes

PowerExchange Adapters for Informatica Release Notes

PowerExchange for HDFS Fixed Limitations (10.4.0)

PowerExchange for HDFS Fixed Limitations (10.4.0)

Review the Release Notes of previous releases for information about previous fixed limitations.
The following table describes fixed limitations:
Bug
Description
OCON-19803
When you run a mapping on the Spark engine to read data from a local complex file source, if the folders inside the source directory contains files with same names and you use a wildcard pattern to read the entire parent directory, the Data Integration Service reads data from any one of the files with the same names and there is no impact in reading from the other files.
OCON-18292
When you use the FileName port in a complex file target in Binary format and run a mapping in the native environment, the Data Integration Service writes all the files to a single folder.
OCON-18208
When you run a mapping to write data to a complex file using
Create target
option for Avro/Parquet formats with mapping flow enabled, the schema is created with primitive datatypes and the Data Integration Service skips the rows having null values.
OCON-18169
When you run a mapping on the Spark engine to read data from a complex file with ORC (Optimized Row Columnar) dayatype using the wildcard character questionmark '?' in MapR distro, the mapping fails.
BDM-14811
When you validate a mapping and select a connection parameter type that is not valid, the parameter name appears incorrectly.
This error occurs when you import a flat file from the Hadoop environment, parameterize the connection name, and change the parameter type to a type that is not valid.

0 COMMENTS

We’d like to hear from you!