PowerExchange Adapters for Informatica
- PowerExchange Adapters for Informatica 10.4.0
- All Products
When the source data contains a slash
'\'and you create a mapping to read the source data and retain the default quote character,
0x1f, the mapping fails with the following error on the Databricks Spark engine:
When you use an ODBC data source to connect to Microsoft Azure SQL Data Warehouse and if the imported metadata contains special characters, the mapping fails.
When you run a dynamic mapping that reads date, datetime, or smalldatetime fields from a source and creates an Azure SQL Data Warehouse target at run time, the mapping fails to run on the Databricks Spark engine.
When you use the default values for field delimiter or quote character in advance source or target properties, mappings fail with incompatible data types or precision errors or the Data Integration Service writes incorrect data to the target.
When you run a mapping that contains unconnected ports on the Hive engine, the mapping might fail or data corruption might happen.
When you upsert or update data to Microsoft Azure SQL Data Warehouse and more than one column in the source table contains same value as the target column on which the primary key is defined, the Data Integration Service updates data incorrectly.
When you run a mapping on the Hive engine to read data from or write data to Microsoft Azure SQL Data Warehouse, the intermediate files get downloaded in the staging directory even if you cancel the mapping.