Table of Contents

Search

  1. Preface
  2. Analyst Service
  3. Catalog Service
  4. Content Management Service
  5. Data Integration Service
  6. Data Integration Service Architecture
  7. Data Integration Service Management
  8. Data Integration Service Grid
  9. Data Integration Service REST API
  10. Data Integration Service Applications
  11. Enterprise Data Preparation Service
  12. Interactive Data Preparation Service
  13. Informatica Cluster Service
  14. Mass Ingestion Service
  15. Metadata Access Service
  16. Metadata Manager Service
  17. Model Repository Service
  18. PowerCenter Integration Service
  19. PowerCenter Integration Service Architecture
  20. High Availability for the PowerCenter Integration Service
  21. PowerCenter Repository Service
  22. PowerCenter Repository Management
  23. PowerExchange Listener Service
  24. PowerExchange Logger Service
  25. SAP BW Service
  26. Search Service
  27. System Services
  28. Test Data Manager Service
  29. Test Data Warehouse Service
  30. Web Services Hub
  31. Application Service Upgrade
  32. Application Service Databases
  33. Connecting to Databases from Windows
  34. Connecting to Databases from UNIX or Linux
  35. Updating the DynamicSections Parameter of a DB2 Database

Data Preparation Storage Options

Data Preparation Storage Options

Data preparation storage options enables you to specify the local storage and HDFS location for data persistence.
To edit the data preparation storage options, click the pencil icon in the Data Preparation Storage Options area. In the
Edit Data Preparation Storage Options
window, edit the required fields.
The following table describes the data preparation storage options:
Property
Description
Local Storage Location
Directory for data preparation file storage on the node where the service runs.
HDFS Connection
HDFS connection for data preparation file storage.
HDFS Storage Location
HDFS location for data preparation file storage. If the connection to the local storage fails, the service recovers data preparation files from the HDFS location.
If the Hadoop cluster uses Kerberos authentication, the impersonation user name must have read, write and execute permission on the HDFS storage location directory. The default location is: /datalake/dps_durable_storage.