Table of Contents

Search

  1. Preface
  2. Analyst Service
  3. Catalog Service
  4. Content Management Service
  5. Data Integration Service
  6. Data Integration Service Architecture
  7. Data Integration Service Management
  8. Data Integration Service Grid
  9. Data Integration Service Applications
  10. Interactive Data Preparation Service
  11. Enterprise Data Preparation Service
  12. Informatica Cluster Service
  13. Mass Ingestion Service
  14. Metadata Access Service
  15. Metadata Manager Service
  16. Model Repository Service
  17. PowerCenter Integration Service
  18. PowerCenter Integration Service Architecture
  19. High Availability for the PowerCenter Integration Service
  20. PowerCenter Repository Service
  21. PowerCenter Repository Management
  22. PowerExchange Listener Service
  23. PowerExchange Logger Service
  24. SAP BW Service
  25. Search Service
  26. System Services
  27. Test Data Manager Service
  28. Test Data Warehouse Service
  29. Web Services Hub
  30. Application Service Upgrade
  31. Appendix A: Application Service Databases
  32. Appendix B: Connecting to Databases from Windows
  33. Appendix C: Connecting to Databases
  34. Appendix D: Updating the DynamicSections Parameter of a DB2 Database

Rules and Guidelines when Jobs Run in Separate Processes

Rules and Guidelines when Jobs Run in Separate Processes

Consider the following rules and guidelines when you configure the Data Integration Service to run jobs in separate local or remote processes:
  • You cannot use the
    Maximum Memory Size
    property for the Data Integration Service to limit the amount of memory that the service allocates to run jobs. If you set the maximum memory size, the Data Integration Service ignores it.
  • If the Data Integration Service runs on UNIX, the host file on each node with the compute role and on each node with both the service and compute roles must contain a localhost entry. If the host file does not contain a localhost entry, jobs that run in separate processes fail. Windows does not require a localhost entry in the host file.
  • If you configure connection pooling, each DTM process maintains its own connection pool library. All DTM instances running in the DTM process can use the connection pool library. The number of connection pool libraries depends on the number of running DTM processes.