Table of Contents

Search

  1. Preface
  2. Advanced clusters
  3. Setting up AWS
  4. Setting up Google Cloud
  5. Setting up Microsoft Azure
  6. Setting up a self-service cluster
  7. Setting up a local cluster
  8. Advanced configurations
  9. Troubleshooting
  10. Appendix A: Command reference

Advanced Clusters

Advanced Clusters

Reconfiguring resource requirements

Reconfiguring resource requirements

If you cannot provision enough resources to fulfill the default requirements, you can reconfigure some of the requirements.
You can reconfigure the requirements for the following components:
Spark shuffle service
If you disable the shuffle service, the
Spark engine
cannot use dynamic allocation. For more details, contact Informatica Global Customer Support.
Spark driver
To reconfigure the amount of memory for the Spark driver, use the Spark session property
spark.driver.memory
in the
mapping
task. To set the memory in terms of GB, use a value such as 2G. To set the memory in terms of MB, use a value such as 1500m.
For information about reconfiguring the CPU requirement for the Spark driver, contact Informatica Global Customer Support.
Spark executor
To reconfigure the amount of memory for the Spark executor, use the Spark session property
spark.executor.memory
in the
mapping
task. Similar to the memory value for the Spark driver, you can specify the memory in GB or MB.
You can also change the number of Spark executor cores using the Spark session property
spark.executor.cores
. The default number of cores for GPU-enabled clusters is 4. The default number of cores for all other clusters is 2.
If you edit the number of cores, you change the number of Spark tasks that run concurrently. For example, two Spark tasks can run concurrently inside each Spark executor when you set spark.executor.cores=2.
For information about reconfiguring the CPU requirement for Spark executors, contact Informatica Global Customer Support.
If you reduce the memory too low for the Spark driver and Spark executor, these components might encounter an OutOfMemoryException.
You cannot edit the resource requirements for the Kubernetes system. The resources are required to maintain a functional Kubernetes system.
For more information about the Spark session properties, see
Tasks
in the Data Integration help.

0 COMMENTS

We’d like to hear from you!