Table of Contents

Search

  1. Preface
  2. Introduction to Big Data Management Administration
  3. Big Data Management Engines
  4. Authentication and Authorization
  5. Running Mappings on a Cluster with Kerberos Authentication
  6. Configuring Access to an SSL/TLS-Enabled Cluster
  7. Cluster Configuration
  8. Cluster Configuration Privileges and Permissions
  9. Cloud Provisioning Configuration
  10. Queuing
  11. Tuning for Big Data Processing
  12. Connections
  13. Multiple Blaze Instances on a Cluster

Big Data Management Administrator Guide

Big Data Management Administrator Guide

Queuing Process

Queuing Process

The Data Integration Service queues deployed jobs before running them in the native or Hadoop batch pool. On-demand jobs run immediately in the on-demand pool.
The following diagram shows the overall queuing and execution process:
The diagram shows a deployed job and an on-demand job. The deployed job goes to the queue, and then to the native or Hadoop batch pool. The on-demand job goes directly to the on-demand queue. Both jobs are completed after they are released from the pool.
When you deploy a mapping job or workflow mapping task, the Data Integration Service moves the job directly to the persisted queue for that node. If the queue is full, the Data Integration Service marks the job as failed.
You can cancel a job in the queue. A job is aborted if the node shuts down unexpectedly and the Data Integration Service is configured to discard all jobs in the queue upon restart.
When resources are available, the Data Integration Service moves the job to the execution pool and starts running the job. A deployed job runs in one of the following execution pools:
Native Batch Pool
Runs deployed native jobs.
Hadoop Batch Pool
Runs deployed Hadoop jobs.
You can cancel a running job, or the job may be aborted if the node shuts down unexpectedly. A job can also fail while running.
The Data Integration Service marks successful jobs as completed.
The Data Integration Service immediately starts running on-demand jobs. If you run more jobs than the
On-Demand Pool
can run concurrently, the extra jobs fail. You must manually run the jobs again when space is available.
The following table describes the mapping job states in the Administrator tool contents panel:
Job Status
Rules and Guidelines
Queued
The job is in the queue.
Running
The Data Integration Service is running the job.
Completed
The job ran successfully.
Aborted
The job was flushed from the queue at restart or the node shut down unexpectedly while the job was running.
Failed
The job failed while running or the queue is full.
Canceled
The job was deleted from the queue or cancelled while running.
Unknown
The job status is unknown.

0 COMMENTS

We’d like to hear from you!