Dataproc Serverless has API quota limits that are enforced at the project and region level. The quotas reset every sixty seconds (one-minute).
The following table lists the specific and default per-project Dataproc Serverless API quota types, quota limits, and the methods to which they apply.
Quota Type | Limit | API Methods or Description |
---|---|---|
ClusterOperationRequestsPerMinutePerProjectPerRegion | 200 | CancelOperation (cancelling batch operation quota is shared with cancelling cluster operation quota). |
BatchOperationRequestsPerMinutePerProjectPerRegion | 200 | CreateBatch, DeleteBatch |
SessionOperationRequestsPerMinutePerProjectPerRegion | 200 | CreateSession, DeleteSession, TerminateSession |
DefaultRequestsPerMinutePerProjectPerRegion | 7500 | GetBatch, ListBatches, GetSession, ListSessions |
ActiveOperationsPerProjectPerRegion | 5000 | Limit on the total number of concurrent active operations of all types in a project in a region. |
Other Google Cloud quotas
Dataproc Serverless batches utilize other Google Cloud products. These products have project-level quotas, which include quotas that apply to Dataproc Serverless use. Some services are required to use Dataproc Serverless, such as Compute Engine and Cloud Storage. Other services, such as BigQuery and Bigtable, can optionally be used with Dataproc Serverless.
Required services
The following services, which enforce quota limits, are required to create Dataproc Serverless batches.
Compute Engine
Dataproc Serverless batches consume the following Compute Engine resource quotas:
Compute Tier | Quota |
---|---|
Standard | CPUS |
Premium | N2_CPUS |
Disk Tier | Quota |
Standard | DISKS_TOTAL_GB |
Premium | LOCAL_SSD_TOTAL_GB_PER_VM_FAMILY with N2 VM Family |
GPU accelerator | Quota |
L4 | NVIDIA_L4_GPUS |
A100 40GB | NVIDIA_A100_GPUS |
A100 80GB | NVIDIA_A100_80GB_GPUS |
The Compute Engine quotas
are split into regional and global limits. These limits apply to batches you
create. For example, to run a Spark batch with 4 driver cores
(spark.driver.cores=4
) and two executors with 4 cores each (spark.executor.cores=4
),
you will use 12 virtual CPUs (4 * 3
).
This batch usage will count against the regional quota limit of 24 virtual CPUs.
Default batch resources
When you create a batch with default settings, the following Compute Engine resources are used:
Item | Resources used |
---|---|
Virtual CPUs | 12 |
Virtual Machine (VM) Instances | 3 |
Persistent disk | 1200 GiB |
Cloud Logging
Dataproc Serverless saves batch output and logs in Cloud Logging. The Cloud Logging quota applies to your Dataproc Serverless batches.
Optional services
The following services, which have quota limits, can optionally be used with Dataproc Serverless batches.
BigQuery
When reading or writing data into BigQuery, the BigQuery quota applies.
Bigtable
When reading or writing data into Bigtable, the Bigtable quota applies.