Hi @jeyob ,
1. Using Another Patch Version
The error message you’re encountering when trying to specify a subminor version (like 2.1.23 instead of 2.1.24) suggests that Google Cloud Dataproc Serverless does not support specifying such detailed version granularity for the runtime. This limitation means you can’t directly choose a specific patch version if it’s not explicitly offered by Google Cloud.
Possible Solution: Since you can’t specify subminor versions, you’re limited to the versions explicitly offered by Google Cloud. You can check the available versions in the Google Cloud Console or using the gcloud command-line tool to see if there are alternative versions you can use.
2. Creating a New Custom Image
Regarding creating a new custom image, you’re correct that the Spark runtime is mounted into the custom image in Dataproc Serverless. When I suggested creating a new custom image, it was under the assumption that there might be additional configurations or dependencies that you could include in your custom container to ensure compatibility with the autoscaling feature. However, if the runtime itself (which is mounted) does not support certain features, then customizing the container might not resolve the issue.
Clarification: If the limitation lies within the Spark runtime version itself, then customizing the container won’t help. In this case, you’re dependent on the versions and features supported by Google Cloud’s provided runtimes.
3. Understanding Control Plane Decommissioning
The concept of “control plane decommissioning” in the context of Dataproc Serverless autoscaling is a bit complex. It generally refers to the ability of the control plane (the management layer of the cluster) to dynamically scale down resources when they are no longer needed, in a way that doesn’t disrupt running jobs. This is a key feature for efficient autoscaling.
4. Verifying Autoscaling Version 2 is Enabled
To verify that autoscaling version 2 is enabled, you can:
Check Job Configuration: When you submit a job, ensure that the property spark.dataproc.scaling.version is set to 2. This should be part of your job submission command or configuration file.
Monitor the Job: Once the job is running, you can monitor its behavior in the Google Cloud Console. Autoscaling version 2 should exhibit more dynamic scaling behavior compared to version 1, particularly in how it scales down resources.
Logs and Metrics: Check the logs and metrics of your Dataproc Serverless job. There might be specific logs or metrics that indicate which version of autoscaling is being used.
Note: The exact steps to verify this might vary based on the tools and interfaces you are using (Google Cloud Console, gcloud CLI, etc.).
Given these points, your best course of action might be to work with the versions available to you and monitor Google Cloud’s updates for any changes in supported runtime versions or features. If the issue persists or if you need more specific guidance, contacting Google Cloud support could provide more tailored assistance.