spark-operator icon indicating copy to clipboard operation
spark-operator copied to clipboard

Reject spark application submit when cluster capacity is reached

Open hiradha opened this issue 1 year ago • 1 comments

We use spark operator to submit spark jobs on our k8s cluster. However many times the cluster will have no capacity or will have capacity to run only one of out the three requested executors.

We need the functionality to reject such spark job submissions - rather than accepting them and the executors stay Pending.

Is such a functionality available? If not, which files do I need to start taking a look at in case I can contribute a PR.

hiradha avatar Oct 25 '23 20:10 hiradha

Take a look at the resource quota enforcement built into the operator https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/user-guide.md#enabling-resource-quota-enforcement. Although I'd also encourage you to take a look at alternative Kubernetes schedulers that fit better with batch workloads such as Volcano or Yunikorn to see if they fit your requirements.

jacobsalway avatar Nov 02 '23 11:11 jacobsalway

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

github-actions[bot] avatar Aug 14 '24 04:08 github-actions[bot]