Description
We've been using spark operator and volcano for a long time in production env, however, there are some problems with the calculation of resource usage for volcano podGroup when the sparkapp is submitted.
The spark.dynamicAllocation.* & spark.kubernetes.memoryOverheadFactor params of spark are not taken into account when calculating memory of minResources for volcano podGroup. As a result, the calculated minResources maybe smaller than real usage of sparkapp, and the gang scheduling maybe fail.
Reproduction Code [Required]
Expected behavior
Actual behavior
Environment & Versions
- Spark Operator App version: 2.0.1
- Helm Chart Version: 2.0.1
- Kubernetes Version: 1.25.7
- Apache Spark version: 3.4.3
Additional context