Skip to content

[BUG] minResources of volcano podGroup didn't take into account dynamicAllocation and memoryOverheadFactor #2244

@kaka-zb

Description

@kaka-zb

Description

We've been using spark operator and volcano for a long time in production env, however, there are some problems with the calculation of resource usage for volcano podGroup when the sparkapp is submitted.

The spark.dynamicAllocation.* & spark.kubernetes.memoryOverheadFactor params of spark are not taken into account when calculating memory of minResources for volcano podGroup. As a result, the calculated minResources maybe smaller than real usage of sparkapp, and the gang scheduling maybe fail.

  • ✋ I have searched the open/closed issues and my issue is not listed.

Reproduction Code [Required]

Expected behavior

Actual behavior

Environment & Versions

  • Spark Operator App version: 2.0.1
  • Helm Chart Version: 2.0.1
  • Kubernetes Version: 1.25.7
  • Apache Spark version: 3.4.3

Additional context

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions