Skip to content

Sync upstream v1.31.1 #339

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Conversation

aaronfern
Copy link

What this PR does / why we need it:
Sync with upstream autoscaler v1.31.1

Which issue(s) this PR fixes:
Fixes #320

Special notes for your reviewer:

Release note:

Synced changes till v1.31.1 of upstream autoscaler

k8s-ci-robot and others added 30 commits June 27, 2024 16:47
Update Azure cluster-autoscaler e2e cluster template
…ogging

vpa-recommender: Log object's namespace
Remove not yet supported parameter from admission-controller deployment
…bleDiskTypes-parsing

Fix: GCE FetchAvailableDiskTypes zones parsing
Add Azure cluster-autoscaler e2e test
Update Azure cluster-autoscaler e2e test
Change GCE AnnotationNodeInfoProvider nested provider to interface instead of concrete type
This reverts commit 7a1aea1.

As per
kubernetes#6788 (comment)
and discussion in DM. The preference is to split the string inside the
main.go file.
Since tests will take 10 seconds on each pass to run
cluster-autoscaler Azure e2e: move Helm to ginkgo
@gardener-robot-ci-3 gardener-robot-ci-3 added the reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) label Dec 3, 2024
@gardener-robot-ci-2 gardener-robot-ci-2 removed the reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) label Dec 3, 2024
@gardener-robot-ci-2 gardener-robot-ci-2 added reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) and removed reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) labels Dec 3, 2024
@gardener-robot-ci-2 gardener-robot-ci-2 added reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) and removed reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) labels Dec 3, 2024
@gardener-robot-ci-1 gardener-robot-ci-1 added reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) and removed reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) labels Dec 9, 2024
@gardener-robot-ci-2 gardener-robot-ci-2 added reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) and removed reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) labels Mar 5, 2025
@gardener-robot-ci-1 gardener-robot-ci-1 added the reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) label Mar 5, 2025
@gardener-robot-ci-3 gardener-robot-ci-3 removed the reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) label Mar 5, 2025
Copy link

@elankath elankath left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

/lgtm

@gardener-robot gardener-robot added reviewed/lgtm Has approval for merging and removed needs/review Needs review needs/second-opinion Needs second review by someone else labels Mar 7, 2025
@gardener-robot-ci-3 gardener-robot-ci-3 added the reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) label Mar 7, 2025
@aaronfern
Copy link
Author

Integration test run

=========================================================================================================
Random Seed: 1741241542

Will run 12 of 12 specs
------------------------------
[BeforeSuite]
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:30
Scaling Cluster Autoscaler to 0 replicas
  STEP: Adjusting node groups to initial required size @ 03/06/25 11:42:29.728
  STEP: Marking nodes present before the tests as unschedulable @ 03/06/25 11:42:30.129
  STEP: Taint node ip-10-180-24-84.eu-west-1.compute.internal @ 03/06/25 11:42:30.327
  STEP: Starting Cluster Autoscaler.... @ 03/06/25 11:42:31.085
[BeforeSuite] PASSED [4.827 seconds]
------------------------------
Cluster Autoscaler test Scale up and down nodes by deploying new workload requesting more resources should not lead to any errors and add 1 more node in target cluster
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:100
  STEP: Checking autoscaler process is running @ 03/06/25 11:42:31.097
  STEP: Deploying workload... @ 03/06/25 11:42:31.097
  STEP: Validating Scale up @ 03/06/25 11:42:31.3
• [118.031 seconds]
------------------------------
Cluster Autoscaler test Scale up and down nodes by scaling deployed workload to 3 replicas should not lead to any errors and add 3 more node in target cluster
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:115
  STEP: Checking autoscaler process is running @ 03/06/25 11:44:29.126
  STEP: Scaling up workload to 3 replicas... @ 03/06/25 11:44:29.126
  STEP: Validating Scale up @ 03/06/25 11:44:29.525
• [96.303 seconds]
------------------------------
Cluster Autoscaler test Scale up and down nodes by scaling down the deployed workload to 0 should not lead to any errors and 3 nodes to be removed from the target cluster
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:130
  STEP: Checking autoscaler process is running @ 03/06/25 11:46:05.525
  STEP: Scaling down workload to zero... @ 03/06/25 11:46:05.525
  STEP: Validating Scale down @ 03/06/25 11:46:05.921
  STEP: Deleting workload @ 03/06/25 11:47:16.083
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:47:16.283
• [70.959 seconds]
------------------------------
Cluster Autoscaler test testing annotation to skip scaledown by adding annotation and then scaling the workload to zero should not scale down the extra node and should log correspondingly
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:147
  STEP: Checking autoscaler process is running @ 03/06/25 11:47:16.484
  STEP: adding the annotation after deploy workload to 1 @ 03/06/25 11:47:16.484
  STEP: Validating Scale up @ 03/06/25 11:47:16.685
  STEP: getting the latest added node and adding annotation to it. @ 03/06/25 11:48:40.957
  STEP: Scaling down workload to zero... @ 03/06/25 11:48:41.366
• [103.366 seconds]
------------------------------
Cluster Autoscaler test testing annotation to skip scaledown by adding annotation and then scaling the workload to zero Should remove the unwanted node once scale down disable annotation is removed
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:172
  STEP: Checking autoscaler process is running @ 03/06/25 11:48:59.849
  STEP: getting the latest added node and removing annotation from it. @ 03/06/25 11:49:00.444
  STEP: Validating Scale down @ 03/06/25 11:49:00.855
  STEP: Deleting workload @ 03/06/25 11:49:52.795
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:49:52.995
• [53.353 seconds]
------------------------------
Cluster Autoscaler test testing min and max limit for Cluster autoscaler by increasing the workload to above max shouldn't scale beyond max number of workers
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:193
  STEP: Checking autoscaler process is running @ 03/06/25 11:49:53.202
  STEP: Deploying workload with replicas = max+4 @ 03/06/25 11:49:53.202
  STEP: Validating Scale up @ 03/06/25 11:49:53.403
• [93.873 seconds]
------------------------------
Cluster Autoscaler test testing min and max limit for Cluster autoscaler by decreasing the workload to below min shouldn't scale down beyond min number of workers
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:205
  STEP: Checking autoscaler process is running @ 03/06/25 11:51:27.074
  STEP: Scaling down workload to zero... @ 03/06/25 11:51:27.075
  STEP: Validating Scale down @ 03/06/25 11:51:27.495
  STEP: Deleting workload @ 03/06/25 11:52:35.361
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:52:35.565
• [68.695 seconds]
------------------------------
Cluster Autoscaler test testing scaling due to taints make current nodes unschedulable should spawn more nodes for accommodating new workload
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:221
  STEP: Checking autoscaler process is running @ 03/06/25 11:52:35.769
  STEP: making the only node available, to be unschedulable @ 03/06/25 11:52:35.769
  STEP: Taint node ip-10-180-79-175.eu-west-1.compute.internal @ 03/06/25 11:52:35.968
  STEP: Increasing the workload @ 03/06/25 11:52:36.375
  STEP: Validating Scale up @ 03/06/25 11:52:36.577
• [99.505 seconds]
------------------------------
Cluster Autoscaler test testing scaling due to taints make current nodes unschedulable should remove the node as the taint has been removed and node has low utilization
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:238
  STEP: Checking autoscaler process is running @ 03/06/25 11:54:15.274
  STEP: making the node available to be schedulable @ 03/06/25 11:54:15.274
  STEP: Removing taint(s) from node ip-10-180-79-175.eu-west-1.compute.internal @ 03/06/25 11:54:15.476
  STEP: Validating Scale down @ 03/06/25 11:54:15.886
  STEP: Deleting workload @ 03/06/25 11:55:22.872
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:55:23.072
• [68.000 seconds]
------------------------------
Cluster Autoscaler test testing scaling due to volume pending create a volume in a zone with no node and a pod requesting it should create a node in the zone with volume and hence scale up
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:257
  STEP: Checking autoscaler process is running @ 03/06/25 11:55:23.274
  STEP: Creating StorageClass with topology restrictions @ 03/06/25 11:55:23.274
  STEP: deploying PVC in zone with no nodes @ 03/06/25 11:55:23.471
  STEP: deploying the workload which requires the volume @ 03/06/25 11:55:23.673
  STEP: Validation scale up to +1 in a new Zone @ 03/06/25 11:55:23.872
  STEP: Removing pvc created earlier @ 03/06/25 11:56:55.585
  STEP: Removing storage Class created earlier @ 03/06/25 11:56:55.784
  STEP: Deleting workload @ 03/06/25 11:56:55.984
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:56:56.185
• [157.554 seconds]
------------------------------
Cluster Autoscaler test testing not able to scale due to excess demand create a pod requiring more resources than a single machine can provide shouldn't scale up and log the error
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:325
  STEP: Checking autoscaler process is running @ 03/06/25 11:58:00.828
  STEP: Deploying the workload @ 03/06/25 11:58:00.828
  STEP: checking that scale up didn't trigger because of no machine satisfying the requirement @ 03/06/25 11:58:01.038
  STEP: Deleting workload @ 03/06/25 11:58:11.093
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:58:11.294
• [10.865 seconds]
------------------------------
Cluster Autoscaler test testing CA behaviour when MCM is offline When the available replicas of MCM are zero. The CA should suspend it's operations as long as MCM is offline
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:402
  STEP: Checking autoscaler process is running @ 03/06/25 11:58:11.693
  STEP: Scaling down the MCM @ 03/06/25 11:58:11.693
  STEP: Deleting workload @ 03/06/25 11:58:21.833
  STEP: Checking that number of Ready nodes is equal to initial @ 03/06/25 11:58:22.03
• [10.736 seconds]
------------------------------
[AfterSuite]
/Users/i544000/go/src/k8s.io/autoscaler/cluster-autoscaler/integration/integration_test.go:31
  STEP: Running CleanUp @ 03/06/25 11:58:22.429
  STEP: Deleting workload @ 03/06/25 11:58:22.429
  STEP: Adjusting node groups to initial required size @ 03/06/25 11:58:23.961
  STEP: Deleting workload @ 03/06/25 11:58:24.362
  STEP: Turning nodes present before the tests, back to schedulable @ 03/06/25 11:58:24.557
  STEP: Removing taint(s) from node ip-10-180-24-84.eu-west-1.compute.internal @ 03/06/25 11:58:24.759
  STEP: Removing taint(s) from node ip-10-180-79-175.eu-west-1.compute.internal @ 03/06/25 11:58:25.174
  STEP: Scaling CA back up to 1 in the Shoot namespace @ 03/06/25 11:58:25.592
Scaling Cluster Autoscaler to 1 replicas
[AfterSuite] PASSED [3.813 seconds]
------------------------------

Ran 12 of 12 Specs in 959.884 seconds
SUCCESS! -- 12 Passed | 0 Failed | 0 Pending | 0 Skipped
PASS

Ginkgo ran 1 suite in 16m4.162160584s
Test Suite Passed

@gardener-robot gardener-robot added the needs/second-opinion Needs second review by someone else label Mar 7, 2025
@gardener-robot-ci-2 gardener-robot-ci-2 removed the reviewed/ok-to-test Has approval for testing (check PR in detail before setting this label because PR is run on CI/CD) label Mar 7, 2025
@gardener-robot gardener-robot removed the reviewed/lgtm Has approval for merging label Mar 7, 2025
@aaronfern aaronfern merged commit e7353a5 into gardener:machine-controller-manager-provider Mar 11, 2025
10 of 11 checks passed
@gardener-robot gardener-robot added the status/closed Issue is closed (either delivered or triaged) label Mar 11, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
needs/ok-to-test Needs approval for testing (check PR in detail before setting this label because PR is run on CI/CD) needs/second-opinion Needs second review by someone else size/xl Size of pull request is huge (see gardener-robot robot/bots/size.py) status/closed Issue is closed (either delivered or triaged)
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Issue for - Sync with Upstream v1.31