Skip to content

Alpha e2e tests fail often #9052

@afrittoli

Description

@afrittoli

Expected Behavior

Tests are successful most of the times

Actual Behavior

From the latest failure:

arendelle-dx92j              9m55s       Normal    PullImageFailed                taskrun/pipelinerun-consume-tasks-artifactswsxvb-consume-artifacts        build step "step-artifacts-consumer-python" is pending with reason "failed to pull and unpack image \"docker.io/library/python:latest\": failed to extract layer sha256:08e14ec5b7497da231d70d47d1d80440ba7d9997d43c0796a8394923bbc98183: write /var/lib/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/631/fs/usr/lib/x86_64-linux-gnu/libicui18n.a: no space left on device"
arendelle-dx92j              9m41s       Normal    PullImageFailed                taskrun/pipelinerun-consume-tasks-artifactswsxvb-consume-artifacts        build step "step-artifacts-consumer-python" is pending with reason "Back-off pulling image \"python:latest\": ErrImagePull: failed to pull and unpack image \"docker.io/library/python:latest\": failed to extract layer sha256:08e14ec5b7497da231d70d47d1d80440ba7d9997d43c0796a8394923bbc98183: write /var/lib/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/631/fs/usr/lib/x86_64-linux-gnu/libicui18n.a: no space left on device"
arendelle-dx92j              9m41s       Warning   Failed                         taskrun/pipelinerun-consume-tasks-artifactswsxvb-consume-artifacts        the step "artifacts-consumer-python" in TaskRun "pipelinerun-consume-tasks-artifactswsxvb-consume-artifacts" failed to pull the image "". The pod errored with the message: "Back-off pulling image "python:latest": ErrImagePull: failed to pull and unpack image "docker.io/library/python:latest": failed to extract layer sha256:08e14ec5b7497da231d70d47d1d80440ba7d9997d43c0796a8394923bbc98183: write /var/lib/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/631/fs/usr/lib/x86_64-linux-gnu/libicui18n.a: no space left on device."

Since this seems to happen mainly on alpha, I imagine the test in question only runs with alpha.
We could change the specific test to use a smaller image and/or change the configuration of the kind cluster to increase the disk space for the cluster nodes (to be verified if possible).

Steps to Reproduce the Problem

  1. Create a PR or check https://github.com/tektoncd/pipeline/actions/runs/18131929619/job/51630667788?pr=9050

Additional Info

  • Kubernetes version:

Happens on various k8s versions

  • Tekton Pipeline version:

main

Metadata

Metadata

Assignees

Labels

area/testingIssues or PRs related to testingkind/flakeCategorizes issue or PR as related to a flakey testpriority/critical-urgentHighest priority. Must be actively worked on as someone's top priority right now.

Type

No type

Projects

Status

Todo

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions