Skip to content

Change GenAI OSS runner to fix OOM #4082

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

spcyppt
Copy link
Contributor

@spcyppt spcyppt commented May 6, 2025

Summary:
X-link: https://github.com/facebookresearch/FBGEMM/pull/1166

Switch to runners with larger memory runner to fix OOM for GenAI OSS CI.

The GenAI OSS CI build jobs fail with

The self-hosted runner lost communication with the server. Verify the machine is running and has a healthy network connection. Anything in your workflow that terminates the runner process, starves it for CPU/Memory, or blocks its network access can cause this error.

The runners for CUDA12.8 for python 3.10 3.12 3.13 consistently failed.
For example:
https://github.com/pytorch/FBGEMM/actions/runs/14731 902981/job/41348012076
https://github.com/pytorch/FBGEMM/actions/runs/14727090721/job/41332229097
https://github.com/pytorch/FBGEMM/actions/runs/14722058017/job/41317562384

Cause from huydhn:

  • The error happens when the runner runs out of memory.
  • This is a common bottleneck for the build job.
  • linux.24.large probably spawns too many processes given its higher number of CPU cores and OOM

Differential Revision: D74221851

Summary:
X-link: facebookresearch/FBGEMM#1166


Switch to runners with larger memory runner to fix OOM for GenAI OSS CI.

The GenAI OSS CI build jobs fail with

```
The self-hosted runner lost communication with the server. Verify the machine is running and has a healthy network connection. Anything in your workflow that terminates the runner process, starves it for CPU/Memory, or blocks its network access can cause this error.
```

The runners for CUDA12.8 for python 3.10 3.12 3.13 consistently failed.
For example:
https://github.com/pytorch/FBGEMM/actions/runs/14731 902981/job/41348012076
https://github.com/pytorch/FBGEMM/actions/runs/14727090721/job/41332229097
https://github.com/pytorch/FBGEMM/actions/runs/14722058017/job/41317562384

Cause from huydhn:
- The error happens when the runner runs out of memory. 
- This is a common bottleneck for the build job.
- linux.24.large probably spawns too many processes given its higher number of CPU cores and OOM

Differential Revision: D74221851
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D74221851

Copy link

netlify bot commented May 6, 2025

Deploy Preview for pytorch-fbgemm-docs ready!

Name Link
🔨 Latest commit b7699b3
🔍 Latest deploy log https://app.netlify.com/sites/pytorch-fbgemm-docs/deploys/681975a31c5db50008818df1
😎 Deploy Preview https://deploy-preview-4082--pytorch-fbgemm-docs.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants