Skip to content

Conversation

@luowyang
Copy link

This pull request fixes issue #1721, where single GPU training/inference may error if the worker uses torch.dist. In summary, it makes sure the default process group is always initialized as long as world_size > 0, otherwise it raises a ValueError to indicate illegal argument(s). The existing code should not be affected, as stated in issue #1721.

Rationale: Always initializing the process group is preferred, because when launch is called, the user most likely requests distributed semantics. This fix makes the user code consistent by allowing the users to make torch.dist calls even if there is only one GPU.

@CLAassistant
Copy link

CLAassistant commented Sep 26, 2023

CLA assistant check
All committers have signed the CLA.

vossr pushed a commit to vossr/YOLOX-custom that referenced this pull request Apr 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants