transformers icon indicating copy to clipboard operation
transformers copied to clipboard

[Trainer] Add optional communication backends for torch.distributed when using GPU

Open heya5 opened this issue 2 years ago • 1 comments

What does this PR do?

Add optional backends for torch.distributed when using GPU. I want to use other communication backends according the pytorch_distribution_tutorial, but I found Trainer only uses nccl when self.no_cuda is False . Fixes # (issue)

Before submitting

  • [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • [x] Did you read the contributor guideline, Pull Request section?
  • [ ] Was this discussed/approved via a Github issue or the forum? Please add a link to it if that's the case.
  • [ ] Did you make sure to update the documentation with your changes? Here are the documentation guidelines, and here are tips on formatting docstrings.
  • [ ] Did you write any new necessary tests?

Who can review?

  • trainer: @sgugger

heya5 avatar Mar 18 '23 17:03 heya5

The documentation is not available anymore as the PR was closed or merged.