-
Couldn't load subscription status.
- Fork 560
Adding functionality for metadata communication across hosts #9570
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
|
The approach seems good. However I don't think this should be merged until the corresponding PyTorch PR has been merged and these tests can be activated. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM pending the follow-up from #9570 (comment).
This PR implement pipeline parallelism support for XLA devices with cross-host metadata communication capabilities. And has been tested on NEURON devices.
Key points -
test_basic_pipelining.py)xla_pipeline_stage_coordinator.py) useful for cross-host communications and used during shape_inference of torch pp codeCorresponding PR on pytorch - pytorch/pytorch#161017