Replies: 1 comment 4 replies
-
To clarify, we only use Torch RPC in distributed part. The reason is primarily due to its compatibility with the PyTorch ecosystem and its ability to facilitate communication through protocols like TCP/RDMA. As for IPC and NVLink(not required distributed settings), they are implemented directly by CUDA at lower level in C++, and are not integrated through the Torch RPC framework. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello,
I have a simple question but very important to me. I am studying and researching the torch RPC module recently, and trying to contribute code to the torch community.
What role does the RPC module play in the optimization of this project, and why do we need to use RPC of torch? In this project, can the torch RPC module be further optimized only through encapsulation at the python level?
In essence, is it dependent on the ability of torch RPC or cuda's IPC and NVLink capabilities?
Looking forward to your answer.
Beta Was this translation helpful? Give feedback.
All reactions