Pytorch distributed: Experiences on accelerating data parallel training S Li, Y Zhao, R Varma, O Salpekar, P Noordhuis, T Li, A Paszke, J Smith, ... arXiv preprint arXiv:2006.15704, 2020 | 452 | 2020 |
Pytorch fsdp: experiences on scaling fully sharded data parallel Y Zhao, A Gu, R Varma, L Luo, CC Huang, M Xu, L Wright, H Shojanazeri, ... arXiv preprint arXiv:2304.11277, 2023 | 76 | 2023 |
Shen Li Y Zhao, A Gu, R Varma, L Luo, CC Huang, M Xu, L Wright, H Shojanazeri, ... Pytorch fsdp: Experiences on scaling fully sharded data parallel, 2023 | 7 | 2023 |
PyTorch Distributed: Experiences on accelerating data parallel training. CoRR abs/2006.15704 (2020) S Li, Y Zhao, R Varma, O Salpekar, P Noordhuis, T Li, A Paszke, J Smith, ... arXiv preprint arXiv:2006.15704, 2020 | 7 | 2020 |
Pytorch rpc: Distributed deep learning built on tensor-optimized remote procedure calls P Damania, S Li, A Desmaison, A Azzolini, B Vaughan, E Yang, ... Proceedings of Machine Learning and Systems 5, 2023 | 4 | 2023 |