HSDP: Accelerating Large-scale Model Training Via Efficient Sharded Data Parallelism
International Symposium on Parallel and Distributed Processing with Applications(2024)
关键词
Deep neural network,Sharded data parallelism,Distributed training,Communication optimization
AI 理解论文
溯源树
样例

生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要