Zobrazeno 1 - 2
of 2
pro vyhledávání: '"Yao Chengji"'
Autor:
Chang, Li-Wen, Bao, Wenlei, Hou, Qi, Jiang, Chengquan, Zheng, Ningxin, Zhong, Yinmin, Zhang, Xuanrun, Song, Zuquan, Yao, Chengji, Jiang, Ziheng, Lin, Haibin, Jin, Xin, Liu, Xin
Large deep learning models have demonstrated strong ability to solve many tasks across a wide range of applications. Those large models typically require training and inference to be distributed. Tensor parallelism is a common technique partitioning
Externí odkaz:
http://arxiv.org/abs/2406.06858
Autor:
Junmin Xiao, Ninghui Sun, Hu Zhongzhe, Tian Zhongbo, Zhu Hongrui, Yao Chengji, Guangming Tan, Xiaoyang Zhang
Publikováno v:
ISPA/BDCloud/SocialCom/SustainCom
Large batch distributed synchronous stochastic gradient descent (SGD) has been widely used to train deep neural networks on a distributed memory system with multi-nodes, which can leverage parallel resources to reduce the number of iterative steps an