Search Results for author: Chang-Wei Shi

Found 3 papers, 0 papers with code

On the Optimal Batch Size for Byzantine-Robust Distributed Learning

no code implementations23 May 2023 Yi-Rui Yang, Chang-Wei Shi, Wu-Jun Li

However, for existing BRDL methods, large batch sizes will lead to a drop on model accuracy, even if there is no Byzantine attack.

Stochastic Normalized Gradient Descent with Momentum for Large-Batch Training

no code implementations28 Jul 2020 Shen-Yi Zhao, Chang-Wei Shi, Yin-Peng Xie, Wu-Jun Li

Empirical results on deep learning verify that when adopting the same large batch size, SNGM can achieve better test accuracy than MSGD and other state-of-the-art large-batch training methods.

Global Momentum Compression for Sparse Communication in Distributed Learning

no code implementations30 May 2019 Chang-Wei Shi, Shen-Yi Zhao, Yin-Peng Xie, Hao Gao, Wu-Jun Li

With the rapid growth of data, distributed momentum stochastic gradient descent~(DMSGD) has been widely used in distributed learning, especially for training large-scale deep models.

Cannot find the paper you are looking for? You can Submit a new open access paper.