Ordered Local Momentum for Asynchronous Distributed Learning under Arbitrary Delays
Momentum SGD (MSGD) serves as a foundational optimizer in training deep models due to momentum’s key role in accelerating convergence and enhancing generalization. Meanwhile, asynchronous distributed learning is crucial for training large-scale deep models, especially when the computing capabilities of the workers in the cluster are heterogeneous. To reduce communication frequency, local updates are widely adopted in distributed learning. However, how to implement asynchronous distributed MSGD with local updates remains unexplored. To solve this problem, we propose a […]