Trend-Smooth: Accelerate Asynchronous SGD by Smoothing Parameters Using Parameter Trends

6Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Stochastic gradient descent(SGD) is the fundamental sequential method in training large scale machine learning models. To accelerate the training process, researchers proposed to use the asynchronous stochastic gradient descent (A-SGD) method in model learning. However, due to the stale information when updating parameters, A-SGD converges more slowly than SGD in the same iteration number. Moreover, A-SGD often converges to a high loss value and results in lower model accuracy. In this paper, we propose a novel algorithm called Trend-Smooth which can be adapted to the asynchronous parallel environment to overcome the above problems. Specifically, Trend-Smooth makes use of the parameter trend during the training process to shrink the learning rate of some dimensions where the gradients' directions are opposite to the trends of parameters. Experiments on MNIST and CIFAR-10 datasets confirm that Trend-Smooth can accelerate the convergence speed in asynchronous training process. The test accuracy that Trend-Smooth achieves is shown to be higher than other asynchronous parallel baseline methods, and is very close to the SGD method. Moreover, Trend-Smooth can also be combined with other adaptive learning rate methods(like Momentum, RMSProp and Adam) in the asynchronous parallel environment to promote their performance.

References Powered by Scopus

ImageNet classification with deep convolutional neural networks

23046Citations
N/AReaders
Get full text

MapReduce: Simplified data processing on large clusters

11914Citations
N/AReaders
Get full text

Optimization methods for large-scale machine learning

1984Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Power System Fault Classification and Prediction Based on a Three-Layer Data Mining Structure

20Citations
N/AReaders
Get full text

An Enhanced CoAP Scheme Using Fuzzy Logic with Adaptive Timeout for IoT Congestion Control

12Citations
N/AReaders
Get full text

Penetration Recognition in GTAW Welding Based on Time and Spectrum Images of Arc Sound Using Deep Learning Method

11Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Cui, G., Guo, J., Fan, Y., Lan, Y., & Cheng, X. (2019). Trend-Smooth: Accelerate Asynchronous SGD by Smoothing Parameters Using Parameter Trends. IEEE Access, 7, 156848–156859. https://doi.org/10.1109/ACCESS.2019.2949611

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 7

88%

Lecturer / Post doc 1

13%

Readers' Discipline

Tooltip

Computer Science 8

89%

Engineering 1

11%

Save time finding and organizing research with Mendeley

Sign up for free