AG-SGD: Angle-Based Stochastic Gradient Descent

In the field of neural network, stochastic gradient descent is often employed as an effective method of accelerating the result's convergence. Generating the new gradient from the past gradient is a common method adopted by many existing optimization algorithms. Since the past gradient is not c...

Full description

Bibliographic Details
Main Authors: Chongya Song, Alexander Pons, Kang Yen
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9343305/
_version_ 1819276839492780032
author Chongya Song
Alexander Pons
Kang Yen
author_facet Chongya Song
Alexander Pons
Kang Yen
author_sort Chongya Song
collection DOAJ
description In the field of neural network, stochastic gradient descent is often employed as an effective method of accelerating the result's convergence. Generating the new gradient from the past gradient is a common method adopted by many existing optimization algorithms. Since the past gradient is not computed based on the most updated stochastic gradient descent state, it can introduce a deviation to the new gradient computation, negatively impacting its rate of convergence. To resolve this problem, we propose an algorithm that quantifies this deviation based on the angle between the past and the current gradients, which is then applied to calibrate these two gradients, generating a more accurate new gradient. To demonstrate the broad applicability of the algorithm, the proposed method is implemented into a neural network and a logistic regression classifier which are evaluated on the datasets MNIST and NSL-KDD, respectively. An in-depth analysis is performed to compare our algorithm with nine optimization algorithms in two experiments, demonstrating the advantages in the cost and the error rate reductions from adopting the proposed method.
first_indexed 2024-12-23T23:46:36Z
format Article
id doaj.art-e4904363afe749dbbd704516f78d6aaf
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-23T23:46:36Z
publishDate 2021-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-e4904363afe749dbbd704516f78d6aaf2022-12-21T17:25:29ZengIEEEIEEE Access2169-35362021-01-019230072302410.1109/ACCESS.2021.30559939343305AG-SGD: Angle-Based Stochastic Gradient DescentChongya Song0https://orcid.org/0000-0003-4208-5738Alexander Pons1Kang Yen2Department of Electrical and Computer Engineering, Florida International University, Miami, FL, USADepartment of Electrical and Computer Engineering, Florida International University, Miami, FL, USADepartment of Electrical and Computer Engineering, Florida International University, Miami, FL, USAIn the field of neural network, stochastic gradient descent is often employed as an effective method of accelerating the result's convergence. Generating the new gradient from the past gradient is a common method adopted by many existing optimization algorithms. Since the past gradient is not computed based on the most updated stochastic gradient descent state, it can introduce a deviation to the new gradient computation, negatively impacting its rate of convergence. To resolve this problem, we propose an algorithm that quantifies this deviation based on the angle between the past and the current gradients, which is then applied to calibrate these two gradients, generating a more accurate new gradient. To demonstrate the broad applicability of the algorithm, the proposed method is implemented into a neural network and a logistic regression classifier which are evaluated on the datasets MNIST and NSL-KDD, respectively. An in-depth analysis is performed to compare our algorithm with nine optimization algorithms in two experiments, demonstrating the advantages in the cost and the error rate reductions from adopting the proposed method.https://ieeexplore.ieee.org/document/9343305/Stochastic gradient descentoptimization algorithmgradient deviationneural network
spellingShingle Chongya Song
Alexander Pons
Kang Yen
AG-SGD: Angle-Based Stochastic Gradient Descent
IEEE Access
Stochastic gradient descent
optimization algorithm
gradient deviation
neural network
title AG-SGD: Angle-Based Stochastic Gradient Descent
title_full AG-SGD: Angle-Based Stochastic Gradient Descent
title_fullStr AG-SGD: Angle-Based Stochastic Gradient Descent
title_full_unstemmed AG-SGD: Angle-Based Stochastic Gradient Descent
title_short AG-SGD: Angle-Based Stochastic Gradient Descent
title_sort ag sgd angle based stochastic gradient descent
topic Stochastic gradient descent
optimization algorithm
gradient deviation
neural network
url https://ieeexplore.ieee.org/document/9343305/
work_keys_str_mv AT chongyasong agsgdanglebasedstochasticgradientdescent
AT alexanderpons agsgdanglebasedstochasticgradientdescent
AT kangyen agsgdanglebasedstochasticgradientdescent