Research on three-step accelerated gradient algorithm in deep learning
Gradient descent (GD) algorithm is the widely used optimisation method in training machine learning and deep learning models. In this paper, based on GD, Polyak's momentum (PM), and Nesterov accelerated gradient (NAG), we give the convergence of the algorithms from an initial value to the optim...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Taylor & Francis Group
2022-01-01
|
Series: | Statistical Theory and Related Fields |
Subjects: | |
Online Access: | http://dx.doi.org/10.1080/24754269.2020.1846414 |