-
1
Low-Rank Gradient Descent
Published 2023-01-01“…In this article, we leverage such low-rank structure to reduce the high computational cost of canonical gradient-based methods such as gradient descent (<monospace>GD</monospace>). Our proposed <italic>Low-Rank Gradient Descent</italic> (<monospace>LRGD</monospace>) algorithm finds an <inline-formula><tex-math notation="LaTeX">$\epsilon$</tex-math></inline-formula>-approximate stationary point of a <inline-formula><tex-math notation="LaTeX">$p$</tex-math></inline-formula>-dimensional function by first identifying <inline-formula><tex-math notation="LaTeX">$r \leq p$</tex-math></inline-formula> significant directions, and then estimating the true <inline-formula><tex-math notation="LaTeX">$p$</tex-math></inline-formula>-dimensional gradient at every iteration by computing directional derivatives only along those <inline-formula><tex-math notation="LaTeX">$r$</tex-math></inline-formula> directions. …”
Get full text
Article -
2
On-manifold projected gradient descent
Published 2024-02-01“…The tools are applied to the setting of neural network image classifiers, where we generate novel, on-manifold data samples and implement a projected gradient descent algorithm for on-manifold adversarial training. …”
Get full text
Article -
3
Correspondence between neuroevolution and gradient descent
Published 2021-11-01“…The authors derive, and illustrate numerically, an analytic equivalence between the dynamics of neural network training under conditioned stochastic mutations, and under gradient descent.…”
Get full text
Article -
4
Semi-Stochastic Gradient Descent Methods
Published 2017-05-01“…We propose a new method, S2GD (Semi-Stochastic Gradient Descent), which runs for one or several epochs in each of which a single full gradient and a random number of stochastic gradients is computed, following a geometric law. …”
Get full text
Article -
5
-
6
Carathéodory sampling for stochastic gradient descent
Published 2020“…Many problems require to optimize empirical risk functions over large data sets. Gradient descent methods that calculate the full gradient in every descent step do not scale to such datasets. …”
Internet publication -
7
Carathéodory sampling for stochastic gradient descent
Published 2020“…Many problems require to optimize empirical risk functions over large data sets. Gradient descent methods that calculate the full gradient in every descent step do not scale to such datasets. …”
Internet publication -
8
Dual space preconditioning for gradient descent
Published 2021“…Thus, in principle our method is capable of improving the conditioning of gradient descent on problems with a non-Lipschitz gradient or nonstrongly convex structure. …”
Journal article -
9
Pipelined Stochastic Gradient Descent with Taylor Expansion
Published 2023-10-01Subjects: Get full text
Article -
10
Limited Gradient Descent: Learning With Noisy Labels
Published 2019-01-01“…To solve this problem, we propose a method that can estimate the optimal stopping timing without a clean validation set, called limited gradient descent. We modified the labels of a few samples in a noisy dataset to obtain false labels and to create a reverse pattern. …”
Get full text
Article -
11
Accelerated Gradient Descent Driven by Lévy Perturbations
Published 2024-03-01Subjects: “…accelerated gradient descent…”
Get full text
Article -
12
Gradient Descent Batch Clustering for Image Classification
Published 2023-07-01Subjects: Get full text
Article -
13
Stochastic gradient descent for optimization for nuclear systems
Published 2023-05-01“…ADAM is a gradient descent method that accounts for gradients with a stochastic nature. …”
Get full text
Article -
14
Stochastic gradient descent for wind farm optimization
Published 2023-08-01“…This study presents stochastic gradient descent (SGD) for wind farm optimization, which is an approach that estimates the gradient of the AEP using Monte Carlo simulation, allowing for the consideration of an arbitrarily large number of atmospheric conditions. …”
Get full text
Article -
15
NOMA Codebook Optimization by Batch Gradient Descent
Published 2019-01-01Subjects: “…Batch gradient descent…”
Get full text
Article -
16
Complexity control by gradient descent in deep networks
Published 2020-02-01“…Here, the author demonstrates an implicit regularization in training deep networks, showing that the control of complexity in the training is hidden within the optimization technique of gradient descent.…”
Get full text
Article -
17
Gradient descent optimization in gene regulatory pathways.
Published 2010-01-01Get full text
Article -
18
Complexity control by gradient descent in deep networks
Published 2021“…For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification.…”
Get full text
Article -
19
Complexity control by gradient descent in deep networks
Published 2022“…For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification.…”
Get full text
Article -
20