Showing 1 - 20 results of 2,357 for search '"gradient descent"', query time: 0.27s Refine Results
  1. 1

    Low-Rank Gradient Descent by Romain Cosson, Ali Jadbabaie, Anuran Makur, Amirhossein Reisizadeh, Devavrat Shah

    Published 2023-01-01
    “…In this article, we leverage such low-rank structure to reduce the high computational cost of canonical gradient-based methods such as gradient descent (<monospace>GD</monospace>). Our proposed <italic>Low-Rank Gradient Descent</italic> (<monospace>LRGD</monospace>) algorithm finds an <inline-formula><tex-math notation="LaTeX">$\epsilon$</tex-math></inline-formula>-approximate stationary point of a <inline-formula><tex-math notation="LaTeX">$p$</tex-math></inline-formula>-dimensional function by first identifying <inline-formula><tex-math notation="LaTeX">$r \leq p$</tex-math></inline-formula> significant directions, and then estimating the true <inline-formula><tex-math notation="LaTeX">$p$</tex-math></inline-formula>-dimensional gradient at every iteration by computing directional derivatives only along those <inline-formula><tex-math notation="LaTeX">$r$</tex-math></inline-formula> directions. …”
    Get full text
    Article
  2. 2

    On-manifold projected gradient descent by Aaron Mahler, Tyrus Berry, Tom Stephens, Harbir Antil, Michael Merritt, Jeanie Schreiber, Ioannis Kevrekidis

    Published 2024-02-01
    “…The tools are applied to the setting of neural network image classifiers, where we generate novel, on-manifold data samples and implement a projected gradient descent algorithm for on-manifold adversarial training. …”
    Get full text
    Article
  3. 3

    Correspondence between neuroevolution and gradient descent by Stephen Whitelam, Viktor Selin, Sang-Won Park, Isaac Tamblyn

    Published 2021-11-01
    “…The authors derive, and illustrate numerically, an analytic equivalence between the dynamics of neural network training under conditioned stochastic mutations, and under gradient descent.…”
    Get full text
    Article
  4. 4

    Semi-Stochastic Gradient Descent Methods by Jakub Konečný, Peter Richtárik

    Published 2017-05-01
    “…We propose a new method, S2GD (Semi-Stochastic Gradient Descent), which runs for one or several epochs in each of which a single full gradient and a random number of stochastic gradients is computed, following a geometric law. …”
    Get full text
    Article
  5. 5
  6. 6

    Carathéodory sampling for stochastic gradient descent by Cosentino, F, Oberhauser, H, Abate, A

    Published 2020
    “…Many problems require to optimize empirical risk functions over large data sets. Gradient descent methods that calculate the full gradient in every descent step do not scale to such datasets. …”
    Internet publication
  7. 7

    Carathéodory sampling for stochastic gradient descent by Cosentino, F, Oberhauser, H, Abate, A

    Published 2020
    “…Many problems require to optimize empirical risk functions over large data sets. Gradient descent methods that calculate the full gradient in every descent step do not scale to such datasets. …”
    Internet publication
  8. 8

    Dual space preconditioning for gradient descent by Maddison, CJ, Paulin, D, Teh, YW, Doucet, A

    Published 2021
    “…Thus, in principle our method is capable of improving the conditioning of gradient descent on problems with a non-Lipschitz gradient or nonstrongly convex structure. …”
    Journal article
  9. 9
  10. 10

    Limited Gradient Descent: Learning With Noisy Labels by Yi Sun, Yan Tian, Yiping Xu, Jianxiang Li

    Published 2019-01-01
    “…To solve this problem, we propose a method that can estimate the optimal stopping timing without a clean validation set, called limited gradient descent. We modified the labels of a few samples in a noisy dataset to obtain false labels and to create a reverse pattern. …”
    Get full text
    Article
  11. 11

    Accelerated Gradient Descent Driven by Lévy Perturbations by Yuquan Chen, Zhenlong Wu, Yixiang Lu, Yangquan Chen, Yong Wang

    Published 2024-03-01
    Subjects: “…accelerated gradient descent…”
    Get full text
    Article
  12. 12
  13. 13

    Stochastic gradient descent for optimization for nuclear systems by Austin Williams, Noah Walton, Austin Maryanski, Sandra Bogetic, Wes Hines, Vladimir Sobes

    Published 2023-05-01
    “…ADAM is a gradient descent method that accounts for gradients with a stochastic nature. …”
    Get full text
    Article
  14. 14

    Stochastic gradient descent for wind farm optimization by J. Quick, P.-E. Rethore, M. Mølgaard Pedersen, R. V. Rodrigues, M. Friis-Møller

    Published 2023-08-01
    “…This study presents stochastic gradient descent (SGD) for wind farm optimization, which is an approach that estimates the gradient of the AEP using Monte Carlo simulation, allowing for the consideration of an arbitrarily large number of atmospheric conditions. …”
    Get full text
    Article
  15. 15

    NOMA Codebook Optimization by Batch Gradient Descent by Zhongwei Si, Shaoguo Wen, Bing Dong

    Published 2019-01-01
    Subjects: “…Batch gradient descent…”
    Get full text
    Article
  16. 16

    Complexity control by gradient descent in deep networks by Tomaso Poggio, Qianli Liao, Andrzej Banburski

    Published 2020-02-01
    “…Here, the author demonstrates an implicit regularization in training deep networks, showing that the control of complexity in the training is hidden within the optimization technique of gradient descent.…”
    Get full text
    Article
  17. 17
  18. 18

    Complexity control by gradient descent in deep networks by Poggio, Tomaso, Liao, Qianli, Banburski, Andrzej

    Published 2021
    “…For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification.…”
    Get full text
    Article
  19. 19

    Complexity control by gradient descent in deep networks by Poggio, Tomaso A, Liao, Qianli, Banburski, Andrzej

    Published 2022
    “…For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification.…”
    Get full text
    Article
  20. 20