Escaping Saddle Points with Adaptive Gradient Methods
© 2019 International Machine Learning Society (IMLS). Adaptive methods such as Adam and RMSProp are widely used in deep learning but are not well understood. In this paper, we seek a crisp, clean and precise characterization of their behavior in nonconvex settings. To this end, we first provide a no...
Main Authors: | Staib, Matthew, Reddi, Sashank, Kale, Satyen, Kumar, Sanjiv, Sra, Suvrit |
---|---|
Other Authors: | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science |
Format: | Article |
Language: | English |
Published: |
2021
|
Online Access: | https://hdl.handle.net/1721.1/137532 |
Similar Items
-
Efficiently testing local optimality and escaping saddles for RELu networks
by: Jadbabaie, Ali, et al.
Published: (2021) -
Escaping saddle points in constrained optimization
by: Mokhtari, Aryan, et al.
Published: (2019) -
On stability in the saddle-point sense
by: Levhari, David, et al.
Published: (2011) -
Conditional gradient methods via stochastic path-integrated differential estimator
by: Sra, Suvrit
Published: (2021) -
Saddle point localization of molecular wavefunctions
by: Mellau, Georg Ch., et al.
Published: (2017)