Training a Two-Layer ReLU Network Analytically

Neural networks are usually trained with different variants of gradient descent-based optimization algorithms such as the stochastic gradient descent or the Adam optimizer. Recent theoretical work states that the critical points (where the gradient of the loss is zero) of two-layer ReLU networks wit...

Full description

Bibliographic Details
Main Author: Adrian Barbu
Format: Article
Language:English
Published: MDPI AG 2023-04-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/23/8/4072

Similar Items