Norm-Based Generalization Bounds for Compositionally Sparse Neural Network
In this paper, we investigate the Rademacher complexity of deep sparse neural networks, where each neuron receives a small number of inputs. We prove generalization bounds for multilayered sparse ReLU neural networks, including convolutional neural networks. These bounds differ from previous ones, a...
Main Authors: | Galanti, Tomer, Xu, Mengjia, Galanti, Liane, Poggio, Tomaso |
---|---|
Format: | Article |
Published: |
Center for Brains, Minds and Machines (CBMM)
2023
|
Online Access: | https://hdl.handle.net/1721.1/148230 |
Similar Items
-
SGD Noise and Implicit Low-Rank Bias in Deep Neural Networks
by: Galanti, Tomer, et al.
Published: (2022) -
Formation of Representations in Neural Networks
by: Ziyin, Liu, et al.
Published: (2024) -
SGD and Weight Decay Provably Induce a Low-Rank Bias in Deep Neural Networks
by: Galanti, Tomer, et al.
Published: (2023) -
The Janus effects of SGD vs GD: high noise and low rank
by: Xu, Mengjia, et al.
Published: (2023) -
On Generalization Bounds for Neural Networks with Low Rank Layers
by: Pinto, Andrea, et al.
Published: (2024)