Reexamining the principle of mean-variance preservation for neural network initialization

Before backpropagation training, it is common to randomly initialize a neural network so that mean and variance of activity are uniform across neurons. Classically these statistics were defined over an ensemble of random networks. Alternatively, they can be defined over a random sample of inputs to...

Full description

Bibliographic Details
Main Authors: Kyle Luther, H. Sebastian Seung
Format: Article
Language:English
Published: American Physical Society 2020-07-01
Series:Physical Review Research
Online Access:http://doi.org/10.1103/PhysRevResearch.2.033135