Theory I: Why and When Can Deep Networks Avoid the Curse of Dimensionality?
[formerly titled "Why and When Can Deep – but Not Shallow – Networks Avoid the Curse of Dimensionality: a Review"] The paper reviews and extends an emerging body of theoretical results on deep learning including the conditions under which it can be exponentially better than shallow learni...
Main Authors: | , , , , |
---|---|
Format: | Technical Report |
Language: | en_US |
Published: |
Center for Brains, Minds and Machines (CBMM), arXiv
2016
|
Subjects: | |
Online Access: | http://hdl.handle.net/1721.1/105443 |