Cross-validation Stability of Deep Networks
Recent theoretical results show that gradient descent on deep neural networks under exponential loss functions locally maximizes classification margin, which is equivalent to minimizing the norm of the weight matrices under margin constraints. This property of the solution however does not fully ch...
Main Authors: | , , , , |
---|---|
Format: | Technical Report |
Published: |
Center for Brains, Minds and Machines (CBMM)
2021
|
Online Access: | https://hdl.handle.net/1721.1/129744 |
_version_ | 1811081815309418496 |
---|---|
author | Banburski, Andrzej De La Torre, Fernanda Plant, Nishka Shastri, Ishana Poggio, Tomaso |
author_facet | Banburski, Andrzej De La Torre, Fernanda Plant, Nishka Shastri, Ishana Poggio, Tomaso |
author_sort | Banburski, Andrzej |
collection | MIT |
description | Recent theoretical results show that gradient descent on deep neural networks under exponential loss functions locally maximizes classification margin, which is equivalent to minimizing the norm of the weight matrices under margin constraints. This property of the solution however does not fully characterize the generalization performance. We motivate theoretically and show empirically that the area under the curve of the margin distribution on the training set is in fact a good measure of generalization. We then show that, after data separation is achieved, it is possible to dynamically reduce the training set by more than 99% without significant loss of performance. Interestingly, the resulting subset of “high capacity” features is not consistent across different training runs, which is consistent with the theoretical claim that all training points should converge to the same asymptotic margin under SGD and in the presence of both batch normalization and weight decay. |
first_indexed | 2024-09-23T11:52:54Z |
format | Technical Report |
id | mit-1721.1/129744 |
institution | Massachusetts Institute of Technology |
last_indexed | 2024-09-23T11:52:54Z |
publishDate | 2021 |
publisher | Center for Brains, Minds and Machines (CBMM) |
record_format | dspace |
spelling | mit-1721.1/1297442021-02-12T03:21:17Z Cross-validation Stability of Deep Networks Banburski, Andrzej De La Torre, Fernanda Plant, Nishka Shastri, Ishana Poggio, Tomaso Recent theoretical results show that gradient descent on deep neural networks under exponential loss functions locally maximizes classification margin, which is equivalent to minimizing the norm of the weight matrices under margin constraints. This property of the solution however does not fully characterize the generalization performance. We motivate theoretically and show empirically that the area under the curve of the margin distribution on the training set is in fact a good measure of generalization. We then show that, after data separation is achieved, it is possible to dynamically reduce the training set by more than 99% without significant loss of performance. Interestingly, the resulting subset of “high capacity” features is not consistent across different training runs, which is consistent with the theoretical claim that all training points should converge to the same asymptotic margin under SGD and in the presence of both batch normalization and weight decay. This material is based upon work supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF-1231216. 2021-02-11T16:59:01Z 2021-02-11T16:59:01Z 2021-02-09 Technical Report Working Paper Other https://hdl.handle.net/1721.1/129744 CBMM Memo;115 application/pdf Center for Brains, Minds and Machines (CBMM) |
spellingShingle | Banburski, Andrzej De La Torre, Fernanda Plant, Nishka Shastri, Ishana Poggio, Tomaso Cross-validation Stability of Deep Networks |
title | Cross-validation Stability of Deep Networks |
title_full | Cross-validation Stability of Deep Networks |
title_fullStr | Cross-validation Stability of Deep Networks |
title_full_unstemmed | Cross-validation Stability of Deep Networks |
title_short | Cross-validation Stability of Deep Networks |
title_sort | cross validation stability of deep networks |
url | https://hdl.handle.net/1721.1/129744 |
work_keys_str_mv | AT banburskiandrzej crossvalidationstabilityofdeepnetworks AT delatorrefernanda crossvalidationstabilityofdeepnetworks AT plantnishka crossvalidationstabilityofdeepnetworks AT shastriishana crossvalidationstabilityofdeepnetworks AT poggiotomaso crossvalidationstabilityofdeepnetworks |