Beyond Memorization: Exploring the Dynamics of Grokking in Sparse Neural Networks
In the domain of machine learning, "grokking" is a phenomenon where neural network models demonstrate a sudden improvement in generalization, distinct from traditional learning phases, long after the initial training appears complete. This behavior was first identified by Power et al. (202...
Main Author: | Fuangkawinsombut, Siwakorn |
---|---|
Other Authors: | Raghuraman, Srinivasan |
Format: | Thesis |
Published: |
Massachusetts Institute of Technology
2024
|
Online Access: | https://hdl.handle.net/1721.1/156751 |
Similar Items
-
Matching and Grokking: Approaches to Personalized Crowdsourcing
by: Organisciak, Peter, et al.
Published: (2017) -
The neural and psychophysical bases of memorability
by: Bainbridge, Wilma A
Published: (2017) -
Strongly-transferring memorized examples in deep neural networks
by: Venigalla, Abhinav S.
Published: (2019) -
Sparse dynamic reorganising fuzzy neural networks
by: Zhou, Jair Weigui
Published: (2017) -
Small ReLU networks are powerful memorizers: A tight analysis of memorization capacity
by: Yun, Chulhee, et al.
Published: (2021)