Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory
How are sequences of events that are temporarily stored in a cognitive working memory unitized, or chunked, through learning? Such sequential learning is needed by the brain in order to enable language, spatial understanding, and motor skills to develop. In particular, how does the brain learn categ...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Frontiers Media S.A.
2014-10-01
|
Series: | Frontiers in Psychology |
Subjects: | |
Online Access: | http://journal.frontiersin.org/Journal/10.3389/fpsyg.2014.01053/full |
_version_ | 1818334176279003136 |
---|---|
author | Stephen eGrossberg Sohrob eKazerounian |
author_facet | Stephen eGrossberg Sohrob eKazerounian |
author_sort | Stephen eGrossberg |
collection | DOAJ |
description | How are sequences of events that are temporarily stored in a cognitive working memory unitized, or chunked, through learning? Such sequential learning is needed by the brain in order to enable language, spatial understanding, and motor skills to develop. In particular, how does the brain learn categories, or list chunks, that become selectively tuned to different temporal sequences of items in lists of variable length as they are stored in working memory, and how does this learning process occur in real time? The present article introduces a neural model that simulates learning of such list chunks. In this model, sequences of items are temporarily stored in an Item-and-Order, or competitive queuing, working memory before learning categorizes them using a categorization network, called a Masking Field, which is a self-similar, multiple-scale, recurrent on-center off-surround network that can weigh the evidence for variable-length sequences of items as they are stored in the working memory through time. A Masking Field hereby activates the learned list chunks that represent the most predictive item groupings at any time, while suppressing less predictive chunks. In a network with a given number of input items, all possible ordered sets of these item sequences, up to a fixed length, can be learned with unsupervised or supervised learning. The self-similar multiple-scale properties of Masking Fields interacting with an Item-and-Order working memory provide a natural explanation of George Miller's Magical Number Seven and Nelson Cowan's Magical Number Four. The article explains why linguistic, spatial, and action event sequences may all be stored by Item-and-Order working memories that obey similar design principles, and thus how the current results may apply across modalities. Item-and-Order properties may readily be extended to Item-Order-Rank working memories in which the same item can be stored in multiple list positions, or ranks, as in the list ABADBD. Comparisons |
first_indexed | 2024-12-13T14:03:22Z |
format | Article |
id | doaj.art-0ad34683ee6545e0be5bc07961d0770c |
institution | Directory Open Access Journal |
issn | 1664-1078 |
language | English |
last_indexed | 2024-12-13T14:03:22Z |
publishDate | 2014-10-01 |
publisher | Frontiers Media S.A. |
record_format | Article |
series | Frontiers in Psychology |
spelling | doaj.art-0ad34683ee6545e0be5bc07961d0770c2022-12-21T23:42:40ZengFrontiers Media S.A.Frontiers in Psychology1664-10782014-10-01510.3389/fpsyg.2014.0105399534Real-time learning of predictive recognition categories that chunk sequences of items stored in working memoryStephen eGrossberg0Sohrob eKazerounian1Boston UniversityBoston UniversityHow are sequences of events that are temporarily stored in a cognitive working memory unitized, or chunked, through learning? Such sequential learning is needed by the brain in order to enable language, spatial understanding, and motor skills to develop. In particular, how does the brain learn categories, or list chunks, that become selectively tuned to different temporal sequences of items in lists of variable length as they are stored in working memory, and how does this learning process occur in real time? The present article introduces a neural model that simulates learning of such list chunks. In this model, sequences of items are temporarily stored in an Item-and-Order, or competitive queuing, working memory before learning categorizes them using a categorization network, called a Masking Field, which is a self-similar, multiple-scale, recurrent on-center off-surround network that can weigh the evidence for variable-length sequences of items as they are stored in the working memory through time. A Masking Field hereby activates the learned list chunks that represent the most predictive item groupings at any time, while suppressing less predictive chunks. In a network with a given number of input items, all possible ordered sets of these item sequences, up to a fixed length, can be learned with unsupervised or supervised learning. The self-similar multiple-scale properties of Masking Fields interacting with an Item-and-Order working memory provide a natural explanation of George Miller's Magical Number Seven and Nelson Cowan's Magical Number Four. The article explains why linguistic, spatial, and action event sequences may all be stored by Item-and-Order working memories that obey similar design principles, and thus how the current results may apply across modalities. Item-and-Order properties may readily be extended to Item-Order-Rank working memories in which the same item can be stored in multiple list positions, or ranks, as in the list ABADBD. Comparisonshttp://journal.frontiersin.org/Journal/10.3389/fpsyg.2014.01053/fullSpeech Perceptionworking memorycategory learningAdaptive Resonance Theorycapacity limitsMasking Field |
spellingShingle | Stephen eGrossberg Sohrob eKazerounian Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory Frontiers in Psychology Speech Perception working memory category learning Adaptive Resonance Theory capacity limits Masking Field |
title | Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory |
title_full | Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory |
title_fullStr | Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory |
title_full_unstemmed | Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory |
title_short | Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory |
title_sort | real time learning of predictive recognition categories that chunk sequences of items stored in working memory |
topic | Speech Perception working memory category learning Adaptive Resonance Theory capacity limits Masking Field |
url | http://journal.frontiersin.org/Journal/10.3389/fpsyg.2014.01053/full |
work_keys_str_mv | AT stephenegrossberg realtimelearningofpredictiverecognitioncategoriesthatchunksequencesofitemsstoredinworkingmemory AT sohrobekazerounian realtimelearningofpredictiverecognitioncategoriesthatchunksequencesofitemsstoredinworkingmemory |