Learning of invariant object recognition in hierarchical neural networks using temporal continuity
A lot of progress in the field of invariant object recognition has been made in recent years using so called deep neural networks with several layers to be trained which can learn patterns of increasing complexity. This architectural feature can alreay be found in older neural models as, e.g.,...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
Computer Vision Center Press
2015-12-01
|
Series: | ELCVIA Electronic Letters on Computer Vision and Image Analysis |
Subjects: | |
Online Access: | https://elcvia.cvc.uab.es/article/view/719 |
_version_ | 1819122434262958080 |
---|---|
author | Markus Lessmann |
author_facet | Markus Lessmann |
author_sort | Markus Lessmann |
collection | DOAJ |
description | A lot of progress in the field of invariant object recognition has been
made in recent years using so called deep neural networks with several
layers to be trained which can learn patterns of increasing complexity. This
architectural feature can alreay be found in older neural models as, e.g.,
the Neocognitron and HMAX but also newer ones as Convolutional Nets. Additionally
researchers emphasized the importance of temporal continuity in input data
and devised learning rules utilizing it (e.g. the trace rule by F\"oldiak
used by Rolls in VisNet). Finally Jeff Hawkins collected a lot of these ideas
concerning functioning of the neocortex in a coherent framework and proposed
three basic principles for neocortical computations (later implemented in HTM):• Learning of temporal sequences for creating invariance to transformations contained in the training data.
• Learning in a hierarchical structure, in which lower level knowledge can be reused in higher level context and thereby makes memory usage efficient.
• Prediction of future signals for disambiguation of noisy input by feedback.
In my thesis I developed two related systems: the \emph{Temporal Correlation Graph} (TCG)
and the \emph{Temporal Correlation Net} (TCN). Both make use of these principles and
implement them in an efficient manner. The main aim was to create systems that are trained
mostly unsupervised (both) and can be trained online, which is possible with TCN. Both achieve
very good performance on several standard datasets for object recognition. |
first_indexed | 2024-12-22T06:52:23Z |
format | Article |
id | doaj.art-93114f4725734b288fa3b94c4b579b65 |
institution | Directory Open Access Journal |
issn | 1577-5097 |
language | English |
last_indexed | 2024-12-22T06:52:23Z |
publishDate | 2015-12-01 |
publisher | Computer Vision Center Press |
record_format | Article |
series | ELCVIA Electronic Letters on Computer Vision and Image Analysis |
spelling | doaj.art-93114f4725734b288fa3b94c4b579b652022-12-21T18:35:06ZengComputer Vision Center PressELCVIA Electronic Letters on Computer Vision and Image Analysis1577-50972015-12-0114310.5565/rev/elcvia.719272Learning of invariant object recognition in hierarchical neural networks using temporal continuityMarkus Lessmann0Ruhr-University Bochum, GermanyA lot of progress in the field of invariant object recognition has been made in recent years using so called deep neural networks with several layers to be trained which can learn patterns of increasing complexity. This architectural feature can alreay be found in older neural models as, e.g., the Neocognitron and HMAX but also newer ones as Convolutional Nets. Additionally researchers emphasized the importance of temporal continuity in input data and devised learning rules utilizing it (e.g. the trace rule by F\"oldiak used by Rolls in VisNet). Finally Jeff Hawkins collected a lot of these ideas concerning functioning of the neocortex in a coherent framework and proposed three basic principles for neocortical computations (later implemented in HTM):• Learning of temporal sequences for creating invariance to transformations contained in the training data. • Learning in a hierarchical structure, in which lower level knowledge can be reused in higher level context and thereby makes memory usage efficient. • Prediction of future signals for disambiguation of noisy input by feedback. In my thesis I developed two related systems: the \emph{Temporal Correlation Graph} (TCG) and the \emph{Temporal Correlation Net} (TCN). Both make use of these principles and implement them in an efficient manner. The main aim was to create systems that are trained mostly unsupervised (both) and can be trained online, which is possible with TCN. Both achieve very good performance on several standard datasets for object recognition.https://elcvia.cvc.uab.es/article/view/719Computer VisionObject Description and RecognitionMachine Learning and Data MiningClassification and ClusteringInvariances in Recognition |
spellingShingle | Markus Lessmann Learning of invariant object recognition in hierarchical neural networks using temporal continuity ELCVIA Electronic Letters on Computer Vision and Image Analysis Computer Vision Object Description and Recognition Machine Learning and Data Mining Classification and Clustering Invariances in Recognition |
title | Learning of invariant object recognition in hierarchical neural networks using temporal continuity |
title_full | Learning of invariant object recognition in hierarchical neural networks using temporal continuity |
title_fullStr | Learning of invariant object recognition in hierarchical neural networks using temporal continuity |
title_full_unstemmed | Learning of invariant object recognition in hierarchical neural networks using temporal continuity |
title_short | Learning of invariant object recognition in hierarchical neural networks using temporal continuity |
title_sort | learning of invariant object recognition in hierarchical neural networks using temporal continuity |
topic | Computer Vision Object Description and Recognition Machine Learning and Data Mining Classification and Clustering Invariances in Recognition |
url | https://elcvia.cvc.uab.es/article/view/719 |
work_keys_str_mv | AT markuslessmann learningofinvariantobjectrecognitioninhierarchicalneuralnetworksusingtemporalcontinuity |