Learning of invariant object recognition in hierarchical neural networks using temporal continuity

A lot of progress in the field of invariant object recognition has been made in recent years using so called deep neural networks with several layers to be trained which can learn patterns of increasing complexity. This architectural feature can alreay be found in older neural models as, e.g.,...

Full description

Bibliographic Details
Main Author: Markus Lessmann
Format: Article
Language:English
Published: Computer Vision Center Press 2015-12-01
Series:ELCVIA Electronic Letters on Computer Vision and Image Analysis
Subjects:
Online Access:https://elcvia.cvc.uab.es/article/view/719
_version_ 1819122434262958080
author Markus Lessmann
author_facet Markus Lessmann
author_sort Markus Lessmann
collection DOAJ
description A lot of progress in the field of invariant object recognition has been made in recent years using so called deep neural networks with several layers to be trained which can learn patterns of increasing complexity. This architectural feature can alreay be found in older neural models as, e.g., the Neocognitron and HMAX but also newer ones as Convolutional Nets. Additionally researchers emphasized the importance of temporal continuity in input data and devised learning rules utilizing it (e.g. the trace rule by F\"oldiak used by Rolls in VisNet). Finally Jeff Hawkins collected a lot of these ideas concerning functioning of the neocortex in a coherent framework and proposed three basic principles for neocortical computations (later implemented in HTM):• Learning of temporal sequences for creating invariance to transformations contained in the training data. • Learning in a hierarchical structure, in which lower level knowledge can be reused in higher level context and thereby makes memory usage efficient. • Prediction of future signals for disambiguation of noisy input by feedback. In my thesis I developed two related systems: the \emph{Temporal Correlation Graph} (TCG) and the \emph{Temporal Correlation Net} (TCN). Both make use of these principles and implement them in an efficient manner. The main aim was to create systems that are trained mostly unsupervised (both) and can be trained online, which is possible with TCN. Both achieve very good performance on several standard datasets for object recognition.
first_indexed 2024-12-22T06:52:23Z
format Article
id doaj.art-93114f4725734b288fa3b94c4b579b65
institution Directory Open Access Journal
issn 1577-5097
language English
last_indexed 2024-12-22T06:52:23Z
publishDate 2015-12-01
publisher Computer Vision Center Press
record_format Article
series ELCVIA Electronic Letters on Computer Vision and Image Analysis
spelling doaj.art-93114f4725734b288fa3b94c4b579b652022-12-21T18:35:06ZengComputer Vision Center PressELCVIA Electronic Letters on Computer Vision and Image Analysis1577-50972015-12-0114310.5565/rev/elcvia.719272Learning of invariant object recognition in hierarchical neural networks using temporal continuityMarkus Lessmann0Ruhr-University Bochum, GermanyA lot of progress in the field of invariant object recognition has been made in recent years using so called deep neural networks with several layers to be trained which can learn patterns of increasing complexity. This architectural feature can alreay be found in older neural models as, e.g., the Neocognitron and HMAX but also newer ones as Convolutional Nets. Additionally researchers emphasized the importance of temporal continuity in input data and devised learning rules utilizing it (e.g. the trace rule by F\"oldiak used by Rolls in VisNet). Finally Jeff Hawkins collected a lot of these ideas concerning functioning of the neocortex in a coherent framework and proposed three basic principles for neocortical computations (later implemented in HTM):• Learning of temporal sequences for creating invariance to transformations contained in the training data. • Learning in a hierarchical structure, in which lower level knowledge can be reused in higher level context and thereby makes memory usage efficient. • Prediction of future signals for disambiguation of noisy input by feedback. In my thesis I developed two related systems: the \emph{Temporal Correlation Graph} (TCG) and the \emph{Temporal Correlation Net} (TCN). Both make use of these principles and implement them in an efficient manner. The main aim was to create systems that are trained mostly unsupervised (both) and can be trained online, which is possible with TCN. Both achieve very good performance on several standard datasets for object recognition.https://elcvia.cvc.uab.es/article/view/719Computer VisionObject Description and RecognitionMachine Learning and Data MiningClassification and ClusteringInvariances in Recognition
spellingShingle Markus Lessmann
Learning of invariant object recognition in hierarchical neural networks using temporal continuity
ELCVIA Electronic Letters on Computer Vision and Image Analysis
Computer Vision
Object Description and Recognition
Machine Learning and Data Mining
Classification and Clustering
Invariances in Recognition
title Learning of invariant object recognition in hierarchical neural networks using temporal continuity
title_full Learning of invariant object recognition in hierarchical neural networks using temporal continuity
title_fullStr Learning of invariant object recognition in hierarchical neural networks using temporal continuity
title_full_unstemmed Learning of invariant object recognition in hierarchical neural networks using temporal continuity
title_short Learning of invariant object recognition in hierarchical neural networks using temporal continuity
title_sort learning of invariant object recognition in hierarchical neural networks using temporal continuity
topic Computer Vision
Object Description and Recognition
Machine Learning and Data Mining
Classification and Clustering
Invariances in Recognition
url https://elcvia.cvc.uab.es/article/view/719
work_keys_str_mv AT markuslessmann learningofinvariantobjectrecognitioninhierarchicalneuralnetworksusingtemporalcontinuity