A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition

The HMAX model has recently been proposed by Riesenhuber & Poggio as a hierarchical model of position- and size-invariant object recognition in visual cortex. It has also turned out to model successfully a number of other properties of the ventral visual stream (the visual pathway thought...

Full description

Bibliographic Details
Main Authors: Schneider, Robert, Riesenhuber, Maximilian
Language:en_US
Published: 2004
Subjects:
AI
Online Access:http://hdl.handle.net/1721.1/7178
_version_ 1826198906484031488
author Schneider, Robert
Riesenhuber, Maximilian
author_facet Schneider, Robert
Riesenhuber, Maximilian
author_sort Schneider, Robert
collection MIT
description The HMAX model has recently been proposed by Riesenhuber & Poggio as a hierarchical model of position- and size-invariant object recognition in visual cortex. It has also turned out to model successfully a number of other properties of the ventral visual stream (the visual pathway thought to be crucial for object recognition in cortex), and particularly of (view-tuned) neurons in macaque inferotemporal cortex, the brain area at the top of the ventral stream. The original modeling study only used ``paperclip'' stimuli, as in the corresponding physiology experiment, and did not explore systematically how model units' invariance properties depended on model parameters. In this study, we aimed at a deeper understanding of the inner workings of HMAX and its performance for various parameter settings and ``natural'' stimulus classes. We examined HMAX responses for different stimulus sizes and positions systematically and found a dependence of model units' responses on stimulus position for which a quantitative description is offered. Interestingly, we find that scale invariance properties of hierarchical neural models are not independent of stimulus class, as opposed to translation invariance, even though both are affine transformations within the image plane.
first_indexed 2024-09-23T11:11:49Z
id mit-1721.1/7178
institution Massachusetts Institute of Technology
language en_US
last_indexed 2024-09-23T11:11:49Z
publishDate 2004
record_format dspace
spelling mit-1721.1/71782019-04-10T11:52:40Z A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition Schneider, Robert Riesenhuber, Maximilian AI The HMAX model has recently been proposed by Riesenhuber & Poggio as a hierarchical model of position- and size-invariant object recognition in visual cortex. It has also turned out to model successfully a number of other properties of the ventral visual stream (the visual pathway thought to be crucial for object recognition in cortex), and particularly of (view-tuned) neurons in macaque inferotemporal cortex, the brain area at the top of the ventral stream. The original modeling study only used ``paperclip'' stimuli, as in the corresponding physiology experiment, and did not explore systematically how model units' invariance properties depended on model parameters. In this study, we aimed at a deeper understanding of the inner workings of HMAX and its performance for various parameter settings and ``natural'' stimulus classes. We examined HMAX responses for different stimulus sizes and positions systematically and found a dependence of model units' responses on stimulus position for which a quantitative description is offered. Interestingly, we find that scale invariance properties of hierarchical neural models are not independent of stimulus class, as opposed to translation invariance, even though both are affine transformations within the image plane. 2004-10-20T20:48:51Z 2004-10-20T20:48:51Z 2002-08-01 AIM-2002-011 CBCL-218 http://hdl.handle.net/1721.1/7178 en_US AIM-2002-011 CBCL-218 12 p. 2137337 bytes 1062341 bytes application/postscript application/pdf application/postscript application/pdf
spellingShingle AI
Schneider, Robert
Riesenhuber, Maximilian
A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition
title A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition
title_full A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition
title_fullStr A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition
title_full_unstemmed A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition
title_short A Detailed Look at Scale and Translation Invariance in a Hierarchical Neural Model of Visual Object Recognition
title_sort detailed look at scale and translation invariance in a hierarchical neural model of visual object recognition
topic AI
url http://hdl.handle.net/1721.1/7178
work_keys_str_mv AT schneiderrobert adetailedlookatscaleandtranslationinvarianceinahierarchicalneuralmodelofvisualobjectrecognition
AT riesenhubermaximilian adetailedlookatscaleandtranslationinvarianceinahierarchicalneuralmodelofvisualobjectrecognition
AT schneiderrobert detailedlookatscaleandtranslationinvarianceinahierarchicalneuralmodelofvisualobjectrecognition
AT riesenhubermaximilian detailedlookatscaleandtranslationinvarianceinahierarchicalneuralmodelofvisualobjectrecognition