Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets

The complexity of a learning task is increased by transformations in the input space that preserve class identity. Visual object recognition for example is affected by changes in viewpoint, scale, illumination or planar transformations. While drastically altering the visual appearance, these changes...

Full description

Bibliographic Details
Main Authors: Tachetti, Andrea, Voinea, Stephen, Evangelopoulos, Georgios
Format: Technical Report
Language:en_US
Published: Center for Brains, Minds and Machines (CBMM), arXiv 2017
Subjects:
Online Access:http://hdl.handle.net/1721.1/107446
_version_ 1811088706023456768
author Tachetti, Andrea
Voinea, Stephen
Evangelopoulos, Georgios
author_facet Tachetti, Andrea
Voinea, Stephen
Evangelopoulos, Georgios
author_sort Tachetti, Andrea
collection MIT
description The complexity of a learning task is increased by transformations in the input space that preserve class identity. Visual object recognition for example is affected by changes in viewpoint, scale, illumination or planar transformations. While drastically altering the visual appearance, these changes are orthogonal to recognition and should not be reflected in the representation or feature encoding used for learning. We introduce a framework for weakly supervised learning of image embeddings that are robust to transformations and selective to the class distribution, using sets of transforming examples (orbit sets), deep parametrizations and a novel orbit-based loss. The proposed loss combines a discriminative, contrastive part for orbits with a reconstruction error that learns to rectify orbit transformations. The learned embeddings are evaluated in distance metric-based tasks, such as one-shot classification under geometric transformations, as well as face verification and retrieval under more realistic visual variability. Our results suggest that orbit sets, suitably computed or observed, can be used for efficient, weakly-supervised learning of semantically relevant image embeddings.
first_indexed 2024-09-23T14:06:15Z
format Technical Report
id mit-1721.1/107446
institution Massachusetts Institute of Technology
language en_US
last_indexed 2024-09-23T14:06:15Z
publishDate 2017
publisher Center for Brains, Minds and Machines (CBMM), arXiv
record_format dspace
spelling mit-1721.1/1074462019-04-11T13:40:28Z Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets Tachetti, Andrea Voinea, Stephen Evangelopoulos, Georgios supervised learning object recognition machine learning The complexity of a learning task is increased by transformations in the input space that preserve class identity. Visual object recognition for example is affected by changes in viewpoint, scale, illumination or planar transformations. While drastically altering the visual appearance, these changes are orthogonal to recognition and should not be reflected in the representation or feature encoding used for learning. We introduce a framework for weakly supervised learning of image embeddings that are robust to transformations and selective to the class distribution, using sets of transforming examples (orbit sets), deep parametrizations and a novel orbit-based loss. The proposed loss combines a discriminative, contrastive part for orbits with a reconstruction error that learns to rectify orbit transformations. The learned embeddings are evaluated in distance metric-based tasks, such as one-shot classification under geometric transformations, as well as face verification and retrieval under more realistic visual variability. Our results suggest that orbit sets, suitably computed or observed, can be used for efficient, weakly-supervised learning of semantically relevant image embeddings. This material is based upon work supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF-1231216. 2017-03-16T19:50:32Z 2017-03-16T19:50:32Z 2017-03-13 Technical Report Working Paper Other http://hdl.handle.net/1721.1/107446 arXiv:1703.04775v1 en_US CBMM Memo Series;062 Attribution-NonCommercial-ShareAlike 3.0 United States http://creativecommons.org/licenses/by-nc-sa/3.0/us/ application/pdf Center for Brains, Minds and Machines (CBMM), arXiv
spellingShingle supervised learning
object recognition
machine learning
Tachetti, Andrea
Voinea, Stephen
Evangelopoulos, Georgios
Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets
title Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets
title_full Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets
title_fullStr Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets
title_full_unstemmed Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets
title_short Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets
title_sort discriminate and rectify encoders learning from image transformation sets
topic supervised learning
object recognition
machine learning
url http://hdl.handle.net/1721.1/107446
work_keys_str_mv AT tachettiandrea discriminateandrectifyencoderslearningfromimagetransformationsets
AT voineastephen discriminateandrectifyencoderslearningfromimagetransformationsets
AT evangelopoulosgeorgios discriminateandrectifyencoderslearningfromimagetransformationsets