Learning latent permutations with Gumbel-Sinkhorn networks

Permutations and matchings are core building blocks in a variety of latent variable models, as they allow us to align, canonicalize, and sort data. Learning in such models is difficult, however, because exact marginalization over these combinatorial objects is intractable. In response, this paper in...

Full description

Bibliographic Details
Main Authors: Mena, G, Snoek, J, Linderman, S, Belanger, D
Format: Conference item
Language:English
Published: OpenReview 2018
_version_ 1797104287076057088
author Mena, G
Snoek, J
Linderman, S
Belanger, D
author_facet Mena, G
Snoek, J
Linderman, S
Belanger, D
author_sort Mena, G
collection OXFORD
description Permutations and matchings are core building blocks in a variety of latent variable models, as they allow us to align, canonicalize, and sort data. Learning in such models is difficult, however, because exact marginalization over these combinatorial objects is intractable. In response, this paper introduces a collection of new methods for end-to-end learning in such models that approximate discrete maximum-weight matching using the continuous Sinkhorn operator. Sinkhorn iteration is attractive because it functions as a simple, easy-to-implement analog of the softmax operator. With this, we can define the Gumbel-Sinkhorn method, an extension of the Gumbel-Softmax method (Jang et al. 2016, Maddison2016 et al. 2016) to distributions over latent matchings. We demonstrate the effectiveness of our method by outperforming competitive baselines on a range of qualitatively different tasks: sorting numbers, solving jigsaw puzzles, and identifying neural signals in worms.
first_indexed 2024-03-07T06:31:43Z
format Conference item
id oxford-uuid:f6331698-2e7e-48ad-b1dc-a8da2e45f09b
institution University of Oxford
language English
last_indexed 2024-03-07T06:31:43Z
publishDate 2018
publisher OpenReview
record_format dspace
spelling oxford-uuid:f6331698-2e7e-48ad-b1dc-a8da2e45f09b2022-03-27T12:33:25ZLearning latent permutations with Gumbel-Sinkhorn networksConference itemhttp://purl.org/coar/resource_type/c_5794uuid:f6331698-2e7e-48ad-b1dc-a8da2e45f09bEnglishSymplectic ElementsOpenReview2018Mena, GSnoek, JLinderman, SBelanger, DPermutations and matchings are core building blocks in a variety of latent variable models, as they allow us to align, canonicalize, and sort data. Learning in such models is difficult, however, because exact marginalization over these combinatorial objects is intractable. In response, this paper introduces a collection of new methods for end-to-end learning in such models that approximate discrete maximum-weight matching using the continuous Sinkhorn operator. Sinkhorn iteration is attractive because it functions as a simple, easy-to-implement analog of the softmax operator. With this, we can define the Gumbel-Sinkhorn method, an extension of the Gumbel-Softmax method (Jang et al. 2016, Maddison2016 et al. 2016) to distributions over latent matchings. We demonstrate the effectiveness of our method by outperforming competitive baselines on a range of qualitatively different tasks: sorting numbers, solving jigsaw puzzles, and identifying neural signals in worms.
spellingShingle Mena, G
Snoek, J
Linderman, S
Belanger, D
Learning latent permutations with Gumbel-Sinkhorn networks
title Learning latent permutations with Gumbel-Sinkhorn networks
title_full Learning latent permutations with Gumbel-Sinkhorn networks
title_fullStr Learning latent permutations with Gumbel-Sinkhorn networks
title_full_unstemmed Learning latent permutations with Gumbel-Sinkhorn networks
title_short Learning latent permutations with Gumbel-Sinkhorn networks
title_sort learning latent permutations with gumbel sinkhorn networks
work_keys_str_mv AT menag learninglatentpermutationswithgumbelsinkhornnetworks
AT snoekj learninglatentpermutationswithgumbelsinkhornnetworks
AT lindermans learninglatentpermutationswithgumbelsinkhornnetworks
AT belangerd learninglatentpermutationswithgumbelsinkhornnetworks