New Directions in Vector Space Models of Meaning

Symbolic approaches have dominated NLP as a means to model syntactic and semantic aspects of natural language. While powerful inferential tools exist for such models, they suffer from an inability to capture correlation between words and to provide a continuous model for word, phrase, and document s...

Full description

Bibliographic Details
Main Authors: Grefenstette, E, Hermann, K, Dinu, G, Blunsom, P
Format: Journal article
Published: 2014
_version_ 1826306883646914560
author Grefenstette, E
Hermann, K
Dinu, G
Blunsom, P
author_facet Grefenstette, E
Hermann, K
Dinu, G
Blunsom, P
author_sort Grefenstette, E
collection OXFORD
description Symbolic approaches have dominated NLP as a means to model syntactic and semantic aspects of natural language. While powerful inferential tools exist for such models, they suffer from an inability to capture correlation between words and to provide a continuous model for word, phrase, and document similarity. Distributed representations are one mechanism to overcome these constraints. This tutorial will supply NLP researchers with the mathematical and conceptual background to make use of vector-based models of meaning in their own research. We will begin by motivating the need for a transition from symbolic representations to distributed ones. We will briefly cover how collocational (distributional) vectors can be used and manipulated to model word meaning. We will discuss the progress from distributional to distributed representations, and how neural networks allow us to learn word vectors and condition them on metadata such as parallel texts, topic labels, or sentiment labels. Finally, we will present various forms of semantic vector composition, and discuss their relative strengths and weaknesses, and their application to problems such as language modelling, paraphrasing, machine translation and document classification. This tutorial aims to bring researchers up to speed with recent developments in this fast-moving field. It aims to strike a balance between providing a general introduction to vector-based models of meaning, an analysis of diverging strands of research in the field, and also being a hands-on tutorial to equip NLP researchers with the necessary tools and background knowledge to start working on such models. Attendees should be comfortable with basic probability, linear algebra, and continuous mathematics. No substantial knowledge of machine learning is required.
first_indexed 2024-03-07T06:54:39Z
format Journal article
id oxford-uuid:fdb6d942-ab66-41c5-a7f6-4ff031023b73
institution University of Oxford
last_indexed 2024-03-07T06:54:39Z
publishDate 2014
record_format dspace
spelling oxford-uuid:fdb6d942-ab66-41c5-a7f6-4ff031023b732022-03-27T13:30:55ZNew Directions in Vector Space Models of MeaningJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:fdb6d942-ab66-41c5-a7f6-4ff031023b73Department of Computer Science2014Grefenstette, EHermann, KDinu, GBlunsom, PSymbolic approaches have dominated NLP as a means to model syntactic and semantic aspects of natural language. While powerful inferential tools exist for such models, they suffer from an inability to capture correlation between words and to provide a continuous model for word, phrase, and document similarity. Distributed representations are one mechanism to overcome these constraints. This tutorial will supply NLP researchers with the mathematical and conceptual background to make use of vector-based models of meaning in their own research. We will begin by motivating the need for a transition from symbolic representations to distributed ones. We will briefly cover how collocational (distributional) vectors can be used and manipulated to model word meaning. We will discuss the progress from distributional to distributed representations, and how neural networks allow us to learn word vectors and condition them on metadata such as parallel texts, topic labels, or sentiment labels. Finally, we will present various forms of semantic vector composition, and discuss their relative strengths and weaknesses, and their application to problems such as language modelling, paraphrasing, machine translation and document classification. This tutorial aims to bring researchers up to speed with recent developments in this fast-moving field. It aims to strike a balance between providing a general introduction to vector-based models of meaning, an analysis of diverging strands of research in the field, and also being a hands-on tutorial to equip NLP researchers with the necessary tools and background knowledge to start working on such models. Attendees should be comfortable with basic probability, linear algebra, and continuous mathematics. No substantial knowledge of machine learning is required.
spellingShingle Grefenstette, E
Hermann, K
Dinu, G
Blunsom, P
New Directions in Vector Space Models of Meaning
title New Directions in Vector Space Models of Meaning
title_full New Directions in Vector Space Models of Meaning
title_fullStr New Directions in Vector Space Models of Meaning
title_full_unstemmed New Directions in Vector Space Models of Meaning
title_short New Directions in Vector Space Models of Meaning
title_sort new directions in vector space models of meaning
work_keys_str_mv AT grefenstettee newdirectionsinvectorspacemodelsofmeaning
AT hermannk newdirectionsinvectorspacemodelsofmeaning
AT dinug newdirectionsinvectorspacemodelsofmeaning
AT blunsomp newdirectionsinvectorspacemodelsofmeaning