Augmenting Transformers for Open Domain Procedural Text Comprehesion

Recent advances in deep learning model architectures have permitted state-of-the-art results in various fields such as NLP and CV. Although these systems have matched and, in some cases, surpassed human performance, many of them are still treated as black boxes, with sometimes unpredictable results....

Full description

Bibliographic Details
Main Author: Pei, Yixuan
Other Authors: Shrobe, Howard
Format: Thesis
Published: Massachusetts Institute of Technology 2022
Online Access:https://hdl.handle.net/1721.1/139980
_version_ 1826196575695667200
author Pei, Yixuan
author2 Shrobe, Howard
author_facet Shrobe, Howard
Pei, Yixuan
author_sort Pei, Yixuan
collection MIT
description Recent advances in deep learning model architectures have permitted state-of-the-art results in various fields such as NLP and CV. Although these systems have matched and, in some cases, surpassed human performance, many of them are still treated as black boxes, with sometimes unpredictable results. To try and shed some light on behaviors of natural language generation models, we examine the task of procedural text comprehension using neuro-symbolic techniques. We use this task as a testbed for exploring the limitations of state-of-the-art systems such as GPT on the task of predicting the resulting state changes from the text description of a procedure. We also experiment with whether and how symbolic augmentations may help these systems with understanding language. We see some promising results in concept-net knowledge injection, and note that other augmentations provide more natural generations.
first_indexed 2024-09-23T10:29:55Z
format Thesis
id mit-1721.1/139980
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T10:29:55Z
publishDate 2022
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1399802022-02-08T03:49:52Z Augmenting Transformers for Open Domain Procedural Text Comprehesion Pei, Yixuan Shrobe, Howard Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Recent advances in deep learning model architectures have permitted state-of-the-art results in various fields such as NLP and CV. Although these systems have matched and, in some cases, surpassed human performance, many of them are still treated as black boxes, with sometimes unpredictable results. To try and shed some light on behaviors of natural language generation models, we examine the task of procedural text comprehension using neuro-symbolic techniques. We use this task as a testbed for exploring the limitations of state-of-the-art systems such as GPT on the task of predicting the resulting state changes from the text description of a procedure. We also experiment with whether and how symbolic augmentations may help these systems with understanding language. We see some promising results in concept-net knowledge injection, and note that other augmentations provide more natural generations. M.Eng. 2022-02-07T15:16:54Z 2022-02-07T15:16:54Z 2021-09 2021-11-03T19:25:25.983Z Thesis https://hdl.handle.net/1721.1/139980 In Copyright - Educational Use Permitted Copyright MIT http://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Pei, Yixuan
Augmenting Transformers for Open Domain Procedural Text Comprehesion
title Augmenting Transformers for Open Domain Procedural Text Comprehesion
title_full Augmenting Transformers for Open Domain Procedural Text Comprehesion
title_fullStr Augmenting Transformers for Open Domain Procedural Text Comprehesion
title_full_unstemmed Augmenting Transformers for Open Domain Procedural Text Comprehesion
title_short Augmenting Transformers for Open Domain Procedural Text Comprehesion
title_sort augmenting transformers for open domain procedural text comprehesion
url https://hdl.handle.net/1721.1/139980
work_keys_str_mv AT peiyixuan augmentingtransformersforopendomainproceduraltextcomprehesion