Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation

Tree-LSTM algorithm accommodates tree structure processing to extract information outside the linear sequence pattern. The use of Tree-LSTM in text generation problems requires the help of an external parser at each generation iteration. Developing a good parser demands the representation of complex...

Full description

Bibliographic Details
Main Authors: Hastuti, Rochana Prih, Suyanto, Yohanes, Sari, Anny Kartika
Format: Other
Published: ACM Transactions on Asian and Low-Resource Language Information Processing 2022
Subjects:
_version_ 1797037856806404096
author Hastuti, Rochana Prih
Suyanto, Yohanes
Sari, Anny Kartika
author_facet Hastuti, Rochana Prih
Suyanto, Yohanes
Sari, Anny Kartika
author_sort Hastuti, Rochana Prih
collection UGM
description Tree-LSTM algorithm accommodates tree structure processing to extract information outside the linear sequence pattern. The use of Tree-LSTM in text generation problems requires the help of an external parser at each generation iteration. Developing a good parser demands the representation of complex features and relies heavily on the grammar of the corpus. The limited corpus results in an insufficient number of vocabs for a grammar-based parser, making it less natural to link the text generation process. This research aims to solve the problem of limited corpus by proposing the use of a Reinforcement Learning algorithm in the formation of constituency trees, which link the sentence generation process given a seed phrase as the input in the Tree-LSTM model. The tree production process is modeled as a Markov's decision process, where a set of states consists of word embedding vectors, and a set of actions of {Shift, Reduce}. The Deep Q-Network model as an approximator of the Q-Learning algorithm is trained to obtain optimal weights in representing the Q-value function.The test results on perplexity-based evaluation show that the proposed Tree-LSTM and Q-Learning combination model achieves values 9.60 and 4.60 for two kinds of corpus with 205 and 1,000 sentences, respectively, better than the Shift-All model. Human evaluation of Friedman test and posthoc analysis showed that all five respondents tended to give the same assessment for the combination model of Tree-LSTM and Q-Learning, which on average outperforms two other nongrammar models, i.e., Shift-All and Reduce-All. © 2022 Copyright held by the owner/author(s). Publication rights licensed to ACM.
first_indexed 2024-03-14T00:10:28Z
format Other
id oai:generic.eprints.org:284475
institution Universiti Gadjah Mada
last_indexed 2024-03-14T00:10:28Z
publishDate 2022
publisher ACM Transactions on Asian and Low-Resource Language Information Processing
record_format dspace
spelling oai:generic.eprints.org:2844752024-01-02T08:46:46Z https://repository.ugm.ac.id/284475/ Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation Hastuti, Rochana Prih Suyanto, Yohanes Sari, Anny Kartika Information and Computing Sciences Tree-LSTM algorithm accommodates tree structure processing to extract information outside the linear sequence pattern. The use of Tree-LSTM in text generation problems requires the help of an external parser at each generation iteration. Developing a good parser demands the representation of complex features and relies heavily on the grammar of the corpus. The limited corpus results in an insufficient number of vocabs for a grammar-based parser, making it less natural to link the text generation process. This research aims to solve the problem of limited corpus by proposing the use of a Reinforcement Learning algorithm in the formation of constituency trees, which link the sentence generation process given a seed phrase as the input in the Tree-LSTM model. The tree production process is modeled as a Markov's decision process, where a set of states consists of word embedding vectors, and a set of actions of {Shift, Reduce}. The Deep Q-Network model as an approximator of the Q-Learning algorithm is trained to obtain optimal weights in representing the Q-value function.The test results on perplexity-based evaluation show that the proposed Tree-LSTM and Q-Learning combination model achieves values 9.60 and 4.60 for two kinds of corpus with 205 and 1,000 sentences, respectively, better than the Shift-All model. Human evaluation of Friedman test and posthoc analysis showed that all five respondents tended to give the same assessment for the combination model of Tree-LSTM and Q-Learning, which on average outperforms two other nongrammar models, i.e., Shift-All and Reduce-All. © 2022 Copyright held by the owner/author(s). Publication rights licensed to ACM. ACM Transactions on Asian and Low-Resource Language Information Processing 2022 Other NonPeerReviewed Hastuti, Rochana Prih and Suyanto, Yohanes and Sari, Anny Kartika (2022) Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation. ACM Transactions on Asian and Low-Resource Language Information Processing. https://dl.acm.org/doi/abs/10.1145/3490501 10.1145/3490501
spellingShingle Information and Computing Sciences
Hastuti, Rochana Prih
Suyanto, Yohanes
Sari, Anny Kartika
Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation
title Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation
title_full Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation
title_fullStr Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation
title_full_unstemmed Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation
title_short Q-Learning for Shift-Reduce Parsing in Indonesian Tree-LSTM-Based Text Generation
title_sort q learning for shift reduce parsing in indonesian tree lstm based text generation
topic Information and Computing Sciences
work_keys_str_mv AT hastutirochanaprih qlearningforshiftreduceparsinginindonesiantreelstmbasedtextgeneration
AT suyantoyohanes qlearningforshiftreduceparsinginindonesiantreelstmbasedtextgeneration
AT sariannykartika qlearningforshiftreduceparsinginindonesiantreelstmbasedtextgeneration