Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model

The task of multi-hop question generation (QG) seeks to generate questions that require a complex reasoning process that spans multiple sentences and answers. Beyond the conventional challenges of what to ask and how to ask, multi-hop QG necessitates sophisticated reasoning from dispersed evidence a...

Full description

Bibliographic Details
Main Authors: Zhenping Li, Zhen Cao, Pengfei Li, Yong Zhong, Shaobo Li
Format: Article
Language:English
Published: MDPI AG 2023-05-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/9/5765
_version_ 1797602946820603904
author Zhenping Li
Zhen Cao
Pengfei Li
Yong Zhong
Shaobo Li
author_facet Zhenping Li
Zhen Cao
Pengfei Li
Yong Zhong
Shaobo Li
author_sort Zhenping Li
collection DOAJ
description The task of multi-hop question generation (QG) seeks to generate questions that require a complex reasoning process that spans multiple sentences and answers. Beyond the conventional challenges of what to ask and how to ask, multi-hop QG necessitates sophisticated reasoning from dispersed evidence across multiple sentences. To address these challenges, a knowledge graph-enhanced language model (KGEL) has been developed to imitate human reasoning for multi-hop questions.The initial step in KGEL involves encoding the input sentence with a pre-trained GPT-2 language model to obtain a comprehensive semantic context representation. Next, a knowledge graph is constructed using the entities identified within the context. The critical information in the graph that is related to the answer is then utilized to update the context representations through an answer-aware graph attention network (GAT). Finally, the multi-head attention generation module (MHAG) is performed over the updated latent representations of the context to generate coherent questions. Human evaluations demonstrate that KGEL generates more logical and fluent multi-hop questions compared to GPT-2. Furthermore, KGEL outperforms five prominent baselines in automatic evaluations, with a BLEU-4 score that is 27% higher than that of GPT-2.
first_indexed 2024-03-11T04:23:43Z
format Article
id doaj.art-eff6ab4e02114366bbf33b0deb4c2165
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-11T04:23:43Z
publishDate 2023-05-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-eff6ab4e02114366bbf33b0deb4c21652023-11-17T22:38:30ZengMDPI AGApplied Sciences2076-34172023-05-01139576510.3390/app13095765Multi-Hop Question Generation with Knowledge Graph-Enhanced Language ModelZhenping Li0Zhen Cao1Pengfei Li2Yong Zhong3Shaobo Li4Chengdu Institute of Computer Applications, Chinese Academy of Sciences, Chengdu 610041, ChinaSchool of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 639798, SingaporeSchool of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 639798, SingaporeChengdu Institute of Computer Applications, Chinese Academy of Sciences, Chengdu 610041, ChinaChengdu Institute of Computer Applications, Chinese Academy of Sciences, Chengdu 610041, ChinaThe task of multi-hop question generation (QG) seeks to generate questions that require a complex reasoning process that spans multiple sentences and answers. Beyond the conventional challenges of what to ask and how to ask, multi-hop QG necessitates sophisticated reasoning from dispersed evidence across multiple sentences. To address these challenges, a knowledge graph-enhanced language model (KGEL) has been developed to imitate human reasoning for multi-hop questions.The initial step in KGEL involves encoding the input sentence with a pre-trained GPT-2 language model to obtain a comprehensive semantic context representation. Next, a knowledge graph is constructed using the entities identified within the context. The critical information in the graph that is related to the answer is then utilized to update the context representations through an answer-aware graph attention network (GAT). Finally, the multi-head attention generation module (MHAG) is performed over the updated latent representations of the context to generate coherent questions. Human evaluations demonstrate that KGEL generates more logical and fluent multi-hop questions compared to GPT-2. Furthermore, KGEL outperforms five prominent baselines in automatic evaluations, with a BLEU-4 score that is 27% higher than that of GPT-2.https://www.mdpi.com/2076-3417/13/9/5765multi-hop question generationgraph neural networknatural language processingreasoning chain
spellingShingle Zhenping Li
Zhen Cao
Pengfei Li
Yong Zhong
Shaobo Li
Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model
Applied Sciences
multi-hop question generation
graph neural network
natural language processing
reasoning chain
title Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model
title_full Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model
title_fullStr Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model
title_full_unstemmed Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model
title_short Multi-Hop Question Generation with Knowledge Graph-Enhanced Language Model
title_sort multi hop question generation with knowledge graph enhanced language model
topic multi-hop question generation
graph neural network
natural language processing
reasoning chain
url https://www.mdpi.com/2076-3417/13/9/5765
work_keys_str_mv AT zhenpingli multihopquestiongenerationwithknowledgegraphenhancedlanguagemodel
AT zhencao multihopquestiongenerationwithknowledgegraphenhancedlanguagemodel
AT pengfeili multihopquestiongenerationwithknowledgegraphenhancedlanguagemodel
AT yongzhong multihopquestiongenerationwithknowledgegraphenhancedlanguagemodel
AT shaoboli multihopquestiongenerationwithknowledgegraphenhancedlanguagemodel