DiffuD2T: Empowering Data-to-Text Generation with Diffusion
Surrounded by structured data, such as medical data, financial data, knowledge bases, etc., data-to-text generation has become an important natural language processing task that can help people better understand the meaning of those data by providing them with user-friendly text. Existing methods fo...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2023-05-01
|
Series: | Electronics |
Subjects: | |
Online Access: | https://www.mdpi.com/2079-9292/12/9/2136 |
_version_ | 1797602791812759552 |
---|---|
author | Heng Gong Xiaocheng Feng Bing Qin |
author_facet | Heng Gong Xiaocheng Feng Bing Qin |
author_sort | Heng Gong |
collection | DOAJ |
description | Surrounded by structured data, such as medical data, financial data, knowledge bases, etc., data-to-text generation has become an important natural language processing task that can help people better understand the meaning of those data by providing them with user-friendly text. Existing methods for data-to-text generation show promising results in tackling two major challenges: content planning and surface realization, which transform structured data into fluent text. However, they lack an iterative refinement process for generating text, which can enable the model to perfect the text step-by-step while accepting control over the process. In this paper, we explore enhancing data-to-text generation with an iterative refinement process via diffusion. We have four main contributions: (1) we use the diffusion model to improve the prefix tuning for data-to-text generation; (2) we propose a look-ahead guiding loss to supervise the iterative refinement process for better text generation; (3) we extract content plans from reference text and propose a planning-then-writing pipeline to give the model content planning ability; and (4) we conducted experiments on three data-to-text generation datasets and both automatic evaluation criteria (BLEU, NIST, METEOR, <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><msub><mi>ROUGE</mi><mi>L</mi></msub></semantics></math></inline-formula>, CIDEr, TER, MoverScore, BLEURT, and BERTScore) and human evaluation criteria (Quality and Naturalness) show the effectiveness of our model. Our model can improve the competitive prefix tuning method by 2.19% in terms of a widely-used automatic evaluation criterion BLEU (BiLingual Evaluation Understudy) on WebNLG dataset with GPT-2 Large as the pretrained language model backbone. Human evaluation criteria also show that our model can improve the quality and naturalness of the generated text across all three datasets. |
first_indexed | 2024-03-11T04:20:29Z |
format | Article |
id | doaj.art-c100428eef0f4a068c6e7d7b34236ac1 |
institution | Directory Open Access Journal |
issn | 2079-9292 |
language | English |
last_indexed | 2024-03-11T04:20:29Z |
publishDate | 2023-05-01 |
publisher | MDPI AG |
record_format | Article |
series | Electronics |
spelling | doaj.art-c100428eef0f4a068c6e7d7b34236ac12023-11-17T22:49:09ZengMDPI AGElectronics2079-92922023-05-01129213610.3390/electronics12092136DiffuD2T: Empowering Data-to-Text Generation with DiffusionHeng Gong0Xiaocheng Feng1Bing Qin2School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, ChinaSchool of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, ChinaSchool of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, ChinaSurrounded by structured data, such as medical data, financial data, knowledge bases, etc., data-to-text generation has become an important natural language processing task that can help people better understand the meaning of those data by providing them with user-friendly text. Existing methods for data-to-text generation show promising results in tackling two major challenges: content planning and surface realization, which transform structured data into fluent text. However, they lack an iterative refinement process for generating text, which can enable the model to perfect the text step-by-step while accepting control over the process. In this paper, we explore enhancing data-to-text generation with an iterative refinement process via diffusion. We have four main contributions: (1) we use the diffusion model to improve the prefix tuning for data-to-text generation; (2) we propose a look-ahead guiding loss to supervise the iterative refinement process for better text generation; (3) we extract content plans from reference text and propose a planning-then-writing pipeline to give the model content planning ability; and (4) we conducted experiments on three data-to-text generation datasets and both automatic evaluation criteria (BLEU, NIST, METEOR, <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><msub><mi>ROUGE</mi><mi>L</mi></msub></semantics></math></inline-formula>, CIDEr, TER, MoverScore, BLEURT, and BERTScore) and human evaluation criteria (Quality and Naturalness) show the effectiveness of our model. Our model can improve the competitive prefix tuning method by 2.19% in terms of a widely-used automatic evaluation criterion BLEU (BiLingual Evaluation Understudy) on WebNLG dataset with GPT-2 Large as the pretrained language model backbone. Human evaluation criteria also show that our model can improve the quality and naturalness of the generated text across all three datasets.https://www.mdpi.com/2079-9292/12/9/2136diffusiondata-to-text generationnatural language processingartificial intelligence |
spellingShingle | Heng Gong Xiaocheng Feng Bing Qin DiffuD2T: Empowering Data-to-Text Generation with Diffusion Electronics diffusion data-to-text generation natural language processing artificial intelligence |
title | DiffuD2T: Empowering Data-to-Text Generation with Diffusion |
title_full | DiffuD2T: Empowering Data-to-Text Generation with Diffusion |
title_fullStr | DiffuD2T: Empowering Data-to-Text Generation with Diffusion |
title_full_unstemmed | DiffuD2T: Empowering Data-to-Text Generation with Diffusion |
title_short | DiffuD2T: Empowering Data-to-Text Generation with Diffusion |
title_sort | diffud2t empowering data to text generation with diffusion |
topic | diffusion data-to-text generation natural language processing artificial intelligence |
url | https://www.mdpi.com/2079-9292/12/9/2136 |
work_keys_str_mv | AT henggong diffud2tempoweringdatatotextgenerationwithdiffusion AT xiaochengfeng diffud2tempoweringdatatotextgenerationwithdiffusion AT bingqin diffud2tempoweringdatatotextgenerationwithdiffusion |