Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization
Recent advances in low-resource abstractive summarization were largely made through the adoption of specialized pre-training, pseudo-summarization, that integrates the content selection knowledge through various centrality-based sentence recovery tasks. However, despite the substantial results, ther...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2024-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10474365/ |
_version_ | 1797221994343694336 |
---|---|
author | Daniil Chernyshev Boris Dobrov |
author_facet | Daniil Chernyshev Boris Dobrov |
author_sort | Daniil Chernyshev |
collection | DOAJ |
description | Recent advances in low-resource abstractive summarization were largely made through the adoption of specialized pre-training, pseudo-summarization, that integrates the content selection knowledge through various centrality-based sentence recovery tasks. However, despite the substantial results, there are several cases where the predecessor general-purpose pre-trained language model BART outperforms the summarization-specialized counterparts in both few-shot and fine-tuned scenarios. In this work, we investigate these performance irregularities and shed some light on the effect of pseudo-summarization pre-training in low-resource settings. We benchmarked five pre-trained abstractive summarization models on five datasets of diverse domains and analyzed their behavior in terms of extractive intuition and attention patterns. Despite that all models exhibit extractive behavior, some lack the prediction confidence to copy longer text fragments and have a misaligned attention distribution with the structure of the real-world texts. The latter happens to be the major factor of underperformance in fiction, news, and scientific articles domains as the better initial attention alignment of BART leads to the best benchmark results in all few-shot settings. A further examination reveals that BART summarization capabilities are the side-effect of the combination of sentence permutation task and specificities of the pre-training dataset. Based on the discovery we introduce Pegasus-SP, an improved pre-trained abstractive summarization model that unifies pseudo-summarization with sentence permutation. The new model outperforms the existing counterparts in low-resource settings and demonstrates superior adaptability. Additionally, we show that all pre-trained summarization models benefit from data-wise attention correction, achieving up to 10% relative ROUGE improvement on model-data pairs with the largest distribution discrepancies. |
first_indexed | 2024-04-24T13:14:16Z |
format | Article |
id | doaj.art-b870b80545c042528dd2d77cb21fc0f7 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-04-24T13:14:16Z |
publishDate | 2024-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-b870b80545c042528dd2d77cb21fc0f72024-04-04T23:00:42ZengIEEEIEEE Access2169-35362024-01-0112472194723010.1109/ACCESS.2024.337913910474365Investigating the Pre-Training Bias in Low-Resource Abstractive SummarizationDaniil Chernyshev0https://orcid.org/0009-0001-6847-2122Boris Dobrov1Research Computing Center, Lomonosov Moscow State University, Moscow, RussiaResearch Computing Center, Lomonosov Moscow State University, Moscow, RussiaRecent advances in low-resource abstractive summarization were largely made through the adoption of specialized pre-training, pseudo-summarization, that integrates the content selection knowledge through various centrality-based sentence recovery tasks. However, despite the substantial results, there are several cases where the predecessor general-purpose pre-trained language model BART outperforms the summarization-specialized counterparts in both few-shot and fine-tuned scenarios. In this work, we investigate these performance irregularities and shed some light on the effect of pseudo-summarization pre-training in low-resource settings. We benchmarked five pre-trained abstractive summarization models on five datasets of diverse domains and analyzed their behavior in terms of extractive intuition and attention patterns. Despite that all models exhibit extractive behavior, some lack the prediction confidence to copy longer text fragments and have a misaligned attention distribution with the structure of the real-world texts. The latter happens to be the major factor of underperformance in fiction, news, and scientific articles domains as the better initial attention alignment of BART leads to the best benchmark results in all few-shot settings. A further examination reveals that BART summarization capabilities are the side-effect of the combination of sentence permutation task and specificities of the pre-training dataset. Based on the discovery we introduce Pegasus-SP, an improved pre-trained abstractive summarization model that unifies pseudo-summarization with sentence permutation. The new model outperforms the existing counterparts in low-resource settings and demonstrates superior adaptability. Additionally, we show that all pre-trained summarization models benefit from data-wise attention correction, achieving up to 10% relative ROUGE improvement on model-data pairs with the largest distribution discrepancies.https://ieeexplore.ieee.org/document/10474365/Abstractive summarizationattention mechanismlow-resource text processingpre-trained language modelsmodel probing |
spellingShingle | Daniil Chernyshev Boris Dobrov Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization IEEE Access Abstractive summarization attention mechanism low-resource text processing pre-trained language models model probing |
title | Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization |
title_full | Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization |
title_fullStr | Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization |
title_full_unstemmed | Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization |
title_short | Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization |
title_sort | investigating the pre training bias in low resource abstractive summarization |
topic | Abstractive summarization attention mechanism low-resource text processing pre-trained language models model probing |
url | https://ieeexplore.ieee.org/document/10474365/ |
work_keys_str_mv | AT daniilchernyshev investigatingthepretrainingbiasinlowresourceabstractivesummarization AT borisdobrov investigatingthepretrainingbiasinlowresourceabstractivesummarization |