An Empirical Study on Automatic Post Editing for Neural Machine Translation

Automatic post editing (APE) researches aim to correct errors in the machine translation results. Recently, APE research has mainly been conducted in two directions: noise-based APE and adapter-based APE. This study poses three questions based on existing APE studies and conducted a verification. Th...

Full description

Bibliographic Details
Main Authors: Hyeonseok Moon, Chanjun Park, Sugyeong Eo, Jaehyung Seo, Heuiseok Lim
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9528385/
_version_ 1819082802788827136
author Hyeonseok Moon
Chanjun Park
Sugyeong Eo
Jaehyung Seo
Heuiseok Lim
author_facet Hyeonseok Moon
Chanjun Park
Sugyeong Eo
Jaehyung Seo
Heuiseok Lim
author_sort Hyeonseok Moon
collection DOAJ
description Automatic post editing (APE) researches aim to correct errors in the machine translation results. Recently, APE research has mainly been conducted in two directions: noise-based APE and adapter-based APE. This study poses three questions based on existing APE studies and conducted a verification. The first is a question about the optimal APE research direction, and this has been figured out through a comparative analysis of the previous studies on noise-based APE and adapter-based APE. The second is about the substantial effectiveness of the bottleneck adapter layer (BAL) in adapter based APE. For the verification, various experiments on the different size of BAL has been conducted, and through these experiments, optimal approaches in adapter based APE has been proposed. For the last, this work raises a question about the reason why leveraging external knowledge is influential in APE. In this regard, we conducted several comparative experiments on the method of utilizing external data to APE training to achieve a better performance. The results revealed that the performance can be improved by applying the method of concatenating the external data with the existing data when training the APE model.Through deep analysis on these experiments, this work propose the optimal research direction in APE.
first_indexed 2024-12-21T20:22:28Z
format Article
id doaj.art-e6f19f7523014d83bf9bd10b82a1ce08
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-21T20:22:28Z
publishDate 2021-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-e6f19f7523014d83bf9bd10b82a1ce082022-12-21T18:51:28ZengIEEEIEEE Access2169-35362021-01-01912375412376310.1109/ACCESS.2021.31099039528385An Empirical Study on Automatic Post Editing for Neural Machine TranslationHyeonseok Moon0https://orcid.org/0000-0002-0841-4262Chanjun Park1https://orcid.org/0000-0002-7200-9632Sugyeong Eo2Jaehyung Seo3https://orcid.org/0000-0002-4761-9818Heuiseok Lim4https://orcid.org/0000-0002-9269-1157Department of Computer Science and Engineering, Korea University, Seoul, Republic of KoreaDepartment of Computer Science and Engineering, Korea University, Seoul, Republic of KoreaDepartment of Computer Science and Engineering, Korea University, Seoul, Republic of KoreaDepartment of Computer Science and Engineering, Korea University, Seoul, Republic of KoreaDepartment of Computer Science and Engineering, Korea University, Seoul, Republic of KoreaAutomatic post editing (APE) researches aim to correct errors in the machine translation results. Recently, APE research has mainly been conducted in two directions: noise-based APE and adapter-based APE. This study poses three questions based on existing APE studies and conducted a verification. The first is a question about the optimal APE research direction, and this has been figured out through a comparative analysis of the previous studies on noise-based APE and adapter-based APE. The second is about the substantial effectiveness of the bottleneck adapter layer (BAL) in adapter based APE. For the verification, various experiments on the different size of BAL has been conducted, and through these experiments, optimal approaches in adapter based APE has been proposed. For the last, this work raises a question about the reason why leveraging external knowledge is influential in APE. In this regard, we conducted several comparative experiments on the method of utilizing external data to APE training to achieve a better performance. The results revealed that the performance can be improved by applying the method of concatenating the external data with the existing data when training the APE model.Through deep analysis on these experiments, this work propose the optimal research direction in APE.https://ieeexplore.ieee.org/document/9528385/Automatic post editingneural machine translationadaptermultilingual pretrained language modelexternal knowledge
spellingShingle Hyeonseok Moon
Chanjun Park
Sugyeong Eo
Jaehyung Seo
Heuiseok Lim
An Empirical Study on Automatic Post Editing for Neural Machine Translation
IEEE Access
Automatic post editing
neural machine translation
adapter
multilingual pretrained language model
external knowledge
title An Empirical Study on Automatic Post Editing for Neural Machine Translation
title_full An Empirical Study on Automatic Post Editing for Neural Machine Translation
title_fullStr An Empirical Study on Automatic Post Editing for Neural Machine Translation
title_full_unstemmed An Empirical Study on Automatic Post Editing for Neural Machine Translation
title_short An Empirical Study on Automatic Post Editing for Neural Machine Translation
title_sort empirical study on automatic post editing for neural machine translation
topic Automatic post editing
neural machine translation
adapter
multilingual pretrained language model
external knowledge
url https://ieeexplore.ieee.org/document/9528385/
work_keys_str_mv AT hyeonseokmoon anempiricalstudyonautomaticposteditingforneuralmachinetranslation
AT chanjunpark anempiricalstudyonautomaticposteditingforneuralmachinetranslation
AT sugyeongeo anempiricalstudyonautomaticposteditingforneuralmachinetranslation
AT jaehyungseo anempiricalstudyonautomaticposteditingforneuralmachinetranslation
AT heuiseoklim anempiricalstudyonautomaticposteditingforneuralmachinetranslation
AT hyeonseokmoon empiricalstudyonautomaticposteditingforneuralmachinetranslation
AT chanjunpark empiricalstudyonautomaticposteditingforneuralmachinetranslation
AT sugyeongeo empiricalstudyonautomaticposteditingforneuralmachinetranslation
AT jaehyungseo empiricalstudyonautomaticposteditingforneuralmachinetranslation
AT heuiseoklim empiricalstudyonautomaticposteditingforneuralmachinetranslation