The neural machine translation models for the low-resource Kazakh–English language pair
The development of the machine translation field was driven by people’s need to communicate with each other globally by automatically translating words, sentences, and texts from one language into another. The neural machine translation approach has become one of the most significant in recent years...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
PeerJ Inc.
2023-02-01
|
Series: | PeerJ Computer Science |
Subjects: | |
Online Access: | https://peerj.com/articles/cs-1224.pdf |
_version_ | 1811166561597128704 |
---|---|
author | Vladislav Karyukin Diana Rakhimova Aidana Karibayeva Aliya Turganbayeva Asem Turarbek |
author_facet | Vladislav Karyukin Diana Rakhimova Aidana Karibayeva Aliya Turganbayeva Asem Turarbek |
author_sort | Vladislav Karyukin |
collection | DOAJ |
description | The development of the machine translation field was driven by people’s need to communicate with each other globally by automatically translating words, sentences, and texts from one language into another. The neural machine translation approach has become one of the most significant in recent years. This approach requires large parallel corpora not available for low-resource languages, such as the Kazakh language, which makes it difficult to achieve the high performance of the neural machine translation models. This article explores the existing methods for dealing with low-resource languages by artificially increasing the size of the corpora and improving the performance of the Kazakh–English machine translation models. These methods are called forward translation, backward translation, and transfer learning. Then the Sequence-to-Sequence (recurrent neural network and bidirectional recurrent neural network) and Transformer neural machine translation architectures with their features and specifications are concerned for conducting experiments in training models on parallel corpora. The experimental part focuses on building translation models for the high-quality translation of formal social, political, and scientific texts with the synthetic parallel sentences from existing monolingual data in the Kazakh language using the forward translation approach and combining them with the parallel corpora parsed from the official government websites. The total corpora of 380,000 parallel Kazakh–English sentences are trained on the recurrent neural network, bidirectional recurrent neural network, and Transformer models of the OpenNMT framework. The quality of the trained model is evaluated with the BLEU, WER, and TER metrics. Moreover, the sample translations were also analyzed. The RNN and BRNN models showed a more precise translation than the Transformer model. The Byte-Pair Encoding tokenization technique showed better metrics scores and translation than the word tokenization technique. The Bidirectional recurrent neural network with the Byte-Pair Encoding technique showed the best performance with 0.49 BLEU, 0.51 WER, and 0.45 TER. |
first_indexed | 2024-04-10T15:54:33Z |
format | Article |
id | doaj.art-0e7d121028c54c5caf59ed5a18e59c3c |
institution | Directory Open Access Journal |
issn | 2376-5992 |
language | English |
last_indexed | 2024-04-10T15:54:33Z |
publishDate | 2023-02-01 |
publisher | PeerJ Inc. |
record_format | Article |
series | PeerJ Computer Science |
spelling | doaj.art-0e7d121028c54c5caf59ed5a18e59c3c2023-02-10T15:05:12ZengPeerJ Inc.PeerJ Computer Science2376-59922023-02-019e122410.7717/peerj-cs.1224The neural machine translation models for the low-resource Kazakh–English language pairVladislav Karyukin0Diana Rakhimova1Aidana Karibayeva2Aliya Turganbayeva3Asem Turarbek4Department of Information Systems, Al-Farabi Kazakh National University, Almaty, KazakhstanDepartment of Information Systems, Al-Farabi Kazakh National University, Almaty, KazakhstanDepartment of Information Systems, Al-Farabi Kazakh National University, Almaty, KazakhstanDepartment of Information Systems, Al-Farabi Kazakh National University, Almaty, KazakhstanDepartment of Information Systems, Al-Farabi Kazakh National University, Almaty, KazakhstanThe development of the machine translation field was driven by people’s need to communicate with each other globally by automatically translating words, sentences, and texts from one language into another. The neural machine translation approach has become one of the most significant in recent years. This approach requires large parallel corpora not available for low-resource languages, such as the Kazakh language, which makes it difficult to achieve the high performance of the neural machine translation models. This article explores the existing methods for dealing with low-resource languages by artificially increasing the size of the corpora and improving the performance of the Kazakh–English machine translation models. These methods are called forward translation, backward translation, and transfer learning. Then the Sequence-to-Sequence (recurrent neural network and bidirectional recurrent neural network) and Transformer neural machine translation architectures with their features and specifications are concerned for conducting experiments in training models on parallel corpora. The experimental part focuses on building translation models for the high-quality translation of formal social, political, and scientific texts with the synthetic parallel sentences from existing monolingual data in the Kazakh language using the forward translation approach and combining them with the parallel corpora parsed from the official government websites. The total corpora of 380,000 parallel Kazakh–English sentences are trained on the recurrent neural network, bidirectional recurrent neural network, and Transformer models of the OpenNMT framework. The quality of the trained model is evaluated with the BLEU, WER, and TER metrics. Moreover, the sample translations were also analyzed. The RNN and BRNN models showed a more precise translation than the Transformer model. The Byte-Pair Encoding tokenization technique showed better metrics scores and translation than the word tokenization technique. The Bidirectional recurrent neural network with the Byte-Pair Encoding technique showed the best performance with 0.49 BLEU, 0.51 WER, and 0.45 TER.https://peerj.com/articles/cs-1224.pdfNeural machine translationForward translationBackward translationSeq2SeqRNNBRNN |
spellingShingle | Vladislav Karyukin Diana Rakhimova Aidana Karibayeva Aliya Turganbayeva Asem Turarbek The neural machine translation models for the low-resource Kazakh–English language pair PeerJ Computer Science Neural machine translation Forward translation Backward translation Seq2Seq RNN BRNN |
title | The neural machine translation models for the low-resource Kazakh–English language pair |
title_full | The neural machine translation models for the low-resource Kazakh–English language pair |
title_fullStr | The neural machine translation models for the low-resource Kazakh–English language pair |
title_full_unstemmed | The neural machine translation models for the low-resource Kazakh–English language pair |
title_short | The neural machine translation models for the low-resource Kazakh–English language pair |
title_sort | neural machine translation models for the low resource kazakh english language pair |
topic | Neural machine translation Forward translation Backward translation Seq2Seq RNN BRNN |
url | https://peerj.com/articles/cs-1224.pdf |
work_keys_str_mv | AT vladislavkaryukin theneuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT dianarakhimova theneuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT aidanakaribayeva theneuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT aliyaturganbayeva theneuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT asemturarbek theneuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT vladislavkaryukin neuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT dianarakhimova neuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT aidanakaribayeva neuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT aliyaturganbayeva neuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair AT asemturarbek neuralmachinetranslationmodelsforthelowresourcekazakhenglishlanguagepair |