Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion

One of the main tasks in the field of natural language processing (NLP) is the analysis of affective states (sentiment and emotional) based on written text, and attempts have improved dramatically in recent years. However, in studies on the Arabic language, machine learning or deep learning algorith...

Full description

Bibliographic Details
Main Authors: Wafa Alshehri, Nora Al-Twairesh, Abdulrahman Alothaim
Format: Article
Language:English
Published: MDPI AG 2023-05-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/9/5609
_version_ 1797602978111160320
author Wafa Alshehri
Nora Al-Twairesh
Abdulrahman Alothaim
author_facet Wafa Alshehri
Nora Al-Twairesh
Abdulrahman Alothaim
author_sort Wafa Alshehri
collection DOAJ
description One of the main tasks in the field of natural language processing (NLP) is the analysis of affective states (sentiment and emotional) based on written text, and attempts have improved dramatically in recent years. However, in studies on the Arabic language, machine learning or deep learning algorithms were utilised to analyse sentiment and emotion more often than current pre-trained language models. Additionally, further pre-training the language model on specific tasks (i.e., within-task and cross-task adaptation) has not yet been investigated for Arabic in general, and for the sentiment and emotion task in particular. In this paper, we adapt a BERT-based Arabic pretrained language model for the sentiment and emotion tasks by further pre-training it on a sentiment and emotion corpus. Hence, we developed five new Arabic models: QST, QSR, QSRT, QE3, and QE6. Five sentiment and two emotion datasets spanning both small- and large-resource settings were used to evaluate the developed models. The adaptation approaches significantly enhanced the performance of seven Arabic sentiment and emotion datasets. The developed models showed excellent improvements over the sentiment and emotion datasets, which ranged from 0.15–4.71%.
first_indexed 2024-03-11T04:23:23Z
format Article
id doaj.art-8e0a74bca3b1481eabf210ff8f05f552
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-11T04:23:23Z
publishDate 2023-05-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-8e0a74bca3b1481eabf210ff8f05f5522023-11-17T22:36:23ZengMDPI AGApplied Sciences2076-34172023-05-01139560910.3390/app13095609Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and EmotionWafa Alshehri0Nora Al-Twairesh1Abdulrahman Alothaim2STC’s Artificial Intelligence Chair, College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi ArabiaSTC’s Artificial Intelligence Chair, College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi ArabiaSTC’s Artificial Intelligence Chair, College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi ArabiaOne of the main tasks in the field of natural language processing (NLP) is the analysis of affective states (sentiment and emotional) based on written text, and attempts have improved dramatically in recent years. However, in studies on the Arabic language, machine learning or deep learning algorithms were utilised to analyse sentiment and emotion more often than current pre-trained language models. Additionally, further pre-training the language model on specific tasks (i.e., within-task and cross-task adaptation) has not yet been investigated for Arabic in general, and for the sentiment and emotion task in particular. In this paper, we adapt a BERT-based Arabic pretrained language model for the sentiment and emotion tasks by further pre-training it on a sentiment and emotion corpus. Hence, we developed five new Arabic models: QST, QSR, QSRT, QE3, and QE6. Five sentiment and two emotion datasets spanning both small- and large-resource settings were used to evaluate the developed models. The adaptation approaches significantly enhanced the performance of seven Arabic sentiment and emotion datasets. The developed models showed excellent improvements over the sentiment and emotion datasets, which ranged from 0.15–4.71%.https://www.mdpi.com/2076-3417/13/9/5609sentiment analysisemotion detectionpretrained language modelsmodel adaptationtask-adaptation approach
spellingShingle Wafa Alshehri
Nora Al-Twairesh
Abdulrahman Alothaim
Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
Applied Sciences
sentiment analysis
emotion detection
pretrained language models
model adaptation
task-adaptation approach
title Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
title_full Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
title_fullStr Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
title_full_unstemmed Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
title_short Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
title_sort affect analysis in arabic text further pre training language models for sentiment and emotion
topic sentiment analysis
emotion detection
pretrained language models
model adaptation
task-adaptation approach
url https://www.mdpi.com/2076-3417/13/9/5609
work_keys_str_mv AT wafaalshehri affectanalysisinarabictextfurtherpretraininglanguagemodelsforsentimentandemotion
AT noraaltwairesh affectanalysisinarabictextfurtherpretraininglanguagemodelsforsentimentandemotion
AT abdulrahmanalothaim affectanalysisinarabictextfurtherpretraininglanguagemodelsforsentimentandemotion