A Lite Romanian BERT: ALR-BERT

Large-scale pre-trained language representation and its promising performance in various downstream applications have become an area of interest in the field of natural language processing (NLP). There has been huge interest in further increasing the model’s size in order to outperform the best prev...

Full description

Bibliographic Details
Main Authors: Dragoş Constantin Nicolae, Rohan Kumar Yadav, Dan Tufiş
Format: Article
Language:English
Published: MDPI AG 2022-04-01
Series:Computers
Subjects:
Online Access:https://www.mdpi.com/2073-431X/11/4/57
_version_ 1797436287459786752
author Dragoş Constantin Nicolae
Rohan Kumar Yadav
Dan Tufiş
author_facet Dragoş Constantin Nicolae
Rohan Kumar Yadav
Dan Tufiş
author_sort Dragoş Constantin Nicolae
collection DOAJ
description Large-scale pre-trained language representation and its promising performance in various downstream applications have become an area of interest in the field of natural language processing (NLP). There has been huge interest in further increasing the model’s size in order to outperform the best previously obtained performances. However, at some point, increasing the model’s parameters may lead to reaching its saturation point due to the limited capacity of GPU/TPU. In addition to this, such models are mostly available in English or a shared multilingual structure. Hence, in this paper, we propose a lite BERT trained on a large corpus solely in the Romanian language, which we called “A Lite Romanian BERT (ALR-BERT)”. Based on comprehensive empirical results, ALR-BERT produces models that scale far better than the original Romanian BERT. Alongside presenting the performance on downstream tasks, we detail the analysis of the training process and its parameters. We also intend to distribute our code and model as an open source together with the downstream task.
first_indexed 2024-03-09T10:59:24Z
format Article
id doaj.art-ad4b78fb3c1b43369d7205561c284964
institution Directory Open Access Journal
issn 2073-431X
language English
last_indexed 2024-03-09T10:59:24Z
publishDate 2022-04-01
publisher MDPI AG
record_format Article
series Computers
spelling doaj.art-ad4b78fb3c1b43369d7205561c2849642023-12-01T01:22:30ZengMDPI AGComputers2073-431X2022-04-011145710.3390/computers11040057A Lite Romanian BERT: ALR-BERTDragoş Constantin Nicolae0Rohan Kumar Yadav1Dan Tufiş2Research Institute for Artificial Intelligence, Romanian Academy, 050711 Bucharest, RomaniaDepartment of Information and Communication, University of Agder, 4604 Grimstad, NorwayResearch Institute for Artificial Intelligence, Romanian Academy, 050711 Bucharest, RomaniaLarge-scale pre-trained language representation and its promising performance in various downstream applications have become an area of interest in the field of natural language processing (NLP). There has been huge interest in further increasing the model’s size in order to outperform the best previously obtained performances. However, at some point, increasing the model’s parameters may lead to reaching its saturation point due to the limited capacity of GPU/TPU. In addition to this, such models are mostly available in English or a shared multilingual structure. Hence, in this paper, we propose a lite BERT trained on a large corpus solely in the Romanian language, which we called “A Lite Romanian BERT (ALR-BERT)”. Based on comprehensive empirical results, ALR-BERT produces models that scale far better than the original Romanian BERT. Alongside presenting the performance on downstream tasks, we detail the analysis of the training process and its parameters. We also intend to distribute our code and model as an open source together with the downstream task.https://www.mdpi.com/2073-431X/11/4/57BERTtransformersALBERTNLPRomanian
spellingShingle Dragoş Constantin Nicolae
Rohan Kumar Yadav
Dan Tufiş
A Lite Romanian BERT: ALR-BERT
Computers
BERT
transformers
ALBERT
NLP
Romanian
title A Lite Romanian BERT: ALR-BERT
title_full A Lite Romanian BERT: ALR-BERT
title_fullStr A Lite Romanian BERT: ALR-BERT
title_full_unstemmed A Lite Romanian BERT: ALR-BERT
title_short A Lite Romanian BERT: ALR-BERT
title_sort lite romanian bert alr bert
topic BERT
transformers
ALBERT
NLP
Romanian
url https://www.mdpi.com/2073-431X/11/4/57
work_keys_str_mv AT dragosconstantinnicolae aliteromanianbertalrbert
AT rohankumaryadav aliteromanianbertalrbert
AT dantufis aliteromanianbertalrbert
AT dragosconstantinnicolae literomanianbertalrbert
AT rohankumaryadav literomanianbertalrbert
AT dantufis literomanianbertalrbert