Sequence-to-sequence pretraining for a less-resourced Slovenian language

IntroductionLarge pretrained language models have recently conquered the area of natural language processing. As an alternative to predominant masked language modeling introduced in BERT, the T5 model has introduced a more general training objective, namely sequence to sequence transformation, which...

Full description

Bibliographic Details
Main Authors: Matej Ulčar, Marko Robnik-Šikonja
Format: Article
Language:English
Published: Frontiers Media S.A. 2023-03-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2023.932519/full