From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough

With the recent advances in deep learning, different approaches to improving pre-trained language models (PLMs) have been proposed. PLMs have advanced state-of-the-art (SOTA) performance on various natural language processing (NLP) tasks such as machine translation, text classification, question ans...

Full description

Bibliographic Details
Main Author: Mourad Mars
Format: Article
Language:English
Published: MDPI AG 2022-09-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/12/17/8805
_version_ 1797496260732649472
author Mourad Mars
author_facet Mourad Mars
author_sort Mourad Mars
collection DOAJ
description With the recent advances in deep learning, different approaches to improving pre-trained language models (PLMs) have been proposed. PLMs have advanced state-of-the-art (SOTA) performance on various natural language processing (NLP) tasks such as machine translation, text classification, question answering, text summarization, information retrieval, recommendation systems, named entity recognition, etc. In this paper, we provide a comprehensive review of prior embedding models as well as current breakthroughs in the field of PLMs. Then, we analyse and contrast the various models and provide an analysis of the way they have been built (number of parameters, compression techniques, etc.). Finally, we discuss the major issues and future directions for each of the main points.
first_indexed 2024-03-10T03:01:07Z
format Article
id doaj.art-bfe5448ccfc141ecb2e01a278b3f1985
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-10T03:01:07Z
publishDate 2022-09-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-bfe5448ccfc141ecb2e01a278b3f19852023-11-23T12:47:09ZengMDPI AGApplied Sciences2076-34172022-09-011217880510.3390/app12178805From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art WalkthroughMourad Mars0College of Computers and Information Systems, Umm Al-Qura University, Mecca 24382, Saudi ArabiaWith the recent advances in deep learning, different approaches to improving pre-trained language models (PLMs) have been proposed. PLMs have advanced state-of-the-art (SOTA) performance on various natural language processing (NLP) tasks such as machine translation, text classification, question answering, text summarization, information retrieval, recommendation systems, named entity recognition, etc. In this paper, we provide a comprehensive review of prior embedding models as well as current breakthroughs in the field of PLMs. Then, we analyse and contrast the various models and provide an analysis of the way they have been built (number of parameters, compression techniques, etc.). Finally, we discuss the major issues and future directions for each of the main points.https://www.mdpi.com/2076-3417/12/17/8805artificial intelligenceNLPpre-trained language model
spellingShingle Mourad Mars
From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
Applied Sciences
artificial intelligence
NLP
pre-trained language model
title From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
title_full From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
title_fullStr From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
title_full_unstemmed From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
title_short From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
title_sort from word embeddings to pre trained language models a state of the art walkthrough
topic artificial intelligence
NLP
pre-trained language model
url https://www.mdpi.com/2076-3417/12/17/8805
work_keys_str_mv AT mouradmars fromwordembeddingstopretrainedlanguagemodelsastateoftheartwalkthrough