A survey of transformers

Transformers have achieved great success in many artificial intelligence fields, such as natural language processing, computer vision, and audio processing. Therefore, it is natural to attract lots of interest from academic and industry researchers. Up to the present, a great variety of Transformer...

Full description

Bibliographic Details
Main Authors: Tianyang Lin, Yuxin Wang, Xiangyang Liu, Xipeng Qiu
Format: Article
Language:English
Published: KeAi Communications Co. Ltd. 2022-01-01
Series:AI Open
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S2666651022000146
_version_ 1828087385067356160
author Tianyang Lin
Yuxin Wang
Xiangyang Liu
Xipeng Qiu
author_facet Tianyang Lin
Yuxin Wang
Xiangyang Liu
Xipeng Qiu
author_sort Tianyang Lin
collection DOAJ
description Transformers have achieved great success in many artificial intelligence fields, such as natural language processing, computer vision, and audio processing. Therefore, it is natural to attract lots of interest from academic and industry researchers. Up to the present, a great variety of Transformer variants (a.k.a. X-formers) have been proposed, however, a systematic and comprehensive literature review on these Transformer variants is still missing. In this survey, we provide a comprehensive review of various X-formers. We first briefly introduce the vanilla Transformer and then propose a new taxonomy of X-formers. Next, we introduce the various X-formers from three perspectives: architectural modification, pre-training, and applications. Finally, we outline some potential directions for future research.
first_indexed 2024-04-11T05:11:35Z
format Article
id doaj.art-26c69b88a20e47fb8f18f209ce250005
institution Directory Open Access Journal
issn 2666-6510
language English
last_indexed 2024-04-11T05:11:35Z
publishDate 2022-01-01
publisher KeAi Communications Co. Ltd.
record_format Article
series AI Open
spelling doaj.art-26c69b88a20e47fb8f18f209ce2500052022-12-25T04:19:31ZengKeAi Communications Co. Ltd.AI Open2666-65102022-01-013111132A survey of transformersTianyang Lin0Yuxin Wang1Xiangyang Liu2Xipeng Qiu3School of Computer Science, Fudan University, Shanghai, 200433, China; Shanghai Key Laboratory of Intelligent Information Processing, Shanghai, 200433, ChinaSchool of Computer Science, Fudan University, Shanghai, 200433, China; Shanghai Key Laboratory of Intelligent Information Processing, Shanghai, 200433, ChinaSchool of Computer Science, Fudan University, Shanghai, 200433, China; Shanghai Key Laboratory of Intelligent Information Processing, Shanghai, 200433, ChinaCorresponding author at: School of Computer Science, Fudan University, Shanghai, 200433, China; School of Computer Science, Fudan University, Shanghai, 200433, China; Shanghai Key Laboratory of Intelligent Information Processing, Shanghai, 200433, ChinaTransformers have achieved great success in many artificial intelligence fields, such as natural language processing, computer vision, and audio processing. Therefore, it is natural to attract lots of interest from academic and industry researchers. Up to the present, a great variety of Transformer variants (a.k.a. X-formers) have been proposed, however, a systematic and comprehensive literature review on these Transformer variants is still missing. In this survey, we provide a comprehensive review of various X-formers. We first briefly introduce the vanilla Transformer and then propose a new taxonomy of X-formers. Next, we introduce the various X-formers from three perspectives: architectural modification, pre-training, and applications. Finally, we outline some potential directions for future research.http://www.sciencedirect.com/science/article/pii/S2666651022000146TransformerSelf-attentionPre-trained modelsDeep learning
spellingShingle Tianyang Lin
Yuxin Wang
Xiangyang Liu
Xipeng Qiu
A survey of transformers
AI Open
Transformer
Self-attention
Pre-trained models
Deep learning
title A survey of transformers
title_full A survey of transformers
title_fullStr A survey of transformers
title_full_unstemmed A survey of transformers
title_short A survey of transformers
title_sort survey of transformers
topic Transformer
Self-attention
Pre-trained models
Deep learning
url http://www.sciencedirect.com/science/article/pii/S2666651022000146
work_keys_str_mv AT tianyanglin asurveyoftransformers
AT yuxinwang asurveyoftransformers
AT xiangyangliu asurveyoftransformers
AT xipengqiu asurveyoftransformers
AT tianyanglin surveyoftransformers
AT yuxinwang surveyoftransformers
AT xiangyangliu surveyoftransformers
AT xipengqiu surveyoftransformers