Pruning Adapters with Lottery Ticket

Massively pre-trained transformer models such as BERT have gained great success in many downstream NLP tasks. However, they are computationally expensive to fine-tune, slow for inference, and have large storage requirements. So, transfer learning with adapter modules has been introduced and has beco...

Full description

Bibliographic Details
Main Authors: Jiarun Wu, Qingliang Chen
Format: Article
Language:English
Published: MDPI AG 2022-02-01
Series:Algorithms
Subjects:
Online Access:https://www.mdpi.com/1999-4893/15/2/63