Enhancing Local Dependencies for Transformer-Based Text-to-Speech via Hybrid Lightweight Convolution
Owing to the powerful self-attention mechanism, the Transformer network has achieved considerable successes across many sequence modeling tasks and has become one of the most popular methods in text-to-speech (TTS). The vanilla self-attention excels in capturing long-range dependencies but suffers i...
Main Authors: | Wei Zhao, Ting He, Li Xu |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9376936/ |
Similar Items
-
Transformer and Graph Convolutional Network for Text Classification
by: Boting Liu, et al.
Published: (2023-10-01) -
NSE-CATNet: Deep Neural Speech Enhancement Using Convolutional Attention Transformer Network
by: Nasir Saleem, et al.
Published: (2023-01-01) -
Lightweight Scene Text Recognition Based on Transformer
by: Xin Luan, et al.
Published: (2023-05-01) -
A Lightweight Convolutional Neural Network Based on Group-Wise Hybrid Attention for Remote Sensing Scene Classification
by: Cuiping Shi, et al.
Published: (2021-12-01) -
A Speech Recognition Model Building Method Combined Dynamic Convolution and Multi-Head Self-Attention Mechanism
by: Wei Liu, et al.
Published: (2022-05-01)