Understanding Self-Supervised Learning of Speech Representation via Invariance and Redundancy Reduction
Self-supervised learning (SSL) has emerged as a promising paradigm for learning flexible speech representations from <i>unlabeled</i> data. By designing <i>pretext tasks</i> that exploit statistical regularities, SSL models can capture <i>useful</i> representation...
Main Authors: | Yusuf Brima, Ulf Krumnack, Simone Pika, Gunther Heidemann |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2024-02-01
|
Series: | Information |
Subjects: | |
Online Access: | https://www.mdpi.com/2078-2489/15/2/114 |
Similar Items
-
Learning An Invariant Speech Representation
by: Evangelopoulos, Georgios, et al.
Published: (2015) -
Evaluating Self-Supervised Speech Representations for Speech Emotion Recognition
by: Bagus Tris Atmaja, et al.
Published: (2022-01-01) -
Evaluating and Improving Domain Invariance in Contrastive Self-Supervised Learning by Extrapolating the Loss Function
by: Samira Zare, et al.
Published: (2023-01-01) -
Spectral Salt-and-Pepper Patch Masking for Self-Supervised Speech Representation Learning
by: June-Woo Kim, et al.
Published: (2023-08-01) -
Length-Normalized Representation Learning for Speech Signals
by: Kyungguen Byun, et al.
Published: (2022-01-01)