Self-supervised learning for automatic speech recognition In low-resource environments

Fatehi, Kavan (2024) Self-supervised learning for automatic speech recognition In low-resource environments. PhD thesis, University of Nottingham.

[thumbnail of Fatehi,Kavan,20167617,corrections.pdf]
Preview
PDF (Thesis - as examined) - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
Available under Licence Creative Commons Attribution.
Download (3MB) | Preview

Abstract

Supervised deep neural networks trained with substantial amounts of annotated speech data have demonstrated impressive performance across a spectrum of spoken language processing applications, frequently establishing themselves as the leading models in respective competitions. Nonetheless, a significant challenge arises from the heavy reliance on extensive annotated data for training these systems. This reliance poses a significant scalability limitation, hindering the continual enhancement of state-of-the-art performance. Moreover, it presents a more fundamental obstacle for deploying deep neural networks in speech-related domains where acquiring labeled data is inherently arduous, expensive, or time-intensive, which are considered as low-resource ASR problems in this thesis.

Unlike annotated speech data, collecting untranscribed audio is typically more cost-effective. In this thesis, we investigate the application of self-supervised learning in low-resource tasks, a learning approach where the learning objective is derived directly from the input data itself. We employ this method to harness the scalability and affordability of untranscribed audio resources in problems where we do not have enough training data, with the goal of enhancing the performance of spoken language technology. In particular, we propose three self-supervised methodologies. One model is based on the concept of two-fine-tuning steps, while the other two revolve around the notion of identifying an improved hidden unit. These approaches are designed to learn contextualized speech representations from speech data lacking annotations. We demonstrate the capacity of our self-supervised techniques to learn representations that convert the higher-level characteristics of speech signals more effectively than conventional acoustic features. Additionally, we present how these representations enhance the performance of deep neural networks on ASR tasks with limited resources. Beyond introducing novel learning algorithms, we conduct in-depth analyses to comprehend the properties of the acquired self-supervised representations and elucidate the distinct design elements that separate one self-supervised model from another.

Item Type: Thesis (University of Nottingham only) (PhD)
Supervisors: Kucukyilmaz, Ayse
French, Andrew
Keywords: Automatic Speech Recognition, Low-resource Environment, Self-Supervised Learning
Subjects: Q Science > QA Mathematics > QA 75 Electronic computers. Computer science
T Technology > TK Electrical engineering. Electronics Nuclear engineering
Faculties/Schools: UK Campuses > Faculty of Science > School of Computer Science
Item ID: 77884
Depositing User: Fatehi, Kavan
Date Deposited: 23 Jul 2024 04:40
Last Modified: 23 Jul 2024 04:40
URI: https://eprints.nottingham.ac.uk/id/eprint/77884

Actions (Archive Staff Only)

Edit View Edit View