Keywords: unsupervised speech recognition, syllabic unit discovery, self-supervised learning, speech tokenization
TL;DR: A syllable-level automatic speech recognizer trained with unpaired speech and text for langauge-universal speech technology
Abstract: Training speech recognizers with unpaired speech and text -- known as unsupervised speech recognition (UASR) -- is a crucial step toward extending ASR to low-resource languages in the long-tail distribution and enabling multimodal learning from non-parallel data. However, existing approaches based on phones often rely on costly resources such as grapheme-to-phoneme converters (G2Ps) and struggle to generalize to languages with ambiguous phoneme boundaries due to training instability. In this paper, we address both challenges by introducing a syllable-level UASR framework based on masked language modeling, which avoids the need for G2P and the instability of GAN-based methods. Our approach achieves up to a 40\% relative reduction in character error rate (CER) on LibriSpeech and generalizes effectively to Mandarin, a language that has remained particularly difficult for prior methods. Code will be released upon acceptance.
Primary Area: unsupervised, self-supervised, semi-supervised, and supervised representation learning
Submission Number: 3426
Loading