A perceptual similarity space for speech based on self-supervised speech representations

Bronya R. Chernyak, Ann R. Bradlow, Joseph Keshet, Matthew Goldrick

Research output: Contribution to journalArticlepeer-review

Abstract

Speech recognition by both humans and machines frequently fails in non-optimal yet common situations. For example, word recognition error rates for second-language (L2) speech can be high, especially under conditions involving background noise. At the same time, both human and machine speech recognition sometimes shows remarkable robustness against signal- and noise-related degradation. Which acoustic features of speech explain this substantial variation in intelligibility? Current approaches align speech to text to extract a small set of pre-defined spectro-temporal properties from specific sounds in particular words. However, variation in these properties leaves much cross-talker variation in intelligibility unexplained. We examine an alternative approach utilizing a perceptual similarity space acquired using self-supervised learning. This approach encodes distinctions between speech samples without requiring pre-defined acoustic features or speech-to-text alignment. We show that L2 English speech samples are less tightly clustered in the space than L1 samples reflecting variability in English proficiency among L2 talkers. Critically, distances in this similarity space are perceptually meaningful: L1 English listeners have lower recognition accuracy for L2 speakers whose speech is more distant in the space from L1 speech. These results indicate that perceptual similarity may form the basis for an entirely new speech and language analysis approach.

Original languageEnglish
Pages (from-to)3915-3929
Number of pages15
JournalJournal of the Acoustical Society of America
Volume155
Issue number6
DOIs
StatePublished - 1 Jun 2024
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2024 Author(s).

Fingerprint

Dive into the research topics of 'A perceptual similarity space for speech based on self-supervised speech representations'. Together they form a unique fingerprint.

Cite this