BabySLM : language-acquisition-friendly benchmark of self-supervised spoken language models
Lavechin, Marvin; Sy, Yaya; Titeux, Hadrien; Blandón, María Andrea Cruz; Räsänen, Okko; Bredin, Hervé; Dupoux, Emmanuel; Cristia, Alejandrina (2023)
Lavechin, Marvin
Sy, Yaya
Titeux, Hadrien
Blandón, María Andrea Cruz
Räsänen, Okko
Bredin, Hervé
Dupoux, Emmanuel
Cristia, Alejandrina
International Speech Communication Association
2023
This publication is copyrighted. You may download, display and print it for Your own personal use. Commercial use is prohibited
Julkaisun pysyvä osoite on
https://urn.fi/URN:NBN:fi:tuni-2023112810299
https://urn.fi/URN:NBN:fi:tuni-2023112810299
Kuvaus
Peer reviewed
Tiivistelmä
Self-supervised techniques for learning speech representations have been shown to develop linguistic competence from exposure to speech without the need for human labels. In order to fully realize the potential of these approaches and further our understanding of how infants learn language, simulations must closely emulate real-life situations by training on developmentally plausible corpora and benchmarking against appropriate test sets. To this end, we propose a language-acquisition-friendly benchmark to probe spoken language models at the lexical and syntactic levels, both of which are compatible with the vocabulary typical of children's language experiences. This paper introduces the benchmark and summarizes a range of experiments showing its usefulness. In addition, we highlight two exciting challenges that need to be addressed for further progress: bridging the gap between text and speech and between clean speech and in-the-wild speech.
Kokoelmat
- TUNICRIS-julkaisut [19195]