Timezone: »
The Zero Resource Speech Challenge is a series that has been running since 2015, which aims to advance research in unsupervised training of speech and dialogue tools, with an application in speech technology for under-resourced languages. This year, we are running an "enhanced" version of the newest challenge task, language modelling from speech. This task asks participants to learn a sequential model that can assign probabilities to sequences---like a typical language model---but which must be trained, and operate, without any text. Assessing and improving on our ability to build such a model is critical to expanding applications such as speech recognition and machine translation to languages without textual resources. The "enhanced" version makes two modifications: it expands the call for submissions to the "high GPU budget" category, encouraging very large models in addition to the smaller, "lower-budget" ones experimented with up to now; and it includes a new, experimental "multi-modal" track, which allows participants to assess the performance of models that include images in training, in addition to audio. Baseline models are already prepared and evaluated for the high-budget and multi-modal settings.
Author Information
Ewan Dunbar (University of Toronto)
Alejandrina Cristia (École Normale Supérieure)
Okko Räsänen (Tampere University)
Bertrand Higy (Tilburg University)
Marvin Lavechin (École Normale Supérieure)
Grzegorz Chrupała (Tilburg University)
Afra Alishahi (Tilburg University)
Chen Yu (Indiana University)
Maureen De Seyssel (École Normale Supérieure)
Tu Anh Nguyen (INRIA, Paris, France)
Mathieu Bernard (Cognitive Machine Learning, EHESS, ENS-PSL University, CNRS, INRIA)
Nicolas Hamilakis (Cognitive Machine Learning, EHESS, ENS-PSL University, CNRS, INRIA)
Emmanuel Dupoux (Facebook)
More from the Same Authors
-
2022 Poster: Emergent Communication: Generalization and Overfitting in Lewis Games »
Mathieu Rita · Corentin Tallec · Paul Michel · Jean-Bastien Grill · Olivier Pietquin · Emmanuel Dupoux · Florian Strub -
2022 Poster: Toward a realistic model of speech processing in the brain with self-supervised learning »
Juliette MILLET · Charlotte Caucheteux · pierre orhan · Yves Boubenec · Alexandre Gramfort · Ewan Dunbar · Christophe Pallier · Jean-Remi King -
2020 : The Zero Resource Speech Benchmark 2021. Metrics and baselines for unsupervised spoken language modeling »
Tu Anh Nguyen -
2020 Workshop: BabyMind: How Babies Learn and How Machines Can Imitate »
Byoung-Tak Zhang · Gary Marcus · Angelo Cangelosi · Pia Knoeferle · Klaus Obermayer · David Vernon · Chen Yu -
2019 Poster: A Self Validation Network for Object-Level Human Attention Estimation »
Zehua Zhang · Chen Yu · David Crandall -
2019 Poster: Anti-efficient encoding in emergent communication »
Rahma Chaabouni · Eugene Kharitonov · Emmanuel Dupoux · Marco Baroni -
2018 Workshop: Modeling the Physical World: Learning, Perception, and Control »
Jiajun Wu · Kelsey Allen · Kevin Smith · Jessica Hamrick · Emmanuel Dupoux · Marc Toussaint · Josh Tenenbaum -
2018 Poster: Toddler-Inspired Visual Object Learning »
Sven Bambach · David Crandall · Linda Smith · Chen Yu -
2017 : Panel Discussion »
Felix Hill · Olivier Pietquin · Jack Gallant · Raymond Mooney · Sanja Fidler · Chen Yu · Devi Parikh -
2017 : How infant learn to speak by interacting with the visual world? »
Chen Yu -
2016 : Datasets, Methodology, and Challenges in Intuitive Physics »
Emmanuel Dupoux · Josh Tenenbaum -
2016 : Naive Physics 101: A Tutorial »
Emmanuel Dupoux · Josh Tenenbaum -
2016 Workshop: Intuitive Physics »
Adam Lerer · Jiajun Wu · Josh Tenenbaum · Emmanuel Dupoux · Rob Fergus