This is the public, feature-limited version of the conference webpage. After Registration and login please visit the full version.

Workshop: Self-Supervised Learning for Speech and Audio Processing

Abdelrahman Mohamed, Hung-yi Lee, Shinji Watanabe, Shang-Wen Li, Tara Sainath, Karen Livescu

Fri, Dec 11th @ 14:50 GMT – Sat, Dec 12th @ 00:25 GMT
Abstract: There is a trend in the machine learning community to adopt self-supervised approaches to pre-train deep networks. Self-supervised learning utilizes proxy supervised learning tasks, for example, distinguishing parts of the input signal from distractors, or generating masked input segments conditioned on the unmasked ones, to obtain training data from unlabeled corpora. These approaches make it possible to use a tremendous amount of unlabeled data on the web to train large networks and solve complicated tasks. ELMo, BERT, and GPT in NLP are famous examples in this direction. Recently self-supervised approaches for speech and audio processing are also gaining attention. These approaches combine methods for utilizing no or partial labels, unpaired text and audio data, contextual text and video supervision, and signals from user interactions. Although the research direction of self-supervised learning is active in speech and audio processing, current works are limited to several problems such as automatic speech recognition, speaker identification, and speech translation, partially due to the diversity of modeling in various speech and audio processing problems. There is still much unexplored territory in the research direction for self-supervised learning.

This workshop will bring concentrated discussions on self-supervision for the field of speech and audio processing via several invited talks, oral and poster sessions with high-quality papers, and a panel of leading researchers from academia and industry. Alongside research work on new self-supervised methods, data, applications, and results, this workshop will call for novel work on understanding, analyzing, and comparing different self-supervision approaches for speech and audio processing. The workshop aims to:
- Review existing and inspire new self-supervised methods and results,
- Motivate the application of self-supervision approaches to more speech and audio processing problems in academia and industry, and encourage discussion amongst experts and practitioners from the two realms,
- Encourage works on studying methods for understanding learned representations, comparing different self-supervision methods and comparing self-supervision to other self-training as well as transfer learning methods that low-resource speech and audio processing have long utilized,
- Facilitate communication within the field of speech and audio processing (e.g., people who attend conferences such as INTERSPEECH and ICASSP) as well as between the field and the whole machine learning community for sharing knowledge, ideas, and data, and encourage future collaboration to inspire innovation in the field and the whole community.

Chat

To ask questions please use rocketchat, available only upon registration and login.

Schedule

14:50 – 15:00 GMT
Opening remarks
Hung-yi Lee
15:00 – 15:35 GMT
Invited talk - 1
Bhuvana Ramabhadran
15:35 – 15:45 GMT
Q&A for invited talk - 1
15:45 – 16:20 GMT
Invited talk - Multimodal Distant Supervision
Mark Hasegawa-Johnson
16:20 – 16:30 GMT
Q&A for invited talk - Multimodal Distant Supervision
16:30 – 16:40 GMT
Self-Supervised Learning using Contrastive Mixtures for Personalized Speech Enhancement
Aswin Sivaraman
16:40 – 16:50 GMT
Self-supervised Pre-training Reduces Label Permutation Instability of Speech Separation
Sung-Feng Huang
16:50 – 17:00 GMT
Augmentation adversarial training for self-supervised speaker recognition
jaesung Huh
17:00 – 17:10 GMT
Neural Composition: Learning to Generate from Multiple Models
Denis Filimonov
17:10 – 17:20 GMT
Towards Semi-Supervised Semantics Understanding from Speech
Cheng-I Lai
17:20 – 17:30 GMT
The Zero Resource Speech Benchmark 2021. Metrics and baselines for unsupervised spoken language modeling
Tu Anh Nguyen
17:30 – 17:45 GMT
Q&A for contributed talks between 11:30 and 12:30
17:45 – 18:00 GMT
Break
18:00 – 18:35 GMT
Invited talk - Speech Processing with Weak Supervision
Dong Yu
18:35 – 18:45 GMT
Q&A for invited talk - Speech Processing with Weak Supervision
18:45 – 18:55 GMT
Towards Localisation of Keywords in Speech Using Weak Supervision
Kayode Olaleye
18:55 – 19:05 GMT
Text-Free Image-to-Speech Synthesis Using Learned Segmental Units
Wei-Ning Hsu
19:05 – 19:15 GMT
Self-Supervised Audio-Visual Separation of On-Screen Sounds from Unlabeled Videos
Efthymios Tzinis
19:15 – 19:25 GMT
Multi-Format Contrastive Learning of Audio Representations
Aaron van den Oord
19:25 – 19:40 GMT
Q&A for contributed talks between 1:45 and 2:25
19:40 – 19:55 GMT
Break
19:55 – 20:30 GMT
Invited talk - Underfitting and Uncertainty in Self-Supervised Predictive Models
Chelsea Finn
20:30 – 20:40 GMT
Q&A for invited talk - Underfitting and Uncertainty in Self-Supervised Predictive Models
20:40 – 21:15 GMT
Invited talk - Towards robust self-supervised learning of speech representations
Mirco Ravanelli
21:15 – 21:25 GMT
Q&A for invited talk - Towards robust self-supervised learning of speech representations
21:25 – 21:35 GMT
Similarity Analysis of Self-Supervised Speech Representations
Yu-An Chung
21:35 – 21:45 GMT
Representation Learning for Sequence Data with Deep Autoencoding Predictive
Junwen Bai
21:45 – 21:55 GMT
Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition
Yu Zhang
21:55 – 22:05 GMT
A Correspondence Variational Autoencoder for Unsupervised Acoustic Word Embedding
Puyuan Peng
22:05 – 22:15 GMT
HUBERT: How much can a bad teacher benefit ASR pre-training?
Wei-Ning Hsu
22:15 – 22:30 GMT
Q&A for contributed talks between 4:25 and 5:15
22:30 – 22:45 GMT
Break
22:45 – 23:20 GMT
Invited talk - Flexible contextualized speech representation learning for diverse downstream tasks
Katrin Kirchhhoff
23:20 – 23:30 GMT
Q&A for invited talk - Flexible contextualized speech representation learning for diverse downstream tasks
Fri, Dec 11th @ 23:30 GMT – Sat, Dec 12th @ 00:05 GMT
Invited talk - De-noising Sequence-to-Sequence Pre-training
Luke Zettlemoyer
00:05 – 00:15 GMT
Q&A for invited talk - De-noising Sequence-to-Sequence Pre-training
00:15 – 00:25 GMT
Closing remark
Abdelrahman Mohamed