-
Interspeech 23
Resource Efficient and Cross-Modal Learning Toward Foundation Modeling Tutorial- Video -
ICASSP 22 Tutorial
Neural Model Reprogramming and Prompting for Speech Modeling- Video | Slide -
ICASSP 23 Tutorial
Parameter-Efficient Learning (PEL) for Speech and NLP: Adapters, Prompts, and Reprogramming- Slide
9:00
- Background of Frozen Model Adaptation
- Neural Adapter, Reprogramming, Prompting, and Low-Rank Adaptation (LoRA)
| Title | Authors | Code | Year |
|---|---|---|---|
| Differentially Private Adapters for Parameter Efficient Acoustic Modeling | C.-W. Ho et al. | code | Interspeech 2023 |
| Parameter-Efficient Learning for Text-to-Speech Accent Adaptation | L.-J. Yang et al. | code | Interspeech 2023 |
| A Parameter-Efficient Learning Approach to Arabic Dialect Identification with Pre-Trained General-Purpose Speech Model | S. Radhakrishnan et al. | code | Interspeech 2023 |
- Reduce to GPU / TPU Memory During the Training (e.g., the Memory of Activation)
- Model Serialization
- Efficient On-Device Learning via Feature Reprogramming (CVPR 2022)
- Ladder-Side Tuning (NeurIPS 2022)
- Universal Approximation Theory (IEEE TIP 1993)
- LogME: Practical Assessment of Pre-trained Models for Transfer Learning (ICML 2021)
- Latent Space Alignment in "Reprogramming Acoustic Models for Time Series Classification" (ICML 2021)
| Title | Authors | Code | Year |
|---|---|---|---|
| How to Estimate Model Transferability of Pre-Trained Speech Models? | Z.-C. Chen et al. | code | Interspeech 2023 |
- Cross-Modal Merging
- Low-Rank Adaptation (LoRA) for Foundation Modeling and Pre-Training
- Special Session in ICASSP 2024: In-Context Learning for Speech and Language Processing
- [email protected]
- How to Train Your Whisper with Neural Adapter and LoRA
11:00 to 11:45
Part 3: Multimodal Pre-Training for Automatic Speech Recognition and Vision Sharing, Dr. Shalini Ghosh
11:45 to 12:20
12:20 to 12:30
12:40 to 12:45