This repository contains code and analysis for the paper Advancing Test-Time Adaptation in Wild Acoustic Test Settings.
Acoustic foundation models, fine-tuned for Automatic Speech Recognition (ASR), suffer from performance degradation in wild acoustic test settings when deployed in real-world scenarios. Stabilizing online Test-Time Adaptation (TTA) under these conditions remains an open and unexplored question. Existing wild vision TTA methods often fail to handle speech data effectively due to the unique characteristics of high-entropy speech frames, which are unreliably filtered out even when containing crucial semantic content. Furthermore, unlike static vision data, speech signals follow short-term consistency, requiring specialized adaptation strategies. In this work, we propose a novel wild acoustic TTA method tailored for ASR fine-tuned acoustic foundation models. Our method, Confidence-Enhanced Adaptation, performs frame-level adaptation using a confidence-aware weight scheme to avoid filtering out essential information in high-entropy frames. Additionally, we apply consistency regularization during test-time optimization to leverage the inherent short-term consistency of speech signals. Our experiments on both synthetic and real-world datasets demonstrate that our approach outperforms existing baselines under various wild acoustic test settings, including Gaussian noise, environmental sounds, accent variations, and sung speech.
cd CEA
pip install -r requirements.txtTo obtained the synthesized dataset LS-P, run the following scripts
python noisyspeech_synthesizer.pyOur main code include ./main_cea.py, ./utils.py.
To run experiments on different datasets, use scripts from scripts/
bash ./scripts/ls-c.shbash ./scripts/ls-p.shbash ./scripts/l2.shbash ./scripts/dsing.shSee more details about the parameters in each script.