Monte Carlo Self-Training for Speech Recognition

ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2024)

引用 0|浏览1
暂无评分
摘要
Self-training in the teacher-student framework generally suffers from the confirmation bias problem, where errors from the teacher are propagated to the student and hence get amplified with multiple iterations. In this paper, we present Monte Carlo Self-training where pseudo labels are generated by sampling from a teacher distribution, as a way to mitigate this problem. We show that Monte Carlo Self-training is an approximation to minimizing label sequence level cross entropy between student and teacher. In our experiments we find that Monte Carlo Self-training always outperforms beam decoder based self-training and is quite robust even when the initial teacher WER is high. We also show that label sampling allows formulating pseudo label confidence in a more natural way and we show that these confidence measures give further improvements in our unsupervised adaptation experiments especially when the initial teacher WER is very high.
更多
查看译文
关键词
self-training,semi-supervised learning,speech recognition,unsupervised adaptation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要