Transformer-Based Transfer Learning And Multi-Task Learning For Improving The Performance Of Speech Emotion Recognition

JOURNAL OF THE ACOUSTICAL SOCIETY OF KOREA(2021)

引用 0|浏览2
暂无评分
摘要
It is hard to prepare sufficient training data for speech emotion recognition due to the difficulty of emotion labeling. In this paper, we apply transfer learning with large-scale training data for speech recognition on a transformer-based model to improve the performance of speech emotion recognition. In addition, we propose a method to utilize context information without decoding by multi-task learning with speech recognition. According to the speech emotion recognition experiments using the IEMOCAP dataset, our model achieves a weighted accuracy of 70.6 % and an unweighted accuracy of 71.6 %, which shows that the proposed method is effective in improving the performance of speech emotion recognition.
更多
查看译文
关键词
Speech emotion recognition, Transformer, Transfer learning, Multi-task learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要