UniMEEC: Towards Unified Multimodal Emotion Recognition and Emotion Cause
CoRR(2024)
Abstract
Multimodal emotion recognition in conversation (MERC) and multimodal
emotion-cause pair extraction (MECPE) has recently garnered significant
attention. Emotions are the expression of affect or feelings; responses to
specific events, thoughts, or situations are known as emotion causes. Both are
like two sides of a coin, collectively describing human behaviors and intents.
However, most existing works treat MERC and MECPE as separate tasks, which may
result in potential challenges in integrating emotion and cause in real-world
applications. In this paper, we propose a Unified Multimodal Emotion
recognition and Emotion-Cause analysis framework (UniMEEC) to explore the
causality and complementarity between emotion and emotion cause. Concretely,
UniMEEC reformulates the MERC and MECPE tasks as two mask prediction problems,
enhancing the interaction between emotion and cause. Meanwhile, UniMEEC shares
the prompt learning among modalities for probing modality-specific knowledge
from the Pre-trained model. Furthermore, we propose a task-specific
hierarchical context aggregation to control the information flow to the task.
Experiment results on four public benchmark datasets verify the model
performance on MERC and MECPE tasks and achieve consistent improvements
compared with state-of-the-art methods.
MoreTranslated text
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined