This paper introduces a system development approach for the SemEval-2024 Task 3 focusing on Multimodal Emotion Cause Analysis. The proposed two-step framework involves employing Llama and GPT models to predict emotions and causes in conversations. The study highlights the importance of integrating multiple modalities like text, audio, and video to enhance emotion cause analysis efficiency. By leveraging instruction-tuning with Llama models and in-context learning with GPT models, the authors achieved significant performance gains, securing rank 4 on the leaderboard. The dataset used contains over 13,000 multimodal utterances from the TV show Friends, annotated with emotion-cause pairs. Through detailed experimentation and analysis, the authors demonstrate the effectiveness of their approaches in addressing the complexities of emotion cause analysis in natural conversation settings.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Arefa,Mohamm... at arxiv.org 03-11-2024
https://arxiv.org/pdf/2403.04798.pdfDeeper Inquiries