TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Tracing Intricate Cues in Dialogue: Joint Graph Structure ...

Tracing Intricate Cues in Dialogue: Joint Graph Structure and Sentiment Dynamics for Multimodal Emotion Recognition

Jiang Li, XiaoPing Wang, Zhigang Zeng

2024-07-31Emotion Recognition in ConversationSentiment AnalysisMultimodal Emotion RecognitionMultimodal Sentiment AnalysisEmotion Recognition
PaperPDFCode(official)

Abstract

Multimodal emotion recognition in conversation (MERC) has garnered substantial research attention recently. Existing MERC methods face several challenges: (1) they fail to fully harness direct inter-modal cues, possibly leading to less-than-thorough cross-modal modeling; (2) they concurrently extract information from the same and different modalities at each network layer, potentially triggering conflicts from the fusion of multi-source data; (3) they lack the agility required to detect dynamic sentimental changes, perhaps resulting in inaccurate classification of utterances with abrupt sentiment shifts. To address these issues, a novel approach named GraphSmile is proposed for tracking intricate emotional cues in multimodal dialogues. GraphSmile comprises two key components, i.e., GSF and SDP modules. GSF ingeniously leverages graph structures to alternately assimilate inter-modal and intra-modal emotional dependencies layer by layer, adequately capturing cross-modal cues while effectively circumventing fusion conflicts. SDP is an auxiliary task to explicitly delineate the sentiment dynamics between utterances, promoting the model's ability to distinguish sentimental discrepancies. Furthermore, GraphSmile is effortlessly applied to multimodal sentiment analysis in conversation (MSAC), forging a unified multimodal affective model capable of executing MERC and MSAC tasks. Empirical results on multiple benchmarks demonstrate that GraphSmile can handle complex emotional and sentimental patterns, significantly outperforming baseline models.

Results

TaskDatasetMetricValueModel
Emotion RecognitionMELD-SentimentAccuracy74.44GraphSmile
Emotion RecognitionMELD-SentimentWeighted F174.31GraphSmile
Emotion RecognitionCMU-MOSEI-SentimentAccuracy46.82GraphSmile
Emotion RecognitionCMU-MOSEI-SentimentWeighted F144.93GraphSmile
Emotion RecognitionIEMOCAP-4Accuracy86.53GraphSmile
Emotion RecognitionIEMOCAP-4Weighted F186.52GraphSmile
Emotion RecognitionMELDAccuracy67.7GraphSmile
Emotion RecognitionMELDWeighted-F166.71GraphSmile
Emotion RecognitionCMU-MOSEI-Sentiment-3Accuracy67.73GraphSmile
Emotion RecognitionCMU-MOSEI-Sentiment-3Weighted F166.73GraphSmile
Emotion RecognitionIEMOCAPAccuracy72.77GraphSmile
Emotion RecognitionIEMOCAPWeighted-F172.81GraphSmile
Emotion RecognitionIEMOCAP-4Accuracy86.53GraphSmile
Emotion RecognitionIEMOCAP-4Weighted F186.52GraphSmile
Emotion RecognitionCMU-MOSEI-Sentiment-3Accuracy67.73GraphSmile
Emotion RecognitionCMU-MOSEI-Sentiment-3Weighted F166.73GraphSmile
Emotion RecognitionCMU-MOSEI-SentimentAccuracy46.82GraphSmile
Emotion RecognitionCMU-MOSEI-SentimentWeighted F144.93GraphSmile
Emotion RecognitionMELDAccuracy67.7GraphSmile
Emotion RecognitionMELDWeighted F166.71GraphSmile
Emotion RecognitionIEMOCAPAccuracy72.77GraphSmile
Emotion RecognitionIEMOCAPWeighted F172.81GraphSmile
Emotion RecognitionMELD-SentimentAccuracy74.44GraphSmile
Emotion RecognitionMELD-SentimentWeighted F174.31GraphSmile
Multimodal Emotion RecognitionIEMOCAP-4Accuracy86.53GraphSmile
Multimodal Emotion RecognitionIEMOCAP-4Weighted F186.52GraphSmile
Multimodal Emotion RecognitionCMU-MOSEI-Sentiment-3Accuracy67.73GraphSmile
Multimodal Emotion RecognitionCMU-MOSEI-Sentiment-3Weighted F166.73GraphSmile
Multimodal Emotion RecognitionCMU-MOSEI-SentimentAccuracy46.82GraphSmile
Multimodal Emotion RecognitionCMU-MOSEI-SentimentWeighted F144.93GraphSmile
Multimodal Emotion RecognitionMELDAccuracy67.7GraphSmile
Multimodal Emotion RecognitionMELDWeighted F166.71GraphSmile
Multimodal Emotion RecognitionIEMOCAPAccuracy72.77GraphSmile
Multimodal Emotion RecognitionIEMOCAPWeighted F172.81GraphSmile
Multimodal Emotion RecognitionMELD-SentimentAccuracy74.44GraphSmile
Multimodal Emotion RecognitionMELD-SentimentWeighted F174.31GraphSmile

Related Papers

Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation2025-07-21AdaptiSent: Context-Aware Adaptive Attention for Multimodal Aspect-Based Sentiment Analysis2025-07-17Camera-based implicit mind reading by capturing higher-order semantic dynamics of human gaze within environmental context2025-07-17AI Wizards at CheckThat! 2025: Enhancing Transformer-Based Embeddings with Sentiment for Subjectivity Detection in News Articles2025-07-15DCR: Quantifying Data Contamination in LLMs Evaluation2025-07-15A Robust Incomplete Multimodal Low-Rank Adaptation Approach for Emotion Recognition2025-07-15SentiDrop: A Multi Modal Machine Learning model for Predicting Dropout in Distance Learning2025-07-14Dynamic Parameter Memory: Temporary LoRA-Enhanced LLM for Long-Sequence Emotion Recognition in Conversation2025-07-11