Skip to yearly menu bar Skip to main content

Workshop: UniReps: Unifying Representations in Neural Models

Deep Multimodal Emotion Recognition using Modality Aware Attention Network for Unifying Representations in Neural Models

Sungpil Woo · MUHAMMAD ZUBAIR · Sunhwan Lim · Daeyoung Kim

[ ] [ Project Page ]
presentation: UniReps: Unifying Representations in Neural Models
Fri 15 Dec 6:15 a.m. PST — 3:15 p.m. PST


This paper introduces a multi-modal emotion recognition system aimed at enhancing emotion recognition by integrating representations from physiological signals. To accomplish this goal, we introduce a modality aware attention network to extract emotion-specific features by influencing and aligning the representation spaces of various modalities into a unified entity. Through a series of experiments and visualizations conducted on the AMIGO dataset, we demonstrate the efficacy of our proposed methodology for emotion classification, highlighting its capability to provide comprehensive representations of physiological signals.

Chat is not available.