M3ER: Multiplicative Multimodal Emotion Recognition using Facial, Textual, and Speech Cues.
Trisha MittalUttaran BhattacharyaRohan ChandraAniket BeraDinesh ManochaPublished in: AAAI (2020)
Keyphrases
- emotion recognition
- audio visual
- multimodal fusion
- multimodal interfaces
- emotional speech
- multi modal
- facial expressions
- affect sensing
- multimedia
- affect detection
- human computer interaction
- multi stream
- multimodal interaction
- visual information
- speaker verification
- facial images
- emotional state
- visual data
- affective states
- face recognition
- visual content
- emotion classification
- metadata
- keywords
- physiological signals
- facial expression recognition
- entity relationship
- audio features
- human faces
- face images
- multiple modalities
- natural language
- low level
- prosodic features
- facial motion
- speech recognition
- facial animation
- landmark detection
- speech synthesis
- relevance feedback
- facial features
- text to speech synthesis