Login / Signup
Using Audio Events to Extend a Multi-modal Public Speaking Database with Reinterpreted Emotional Annotations.
Esther Rituerto-González
Clara Luis-Mingueza
Carmen Peález-Moreno
Published in:
IberSPEECH (2021)
Keyphrases
</>
multi modal
audio visual
database
cross modal
image annotation
emotion recognition
single modality
metadata
multi modality
multimedia
event detection
high dimensional
uni modal
semantic concepts
automatic image annotation
audio features
humanoid robot
video search
keywords
broadcast news
image processing