Uncertainty-Aware Multimodal Emotion Recognition through Dirichlet Parameterization
This lightweight AI could power the next generation of empathetic devices...
Researchers unveiled a new lightweight, privacy-preserving AI framework for Multimodal Emotion Recognition (MER) designed to run on edge devices. It processes speech, text, and facial imagery using efficient backbones like Emotion2Vec and DistilRoBERTa. A novel fusion mechanism based on Dempster-Shafer theory handles uncertainty across modalities without extra training. Validated on five major datasets, it achieves competitive accuracy while being robust to ambiguous or missing inputs, emphasizing real-world feasibility.
Why It Matters
It enables more natural, trustworthy human-computer interaction for healthcare, customer service, and smart devices.