Skip to content

A multimodal SER project combining BERT and ECAPA-TDNN with cross-attention-based fusion on the IEMOCAP dataset.

License

Notifications You must be signed in to change notification settings

nhut-ngnn/Multimodal-Speech-Emotion-Recognition

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Multi-modal with Late Fusion for Speech Emotion Recognition

Please press ⭐ button and/or cite papers if you feel helpful.

Table of Contents

Usage

Dataset

In this study, we use voice dataset from IEMOCAP.

Download in here

Clone this repository

git clone "https://github.com/nhut-ngnn/Multimodal-Speech-Emotion-Recognition.git"

Contact

For any information, please contact the main author:

Nhut Minh Nguyen at FPT University, Vietnam

Email: minhnhut.ngnn@gmail.com

GitHub: https://github.com/nhut-ngnn

About

A multimodal SER project combining BERT and ECAPA-TDNN with cross-attention-based fusion on the IEMOCAP dataset.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages