Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Convert VAD to Ekman #31

Open
mirix opened this issue Jul 31, 2023 · 3 comments
Open

Convert VAD to Ekman #31

mirix opened this issue Jul 31, 2023 · 3 comments

Comments

@mirix
Copy link

mirix commented Jul 31, 2023

Hello,

This model provides VAD values in 3D space.

However, the Ekman model is more intuitive to share the results with users.

I have found papers with 3D representations hinting at how to perform this conversion.

Are you aware of a straightforward approach to perform the conversion between both models?

Ideally in Python, but any hint on the algorithm would also do.

Best,

Ed

@hagenw
Copy link
Member

hagenw commented Aug 2, 2023

The VAD model is only fine-tuned on the MSP-Podcast dataset, which has several shortcomings for a full blown VAD model:

  • Podcast recordings most likely do not contain all possible emotions, e.g. fear
  • The dominance and arousal annotations show a high correlation, that is mimicked by the model, which means we most likely do not cover the 3D space of VAD in a meaningful way

Having this in mind I would propose to be very carefully when trying to map the VAD values to emotional categories.

Another way might be to further fine-tune the model on a given database containing the desired emotional categories, or using the embeddings of the model to train a simple classifier on such a database like we do in the notebook under the "https://github.com/audeering/w2v2-how-to/blob/main/notebook.ipynb" section.

@mirix
Copy link
Author

mirix commented Aug 3, 2023

@hagenw

Thanks a million for the clarifications.

In general, the conversion from VAD to Ekman seems to provide useful results:

https://github.com/mirix/approaches-to-diarisation/tree/main/emotions

However, it is true that fear is never detected.

I will see what other models are available and pay more attention to which datasets were used.

@mirix
Copy link
Author

mirix commented Aug 10, 2023

Hi @hagenw

I have forked MOSEI for SER:

https://huggingface.co/datasets/mirix/messaih

https://github.com/mirix/messaih

Now I will try to train a model and test it in a real-life scenario.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants