Skip to content
/ sea Public

Speech Emotion Analyzer (SEA) - project to determinate human speech emotions from audio using deep learning

License

Notifications You must be signed in to change notification settings

VMois/sea

Repository files navigation

Speech Emotion Analysis

A neural network model for determining human speech emotions from audio recordings.

Project structure

Datasets

id name
1 The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS)
2 SAVEE
3 CREMA-D (Crowd-sourced Emotional Mutimodal Actors Dataset)

Emotions

We are using names for emotions:

  • happy
  • surprise
  • sad
  • angry
  • disgust
  • fear
  • neutral

How to start

Prerequisites

Prepare environment

  1. Install Conda through Miniconda/Anaconda
  2. Clone the repo
  3. Create Conda environment using conda env create -f env.yaml
  4. Activate the enviroment with source activate sea
  5. Install Tensorflow for CPU using conda install tensorflow=1.12.0 (or tensorflow-gpu for GPU support)
  6. Install keras using conda install keras=2.2.4

Prepare data

RAVDESS dataset

  1. Download Audio_Speech_Actors_01-24.zip and Audio_Song_Actors_01-24.zip from https://zenodo.org/record/1188976
  2. Place these zip files in a folder called raw-data in the main directory

Related articles

Credits

About

Speech Emotion Analyzer (SEA) - project to determinate human speech emotions from audio using deep learning

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published