The motivation behind the project lies in the vast potential applications of emotion recognition systems, which can significantly improve human-computer interactions and user experiences across a range of domains. We can ensure that our deep learning models are trained to recognize emotions accurately across different populations and contexts. This is essential for creating a robust and reliable emotion recognition system.
📝 Table of Contents:
1) About the Project
2) Getting Started
3) Built With
4) Methodology
5) Data Collection and Analysis
6) Model Training and Evaluation
7) Observations
8) Conclusion
🌐 1) About the Project
The project aims to develop an advanced emotion detection system using deep learning techniques to enhance applications such as voice-based IoT applications. The system aims to accurately identify and classify emotions conveyed through human speech, and it involves collecting and analysing emotional speech data, extracting relevant features, and applying deep learning methods to build a robust and accurate emotion detection model. The system intends to provide improved performance and capabilities, transforming the way voice-based intelligent systems interact with users and understand their emotions. This project was made in Google collab environment.
🏁 2) Getting Started
git clone https:/github.com/aninda20/emotion-recognition
-Follow the instructions provided in the notebook for setup and execution. It is recommended to run in Google Colab.
⛏️ 3) Built With
1)Python
2) Google Colab
3) Machine Learning Libraries (e.g., Scikit-Learn, TensorFlow)
4) Data Analysis and Visualization Tools (e.g., Pandas, Matplotlib)
📝 4) Methodology
a) Data Collection
b) Feature Engineering
c) Model Development
d) Model Evaluation
5) Data Collection and Analysis
a) RAVDESS Dataset
b) Preprocessing and feature extraction
c) Deep learning techniques: CNNs and RNNs
6) Model Training and Evaluation
- 8 different models tested
- CNN + LSTM model selected
- Confusion matrices, precision, recall, and F1-score
7) Observations
a) Count of Emotions
b) Mel Sectrogram-Female-fear
c) Confusion Matrix
d) Accuracy vs. Epochs
e) Precision, recall, f1-score, support
f) Librosa Waveplot for Aninda’s voice
g) Predicted Probabilities of Uploaded Voice
8) Conclusion
a) Successful development of a speech emotion detection system
b) Potential applications in various fields
c) Future work: advanced machine learning techniques, expanded dataset