Skip to content

Detecting emotions from speech signals using a CNN model. Signal processing and feature extraction were done using Librosa.

Notifications You must be signed in to change notification settings

MeshalAlamr/speech-emotion-recognition

Repository files navigation

speech-emotion-recognition

SDAIA Bootcamp project 5 - Deep Learning.

This project aims to predict emotions from speech signals. The goal of this project is to help school consultancies analyze emotions of children to help them intervene in case a child shows early signs of mental illness.

Table of Contents

MVP

The project MVP can be found here.

Dataset

The dataset of this project is a merged dataset from four different datasets which are CREMA, SAVEE, TESS and RAVDESS. The dataset containing all four datasets can be found here on Kaggle. After loading each dataset and combining the datasets we end up with 12,162 rows.

Analysis and Results

The project's notebook is split into two parts.

Part 1 includes loading the data, exploring it, doing some data augmentation and finally extracting the features. It can be found here. If you'd like to replicate the results from part 1 then the Pickle file for the final dataframe can be downloaded here on Google Drive.

Part 2 is all about modelling using the extracted features from part 1. It can be found here.

Distribution of emotions of the combined dataset:

chrome_7u2LRkIY5G

Results after training the model:

We obtained an accuraccy of 72.14%.

chrome_EPSiPHwzt6 chrome_mjqoFtiz5K

Presentation

The presentation can be found here.

Please note that the presentation was based off an earlier version of the project so the results differ from what is obtained in the project's notebooks.

Authors

About

Detecting emotions from speech signals using a CNN model. Signal processing and feature extraction were done using Librosa.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published