Semester project for the Multimodal Machine Learning class of the MSc in Artificial Intelligence by NCSR Demokritos and University of Piraeus
The project evaluates the performance of traditional Machine Learning techniques and Deep Learning methods over predicting the correct emotion of the given human speech audio example.
The SAVEE and the TESS datasets were selected for the emotion recognition task.
pyAudioAnalysis and Deep Audio Features libraries were used for the experiments that took place.
git clone https://github.com/JoKoum/speech-emotion-detection.git
Google Colab environment was used for the project development.