Skip to content

Predicting various emotion in human speech signal by detecting different speech components affected by human emotion.

Notifications You must be signed in to change notification settings

Ztrimus/speech-emotion-recognition

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Speech Emotion Recognition

Project source can be downloaded from https://github.com/SBZed/speech-emotion-recognition.git

Author & Contributor List

Saurabh Zinjad

All other known bugs and fixes can be sent to "zinjadsaurabh1997@gmail.com" with subject "speech emotion recognition Suggestion".

Reported bugs/fixes will be submitted to correction.

Basic Idea :

Our machine learning model tries to detect and predict various emotion in speech signal or human audio by detecting different features and component of speech affected by human emotion. Emotion detection from the speech is a relatively new field of research. Here, we are trying a new approach or ways to contribute towards emotion recognition research.

Presentation :

To get familiar and comfortable with project concept and outcomes please refer to this presentation.

Video Explanation [YouTube] :

If you want an explanation for the project watch this video.

Want to see result quickly without any installation :

  1. Go to this google drive link.
  2. Open speech_emotion_recognition.ipynb notebook. it's main code for project.
  3. For live demonstration or detecting your Speech emotion, Open live_demo_speech_emotion_recognition.ipynb notebook. Just keep two things in mind. First, You need to record your audio in very silent room without any noise.And secondly, Model is not perfect, Show the great result for Disgust, fear, Anger, Sad. But shows poor result for surprise, Happy.

How to run file :

To implement this project on your Machine

  1. Clone this repository.
  2. Go to terminal and type pip install -r requirements.txt.
  3. Then open ./Code/main_speech_emotion_recognition.ipynb and run it.

Bugs :

Limitation and Further growth :

  1. Need to record audio in clear and silent environment.
  2. Dataset is not accurate and prominant. Even, human ear can't detect accurate emotion for some audios. Need Promising dataset.