Skip to content

eesungkim/Speech_Emotion_Recognition_AAE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Speech Emotion Recognition using Adversarial auto-encoders

For low-level acoustic features, Authors extract a set of 1582 features using the openSMILE toolkit. The set consists of an assembly of spectral prosody and energy based features. Authors use five folder cross validation scheme, but this implementation is used one leave speaker cross validation scheme for speaker-independent manner.

Datasets

  • Interactive Emotional Dyadic Motion Capture (IEMOCAP) database is required to run this code.

Dependencies

  • openSMILE for low-level acoustic features extraction
  • Tensorflow for Adversarial Auto-encoders
  • scikit-learn for classification and performance evaluation

References

About

Adversarial Auto-encoders for Speech Based Emotion Recogntion

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages