Skip to content

Hossein-Molaeian/Speech-Emotion-Recognition

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 

Repository files navigation

Speech Emotion Recognition

This project focuses on enabling machines to understand human emotions from speech.

Key Features:

  • Identifies seven types of emotions: anger, disgust, fear, happiness, pleasant surprise, sadness, and neutral.
  • Long Short-Term Memory (LSTM) layers for handling sequential data.
  • MFCCs for capturing the phonetic properties of speech.
  • Validation Accuracy: 67%

Technologies Used

  • Python
  • Keras: building and training the neural network model
  • Librosa: audio processing and feature extraction
  • Pandas: data manipulation
  • Matplotlib and Seaborn: data visualization

Dataset Information

Summary of the Model

  • 1 LSTM layer with 256 units
  • 2 Dense layers with 128 and 64 units using ReLU activation
  • Dropout layers with a rate of 20% after each Dense layer
  • Output layer with 7 units using softmax activation
  • Optimizer: Adam
  • Loss Function: Categorical Cross-Entropy
  • Training Duration: 50 epochs with a batch size of 64

Thanks for visiting this project!

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published