Skip to content

Latest commit

 

History

History
37 lines (26 loc) · 1.25 KB

README.md

File metadata and controls

37 lines (26 loc) · 1.25 KB

Speech Emotion Recognition

This project focuses on enabling machines to understand human emotions from speech.

Key Features:

  • Identifies seven types of emotions: anger, disgust, fear, happiness, pleasant surprise, sadness, and neutral.
  • Long Short-Term Memory (LSTM) layers for handling sequential data.
  • MFCCs for capturing the phonetic properties of speech.
  • Validation Accuracy: 67%

Technologies Used

  • Python
  • Keras: building and training the neural network model
  • Librosa: audio processing and feature extraction
  • Pandas: data manipulation
  • Matplotlib and Seaborn: data visualization

Dataset Information

Summary of the Model

  • 1 LSTM layer with 256 units
  • 2 Dense layers with 128 and 64 units using ReLU activation
  • Dropout layers with a rate of 20% after each Dense layer
  • Output layer with 7 units using softmax activation
  • Optimizer: Adam
  • Loss Function: Categorical Cross-Entropy
  • Training Duration: 50 epochs with a batch size of 64

Thanks for visiting this project!