-
-
Notifications
You must be signed in to change notification settings - Fork 347
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Emotion Recognition from Audio using Deep Learning #830
Comments
Thank you for creating this issue! We'll look into it as soon as possible. Your contributions are highly appreciated! 😊 |
Assigned @ChethanaPotukanam |
can i work on this ? |
Please share your approach. |
@abhisheks008 could you please assign me this issue? my approach is as follows:- using the ravdess dataset for emotional speech audio
this will be followed by evaluating the model using metrics and visualizing heatmaps of confusion matrices to analyse the error distribution name : Moksh patel |
Hi @T3CH-Pyth0n sorry for replying late. Assigning this issue to you. |
@abhisheks008 ill be altering the approach a bit, but I'll still implement 3-4 models. does that work? |
Deep Learning Simplified Repository (Proposing new issue)
🔴 Project Title :
Emotion Recognition from Audio using Deep Learning
🔴 Aim :
To build a deep learning model that can analyze audio recordings and classify the emotions expressed. This can have applications in areas such as customer service, mental health monitoring, and entertainment.
🔴 Dataset :
Various publicly available datasets for emotion recognition in audio, such as RAVDESS, TESS, CREMA-D, etc.
🔴 Approach : Try to use 3-4 algorithms to implement the models and compare all the algorithms to find out the best fitted algorithm for the model by checking the accuracy scores. Also do not forget to do a exploratory data analysis before creating any model.
📍 Follow the Guidelines to Contribute in the Project :
requirements.txt
- This file will contain the required packages/libraries to run the project in other machines.Model
folder, theREADME.md
file must be filled up properly, with proper visualizations and conclusions.🔴🟡 Points to Note :
✅ To be Mentioned while taking the issue :
Load the Dataset
Exploratory Data Analysis (EDA): Visualise common patterns and features in audio signals.
Feature Extraction: Extract features such as MFCC, Chroma, Mel Spectrogram, etc.
Model Implementation: Convolutional Neural Network (CNN) , Recurrent Neural Network (RNN) , Long Short-Term ,
Memory (LSTM) , Bidirectional LSTM (BiLSTM)
Train and Evaluate Each Model
Compare Performance using accuracy and loss metrics.
Happy Contributing 🚀
All the best. Enjoy your open source journey ahead. 😎
The text was updated successfully, but these errors were encountered: