EmotionAware-AI is an open-source project designed to detect and interpret human emotions through facial expressions and speech analysis. By leveraging advanced machine learning techniques, this system aims to enhance human-computer interactions by providing emotionally intelligent responses.
- Facial Emotion Recognition: Analyzes facial expressions to identify emotions such as happiness, sadness, anger, surprise, and more.
- Speech Emotion Recognition: Processes vocal inputs to detect emotional states conveyed through tone and pitch.
- Real-time Processing: Capable of analyzing and responding to emotional cues in real-time.
- Modular Architecture: Organized into distinct components for facial analysis (
face), speech analysis (speech), backend processing (backend), and frontend interface (frontend).
To set up the EmotionAware-AI project locally, follow these steps:
-
Clone the Repository:
git clone https://github.com/Priyanchew/EmotionAware-AI.git
-
Navigate to the Project Directory:
cd EmotionAware-AI -
Install Dependencies:
-
Backend and Speech Analysis:
Navigate to the respective directories and install the required Python packages:
cd backend pip install -r requirements.txtcd ../speech pip install -r requirements.txt -
Frontend:
Navigate to the frontend directory and install the necessary Node.js packages:
cd ../frontend npm install
-
-
Run the Application:
-
Backend:
cd ../backend python app.py -
Frontend:
In a new terminal window:
cd frontend npm start
-
Once the application is running:
-
Access the Frontend Interface:
Open your web browser and navigate to
http://localhost:3000to interact with the EmotionAware-AI system. -
Emotion Detection:
- Facial Analysis: Use your device's camera to allow the system to analyze your facial expressions.
- Speech Analysis: Provide vocal input through your device's microphone for speech emotion recognition.
-
Real-time Feedback:
The system will display the detected emotions in real-time, providing insights into the emotional state conveyed through facial expressions and speech.
Contributions to EmotionAware-AI are welcome. To contribute:
-
Fork the Repository:
Click on the 'Fork' button at the top right corner of the repository page.
-
Create a New Branch:
git checkout -b feature/your-feature-name
-
Make Your Changes:
Implement your feature or fix and commit the changes with a descriptive message.
-
Push to Your Fork:
git push origin feature/your-feature-name
-
Submit a Pull Request:
Navigate to the original repository and click on 'New Pull Request' to submit your changes for review.
This project is licensed under the MIT License. See the LICENSE file for details.
Special thanks to the contributors and the open-source community for their invaluable support and resources.
Note: For optimal performance, ensure that your device has a functional camera and microphone. The application may require permissions to access these devices.