Skip to content

🎹 Moodify - an emotion-based music recommendation system that uses AI/ML models to analyze text, speech, and facial expressions, providing personalized music recommendations across web and mobile platforms.

License

Notifications You must be signed in to change notification settings

rickynguyxn/Moodify-Emotion-Music-App

Β 
Β 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

62 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Moodify - Emotion-Based Music Recommendation App

With the rise of personalized music streaming services, there is a growing need for systems that can recommend music based on users' emotional states. Realizing this need, Moodify is being developed by Son Nguyen in 2024 to provide personalized music recommendations based on users' detected emotions.

The Moodify project is an integrated emotion-based music recommendation system that combines frontend, backend, AI/ML models, and data analytics to provide personalized music recommendations based on user emotions. The application analyzes text, speech, or facial expressions and suggests music that aligns with the detected emotions.

Supporting both desktop and mobile platforms, Moodify offers a seamless user experience with real-time emotion detection and music recommendations. The project leverages React for the frontend, Django for the backend, and three advanced, self-trained AI/ML models for emotion detection. Data analytics scripts are used to visualize emotion trends and model performance.

Moodify Logo

Table of Contents

🎡 Overview

Moodify provides personalized music recommendations based on users' emotional states detected through text, speech, and facial expressions. It interacts with a Django-based backend, AI/ML models for emotion detection, and utilizes data analytics for visual insights into emotion trends and model performance.

🌟 Features

  • User registration and login functionality.
  • Input analysis through text, speech, and facial expressions.
  • Real-time music recommendations based on emotion detection.
  • Visualization of emotion detection results and user history.
  • Data analytics scripts for emotion trends and model performance.
  • AI/ML models for text, speech, and facial emotion detection.
  • User profile management and customization.
  • Mobile app version for seamless user experience.
  • Progressive Web App (PWA) features for offline support.
  • Admin panel for managing users, recommendations, and data analytics.

πŸ› οΈ Technologies

  • Frontend:
    • React
    • Axios
    • Material UI (MUI)
    • React Router
    • React Redux
    • Jest and React Testing Library
  • Backend:
    • Django
    • Django REST Framework
    • MongoEngine
    • JWT Authentication
    • Spotify APIs
    • Swagger and Redoc for API documentation
  • Databases:
    • MongoDB
    • Redis (for server-side caching)
    • SQLite
  • AI/ML Models:
    • PyTorch
    • TensorFlow
    • Keras
    • HuggingFace Transformers
    • Pandas
    • Scikit-learn
    • NumPy
    • FER (Facial Expression Recognition) Library
  • Data Analytics:
    • Pandas
    • Matplotlib
    • Hadoop
    • Spark
  • Mobile:
    • React Native
    • Expo
    • Expo Go
  • PWA:
    • Progressive Web App features for offline support
  • Containerization:
    • Docker
    • Docker Desktop

πŸ–ΌοΈ User Interface

Landing Page

Landing Page

Home Page

Home Page

Text Input

Text Input

Speech Input

Speech Input

Speech Input

Facial Expression Input

Facial Input

Facial Input

Profile Page

Profile Page

Profile Page

Results/Recommendations Page

Results Page

Login Page

Login Page

Registration Page

Registration Page

404 Not Found Page

404 Not Found Page

Footer

Footer

πŸ“‚ Complete File Structure

The project has a comprehensive file structure combining frontend, backend, AI/ML models, and data analytics components:

Moodify/
β”œβ”€β”€ frontend/                      # React frontend for the web application
β”‚   β”œβ”€β”€ public/
β”‚   β”‚   β”œβ”€β”€ index.html             # Main HTML file
β”‚   β”‚   β”œβ”€β”€ manifest.json          # Web app manifest
β”‚   β”‚   └── favicon.ico            # Favicon for the app
β”‚   β”‚
β”‚   β”œβ”€β”€ src/
β”‚   β”‚   β”œβ”€β”€ components/            # Contains all React components
β”‚   β”‚   β”œβ”€β”€ pages/                 # Contains main pages of the app
β”‚   β”‚   β”œβ”€β”€ styles/                # Contains global styles and themes
β”‚   β”‚   β”œβ”€β”€ App.js                 # Main App component
β”‚   β”‚   β”œβ”€β”€ index.js               # Entry point for React
β”‚   β”‚   └── theme.js               # Material UI theme configuration
β”‚   β”‚ 
β”‚   β”œβ”€β”€ .gitignore                 # Git ignore file
β”‚   β”œβ”€β”€ package.json               # NPM dependencies and scripts
β”‚   └── README.md                  # Project documentation
β”‚ 
β”œβ”€β”€ backend/                       # Django backend for API services and database management
β”‚   β”œβ”€β”€ manage.py                  # Django's command-line utility
β”‚   β”œβ”€β”€ requirements.txt           # Backend dependencies
β”‚   β”œβ”€β”€ backend/
β”‚   β”‚   β”œβ”€β”€ settings.py            # Django settings for the project
β”‚   β”‚   β”œβ”€β”€ urls.py                # URL declarations for the project
β”‚   β”‚   β”œβ”€β”€ users/                 # User management components
β”‚   β”‚   └── api/                   # Emotion detection and recommendation APIs
β”‚   β”‚
β”‚   └── db.sqlite3                 # SQLite database (if used)
β”‚
β”œβ”€β”€ ai_ml/                         # AI/ML models for emotion detection
β”‚   β”œβ”€β”€ data/                      # Datasets for training and testing
β”‚   β”œβ”€β”€ models/                    # Trained models for emotion detection
β”‚   β”œβ”€β”€ src/                       # Source files for emotion detection and recommendation
β”‚   β”‚   β”œβ”€β”€ api/                   # API scripts for running emotion detection services
β”‚   β”‚   β”œβ”€β”€ recommendation/        # Music recommendation logic
β”‚   β”‚   └── data_processing/       # Data preprocessing scripts
β”‚   β”‚
β”‚   └── README.md                 # AI/ML documentation
β”‚
β”œβ”€β”€ data_analytics/                # Data analytics scripts and visualizations
β”‚   β”œβ”€β”€ emotion_distribution.py    # Script for visualizing emotion distribution
β”‚   β”œβ”€β”€ training_visualization.py  # Script for visualizing training and validation metrics
β”‚   β”œβ”€β”€ predictions_analysis.py    # Script for analyzing model predictions
β”‚   β”œβ”€β”€ recommendation_analysis.py # Script for visualizing music recommendations
β”‚   β”œβ”€β”€ spark-hadoop/              # Spark and Hadoop integration scripts
β”‚   └── visualizations/            # Generated visualizations
β”‚
β”œβ”€β”€ mobile/                        # React Native mobile application
β”‚   β”œβ”€β”€ App.js                     # Main entry point for React Native app
β”‚   β”œβ”€β”€ index.js                   # App registry for React Native
β”‚   β”œβ”€β”€ package.json               # NPM dependencies and scripts
β”‚   β”œβ”€β”€ babel.config.js            # Babel configuration for React Native
β”‚   β”œβ”€β”€ 
β”‚   β”œβ”€β”€ components/                # React Native components
β”‚   β”‚   β”œβ”€β”€ Footer.js              # Footer component
β”‚   β”‚   β”œβ”€β”€ Navbar.js              # Header component
β”‚   β”‚   β”œβ”€β”€ Auth/                  # Authentication components (e.g., Login, Register)
β”‚   β”‚   └── Profile/               # Profile-related components
β”‚   β”‚
β”‚   β”œβ”€β”€ pages/                     # Main pages of the app
β”‚   β”‚   β”œβ”€β”€ HomePage.js            # Home page component
β”‚   β”‚   β”œβ”€β”€ ProfilePage.js         # Profile page component
β”‚   β”‚   β”œβ”€β”€ ResultsPage.js         # Results page component
β”‚   β”‚   └── NotFoundPage.js        # 404 page component
β”‚   β”‚
β”‚   β”œβ”€β”€ assets/                    # Images, fonts, and other assets
β”‚   β”œβ”€β”€ styles/                    # Styling files (similar to CSS for web)
β”‚   β”œβ”€β”€ .gitignore                 # Git ignore file
β”‚   β”œβ”€β”€ package.json               # Dependencies and scripts
β”‚   └── README.md                  # Mobile app documentation
β”‚
β”œβ”€β”€ images/                        # Images used in the README documentation 
└── README.md                      # Comprehensive README file for the entire project

πŸ› οΈ Getting Started

Prerequisites

  • Node.js (v14 or higher)
  • Python 3.8 or later
  • MongoDB
  • Virtual Environment (venv)

1. Setup and Train AI/ML Models

Start with setting up and training the AI/ML models, as they will be required for the backend to function properly.

  1. Clone the repository:

    git clone https://github.com/hoangsonww/Moodify-Emotion-Music-App.git
  2. Navigate to the AI/ML directory:

    cd Moodify-Emotion-Music-App/ai_ml
  3. Create and activate a virtual environment:

    python -m venv venv
    source venv/bin/activate   # For macOS/Linux
    .\venv\Scripts\activate    # For Windows
  4. Install dependencies:

    pip install -r requirements.txt
  5. Edit the configurations in the src/config.py file:

    • Visit the src/config.py file and update the configurations as needed, especially your Spotify API keys and configure ALL the paths.
    • Visit the individual model training scripts in the src/models directory and update the paths to the datasets and output paths as needed.
    • Ensure all paths are correctly set before training the models!
  6. Train the text emotion model:

    python src/models/train_text_emotion.py

    Repeat similar commands for other models as needed (e.g., facial and speech emotion models).

  7. Ensure all trained models are placed in the models directory, and that you have trained all necessary models before moving to the next step!

  8. Test the trained AI/ML models as needed:

    • Run the src/models/test_emotion_models.py script to test the trained models.
    • Ensure the models are providing accurate predictions before moving to the next step.

2. Set Up the Backend

Once the AI/ML models are ready, proceed with setting up the backend.

  1. Navigate to the backend directory:

    cd ../backend
  2. Create and activate a virtual environment:

    python -m venv venv
    source venv/bin/activate   # For macOS/Linux
    .\venv\Scripts\activate    # For Windows
  3. Install dependencies:

    pip install -r requirements.txt
  4. Configure your secrets and environment:

    • Create a .env file in the backend directory.
    • Add the following environment variables to the .env file:
      SECRET_KEY=your_secret_key
      DEBUG=True
      ALLOWED_HOSTS=<your_hosts>
      MONGODB_URI=<your_mongodb_uri>
      
    • Visit backend/settings.py and add SECRET_KEY & set DEBUG to True.
    • Important: Ensure these steps are completed before running the backend server.
  5. Run database migrations:

    python manage.py migrate
  6. Start the Django server:

    python manage.py runserver

    The backend server will be running at http://127.0.0.1:8000/.

3. Install and Run the Frontend

Finally, set up the frontend to interact with the backend.

  1. Navigate to the frontend directory:

    cd ../frontend
  2. Install dependencies using Yarn:

    npm install
  3. Start the development server:

    npm start

    The frontend will start at http://localhost:3000.

Note: If you encounter any problems or need my .env file, feel free to contact me.

πŸ“‹ API Endpoints

User Endpoints

HTTP Method Endpoint Description
POST /users/register/ Register a new user
POST /users/login/ Login a user and obtain a JWT token
GET /users/user/profile/ Retrieve the authenticated user's profile
PUT /users/user/profile/update/ Update the authenticated user's profile
DELETE /users/user/profile/delete/ Delete the authenticated user's profile
POST /users/recommendations/ Save recommendations for a user
GET /users/recommendations/<str:username>/ Retrieve recommendations for a user by username
DELETE /users/recommendations/<str:username>/<str:recommendation_id>/ Delete a specific recommendation for a user
DELETE /users/recommendations/<str:username>/ Delete all recommendations for a user
POST /users/mood_history/<str:user_id>/ Add a mood to the user's mood history
GET /users/mood_history/<str:user_id>/ Retrieve mood history for a user
DELETE /users/mood_history/<str:user_id>/ Delete a specific mood from the user's history
POST /users/listening_history/<str:user_id>/ Add a track to the user's listening history
GET /users/listening_history/<str:user_id>/ Retrieve listening history for a user
DELETE /users/listening_history/<str:user_id>/ Delete a specific track from the user's history
POST /users/user_recommendations/<str:user_id>/ Save a user's recommendations
GET /users/user_recommendations/<str:user_id>/ Retrieve a user's recommendations
DELETE /users/user_recommendations/<str:user_id>/ Delete all recommendations for a user

Emotion Detection Endpoints

HTTP Method Endpoint Description
POST /api/text_emotion/ Analyze text for emotional content
POST /api/speech_emotion/ Analyze speech for emotional content
POST /api/facial_emotion/ Analyze facial expressions for emotions
POST /api/music_recommendation/ Get music recommendations based on emotion

Admin Interface Endpoints

HTTP Method Endpoint Description
GET /admin/ Access the Django Admin interface

Admin Interface

  1. Create a superuser:

    python manage.py createsuperuser
  2. Access the admin panel at http://127.0.0.1:8000/admin/

  3. You should see the following login page:

Admin Login

πŸš€ Backend APIs Documentation

Our backend APIs are all well-documented using Swagger UI and Redoc. You can access the API documentation at the following endpoints:

  • Swagger UI: http://127.0.0.1:8000/swagger.
  • Redoc: http://127.0.0.1:8000/redoc.

If you have set up the backend correctly, you should see the following API documentation:

Swagger UI:

Swagger UI

Redoc:

Redoc

πŸ€– About the AI/ML Models

The AI/ML models are built using PyTorch, TensorFlow, Keras, and HuggingFace Transformers. These models are trained on various datasets to detect emotions from text, speech, and facial expressions.

The emotion detection models are used to analyze user inputs and provide real-time music recommendations based on the detected emotions. The models are trained on various datasets to capture the nuances of human emotions and provide accurate predictions.

  • Text Emotion Detection: Detects emotions from text inputs.
  • Speech Emotion Detection: Analyzes emotions from speech inputs.
  • Facial Emotion Detection: Detects emotions from facial expressions.

The models are integrated into the backend API services to provide real-time emotion detection and music recommendations for users.

Note: The models must be trained first before using them in the backend services. Ensure that the models are trained and placed in the models directory before running the backend server. Refer to the (Getting Started)[#getting-started] section for more details.

AI/ML Models Examples of training the text emotion model.

πŸ“± Mobile App Version

There is also a mobile version of the Moodify app built using React Native and Expo. You can find the mobile app in the mobile directory.

  1. Navigate to the mobile directory:

    cd ../mobile
  2. Install dependencies using Yarn:

     yarn install
  3. Start the Expo development server:

    yarn start
  4. Scan the QR code using the Expo Go app on your mobile device to run the app.

If successful, you should see the following home screen:

Mobile Home

Feel free to explore the mobile app and test its functionalities!

πŸ“Š Analytics Scripts

The data_analytics folder provides data analysis and visualization scripts to gain insights into the emotion detection model's performance.

  1. Run All Analytics Scripts:

    python data_analytics/main.py
  2. View generated visualizations in the visualizations folder.

  3. Here are some example visualizations:

Emotion Distribution Emotion Distribution Visualization

Training Loss Curve Visualization Training Loss Curve Visualization

🐳 Containerization

The project can be containerized using Docker for easy deployment and scaling. You can create Docker images for the frontend, backend, and AI/ML models.

  1. Build the Docker images:

    docker compose up --build
  2. The Docker images will be built for the frontend, backend, and AI/ML models. Verify the images using:

    docker images

If you encounter any errors, try to rebuild your image without using the cache since Docker's cache may cause issues.

docker-compose build --no-cache

πŸ”§ Contributing

  • Contributions are welcome! Feel free to fork the repository and submit a pull request.

  • Note that this project is still under active development, and any contributions are appreciated.

  • If you have any suggestions, feature requests, or bug reports, feel free to open an issue here.

πŸ“ License

  • This project is licensed under the MIT License. Please see the LICENSE file for details.

πŸ“§ Contact


Happy Coding and Analyzing! 🎢

πŸ” Back to Top

About

🎹 Moodify - an emotion-based music recommendation system that uses AI/ML models to analyze text, speech, and facial expressions, providing personalized music recommendations across web and mobile platforms.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • JavaScript 59.7%
  • Python 35.3%
  • CSS 2.7%
  • HTML 1.7%
  • Dockerfile 0.6%