Final 4
Final 4
B. E. [Information Technology]
of
Savitribai Phule Pune University
By
2
Project Group ID 6
3 Title of Project
Mood Based Music Recommendation
System
Shelke Akshay:
Sinare Rajeshree:
Sanap Madhuri:
Rasal Rutuja:
Prof. N. L. Bhale
Head, Information Technology Department
SYNOPSIS OF THE WORK
2. Literature review:
We propose a new approach for playing music automatically using facial emotion. Most of
the existing approaches involve playing music manually, using wearable computing devices,
or classifying based on audio features. Instead, we propose to change the manual sorting and
playing. We have used a Convolutional Neural Network for emotion detection. For music
recommendations, Pygame & Tkinter are used. Our proposed sys-tem tends to reduce the
computational time involved in obtaining the results and the overall cost of the designed
system, thereby increasing the system’s overall accuracy. Testing of the system is done on the
FER2013 dataset. Facial expressions are captured using an inbuilt camera. Feature extraction
is performed on input face images to detect emotions such as happy, angry, sad, surprise, and
neutral. Automatically music playlist is generated by identifying the current emotion of the
user. It yields better performance in terms of computational time, as compared to the
algorithm in the existing literature[1].
One of the most significant fields in the man–machine interface is emotion recognition using
facial expressions. Some of the challenges in the emotion recognition area are facial
accessories, non-uniform illuminations, pose variations, etc. Emotion detection using
conventional approaches having the drawback of mutual optimization of feature extraction
and classification. To overcome this problem, researchers are showing more attention toward
deep learning techniques. Nowadays, deep-learning approaches are playing a major role in
classification tasks. This paper deals with emotion recognition by using transfer learning
approaches. In this work pre-trained networks of Resnet50, vgg19, Inception V3, and Mobile
Net are used. The fully connected layers of the pre-trained ConvNets are eliminated, and we
add our fully connected layers that are suitable for the number of instructions in our task.
Finally, the newly added layers are only trainable to update the weights. The experiment was
conducted by using the CK + database and achieved an average accuracy of 96% for emotion
detection problems [2].
In the current era of media and technology, music information retrieval techniques have made
progress in the past years but the music recommendation system is still at a rudimentary
stage. However personalized music recommendation is quotidian, but recommending songs
based on emotions is still an uphill battle. Music has greatly influenced the human brain and
helps dispense an exhilarating and frivolous state of mind because it helps us work more
effectively. Recommending songs based on emotions will comfort the listener by suggesting
music in keeping with the listeners' pervading mental and physical state. Hence, Natural
Language Processing and Deep Learning technologies made it possible for machines to read
and interpret emotions through texts by recognizing patterns and finding correlations. In this
paper, various deep learning models such as Long Short-Term Memory (LSTM),
Convolution Neural network (CNN), CNN-LSTM, and LSTM CNN Architectures were
collated for detecting emotions such as angry, happy, love, and sad, the best model was
integrated into the application. To enhance the application, a CNN model was used to detect
emotions through facial expressions. The application takes text input from the user or a facial
expression input. Depending upon the emotion detected, it recommends the user songs and
playlists[3].
It is often confusing for a person to decide which music he/she have to listen from a massive
collection of existing options. There have been several suggestion frameworks available for
issues like music, dining, and shopping depending upon the mood of user. The main objective
of our music recommendation system is to provide suggestions to the users that fit the user's
preferences. The analysis of the facial expression/user emotion may lead to understanding
the current emotional or mental state of the user. Music and videos are one region where there
is a significant chance to prescribe abundant choices to clients in light of their inclinations
and also recorded information. It is well known that humans make use of facial expressions
to express more clearly what they want to say and the context in which they meant their
words. More than 60 percent of the users believe that at a certain point of time the number of
songs present in their songs library is so large that they are unable to figure out the song
which they have to play. By developing a recommendation system, it could assist a user to
make a decision regarding which music one should listen to helping the user to reduce his/her
stress levels. The user would not have to waste any time in searching or to look up for songs
and the best track matching the user's mood is detected, and songs would be shown to the
user according to his/her mood. The image of the user is captured with the help of a webcam.
The user's picture is taken and then as per the mood/emotion of the user an appropriate song
from the playlist of the user is shown matching the user's requirement[4].
3. Motivation :
The motivation for developing a music recommendation system based on mood detection
stems from a few key areas of interest and practical needs:
5. Objectives:
• To study the literature to understand the challenges in developing and automatic mood
recommendation system.
• To study various techniques to enhance the performance parameters of an automatic mood
recommendation system.
• To design a system that can accurately recommend music based on the real-time detection
of a user's mood.
• To Achieve High Accuracy in Emotion Detection.
• To validate the results of the proposed system.
6. Scope of Project
The “Mood Based Music Recommendation” introduces an innovative approach to music
recommendation, merging advanced image processing and mood recognition with
personalized music selection. Utilizing a Convolutional Neural Network (CNN), the project
focuses on identifying a listener’s current emotional state — Angry, Happy, Sad, or Calm —
through analysing facial expressions in images or videos. This mood recognition aspect is
critical as it forms the basis for a tailored music recommendation system. The project’s unique
feature lies in its ability to align music recommendations with the detected emotional state,
enhancing the relevance and impact of the suggested songs. In the second phase of the project,
songs are categorized into the emotional states based on their intrinsic features, employing
content-based filtering for recommendations. This method ensures that the music selection is
not only aligned with the user’s preferences but also resonates with their current mood. The
system’s aim is to enrich the music listening experience, offering songs that are not just liked
but are emotionally fitting as well. By integrating emotional intelligence into music
recommendation, the project intends to create responsive and empathetic AI systems, ensuring
a more personalized and emotionally engaging user experience.
7. List of required hardware, software, or other equipment for
executing the project:
Software Requirements:
5. Database: SQLite, MySQL, or PostgreSQL for storing user preferences and music data.
6. API/Services: Spotify API or any music streaming service API for song recommendations.
Hardware Requirements:
1. Processor: Intel Core i5 or equivalent for smooth performance during model training and
testing.
2. RAM: Minimum 8GB RAM (16GB recommended) to handle real-time image processing
and model predictions.
3. Storage: At least 256GB of SSD for fast read/write operations, especially for handling
large datasets.
4. Graphics Card: NVIDIA GPU (e.g., GTX 1060 or higher) if using GPU acceleration for
training deep learning models.
Expected Outcomes :
Accurate Mood Detection:
The system should be able to accurately detect the user’s mood using inputs such as facial
expressions, voice, or text. The detection system can analyze emotions like happiness, sadness,
anger, or relaxation and identify the corresponding mood.
Personalized Music Recommendations:
Based on the detected mood, the system will suggest music that fits the emotional state of the
user. For instance, calming music for relaxation, upbeat tracks for happiness, or slower tempos
for sad moods.
Improved User Experience:
The system aims to enhance the music streaming experience by offering emotionally intelligent
recommendations. Users would likely find this more engaging compared to traditional
recommendation systems based on genres or playlists.
Dynamic Music Adjustments:
The system could adapt in real-time as the user’s mood shifts, continuously updating the music
recommendations, ensuring that the music matches their mood throughout the session.
Enhanced Mental Well-being:
By offering mood-enhancing or mood-supporting music, the system could help users manage
stress, anxiety, or even improve focus and productivity, contributing to overall mental well-being.
Data-Driven Insights:
The system can provide data on how music preferences correlate with emotions, which can be
useful for further studies or improvements in recommendation algorithms.
Integration with Music Streaming Platforms:
The system could be integrated with popular platforms like Spotify or Apple Music, allowing
users to receive recommendations seamlessly within their existing music libraries.
Scalability and Customization:
The final system should be scalable, allowing for customization based on user preferences,
cultural background, or even activity (e.g., workout, studying).
1. Aggarwal, K., Mijwil, M. M., Al-Mistarehi, A. H., Alomari, S., Gök, M., Alaabdin, A. M. Z.,
& Abdulrhman, S. H. (2022). Has the future started? The current growth of artificial
intelligence, machine learning and deep learning. Iraqi Journal for Computer Science and
Mathematics, 3(1), 115-123. https://doi.org/10.52866/ijcsm.2022.01.01.013
2. Athavle, M., Mudale, D., Shrivastav, U., & Gupta, M. (2021). Music Recommendation Based
on Face Emotion Recognition. Journal of Informatics Electrical and Electronics Engineering
(JIEEE), 2(2), 1-11. https://doi.org/10.54060/JIEEE/002.02.018
3. Bhattarai, B., & Lee, J. (2019). Automatic music mood detection using transfer learning and
multilayer perceptron. International Journal of Fuzzy Logic and Intelligent Systems, 19(2),
88-96. https://doi.org/10.5391/IJFIS.2019.19.2.88
4. Chowdary, M. K., Nguyen, T. N., & Hemanth, D. J. (2021). Deep learning-based facial
emotion recognition for human computer interaction applications. Neural Computing and
Applications, 1-18. https://doi.org/10.1007/s00521-021-06012-8
5. Chung, T. S., Rust, R. T., & Wedel, M. (2009). My mobile music: An adaptive personalization
system for digital audio players. Marketing Science, 28(1), 52-68.
https://doi.org/10.1287/mksc.1080.0371
6. Florence, S. M., & Uma, M. (2020, August). Emotional detection and music recommendation
system based on user facial expression. In IOP Conference Series: Materials Science and
Engineering (Vol. 912, No. 6, p. 062007). IOP Publishing.
https://doi.org/10.1088/1757899X/912/6/062007
7. Hasib, K. M., Tanzim, A., Shin, J., Faruk, K. O., Al Mahmud, J., & Mridha, M. F. (2022).
BMNet-5: A novel approach of neural network to classify the genre of Bengali music based
on audio features. IEEE Access, 10, 108545-108563.
https://ieeexplore.ieee.org/abstract/document/9916245/
8. Hung, J. C., Lin, K. C., & Lai, N. X. (2019). Recognizing learning emotion based on
convolutional neural networks and transfer learning. Applied Soft Computing, 84, 105724.
https://doi.org/10.1016/j.asoc.2019.105724
9. IFPI. (2018). Global Music Report - Annual state of the industry. Retrieved from IFPI.
https://www.ifpi.org/ifpi-global-music-report-2018/ [Accessed 15 April 2023].
10. James, H. I., Arnold, J. J. A., Ruban, J. M. M., Tamilarasan, M., & Saranya, R. (2019). Emotion
based music recommendation system. Emotion, 6(03).
11. Joshi, S., Jain, T., & Nair, N. (2021, July). Emotion based music recommendation system using
LSTM-CNN architecture. In 2021 12th International Conference on Computing
Communication and Networking Technologies (ICCCNT) (pp. 01-06). IEEE.
https://ieeexplore.ieee.org/abstract/document/9579813
12. Khanzada, A., Bai, C., & Celepcikay, F. T. (2020). Facial expression recognition with deep
learning. Ar Xiv preprint arXiv:2004.11823. https://doi.org/10.48550/arXiv.2004.11823
11.Base papers
1.Tital: A Study on Emotion Analysis and Music Recommendation Using Transfer Learning.
Authors: Krishna Kumar Singh and Payal Dembla.