Academia.eduAcademia.edu

Music Emotion Classification

description72 papers
group70 followers
lightbulbAbout this topic
Music Emotion Classification is the interdisciplinary study of identifying and categorizing the emotional content of music using computational methods, psychological theories, and music theory. It involves analyzing audio features, lyrics, and contextual factors to determine the emotions conveyed by musical pieces, facilitating applications in areas such as music recommendation systems and affective computing.
lightbulbAbout this topic
Music Emotion Classification is the interdisciplinary study of identifying and categorizing the emotional content of music using computational methods, psychological theories, and music theory. It involves analyzing audio features, lyrics, and contextual factors to determine the emotions conveyed by musical pieces, facilitating applications in areas such as music recommendation systems and affective computing.

Key research themes

1. How can dynamic, dimensionally-annotated datasets and benchmarks advance the evaluation and development of music emotion recognition systems?

This research area focuses on creating large, publicly available datasets that provide continuous, time-dependent emotional annotations (primarily in valence-arousal dimensions) for musical excerpts, enabling standardized benchmarking of music emotion recognition (MER) methods. Such datasets tackle challenges of data scarcity, copyright restrictions, and inconsistent annotation schemes, offering a foundation for systematic comparison of feature sets and algorithms in MER. The theme is crucial for developing robust MER systems that capture temporal emotion variations in music and for fostering reproducibility and comparability across studies.

Key finding: Introduced the MediaEval Database for Emotional Analysis in Music (DEAM), the largest dataset with continuous valence and arousal annotations at 2 Hz resolution over 1,802 Creative Commons songs, supporting dynamic music... Read more
Key finding: Created a sizable dataset of 903 audio clips labeled across five mood clusters aligned with MIREX standards, while using multiple audio feature extraction frameworks and support vector machine classifiers. Achieved an... Read more
Key finding: Proposed a biologically inspired cochlear modeling approach combined with convolutional neural networks to extract features from cochleogram images, aligning with human auditory perception. Evaluated on a public 1000-song... Read more

2. What are the effective machine learning approaches for multilabel and multimodal classification of emotions induced or perceived in music?

This theme investigates advanced machine learning methods capable of recognizing multiple simultaneous emotions in music, reflecting the complexity of human emotional responses. It encompasses multilabel classification paradigms, multimodal integration of audio and lyrics or video data, and the use of deep learning architectures like CNNs, LSTMs, and transformer models (e.g., XLNet). Addressing multilabel and multimodal approaches expands recognition accuracy and models emotional nuance, better reflecting real-world scenarios and enhancing applications such as music recommendation and emotion-based interaction.

Key finding: Analyzed Geneva Emotional Music Scale 9 annotations in the Emotify dataset using several machine learning algorithms for multilabel and multiclass classification, emphasizing simultaneous emotions. Findings informed... Read more
Key finding: Developed a multimodal MER system using mel spectrograms with CNN-LSTM for audio and XLNet transformers for lyrics, combining outputs via stacking ensemble and ANN meta-classifier. Achieved state-of-the-art 80.56% accuracy on... Read more
Key finding: Presented a hybrid emotion classification framework combining audio and video features extracted from the SAVEE database, using SVM for classification. The hybrid approach significantly improved accuracy to 99.26% compared to... Read more
Key finding: Reviewed and applied AI algorithms including SVM, RNN, and CNN on audio features such as pitch and Mel-frequency cepstral coefficients, illustrating deep learning’s superiority in modeling sequential emotional patterns from... Read more

3. How do music structural elements and compositional eras influence perceived musical emotions, and how can computational models incorporate these insights?

This theme explores how intrinsic musical features (e.g., tempo, mode, pitch patterns) and historical changes across musical eras affect emotional perception. It investigates score-based analyses combined with perceptual evaluations to reveal changing cue associations (e.g., between major/minor modes and emotional valence/arousal) from Classical to Romantic periods. Integrating these musicological insights with computational models enhances generation of emotionally expressive music and improves classification by accounting for temporal and cultural factors shaping emotional meaning.

Key finding: Combined score-based acoustic cue analyses with behavioral classification of Bach and Chopin excerpts, revealing that Romantic era compositions alter associations between musical mode and affective meanings compared to... Read more
Key finding: Developed EmotionBox, a deep neural network system generating symbolic music guided by music elements tempo and mode derived from music psychology, mapped onto emotional valence-arousal dimensions without requiring labeled... Read more
Key finding: Investigated various feature sets including audio signal processing, chord features, and EEG data to classify music emotion in valence-activation space. Found that combining music-inspired features, frequency modulation... Read more

All papers in Music Emotion Classification

In this thesis we focus on the automatic emotion classification of music samples. We extract a set of features from the music signal and examine their discriminatory capability using various classification techniques. Our goal is to... more
This paper focuses on emotion recognition and understanding in Contemporary Western music. The study seeks to investigate the relationship between perceived emotion and musical features in the fore-mentioned musical genre. A set of 27... more
Emotion judgments and five channels of physiological data were obtained from 60 participants listening to 60 music excerpts. Various machine learning (ML) methods were used to model the emotion judgments inclusive of neural networks,... more
Abstract This project will investigate the link between physical expression and emotions in music. Since the emotive nature of music is already well known, it will focus on the following questions: 1. Is physical expression a natural... more
In this paper, the automated detection of emotion in music is modeled as a multilabel classification task, where a piece of music may belong to more than one class. Four algorithms are evaluated and compared in this task. Furthermore, the... more
Electroencephalography (EEG)-based emotion classification during music listening has gained increasing attention nowadays due to its promise of potential applications such as musical affective brain-computer interface (ABCI),... more
Electroacoustic music with video is now frequently programmed in classical concert venues. There are many different ways in which video and sound can be organized together into an effective work of art. But after attending a number of... more
Müzik kavramı bir lisanı, bu lisanın alfabesini ve bu dilin edebiyatını ifade eder. Tıpkı birincil anlamı ile akla gelen edebiyatta olduğu gibi, müzikte de bir olgu farklı biçimler ve farklı edebi kurulumlarla ifade edilebilir. Müziğin... more
It has been observed from recent studies that corticolimbic Theta rhythm from EEG recordings perceived as fear or threatening scene during neural processing of visual stimuli. In additions, neural oscillations’ patterns in Theta, Alpha... more
In this paper we propose a neural network model for human emotion and gesture classification. We demonstrate that the proposed architecture represents an effective tool for real-time processing of customer's behavior for distributed... more
Speech processing is the study of speech signals, and the methods used to process them. In application such as speech coding, speech synthesis, speech recognition and speaker recognition technology, speech processing is employed. In... more
Spectral and excitation features, commonly used in automatic emotion classification systems, parameterise different aspects of the speech signal. This paper groups these features as speech production cues, broad spectral measures and... more
Many current electroacoustic works are weakened by not having clear structure. Why is this happening, and why now? In the following article on New Music Box I consider a number of reasons why this may be occurring. I begin with... more
In recent years, there are many great successes in using deep architectures for unsupervised feature learning from data, especially for images and speech. In this paper, we introduce recent advanced deep learning models to classify two... more
Continuous emotion prediction in the arousal-valence space is now being used in various modalities: music, facial expressions, gestures, text, etc. In order to be able to compare the work of different research groups effectively, we... more
Enabling computer systems to recognize facial expressions and infer emotions from them in real time presents a challenging research topic. In this paper, we present a real time approach to emotion recognition through facial expression in... more
The following paper presents parameterization of emotional speech using perceptual coefficients as well as a comparison of Mel Frequency Cepstral Coefficients (MFCC), Bark Frequency Cepstral Coefficients (BFCC), Perceptual Linear... more
Özetçe Literatürde birbirinden farklı sınıflandırma algoritmalarının belirli bir problem için performanslarının karşılaştırılması oldukça yaygın bir uygulama olarak karşımıza çıkmaktadır. Ancak bu çalışmalardan elde edilen sonuçların... more
Özetçe Literatürde birbirinden farklı sınıflandırma algoritmalarının belirli bir problem için performanslarının karşılaştırılması oldukça yaygın bir uygulama olarak karşımıza çıkmaktadır. Ancak bu çalışmalardan elde edilen sonuçların... more
In this paper, the automated detection of emotion in music is modeled as a multilabel classification task, where a piece of music may belong to more than one class. Four algorithms are evaluated and compared in this task. Furthermore, the... more
Exploring how people represent natural categories is a key step toward developing a better understanding of how people learn, form memories, and make decisions. Much research on categorization has focused on artificial categories that are... more
During everyday interaction people display various non-verbal signals that convey emotions. These signals are multi-modal and range from facial expressions, shifts in posture, head pose, and non-verbal speech. They are subtle, continuous... more
In this paper, we propose a neural network model for human emotion and gesture classification. We demonstrate that the proposed architecture represents an effective tool for real-time processing of customer's behavior for distributed... more
Emotion recognition has been a research topic in the field of Human Computer Interaction (HCI) during recent years. Computers have become an inseparable part of human life. Users need human-like interaction to better communicate with... more
Music has grown into an important part of people's daily lives. As we move further into the digital age in which a large collection of music is being created daily and becomes easily accessible renders people to spend more time on... more
Download research papers for free!