Key research themes
1. How can dynamic, dimensionally-annotated datasets and benchmarks advance the evaluation and development of music emotion recognition systems?
This research area focuses on creating large, publicly available datasets that provide continuous, time-dependent emotional annotations (primarily in valence-arousal dimensions) for musical excerpts, enabling standardized benchmarking of music emotion recognition (MER) methods. Such datasets tackle challenges of data scarcity, copyright restrictions, and inconsistent annotation schemes, offering a foundation for systematic comparison of feature sets and algorithms in MER. The theme is crucial for developing robust MER systems that capture temporal emotion variations in music and for fostering reproducibility and comparability across studies.
2. What are the effective machine learning approaches for multilabel and multimodal classification of emotions induced or perceived in music?
This theme investigates advanced machine learning methods capable of recognizing multiple simultaneous emotions in music, reflecting the complexity of human emotional responses. It encompasses multilabel classification paradigms, multimodal integration of audio and lyrics or video data, and the use of deep learning architectures like CNNs, LSTMs, and transformer models (e.g., XLNet). Addressing multilabel and multimodal approaches expands recognition accuracy and models emotional nuance, better reflecting real-world scenarios and enhancing applications such as music recommendation and emotion-based interaction.
3. How do music structural elements and compositional eras influence perceived musical emotions, and how can computational models incorporate these insights?
This theme explores how intrinsic musical features (e.g., tempo, mode, pitch patterns) and historical changes across musical eras affect emotional perception. It investigates score-based analyses combined with perceptual evaluations to reveal changing cue associations (e.g., between major/minor modes and emotional valence/arousal) from Classical to Romantic periods. Integrating these musicological insights with computational models enhances generation of emotionally expressive music and improves classification by accounting for temporal and cultural factors shaping emotional meaning.