Music emotion recognition (MER) studies have made great progress in detecting the emotions of music segments and analyzing the emotional dynamics of songs. The overall emotion and depth information of entire songs may be more suitable for real-life applications in certain scenarios. This study focuses on recognizing the overall emotion and depth of entire songs. First, we constructed a public dataset containing 3839 popular songs in China (PSIC3839) by conducting an online experiment to collect the arousal, valence, and depth annotation of each song. Second, we used handcrafted feature-based method to predict the overall emotion and depth values. Support vector regressions using Mel frequency cepstrum coefficients features as inputs achieve good model performance (arousal: R2 = 0.609; valence: R2 = 0.354; and depth: R2 = 0.465). Finally, the groupwise and personalized results were also investigated by training a unique regressor for each group or individual, which provides a reference for future research.
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
Tel.: +1 703 830 6300
Fax: +1 703 830 2300 firstname.lastname@example.org
(Corporate matters and books only) IOS Press c/o Accucoms US, Inc.
For North America Sales and Customer Service
West Point Commons
Lansdale PA 19446
Tel.: +1 866 855 8967
Fax: +1 215 660 5042 email@example.com