ABSTRACT
Mood is an emerging metadata type and access point in music digital libraries (MDL) and online music repositories. In this study, we present a comprehensive investigation of the usefulness of lyrics in music mood classification by evaluating and comparing a wide range of lyric text features including linguistic and text stylistic features. We then combine the best lyric features with features extracted from music audio using two fusion methods. The results show that combining lyrics and audio significantly outperformed systems using audio-only features. In addition, the examination of learning curves shows that the hybrid lyric + audio system needed fewer training samples to achieve the same or better classification accuracies than systems using lyrics or audio singularly. These experiments were conducted on a unique large-scale dataset of 5,296 songs (with both audio and lyrics for each) representing 18 mood categories derived from social tags. The findings push forward the state-of-the-art on lyric sentiment analysis and automatic music mood classification and will help make mood a practical access point in music digital libraries.
- Alm, C.O. 2009. Affect in Text and Speech. VDM Verlag: Saarbrücken.Google Scholar
- Argamon, S., Saric, M., and Stein, S. S. 2003. Style mining of electronic messages for multiple authorship discrimination: first results. In Proceedings of the 9th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. pp. 475--480. Google ScholarDigital Library
- Bischoff, K., Firan, C. S., Nejdl, W., and Paiu, R. 2009. How do you feel about "Dancing Queen"? Deriving mood and theme annotations from user tags. In Proceedings of Joint Conference on Digital Libraries (JCDL'09). Google ScholarDigital Library
- Bischoff, K., Firan, C., Paiu, R., Nejdl, W., Laurier, C., and Sordo, M. 2009. Music mood and theme classification - a hybrid approach. In Proceedings of the 10th International Conference on Music Information Retrieval (ISMIR'09).Google Scholar
- Bradley, M. M. and Lang, P. J. 1999. Affective Norms for English Words (ANEW): Stimuli, Instruction Manual and Affective Ratings. Technical report C-1. University of Florida.Google Scholar
- Chang, C. and Lin. C. 2001. LIBSVM: a library for support vector machines. Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvmGoogle Scholar
- Downie, J. S. 2008. The Music Information Retrieval Evaluation Exchange (2005-2007): A window into music information retrieval research. Acoustical Science and Technology 29 (4): 247--255. Available at: http://dx.doi.org/10.1250/ast.29.247Google ScholarCross Ref
- Fellbaum, C. 1998. WordNet: An Electronic Lexical Database, MIT Press.Google Scholar
- He, H., Jin, J., Xiong, Y., Chen, B., Sun, W., and Zhao, L. 2008. Language feature mining for music emotion classification via supervised learning from lyrics. In Proceedings of Advances in the 3rd International Symposium on Computation and Intelligence (ISICA 2008). Google ScholarDigital Library
- Hu, X. 2010. Music and mood: where theory and reality meet. In Proceedings of iConference 2010.Google Scholar
- Hu, Y., Chen, X. and Yang, D. 2009. Lyric-based song emotion detection with affective lexicon and fuzzy clustering method. In Proceedings of the 10th International Conference on Music Information Retrieval (ISMIR'09).Google Scholar
- Hu, X., Downie, J. S. and Ehmann, A. 2009. Lyric text mining in music mood classification, In Proceedings of the 10th International Conference on Music Information Retrieval (ISMIR'09).Google Scholar
- Laurier, C., Grivolla, J., and Herrera, P. 2008. Multimodal music mood classification using audio and lyrics. In Proceedings of the International Conference on Machine Learning and Applications. Google ScholarDigital Library
- Li, T. and Ogihara, M. 2004. Semi-supervised learning from different information sources. Knowledge and Information Systems, 7 (3): 289--309.Google ScholarCross Ref
- Liu, H., Lieberman, H., and Selker, T. 2003. A model of textual affect sensing using real-world knowledge. In Proceedings of the 8th International Conference on Intelligent User Interfaces, pp. 125--132. Google ScholarDigital Library
- Lu, L., Liu, D., and Zhang, H. 2006. Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio, Speech, and Language Processing, 14(1): 5--18. Google ScholarDigital Library
- Mayer, R., Neumayer, R., and Rauber, A. 2008. Combination of audio and lyrics features for genre classification in digital audio collections. In Proceeding of the 16th ACM International Conference on Multimedia. Google ScholarDigital Library
- Pang, B. and Lee, L. 2008. Opinion mining and sentiment analysis. Foundations and Trends in Information Retrieval, 2(1-2): 1--135. Google ScholarDigital Library
- Pohle, T., Pampalk, E., and Widmer, G. 2005. Evaluation of frequently used audio features for classification of music into perceptual categories. In Proceedings of the 4th International Workshop on Content-Based Multimedia Indexing.Google Scholar
- Russell, J. A. 1980. A circumplex model of affect, Journal of Personality and Social Psychology, 39(6): 1161--1178.Google ScholarCross Ref
- Stone, P. J. 1966. General Inquirer: a Computer Approach to Content Analysis. Cambridge: M.I.T. Press.Google Scholar
- Strapparava, C. and Valitutti, A. 2004. WordNet-Affect: an affective extension of WordNet. In Proceedings of the 4th International Conference on Language Resources and Evaluation (LREC'04) pp. 1083--1086.Google Scholar
- Subasic, P. and Huettner, A. 2001. Affect analysis of text using fuzzy semantic typing. IEEE Transactions on Fuzzy Systems, Special Issue, 9: 483--496. Google ScholarDigital Library
- Tax, D. M. J., van Breukelen, M., Duin, R. P. W., and Kittler, J. 2000. Combining multiple classifiers by averaging or by multiplying. Pattern Recognition, 33: 1475--1485Google ScholarCross Ref
- Trohidis, K., Tsoumakas, G., Kalliris, G., and Vlahavas, I. 2008. Multi-label classification of music into emotions. In Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR'08).Google Scholar
- Tzanetakis, G. 2007. Marsyas submissions to mirex 2007, avaible at http://www.music-ir.org/mirex/2007/abs/AI_CC_GC_MC_AS_tzanetakis.pdfGoogle Scholar
- Vignoli, F. 2004. Digital Music Interaction concepts: a user study. In Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR'04).Google Scholar
- Whitman, B. and Smaragdis, P. 2002. Combining musical and cultural features for intelligent style detection. In Proceedings of the 3rd International Conference on Music Information Retrieval (ISMIR'02)Google Scholar
- Yang, D. and Lee, W. 2004. Disambiguating music emotion using software agents. In Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR'04).Google Scholar
- Yang, Y.-H., Lin, Y.-C., Cheng, H.-T., Liao, I.-B., Ho, Y.-C., and Chen, H. H. 2008. Toward multi-modal music emotion classification. In Proceedings of Pacific Rim Conference on Multimedia (PCM'08). Google ScholarDigital Library
- Yu, B. 2008. An evaluation of text classification methods for literary study, Literary and Linguistic Computing, 23(3): 327--343Google ScholarCross Ref
Index Terms
- Improving mood classification in music digital libraries by combining lyrics and audio
Recommendations
Combination of audio and lyrics features for genre classification in digital audio collections
MM '08: Proceedings of the 16th ACM international conference on MultimediaIn many areas multimedia technology has made its way into mainstream. In the case of digital audio this is manifested in numerous online music stores having turned into profitable businesses. The widespread user adaption of digital audio both on home ...
Unsupervised Approach to Hindi Music Mood Classification
MIKE 2013: Proceedings of the First International Conference on Mining Intelligence and Knowledge Exploration - Volume 8284We often choose to listen to a song that suits our mood at that instant because an intimate relationship presents between music and human emotions. Thus, the automatic methods are needed to classify music by moods that have gained a lot of momentum in ...
Multimodal Music Mood Classification Using Audio and Lyrics
ICMLA '08: Proceedings of the 2008 Seventh International Conference on Machine Learning and ApplicationsIn this paper we present a study on music mood classification using audio and lyrics information. The mood of a song is expressed by means of musical features but a relevant part also seems to be conveyed by the lyrics. We evaluate each factor ...
Comments