skip to main content
10.1145/1816123.1816146acmconferencesArticle/Chapter ViewAbstractPublication PagesjcdlConference Proceedingsconference-collections
research-article

Improving mood classification in music digital libraries by combining lyrics and audio

Authors Info & Claims
Published:21 June 2010Publication History

ABSTRACT

Mood is an emerging metadata type and access point in music digital libraries (MDL) and online music repositories. In this study, we present a comprehensive investigation of the usefulness of lyrics in music mood classification by evaluating and comparing a wide range of lyric text features including linguistic and text stylistic features. We then combine the best lyric features with features extracted from music audio using two fusion methods. The results show that combining lyrics and audio significantly outperformed systems using audio-only features. In addition, the examination of learning curves shows that the hybrid lyric + audio system needed fewer training samples to achieve the same or better classification accuracies than systems using lyrics or audio singularly. These experiments were conducted on a unique large-scale dataset of 5,296 songs (with both audio and lyrics for each) representing 18 mood categories derived from social tags. The findings push forward the state-of-the-art on lyric sentiment analysis and automatic music mood classification and will help make mood a practical access point in music digital libraries.

References

  1. Alm, C.O. 2009. Affect in Text and Speech. VDM Verlag: Saarbrücken.Google ScholarGoogle Scholar
  2. Argamon, S., Saric, M., and Stein, S. S. 2003. Style mining of electronic messages for multiple authorship discrimination: first results. In Proceedings of the 9th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. pp. 475--480. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Bischoff, K., Firan, C. S., Nejdl, W., and Paiu, R. 2009. How do you feel about "Dancing Queen"? Deriving mood and theme annotations from user tags. In Proceedings of Joint Conference on Digital Libraries (JCDL'09). Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Bischoff, K., Firan, C., Paiu, R., Nejdl, W., Laurier, C., and Sordo, M. 2009. Music mood and theme classification - a hybrid approach. In Proceedings of the 10th International Conference on Music Information Retrieval (ISMIR'09).Google ScholarGoogle Scholar
  5. Bradley, M. M. and Lang, P. J. 1999. Affective Norms for English Words (ANEW): Stimuli, Instruction Manual and Affective Ratings. Technical report C-1. University of Florida.Google ScholarGoogle Scholar
  6. Chang, C. and Lin. C. 2001. LIBSVM: a library for support vector machines. Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvmGoogle ScholarGoogle Scholar
  7. Downie, J. S. 2008. The Music Information Retrieval Evaluation Exchange (2005-2007): A window into music information retrieval research. Acoustical Science and Technology 29 (4): 247--255. Available at: http://dx.doi.org/10.1250/ast.29.247Google ScholarGoogle ScholarCross RefCross Ref
  8. Fellbaum, C. 1998. WordNet: An Electronic Lexical Database, MIT Press.Google ScholarGoogle Scholar
  9. He, H., Jin, J., Xiong, Y., Chen, B., Sun, W., and Zhao, L. 2008. Language feature mining for music emotion classification via supervised learning from lyrics. In Proceedings of Advances in the 3rd International Symposium on Computation and Intelligence (ISICA 2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Hu, X. 2010. Music and mood: where theory and reality meet. In Proceedings of iConference 2010.Google ScholarGoogle Scholar
  11. Hu, Y., Chen, X. and Yang, D. 2009. Lyric-based song emotion detection with affective lexicon and fuzzy clustering method. In Proceedings of the 10th International Conference on Music Information Retrieval (ISMIR'09).Google ScholarGoogle Scholar
  12. Hu, X., Downie, J. S. and Ehmann, A. 2009. Lyric text mining in music mood classification, In Proceedings of the 10th International Conference on Music Information Retrieval (ISMIR'09).Google ScholarGoogle Scholar
  13. Laurier, C., Grivolla, J., and Herrera, P. 2008. Multimodal music mood classification using audio and lyrics. In Proceedings of the International Conference on Machine Learning and Applications. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Li, T. and Ogihara, M. 2004. Semi-supervised learning from different information sources. Knowledge and Information Systems, 7 (3): 289--309.Google ScholarGoogle ScholarCross RefCross Ref
  15. Liu, H., Lieberman, H., and Selker, T. 2003. A model of textual affect sensing using real-world knowledge. In Proceedings of the 8th International Conference on Intelligent User Interfaces, pp. 125--132. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Lu, L., Liu, D., and Zhang, H. 2006. Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio, Speech, and Language Processing, 14(1): 5--18. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Mayer, R., Neumayer, R., and Rauber, A. 2008. Combination of audio and lyrics features for genre classification in digital audio collections. In Proceeding of the 16th ACM International Conference on Multimedia. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Pang, B. and Lee, L. 2008. Opinion mining and sentiment analysis. Foundations and Trends in Information Retrieval, 2(1-2): 1--135. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Pohle, T., Pampalk, E., and Widmer, G. 2005. Evaluation of frequently used audio features for classification of music into perceptual categories. In Proceedings of the 4th International Workshop on Content-Based Multimedia Indexing.Google ScholarGoogle Scholar
  20. Russell, J. A. 1980. A circumplex model of affect, Journal of Personality and Social Psychology, 39(6): 1161--1178.Google ScholarGoogle ScholarCross RefCross Ref
  21. Stone, P. J. 1966. General Inquirer: a Computer Approach to Content Analysis. Cambridge: M.I.T. Press.Google ScholarGoogle Scholar
  22. Strapparava, C. and Valitutti, A. 2004. WordNet-Affect: an affective extension of WordNet. In Proceedings of the 4th International Conference on Language Resources and Evaluation (LREC'04) pp. 1083--1086.Google ScholarGoogle Scholar
  23. Subasic, P. and Huettner, A. 2001. Affect analysis of text using fuzzy semantic typing. IEEE Transactions on Fuzzy Systems, Special Issue, 9: 483--496. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Tax, D. M. J., van Breukelen, M., Duin, R. P. W., and Kittler, J. 2000. Combining multiple classifiers by averaging or by multiplying. Pattern Recognition, 33: 1475--1485Google ScholarGoogle ScholarCross RefCross Ref
  25. Trohidis, K., Tsoumakas, G., Kalliris, G., and Vlahavas, I. 2008. Multi-label classification of music into emotions. In Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR'08).Google ScholarGoogle Scholar
  26. Tzanetakis, G. 2007. Marsyas submissions to mirex 2007, avaible at http://www.music-ir.org/mirex/2007/abs/AI_CC_GC_MC_AS_tzanetakis.pdfGoogle ScholarGoogle Scholar
  27. Vignoli, F. 2004. Digital Music Interaction concepts: a user study. In Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR'04).Google ScholarGoogle Scholar
  28. Whitman, B. and Smaragdis, P. 2002. Combining musical and cultural features for intelligent style detection. In Proceedings of the 3rd International Conference on Music Information Retrieval (ISMIR'02)Google ScholarGoogle Scholar
  29. Yang, D. and Lee, W. 2004. Disambiguating music emotion using software agents. In Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR'04).Google ScholarGoogle Scholar
  30. Yang, Y.-H., Lin, Y.-C., Cheng, H.-T., Liao, I.-B., Ho, Y.-C., and Chen, H. H. 2008. Toward multi-modal music emotion classification. In Proceedings of Pacific Rim Conference on Multimedia (PCM'08). Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Yu, B. 2008. An evaluation of text classification methods for literary study, Literary and Linguistic Computing, 23(3): 327--343Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Improving mood classification in music digital libraries by combining lyrics and audio

              Recommendations

              Comments

              Login options

              Check if you have access through your login credentials or your institution to get full access on this article.

              Sign in
              • Published in

                cover image ACM Conferences
                JCDL '10: Proceedings of the 10th annual joint conference on Digital libraries
                June 2010
                424 pages
                ISBN:9781450300858
                DOI:10.1145/1816123

                Copyright © 2010 ACM

                Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                Publisher

                Association for Computing Machinery

                New York, NY, United States

                Publication History

                • Published: 21 June 2010

                Permissions

                Request permissions about this article.

                Request Permissions

                Check for updates

                Qualifiers

                • research-article

                Acceptance Rates

                Overall Acceptance Rate415of1,482submissions,28%

              PDF Format

              View or Download as a PDF file.

              PDF

              eReader

              View online with eReader.

              eReader