Improving mood classification in music digital libraries by combining lyrics and audio

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Mood is an emerging metadata type and access point in music digital libraries (MDL) and online music repositories. In this study, we present a comprehensive investigation of the usefulness of lyrics in music mood classification by evaluating and comparing a wide range of lyric text features including linguistic and text stylistic features. We then combine the best lyric features with features extracted from music audio using two fusion methods. The results show that combining lyrics and audio significantly outperformed systems using audio-only features. In addition, the examination of learning curves shows that the hybrid lyric + audio system needed fewer training samples to achieve the same or better classification accuracies than systems using lyrics or audio singularly. These experiments were conducted on a unique large-scale dataset of 5,296 songs (with both audio and lyrics for each) representing 18 mood categories derived from social tags. The findings push forward the state-of-the-art on lyric sentiment analysis and automatic music mood classification and will help make mood a practical access point in music digital libraries.

Original languageEnglish (US)
Title of host publicationJCDL'10 - Digital Libraries - 10 Years Past, 10 Years Forward, a 2020 Vision
Pages159-168
Number of pages10
DOIs
StatePublished - Aug 5 2010
Event10th Annual Joint Conference on Digital Libraries, JCDL 2010 - Gold Coast, QLD, Australia
Duration: Jun 21 2010Jun 25 2010

Publication series

NameProceedings of the ACM International Conference on Digital Libraries

Other

Other10th Annual Joint Conference on Digital Libraries, JCDL 2010
CountryAustralia
CityGold Coast, QLD
Period6/21/106/25/10

Keywords

  • Experimentation
  • Measurement
  • Performance

ASJC Scopus subject areas

  • Software
  • Information Systems
  • Computer Science Applications
  • Library and Information Sciences

Fingerprint Dive into the research topics of 'Improving mood classification in music digital libraries by combining lyrics and audio'. Together they form a unique fingerprint.

  • Cite this

    Hu, X., & Downie, J. S. (2010). Improving mood classification in music digital libraries by combining lyrics and audio. In JCDL'10 - Digital Libraries - 10 Years Past, 10 Years Forward, a 2020 Vision (pp. 159-168). (Proceedings of the ACM International Conference on Digital Libraries). https://doi.org/10.1145/1816123.1816146