When lyrics outperform audio for music mood classification: A feature analysis

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

This paper builds upon and extends previous work on multi-modal mood classification (i.e., combining audio and lyrics) by analyzing in-depth those feature types that have shown to provide statistically significant improvements in the classification of individual mood categories. The dataset used in this study comprises 5,296 songs (with lyrics and audio for each) divided into 18 mood categories derived from user-generated tags taken from last.fm. These 18 categories show remarkable consistency with the popular Russell's mood model. In seven categories, lyric features significantly outperformed audio spectral features. In one category only, audio outperformed all lyric features types. A fine grained analysis of the significant lyric feature types indicates a strong and obvious semantic association between extracted terms and the categories. No such obvious semantic linkages were evident in the case where audio spectral features proved superior.

Original languageEnglish (US)
Title of host publicationProceedings of the 11th International Society for Music Information Retrieval Conference, ISMIR 2010
Pages619-624
Number of pages6
StatePublished - 2010
Event11th International Society for Music Information Retrieval Conference, ISMIR 2010 - Utrecht, Netherlands
Duration: Aug 9 2010Aug 13 2010

Publication series

NameProceedings of the 11th International Society for Music Information Retrieval Conference, ISMIR 2010

Other

Other11th International Society for Music Information Retrieval Conference, ISMIR 2010
Country/TerritoryNetherlands
CityUtrecht
Period8/9/108/13/10

ASJC Scopus subject areas

  • Music
  • Information Systems

Fingerprint

Dive into the research topics of 'When lyrics outperform audio for music mood classification: A feature analysis'. Together they form a unique fingerprint.

Cite this