Lyrics segmentation via bimodal text-audio representation

  • Michael Fell
  • , Yaroslav Nechaev
  • , Gabriel Meseguer-Brocal
  • , Elena Cabrio
  • , Fabien Gandon
  • , Geoffroy Peeters

Research output: Contribution to journalArticlepeer-review

Abstract

Song lyrics contain repeated patterns that have been proven to facilitate automated lyrics segmentation, with the final goal of detecting the building blocks (e.g., chorus, verse) of a song text. Our contribution in this article is twofold. First, we introduce a convolutional neural network (CNN)-based model that learns to segment the lyrics based on their repetitive text structure. We experiment with novel features to reveal different kinds of repetitions in the lyrics, for instance based on phonetical and syntactical properties. Second, using a novel corpus where the song text is synchronized to the audio of the song, we show that the text and audio modalities capture complementary structure of the lyrics and that combining both is beneficial for lyrics segmentation performance. For the purely text-based lyrics segmentation on a dataset of 103k lyrics, we achieve an F-score of 67.4%, improving on the state of the art (59.2% F-score). On the synchronized text-audio dataset of 4.8k songs, we show that the additional audio features improve segmentation performance to 75.3% F-score, significantly outperforming the purely text-based approaches.

Original languageEnglish
Pages (from-to)317-336
Number of pages20
JournalNatural Language Engineering
Volume28
Issue number3
DOIs
Publication statusPublished - 5 May 2022

Keywords

  • Artificial Intelligence
  • Music Information Retrieval
  • Natural Language Processing
  • Natural Language in Multimodal and Multimedia Systems
  • Text Segmentation

Fingerprint

Dive into the research topics of 'Lyrics segmentation via bimodal text-audio representation'. Together they form a unique fingerprint.

Cite this