Weakly informed audio source separation

Kilian Schulze-Forster, Clement Doire, Gael Richard, Roland Badeau

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Prior information about the target source can improve audio source separation quality but is usually not available with the necessary level of audio alignment. This has limited its usability in the past. We propose a separation model that can nevertheless exploit such weak information for the separation task while aligning it on the mixture as a byproduct using an attention mechanism. We demonstrate the capabilities of the model on a singing voice separation task exploiting artificial side information with different levels of expressiveness. Moreover, we highlight an issue with the common separation quality assessment procedure regarding parts where targets or predictions are silent and refine a previous contribution for a more complete evaluation.

Original languageEnglish
Title of host publication2019 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages273-277
Number of pages5
ISBN (Electronic)9781728111230
DOIs
Publication statusPublished - 1 Oct 2019
Externally publishedYes
Event2019 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2019 - New Paltz, United States
Duration: 20 Oct 201923 Oct 2019

Publication series

NameIEEE Workshop on Applications of Signal Processing to Audio and Acoustics
Volume2019-October
ISSN (Print)1931-1168
ISSN (Electronic)1947-1629

Conference

Conference2019 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2019
Country/TerritoryUnited States
CityNew Paltz
Period20/10/1923/10/19

Keywords

  • attention
  • informed source separation
  • separation evaluation
  • singing voice separation
  • weak labels

Fingerprint

Dive into the research topics of 'Weakly informed audio source separation'. Together they form a unique fingerprint.

Cite this