Synthia's Melody: A Benchmark Framework for Unsupervised Domain Adaptation in Audio

Chia Hsin Lin, Charles Jones, Bjorn W. Schuller, Harry Coppock, Alican Akman

Research output: Contribution to journalConference articlepeer-review

Abstract

Despite significant advancements in deep learning for vision and natural language, unsupervised domain adaptation in audio remains relatively unexplored. We, in part, attribute this to the lack of an appropriate benchmark dataset. To address this gap, we present Synthia's melody, a novel audio data generation framework capable of simulating an infinite variety of 4-second melodies with user-specified confounding structures characterised by musical keys, timbre, and loudness. Unlike existing datasets collected under observational settings, Synthia's melody is free of unobserved biases, ensuring the reproducibility and comparability of experiments. To showcase its utility, we generate two types of distribution shifts - domain shift and sample selection bias - and evaluate the performance of acoustic deep learning models under these shifts. Our evaluations reveal that Synthia's melody provides a robust testbed for examining the susceptibility of these models to varying levels of distribution shift.

Original languageEnglish
Pages (from-to)7450-7454
Number of pages5
JournalICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
DOIs
StatePublished - 2024
Externally publishedYes
Event2024 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 - Seoul, Korea, Republic of
Duration: 14 Apr 202419 Apr 2024

Fingerprint

Dive into the research topics of 'Synthia's Melody: A Benchmark Framework for Unsupervised Domain Adaptation in Audio'. Together they form a unique fingerprint.

Cite this