Abstract
Recent advancements in music source separation have significantly progressed, particularly in isolating vocals, drums, and bass elements from mixed tracks. These developments owe much to the creation and use of large-scale, multitrack datasets dedicated to these specific components. However, the challenge of extracting similarly sounding sources from orchestra recordings has not been extensively explored, largely due to a scarcity of comprehensive and clean (i.e bleed-free) multitrack datasets. In this paper, we introduce a novel multitrack dataset called SynthSOD, developed using a set of simulation techniques to create a realistic, musically motivated, and heterogeneous training set comprising different dynamics, natural tempo changes, styles, and conditions by employing high-quality digital libraries that define virtual instrument sounds for MIDI playback (a.k.a., soundfonts). Moreover, we demonstrate the application of a widely used baseline music separation model trained on our synthesized dataset w.r.t to the well-known EnsembleSet, and evaluate its performance under both synthetic and real-world conditions.
Original language | English |
---|---|
Journal | IEEE Open Journal of Signal Processing |
DOIs | |
Publication status | E-pub ahead of print - 2025 |
Publication type | A1 Journal article-refereed |
Keywords
- Classical music
- dataset
- deep learning
- machine learning
- music source separation
- orchestra music
Publication forum classification
- Publication forum level 1
ASJC Scopus subject areas
- Signal Processing