This webpage is to show some listening examples for our Sams-Net method and the state-of-the-art methods Open-Unmix[1] and Demucs[2]. It enables different contributions of feature interactions, achieves easier parallel computing and has a larger receptive field compared with LSTMs and CNNs respectively. Experiments show that the proposed framework outperforms most of the state-of-the-art DNN-based methods with fewer parameters.
Track | A-R | C-Y | S-J | T-S | J-T | T-E |
---|---|---|---|---|---|---|
Mixture Input | ||||||
SamsNet:Vocals | ||||||
SamsNet:Drums | ||||||
SamsNet:Bass | ||||||
SamsNet:Other | ||||||
OpenUnmix:Vocals | ||||||
OpenUnmix:Drums | ||||||
OpenUnmix:Bass | ||||||
OpenUnmix:Other | ||||||
Demucs:Vocals | ||||||
Demucs:Drums | ||||||
Demucs:Bass | ||||||
Demucs:Other |
[1] F.-R. Stoter, S. Uhlich, A. Liutkus, and Y. Mitsufuji, "Open-unmix-a reference implementation for music source separation," Journal of Open Source Software, 2019.
[2] A. Defossez, N. Usunier, L. Bottou, and F. Bach, "Music Source Separation in the Waveform Domain," HAL, Tech. Rep. 02379796v1, 2019.