TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SCNet: Sparse Compression Network for Music Source Separat...

SCNet: Sparse Compression Network for Music Source Separation

Weinan Tong, Jiaxu Zhu, Jun Chen, Shiyin Kang, Tao Jiang, Yang Li, Zhiyong Wu, Helen Meng

2024-01-24Music Source Separation
PaperPDFCodeCode(official)Code

Abstract

Deep learning-based methods have made significant achievements in music source separation. However, obtaining good results while maintaining a low model complexity remains challenging in super wide-band music source separation. Previous works either overlook the differences in subbands or inadequately address the problem of information loss when generating subband features. In this paper, we propose SCNet, a novel frequency-domain network to explicitly split the spectrogram of the mixture into several subbands and introduce a sparsity-based encoder to model different frequency bands. We use a higher compression ratio on subbands with less information to improve the information density and focus on modeling subbands with more information. In this way, the separation performance can be significantly improved using lower computational consumption. Experiment results show that the proposed model achieves a signal to distortion ratio (SDR) of 9.0 dB on the MUSDB18-HQ dataset without using extra data, which outperforms state-of-the-art methods. Specifically, SCNet's CPU inference time is only 48% of HT Demucs, one of the previous state-of-the-art models.

Results

TaskDatasetMetricValueModel
Music Source SeparationMUSDB18-HQSDR (avg)9.69SCNet-large
Music Source SeparationMUSDB18-HQSDR (bass)9.49SCNet-large
Music Source SeparationMUSDB18-HQSDR (drums)10.98SCNet-large
Music Source SeparationMUSDB18-HQSDR (others)7.44SCNet-large
Music Source SeparationMUSDB18-HQSDR (vocals)10.86SCNet-large
Music Source SeparationMUSDB18-HQSDR (avg)9SCNet
Music Source SeparationMUSDB18-HQSDR (bass)8.82SCNet
Music Source SeparationMUSDB18-HQSDR (drums)10.51SCNet
Music Source SeparationMUSDB18-HQSDR (others)6.76SCNet
Music Source SeparationMUSDB18-HQSDR (vocals)9.89SCNet
2D ClassificationMUSDB18-HQSDR (avg)9.69SCNet-large
2D ClassificationMUSDB18-HQSDR (bass)9.49SCNet-large
2D ClassificationMUSDB18-HQSDR (drums)10.98SCNet-large
2D ClassificationMUSDB18-HQSDR (others)7.44SCNet-large
2D ClassificationMUSDB18-HQSDR (vocals)10.86SCNet-large
2D ClassificationMUSDB18-HQSDR (avg)9SCNet
2D ClassificationMUSDB18-HQSDR (bass)8.82SCNet
2D ClassificationMUSDB18-HQSDR (drums)10.51SCNet
2D ClassificationMUSDB18-HQSDR (others)6.76SCNet
2D ClassificationMUSDB18-HQSDR (vocals)9.89SCNet

Related Papers

Music Source Restoration2025-05-27Training-Free Multi-Step Audio Source Separation2025-05-26Is MixIT Really Unsuitable for Correlated Sources? Exploring MixIT for Unsupervised Pre-training in Music Source Separation2025-05-12Solving Copyright Infringement on Short Video Platforms: Novel Datasets and an Audio Restoration Deep Learning Pipeline2025-04-30Score-informed Music Source Separation: Improving Synthetic-to-real Generalization in Classical Music2025-03-10Separate This, and All of these Things Around It: Music Source Separation via Hyperellipsoidal Queries2025-01-27Sanidha: A Studio Quality Multi-Modal Dataset for Carnatic Music2025-01-12MAJL: A Model-Agnostic Joint Learning Framework for Music Source Separation and Pitch Estimation2025-01-07