Skip to content

Latest commit

 

History

History
 
 

sisnr-estimation

The Blind SI-SNR estimation recipe

  • The goal of this recipe is to train a neural network to be able to estimate the scale-invariant signal-to-noise ratio (SI-SNR) from the separated signals.

  • This model is developed to estimate source separation performance on the REAL-M dataset which consists of real life mixtures.

  • The REAL-M dataset can downloaded from this link.

  • The paper for the REAL-M dataset can be found on this arxiv link.

  • The model is trained with the LibriMix and WHAMR! datasets. You can download LibriMix by following the instructions here. Instructions on WHAMR! can be found here

How to Run

  • To train with dynamic mixing:
python train.py hparams/pool_sisnrestimator.yaml --data_folder /yourLibri2Mixpath --base_folder_dm /yourLibriSpeechpath --rir_path /yourpathforwhamrRIRs --dynamic_mixing True --use_whamr_train True --whamr_data_folder /yourpath/whamr --base_folder_dm_whamr /yourpath/wsj0-processed/si_tr_s

How to run on test sets only

If you want to run it on the test sets only, you can add the flag --test_only to the following command:

python train.py hparams/pool_sisnrestimator.yaml --data_folder /yourLibri2Mixpath --base_folder_dm /yourLibriSpeechpath --rir_path /yourpathforwhamrRIRs --dynamic_mixing True --use_whamr_train True --whamr_data_folder /yourpath/whamr --base_folder_dm_whamr /yourpath/wsj0-processed/si_tr_s --test_only

Results

Release hyperparams file L1-Error (DB) HuggingFace link Full model link GPUs
18-10-21 pool_sisnrestimator.yaml 1.71 HuggingFace Not Available RTX8000 48GB

You can find the output folders with the training logs here. This repository provides the Separator models needed to train a blind SI-SNR estimator.

Training Time

It takes about 5 hours for each epoch on a RTX8000 (48GB).

PreTrained Model + Easy-Inference

You can find the pre-trained model with an easy-inference function on HuggingFace:

About SpeechBrain

Citing REAL-M

Please, cite our paper for the REAL-M dataset, if you use it for your research or business.

@misc{subakan2021realm,
  title={REAL-M: Towards Speech Separation on Real Mixtures},
  author={Cem Subakan and Mirco Ravanelli and Samuele Cornell and François Grondin},
  year={2021},
  eprint={2110.10812},
  archivePrefix={arXiv},
  primaryClass={eess.AS}
}

Citing SpeechBrain

Please, cite SpeechBrain if you use it for your research or business.

@misc{speechbrainV1,
  title={Open-Source Conversational AI with SpeechBrain 1.0},
  author={Mirco Ravanelli and Titouan Parcollet and Adel Moumen and Sylvain de Langen and Cem Subakan and Peter Plantinga and Yingzhi Wang and Pooneh Mousavi and Luca Della Libera and Artem Ploujnikov and Francesco Paissan and Davide Borra and Salah Zaiem and Zeyu Zhao and Shucong Zhang and Georgios Karakasidis and Sung-Lin Yeh and Pierre Champion and Aku Rouhe and Rudolf Braun and Florian Mai and Juan Zuluaga-Gomez and Seyed Mahed Mousavi and Andreas Nautsch and Xuechen Liu and Sangeet Sagar and Jarod Duret and Salima Mdhaffar and Gaelle Laperriere and Mickael Rouvier and Renato De Mori and Yannick Esteve},
  year={2024},
  eprint={2407.00463},
  archivePrefix={arXiv},
  primaryClass={cs.LG},
  url={https://arxiv.org/abs/2407.00463},
}
@misc{speechbrain,
  title={{SpeechBrain}: A General-Purpose Speech Toolkit},
  author={Mirco Ravanelli and Titouan Parcollet and Peter Plantinga and Aku Rouhe and Samuele Cornell and Loren Lugosch and Cem Subakan and Nauman Dawalatabad and Abdelwahab Heba and Jianyuan Zhong and Ju-Chieh Chou and Sung-Lin Yeh and Szu-Wei Fu and Chien-Feng Liao and Elena Rastorgueva and François Grondin and William Aris and Hwidong Na and Yan Gao and Renato De Mori and Yoshua Bengio},
  year={2021},
  eprint={2106.04624},
  archivePrefix={arXiv},
  primaryClass={eess.AS},
  note={arXiv:2106.04624}
}