This is the official implementation of our paper "Boosting Self-Supervised Embeddings for Speech Enhancement"
- pytorch 1.10.2
- torchaudio 0.10.2
- pesq 0.0.3
- pystoi 0.3.3
- numpy 1.20.3
- tensorboardx 2.2
- tqdm 4.60.0
- scikit-learn 0.24.1
- pandas 1.2.4
- fairseq 0.11.0+f97cdf7
You can use pip to install Python depedencies.
pip install -r requirements.txt
The Voice Bank--Demand Dataset is not provided by this repository. Please download the dataset and build your own PyTorch dataloader from here.
For each .wav
file, you need to first convert it into 16kHz format by any audio converter (e.g., sox).
sox <48K.wav> -r 16000 -c 1 -b 16 <16k.wav>
Please download the model weights from here, and make a folder named save_model
then put the weight file under the folder.
Experiment Date | PESQ | CSIG | CBAK | COVL |
---|---|---|---|---|
2022-04-30 | 3.20 | 4.53 | 3.60 | 3.88 |
Please download the pretrained model first if you want to used ssl feature and put the weight under the save_model
folder (e.g, save_model/WavLM-Base+.pt
). The pretrain model can be downloaded by below link.
Model | Pre-training Dataset | Fine-tuning | Model |
---|---|---|---|
WavLM Base+ | 60k hrs Libri-Light + 10k hrs GigaSpeech + 24k hrs VoxPopuli | - | Azure Storage Google Drive |
WavLM Large | 60k hrs Libri-Light + 10k hrs GigaSpeech + 24k hrs VoxPopuli | - | Azure Storage Google Drive |
Wav2Vec 2.0 Base | Librispeech | - | download |
Wav2Vec 2.0 Large | Librispeech | - | download |
HuBERT Base (~95M params) | Librispeech 960 hr | - | download |
HuBERT Large (~316M params) | Libri-Light 60k hr | - | download |
Run the following command to train the speech enhancement model:
python main.py \
--data_folder <root/dir/of/dataset>
--model BLSTM
--ssl_model <wavlm/hubert/wav2vec2>
--feature <raw/ssl/cross>
--size <base/large>
--target IRM
--finetune_SSL <PF/EF/None>
--weighted_sum
add --mode test
in the command line and the rest remain the same to evaluate the speech enhancement model:
python main.py --mode test ...
Please cite the following paper if you find the codes useful in your research.
@article{hung2022boosting,
title={Boosting Self-Supervised Embeddings for Speech Enhancement},
author={Hung, Kuo-Hsuan and Fu, Szu-wei and Tseng, Huan-Hsin and Chiang, Hsin-Tien and Tsao, Yu and Lin, Chii-Wann},
journal={arXiv preprint arXiv:2204.03339},
year={2022}
}
Please cite the following paper if you use the following pretrained ssl model.
WavLM
@article{chen2021wavlm,
title={Wavlm: Large-scale self-supervised pre-training for full stack speech processing},
author={Chen, Sanyuan and Wang, Chengyi and Chen, Zhengyang and Wu, Yu and Liu, Shujie and Chen, Zhuo and Li, Jinyu and Kanda, Naoyuki and Yoshioka, Takuya and Xiao, Xiong and others},
journal={arXiv preprint arXiv:2110.13900},
year={2021}
}
Wav2vec 2.0
@article{baevski2020wav2vec,
title={wav2vec 2.0: A framework for self-supervised learning of speech representations},
author={Baevski, Alexei and Zhou, Yuhao and Mohamed, Abdelrahman and Auli, Michael},
journal={Advances in Neural Information Processing Systems},
volume={33},
pages={12449--12460},
year={2020}
}
HuBert
@article{hsu2021hubert,
title={Hubert: Self-supervised speech representation learning by masked prediction of hidden units},
author={Hsu, Wei-Ning and Bolte, Benjamin and Tsai, Yao-Hung Hubert and Lakhotia, Kushal and Salakhutdinov, Ruslan and Mohamed, Abdelrahman},
journal={IEEE/ACM Transactions on Audio, Speech, and Language Processing},
volume={29},
pages={3451--3460},
year={2021},
publisher={IEEE}
}
This project is licensed under the MIT License - see the LICENSE file for details
- Bio-ASP Lab, CITI, Academia Sinica, Taipei, Taiwan