Skip to content

msaddler/auditory-model-denoising

Repository files navigation

Speech Denoising with Auditory Models (arXiv, audio examples)

This is a TensorFlow implementation of our Speech Denoising with Auditory Models.

Contact: Mark Saddler or Andrew Francl

Citation

If you use our code for research, please cite our paper: Mark R. Saddler*, Andrew Francl*, Jenelle Feather, Kaizhi Qian, Yang Zhang, Josh H. McDermott (2021). Speech Denoising with Auditory Models. Proc. Interspeech 2021, 2681-2685. arXiv:2011.10706.

License

The source code is published under the MIT license. See LICENSE for details. In general, you can use the code for any purpose with proper attribution. If you do something interesting with the code, we'll be happy to know. Feel free to contact us.

Requirements

In order to speed setup and aid reproducibility we provide a Singularity container. This container holds all the libraries and dependencies needed to run the code and allows you to work in the same environment as was originally used. Please see the Singularity Documentation for more details. Download Singularity image: tensorflow-1.13.0-denoising.simg.

Trained Models

We provide model checkpoints for all of our trained audio transforms and deep feature recognition networks. Users must download the audio transform checkpoints to evaluate our denoising algorithms on their own audio. Both sets of checkpoints must be downloaded to run our DEMO Jupyter notebook. Download the entire auditory-model-denoising/models directory here:

Quick Start

We provide a Jupyter notebook that (1) demos how to run our trained denoising models and (2) provides examples of how to compute the auditory model losses used to train the models. A second notebook demos how to train a new audio transform using the auditory model losses.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published