Skip to content
forked from pheepa/DCUnet

Phase-aware speech enchancement with Deep Complex U-Net

Notifications You must be signed in to change notification settings

madhavmk/DCUnet

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Phase-aware speech enchancement with Deep Complex U-Net

For research I chose this article "Phase-aware speech enchancement with Deep Complex U-Net", which describes the architecture and training of a convolutional neural network for improving speech, so-called denoising, and set up an experiment. architecture

The task

The main task is to develop a complex variation of the architecture of the well known UNet network to eliminate unwanted noise from the audio.

Method's features

Its speciality and difference from other networks, such as SegNet, for semantic segmentation (not only that) lies in Skip-Connections and that the values of both input data and all of network parameters (convolution filters, etc.) are complex.

Skip-Connections

The main idea is that the early layers of the Encoder are concatenated with the " parallel " layers of the Decoder.

skip-connection

Mask

As a result of the convolution layers, we get a mask, which we multiply by the input time-frequency signal with noise and get a cleared time-frequency signal, which then passes the inverse Short-time Fourier transform. arch

Alternative solution

The Experiment

For training we will use Noisy speech database for training speech enhancement algorithms and TTS models, which contains a data set for training and testing with 28 and 56 speakers in .wav audio files are 48 KHz. The 10-layer network architecture will be implemented, which looks like this: 10-layers

A graph of changes in the value of the loss function during training and validation will be shown.

The PESQ metric will also be calculated.

Issues

Due to my lack of equipment with proper GPU (a laptop with 2 GB of GPU, so the model does not fit into the given memory, not to mention training) I had to consider alternatives for training:

  • Training on Google Colab or another cloud service. Cloud services have strict session time limits, so it was decided to train on a small number of epochs.

About

Phase-aware speech enchancement with Deep Complex U-Net

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 100.0%