denoising / README.md
BorisovMaksim's picture
Update README.md
7b4befa
---
title: Denoising
emoji: 🤗
colorFrom: red
colorTo: orange
sdk: gradio
sdk_version: 3.28.1
app_file: app.py
pinned: false
---
This is a repo that implements web interface for DEMUCS model proposed in [Real Time Speech Enhancement in the Waveform Domain](https://arxiv.org/abs/2006.12847).
The model was trained from scratch in Pytorch. The proposed model is based on an encoder-decoder architecture with skip-connections. It is optimized on both time and frequency domains, using multiple loss functions.
You can record your voice in noisy conditions and get denoised version using DEMUCS model. There is also Spectral Gating denoiser as baseline.
<center><img src="https://th.bing.com/th/id/OIP.JuP5GDiGXTGkVR-x7Ov0bgHaF8?w=233&h=186&c=7&r=0&o=5&dpr=2&pid=1.7" >
</center>
# Running
Without docker:
<pre><code>pip install -r requirements.txt
python app.py</code></pre>
Using docker:
<pre><code>docker build . --tag python-docker
docker run -p 7860:7860 -e GRADIO_SERVER_NAME=0.0.0.0 -it python-docker:latest</code></pre>
# Data
In the scope of this project [Valentini](https://datashare.ed.ac.uk/handle/10283/2791) dataset in used. It is clean and noisy parallel speech database. The database was designed to train and test speech enhancement methods that operate at 48kHz. There are 56 speakers and ~10 gb of speech data.
For model improvement it is possible to use a bigger training set from [DNS](https://www.bing.com/search?q=dns+challenge&cvid=3773a401b19d40269d725a02faf6f79c&aqs=edge.0.69i59j69i57j0l6j69i60.1021j0j4&FORM=ANAB01&PC=U531) challenge.
# Training
The training process in impemented in Pytorch. The data is (noisy speech, clean speech) pairs that are loaded as 2 second samples, randomly cutted from audio and padded if necessary. Model is optimized using SGD. In terms of loss functions, the L1 loss and MultiResolutionSTFTLoss are used. MultiResolutionSTFTLoss is the sum of STFT loss over different window sizes, hop sizes and fft sizes.
$$L_{STFT}= L_{sc} + L_{mag}$$
$$L_{sc}= \frac{|| |STFT(\tilde{x})| - |STFT(x)| ||_{F}^{1}}{|STFT(x)|}$$
$$L_{mag} = \frac{1}{T}|| log|STFT(\tilde{x})| - log|STFT(x)| ||_{F}^{1}$$
where T is the time points in the waveform.
# Metrics
- Perceptual Evaluation of Speech Quality ([PESQ](https://torchmetrics.readthedocs.io/en/stable/audio/perceptual_evaluation_speech_quality.html))
- Short-Time Objective Intelligibility ([STOI](https://torchmetrics.readthedocs.io/en/stable/audio/short_time_objective_intelligibility.html))
The PESQ metric is used for estimating overall speech quality after denoising and STOI is used for estimating speech intelligibility after denoising.
Intelligibility measure is highly correlated with the intelligibility of degraded speech signals
# Experiments
For tracking experiments local server of [Weights & Biases](https://wandb.ai/site) is used. To manage configs for different experiments [hydra](https://hydra.cc/) is used. It allows an easy way to track configs and override paramaters.
| Experiment | Description | Result |
|--------------|:-----:|--------------------------------------------------------|
| Baseline | Initial experiment with L1 loss | Poor quality |
| Baseline_L1_Multi_STFT_loss | Changed loss to Multi STFT + L1 loss | Better performance |
|L1_Multi_STFT_no_resample | Tried to train without resampling | No impovement, probably because RELU on the last layer |
|Updated_DEMUCS | Used relu in the last layer. Removed it.| Significant improvement |
|wav_normalization | Tried to normalized wav by std during training| Small improvement |
| original_sr| Train with original sample rate | Significant improvement |
|increased_L | Increased number of encoder-decoder pairs from 3 to 5| Performance comparable with original_sr |
| double_sr| Train with double sample rate| Small improvement |
|replicate paper | Lower learning rate and fix bug in dataloader | Massive improvement! |
![img.png](images/plot.png)