You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,14 +3,17 @@
3
3
4
4
DiffWave is a fast, high-quality neural vocoder and waveform synthesizer. It starts with Gaussian noise and converts it into speech via iterative refinement. The speech can be controlled by providing a conditioning signal (e.g. log-scaled Mel spectrogram). The model and architecture details are described in [DiffWave: A Versatile Diffusion Model for Audio Synthesis](https://arxiv.org/pdf/2009.09761.pdf).
5
5
6
+
## What's new (2021-11-09)
7
+
- unconditional waveform synthesis (thanks to [Andrechang](https://github.com/Andrechang)!)
8
+
6
9
## What's new (2021-04-01)
7
10
- fast sampling algorithm based on v3 of the DiffWave paper
8
11
9
12
## What's new (2020-10-14)
10
13
- new pretrained model trained for 1M steps
11
14
- updated audio samples with output from new model
12
15
13
-
## Status (2021-04-01)
16
+
## Status (2021-11-09)
14
17
-[x] fast inference procedure
15
18
-[x] stable training
16
19
-[x] high-quality synthesis
@@ -21,7 +24,7 @@ DiffWave is a fast, high-quality neural vocoder and waveform synthesizer. It sta
21
24
-[x] PyPI package
22
25
-[x] audio samples
23
26
-[x] pretrained models
24
-
-[] unconditional waveform synthesis
27
+
-[x] unconditional waveform synthesis
25
28
26
29
Big thanks to [Zhifeng Kong](https://github.com/FengNiMa) (lead author of DiffWave) for pointers and bug fixes.
0 commit comments