Beronx86 / tensorflow-wavenet

A TensorFlow implementation of DeepMind's WaveNet paper

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

A TensorFlow implementation of DeepMind's WaveNet paper

This is a TensorFlow implementation of the WaveNet generative neural network architecture for audio generation.

The WaveNet architecture directly generates a raw audio waveform, and shows excellent results in TTS and general audio generation (see the DeepMind blog post and paper for examples).

The network is a model of the conditional probability to generate the next sample in the audio waveform, given all previous samples and possibly additional parameters. It is constructed from a stack of causal dilated layers, each of which is a dilated convolution (convolution with holes), which only accesses the current and past audio samples.

The network is implemented in the file wavenet.py.

This implementation is currently still under 🚧

TODO:

  • Add step-by-step generation of audio samples (and make them available in TensorBoard)
  • Add conditioning on extra parameters (tagged speech, speaker, etc.)
  • Generate example outputs

Training the network

The VCTK corpus is currently used. In order to train the network, you need to download the corpus and unpack it in the same directory as the main.py script.

Then, execute

python main.py

to train the network.

You can see documentation on the settings by by running

python main.py --help

Disclaimer: This repository is not affiliated with DeepMind or Google in any way.

About

A TensorFlow implementation of DeepMind's WaveNet paper

License:MIT License


Languages

Language:Python 100.0%