kan-bayashi / PytorchWaveNetVocoder

WaveNet-Vocoder implementation with pytorch.

Home Page:https://kan-bayashi.github.io/WaveNetVocoderSamples/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Update information

kan-bayashi opened this issue · comments

2018/05/01

  • Updated to be compatible with pytorch v0.4
  • Updated to be able to use melspectrogram as auxiliary feature

Due to above update, some parts are changed (see below)

# -------------------- #
# feature path in hdf5 #
# -------------------- #
old -> new
/feat_org -> /world or /melspc
/feat -> no more saving extended featrue (it is replicated when loading)

# ----------------------- #
# statistics path in hdf5 #
# ----------------------- #
old -> new
/mean -> /world/mean or /melspc/mean
/scale -> /world/scale or /melspc/scale

# ----------------------- #
# new options in training #
# ----------------------- #
--feature_type: Auxiliary feature type (world or melspc)
--use_upsampling_layer: Flag to decide whether to use upsampling layer in WaveNet
--upsampling_factor: Changed to be alway needed because feature extension is performed in loading

Note that old model file checkpoint-*.pkl can be used, but it is necessary to modify model.conf file as follows.

# how-to-convert to new config file
import torch
args = torch.load("old_model.conf")
args.use_upsampling_layer = True
args.feature_type = "world"
torch.save(args, "new_model.conf")

2018/05/07

  • Added samples using mel-spectrogram

2018/05/11

  • Implemented noise-shaping stft-based mel-cepstrum
  • All of the recipes can use noise shaping technique
  • Added sample of mel-spectrogram with noise shaping

2018/06/21