[P] A PyTorch implementation of Robust Universal Neural Vocoding.
Just wanted to share my PyTorch implementation of Amazon’s paper Robust Universal Neural Vocoding
Pretrained Models: https://github.com/bshall/UniversalVocoding/releases/tag/v0.1
I found that the model trains relatively quickly with intelligible audio after only 20k steps and decent results after 100k steps. Also seems to work well on out of domain speakers (as advertised in the paper).