[P] A PyTorch implementation of Robust Universal Neural Vocoding.
Just wanted to share my PyTorch implementation of Amazon’s paper Robust Universal Neural Vocoding
Repo: https://github.com/bshall/UniversalVocoding
Samples: https://bshall.github.io/UniversalVocoding/
Pretrained Models: https://github.com/bshall/UniversalVocoding/releases/tag/v0.1
I found that the model trains relatively quickly with intelligible audio after only 20k steps and decent results after 100k steps. Also seems to work well on out of domain speakers (as advertised in the paper).
submitted by /u/b-shall
[link] [comments]