Skip to content

Pre-trained model #17

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
jJohnny342 opened this issue Jul 13, 2021 · 1 comment
Closed

Pre-trained model #17

jJohnny342 opened this issue Jul 13, 2021 · 1 comment

Comments

@jJohnny342
Copy link

Could you please tell us when you are planning to release pre-trained model?
Is it possible for you to provide us some kind of loss graph or just the number of training steps necessary for each module to converge on LibriTTS+VCTK dataset? So we could estimate whether it is possible for mere mortals to train the model without multiple advanced GPUs...
Could you elaborate on audio normalization mentioned in you paper? Is it implemented somewhere in your project or should we process audio files by some other means?
Thank you!

@wookladin
Copy link
Contributor

Hi! We just released pre-trained weights. Please check out!

Unfortunately, our cotatron loss graph has been deleted, so it is difficult to upload.
However, we can provide a loss graph of the synthesizer(VC decoder). Please see the graph below.
image

FYI, we trained cotatron 25k steps only with LibriTTS, and 20k steps with LibriTTS and VCTK.
The scale of cotatron validation's reconstruction loss was about 0.28.

Lastly, we implemented audio normalization in our project. You can use normalization by changing the norm option in https://github.com/mindslab-ai/assem-vc/blob/master/datasets/text_mel_dataset.py#L17 to True.
To use audio normalization in the training process, make sure that the norm option of cotatron.py and synthesizer.py must also be set to True!
https://github.com/mindslab-ai/assem-vc/blob/master/cotatron.py#L136

Since we observed that normalization does not affect the output result, we set the norm option to False in this implementation.
Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants