You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Could you please tell us when you are planning to release pre-trained model?
Is it possible for you to provide us some kind of loss graph or just the number of training steps necessary for each module to converge on LibriTTS+VCTK dataset? So we could estimate whether it is possible for mere mortals to train the model without multiple advanced GPUs...
Could you elaborate on audio normalization mentioned in you paper? Is it implemented somewhere in your project or should we process audio files by some other means?
Thank you!
The text was updated successfully, but these errors were encountered:
Hi! We just released pre-trained weights. Please check out!
Unfortunately, our cotatron loss graph has been deleted, so it is difficult to upload.
However, we can provide a loss graph of the synthesizer(VC decoder). Please see the graph below.
FYI, we trained cotatron 25k steps only with LibriTTS, and 20k steps with LibriTTS and VCTK.
The scale of cotatron validation's reconstruction loss was about 0.28.
Could you please tell us when you are planning to release pre-trained model?
Is it possible for you to provide us some kind of loss graph or just the number of training steps necessary for each module to converge on LibriTTS+VCTK dataset? So we could estimate whether it is possible for mere mortals to train the model without multiple advanced GPUs...
Could you elaborate on audio normalization mentioned in you paper? Is it implemented somewhere in your project or should we process audio files by some other means?
Thank you!
The text was updated successfully, but these errors were encountered: