Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to get a generated speech from the output of the trained Generator? #55

Open
6lyx opened this issue Oct 3, 2021 · 3 comments
Open

Comments

@6lyx
Copy link

6lyx commented Oct 3, 2021

I have trained the Generator model with my own data. However, I found that there may not exist a code for generating the speech from the trained Generator. And I check the code named "demo.ipynb" for founding out the way. It indicates that a trained F0_Converter is needed.
So I would like to ask the author that dose it nessusary to train a F0_Converter first for generating the speech from the trained Generator?(Because I found no code for training F0_Converter)? Or we just need to use the pretrained F0_Converter?

@auspicious3000
Copy link
Owner

If your data is very different from vctk, you probably need to re-train the F0-converter

@6lyx
Copy link
Author

6lyx commented Oct 3, 2021

Many many thanks for your quick answering. I am now using the speech with the sampling rate of 44100hz, does it mean that it is nesscuary to retrain the F0_Converter and the wavegen model? I have found that the speech I generated is much shorter than the original speech....... (Using the trained G model and the pretrained wavegen model obtained in this project).

@auspicious3000
Copy link
Owner

auspicious3000 commented Oct 3, 2021

Yes. In that case, you probably need to tweak other parts of the model as well.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants