Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How long is the training data suitable? The more, the better? #137

Open
ZLZ-Nancy opened this issue Jun 14, 2019 · 1 comment
Open

How long is the training data suitable? The more, the better? #137

ZLZ-Nancy opened this issue Jun 14, 2019 · 1 comment

Comments

@ZLZ-Nancy
Copy link

I have used 170 corpus training, The effect is still good. Is it necessary to add sample data to get a bigger effect?

@unilight
Copy link

In theory, a bigger training dataset leads to a better performance (generalization ability), but I think for GMM-VC about 50-80 utterances should be enough, i.e. the improvement can be marginal if you add in more data after you dataset reach this size.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants