Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to reproduce the result on VCTK dataset? #71

Open
liangshuang1993 opened this issue Dec 21, 2020 · 1 comment
Open

How to reproduce the result on VCTK dataset? #71

liangshuang1993 opened this issue Dec 21, 2020 · 1 comment

Comments

@liangshuang1993
Copy link

I run make_spect.py and make_metadata.py to prepreocess the dataset (I used all speakers in VCTK). And then I used pretrained model of Speaker Encoder to extract speaker embedding and train the model. The final loss is about 0.03. Are there anyone reproduce the result successfully? Could you help me? Thanks!

@ghost
Copy link

ghost commented Jan 7, 2021

After days of working on this project, I tried reproducing the results but all I get is silence, no voice.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant