Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to test on my own data? #108

Open
Ha0Tang opened this issue Jan 16, 2022 · 7 comments
Open

How to test on my own data? #108

Ha0Tang opened this issue Jan 16, 2022 · 7 comments

Comments

@Ha0Tang
Copy link

Ha0Tang commented Jan 16, 2022

How to test on my own data? I have a "Source Speaker / Speech" and a "Target Speaker / Speech", I want to generate the "Conversion", as shown on the demo page https://auspicious3000.github.io/autovc-demo/. Can anyone provide some instructions?

@auspicious3000
Copy link
Owner

You can follow the code in conversion.ipynb

@ljc222
Copy link

ljc222 commented Jan 25, 2022

how to generate the metadata.pkl file?

@lisabecker
Copy link
Contributor

lisabecker commented Apr 27, 2022

I'm also interested in generating a metadata.pkl from my own data for inference. @auspicious3000 do you happen to have the script that produces the metadata.pkl that's used for inference, not for training?

Simply adding the mel spectograms produced by make_spect.py and make_metadata.py to the lists of speakers in make_metadata.py instead of the paths does not produce sensible results with conversion.ipynb, just noise.

@auspicious3000
Copy link
Owner

Each metadata is a list of [filename, speaker embedding, spectrogram]

@lisabecker
Copy link
Contributor

lisabecker commented May 4, 2022

@ljc222 @Ha0Tang if you go through past issues, you'll stumble across this repo/notebook which puzzles it all together to make it work end-to-end: https://github.com/KnurpsBram/AutoVC_WavenetVocoder_GriffinLim_experiments/blob/master/AutoVC_WavenetVocoder_GriffinLim_experiments_17jun2020.ipynb

Hope this helps!

@arthurwolf
Copy link

You can follow the code in conversion.ipynb

I looked at https://github.com/auspicious3000/autovc/blob/master/conversion.ipynb and I have zero idea how that helps with this.

I just need some way to, on the command line, provide the original voice as a wav/mp3, provide the file to change the voice of as wav/mp3, and get the output file with the voice changed written to disk.

How do I do that? How does anyone ever use this if something this basic isn't documented? Am I missing something obvious ?

Thanks a lot to anyone with any information.

@jvel07
Copy link

jvel07 commented May 6, 2024

@ljc222 @Ha0Tang if you go through past issues, you'll stumble across this repo/notebook which puzzles it all together to make it work end-to-end: https://github.com/KnurpsBram/AutoVC_WavenetVocoder_GriffinLim_experiments/blob/master/AutoVC_WavenetVocoder_GriffinLim_experiments_17jun2020.ipynb

@lisabecker hi! How do you generate the metadata for inference?
make_metadata.py generates metadata for train, or in the end, this script is for generating metadata for both train and inference? (just with a different path for the mel-spec)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants