Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does this work with the unseen speech? #10

Open
sbkim052 opened this issue Jul 20, 2020 · 3 comments
Open

Does this work with the unseen speech? #10

sbkim052 opened this issue Jul 20, 2020 · 3 comments

Comments

@sbkim052
Copy link

Thank you for sharing your git.

My question is same above.
Does this work with the unseen speech?

@bshall
Copy link
Owner

bshall commented Jul 20, 2020

Hi @sbkim052,

Yeah, it should work with unseen speech as the input. All the examples here are converted from unseen speech.

If you want to convert to an unseen speaker, you'd have to retrain the model. You could also look into conditioning on x-vectors or other speaker embeddings if you want to do zero-shot conversion.

@sbkim052
Copy link
Author

sbkim052 commented Jul 21, 2020

Hi @sbkim052,

Yeah, it should work with unseen speech as the input. All the examples here are converted from unseen speech.

If you want to convert to an unseen speaker, you'd have to retrain the model. You could also look into conditioning on x-vectors or other speaker embeddings if you want to do zero-shot conversion.

Thank you for answering:)

I have an additional question about your answer.
What do you mean by conditioning on x-vectors or other embeddings for zero-shot conversion?
Could you explain it in more detail on doing the zero-shot conversion?

@bshall
Copy link
Owner

bshall commented Jul 21, 2020

@sbkim052, no problem.

The basic idea is to train a speaker verification/classification model to learn an embedding space for speaker identity. Then, instead of conditioning the decoder on a fixed speaker id (like I did in this repo), you condition on the learned embeddings. At test time you can get the embedding for a new unseen speaker and condition the decoder to generate speech in that voice. For more info, you can take a look at this paper. They use a text-to-speech model instead of an autoencoder but the general idea is the same.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants