Made a tutorial for how to train a model with VITS V6 #1074
Replies: 13 comments 25 replies
-
Hey @LordApplesause, I m having the following issue while running the Training TTS colab.
Here the name of my dataset folder is "kumar" I do have changed the Dataset folder path in the train_vits.py like this 👍
The original notebook contains same lines as :
Can you help me with what could be the issue? Thanks |
Beta Was this translation helpful? Give feedback.
-
Hi @LordApplesause , Very good idea to write and share an up-to-date tutorial. I will provide feedback when I am done reading it. For now could you explain why we need to create a dataset (withthat 500 clips as you wrote) ? I tested the TTS colab and with just a minute of target speaker clips the results are already stunning. What will it bring to create a larger dataset ? Looking forward to reading your reply |
Beta Was this translation helpful? Give feedback.
-
@LordApplesause Thank you for your effort. Is it possible for you to upload the tutorial elsewhere? Medium.com is not accessible from my place. |
Beta Was this translation helpful? Give feedback.
-
Hi, which model are you using as a starting point in your tutorial ? Is it the one from Exp1 in the paper ? Will it work to generate a French voice ? Thanks |
Beta Was this translation helpful? Give feedback.
-
@LordApplesause Hello, can u share ur custom voice dataset for reproduce ur result. It will let me know if I'm doing everything right. |
Beta Was this translation helpful? Give feedback.
-
Is the tutorial still working? I am getting ModuleNotFoundError: No module named 'TTS.trainer' |
Beta Was this translation helpful? Give feedback.
-
@LordApplesause can you please reach me on Gitter, Element, or email. We prepared a small 🎁 for you. |
Beta Was this translation helpful? Give feedback.
-
@rioharper Thank you so much for writing the tutorial. I had a couple of problems using it. First, the SNR binary file could not be executed (I'm using WSL2). Then I can't find the code for synthesis in the notebooks you provide. (Do end-to-end models use vocoders??) Tensorboard also says that there is no info to be read in the log directory, which is not correct! |
Beta Was this translation helpful? Give feedback.
-
e2e models have their own vocoder in them, so you don't have to specify anything! All hail to the awsome vits! |
Beta Was this translation helpful? Give feedback.
-
Any clue on how to run WADASNR locally? My WSL refuses to. It just acts as if it doesn't exist. |
Beta Was this translation helpful? Give feedback.
-
Hi man, I was following your tutorial to train my own model. First of all I must say congratulations on the job, it is not only really helpful but also entertaining. Now I have a few questions:
I'm running my code locally and recording all the modifications I had to do to the Collab notebooks for them to wok on something like the WSL of a PC. If you want, as soon as I have it done, I can share it somehow, maybe I can create a repo and send you the link or something. Once again, thanks for such an amazing tuto. Greetings from Argentina! |
Beta Was this translation helpful? Give feedback.
-
The article mentions that to continue, just change resume_from to your latest checkpoint, but the training notebook doesn't seem to have any resume_from in it. Is it training from scratch rather than fine tuning? There is a spot it downloads a checkpoint, but I don't see it used. |
Beta Was this translation helpful? Give feedback.
-
NOTE: This is a fairly old tutorial, so the training process that works with VITS may not work. I will get around to updating the tutorial soon, but for now refer to the coqui docs for more info!
I've been using this tech for a while now, so thought I should make some kind of contribution. This article details how to make a dataset, configure training values and generate audio all on Google Colab.
If anyone had any input to make this article more detailed and helpful, please let me know, and ill make sure to implement it!
Article
Beta Was this translation helpful? Give feedback.
All reactions