Replies: 2 comments 4 replies
-
Could you share with us those changes? this would really help me know if i"m 'heading in the right direction' or wasting cycles - thanks! |
Beta Was this translation helpful? Give feedback.
-
On a separate note, I'm testing changing I'm on a 1070ti, so testing this is a little bit slow, 1.09s/it with the card limited to 60c. My plan is to do a 5000 step training with amsgrad on and off to see if it indeed converges better. Amsgrad option is from this paper. |
Beta Was this translation helpful? Give feedback.
-
As the tile says, could we get a couple CSV files with the loss logged during training?
File 1: Just the loss for each step
File 2: The loss for each image each time they are processed
Mainly so we can see if the training is converging and if there are any problematic images that are increasing loss.
Edit for a second idea that seemed useful to me:
I just edited my own textual_inversion.py to do the following;
Every N steps, when generating the image, generate one with a random seed and one with a fixed seed.
Random seed helps with visualizing overall understanding of the concept and fixed seed with smaller details of the concept, so you know if your learning rate is too high (Images start bouncing back and forth) or too low (barely any changes) and when to stop training.
3rd edit:
For people wondering about learning rate, You can be really aggressive with it with 0.1to 0.5 for 500 steps, then drop back down to 0.005 which is already a bit strong for Adam. Default for the algorithm is 0.001. You can go lower than that if you're already getting good results but it might be missing details. In any case, the way it works, it should always converge or bounce closer to convergence, it's just a matter of when.
Beta Was this translation helpful? Give feedback.
All reactions