Hypernetwork training #2284
Replies: 52 comments 163 replies
-
I trained an HN for 2500 steps on various midjourney images to see what would come out of that. Here are 2 examples of using the trained hypernetwork. Using same seed with two different prompts: sd1-4 only: sd1-4 + midjourney hypernetwork: |
Beta Was this translation helpful? Give feedback.
-
The wiki page has already been updated with the new info. It recommends 0.000005 or 0.0000005 even. |
Beta Was this translation helpful? Give feedback.
-
Might be interesting to move the hypernetwork selection dropdown to the right of the checkpoint selection dropdown... Would be quicker than phishing for it in the settings. |
Beta Was this translation helpful? Give feedback.
-
It is not yet clear to me what the hypernetwork is for and how to train it. I'm interested to know how it differs from textual inversion, why there are different sizes, which one is good for what and when? I am a beginner in this subject. |
Beta Was this translation helpful? Give feedback.
-
Here are the same prompts as the ones above but using a HN trained with 0.000005 for 4500 steps instead of 0.00005 at 2500 steps: The effect is more subtle... I would say training at 0.00005 provided results that are closer to what I would expect for a midjourney style. HN are really fascinating. I wonder if this is how midjourney is actually applying their "style" on top of other checkpoints,,, quite possible... |
Beta Was this translation helpful? Give feedback.
-
How does hypernetworks training speed compare to textual inversion and dreambooth also the vram usage? |
Beta Was this translation helpful? Give feedback.
-
Is embeddig or not necessary for hypernetwork training? |
Beta Was this translation helpful? Give feedback.
-
Quick output showing the overall impact an hypernetwork has on the model output. Here, I am prompting for a photo of tom cruise with and without the hypernetwork trained on my face: Prompt: So essentially just like dreambooth... |
Beta Was this translation helpful? Give feedback.
-
Testing now with 50 face images and a learning rate of 0.0000005 (< lowest of the recommended) |
Beta Was this translation helpful? Give feedback.
-
what are the modules checkboxes when creating a hypernetwork ? 768 320 640 1280 edit: ok I cannot even create a hypernetwork.. it gives error the pt file not found in models/hypernetwork |
Beta Was this translation helpful? Give feedback.
-
I have a question, why the hypernetwork is just two linear without activation? |
Beta Was this translation helpful? Give feedback.
-
Do I understand this correctly: hypernetworks are trained like embeddings with a very low learning rate and steps under 5000 and then put in /models/hypernetworks as a .pt file? And this might work better than dreambooth? An answer would be appreciated. This sounds... really good! |
Beta Was this translation helpful? Give feedback.
-
I get an error message when trying to load my trained HN: Loading hypernetwork STYLEHN I trained it in the textual inversion tab with the instructions described here and put it in models/hypernetworks |
Beta Was this translation helpful? Give feedback.
-
So I trained for 1000 steps on 6 photos of myself, activated the hypernetwork in the settings but I'm a bit lost as to how to generate photos. Am I using it the same as an embedding: ie: a photo of x, where x is the name of the hypernetwork ? |
Beta Was this translation helpful? Give feedback.
-
When training a hypernetwork, what do I select for "Prompt template file" if I am training for a subject? |
Beta Was this translation helpful? Give feedback.
-
I'm going to train hypernetwork for WD1.3, but in their release note they says that the float32 version can only be used for generation, and the full version for generation and training. For hypernetwork training, should I use only the Full version, or is it possible to train using float32 in this case? It uses less VRAM |
Beta Was this translation helpful? Give feedback.
-
I have some issues when i training a hypernetwork. |
Beta Was this translation helpful? Give feedback.
-
@Heathen |
Beta Was this translation helpful? Give feedback.
-
There is no option to use deepbooru for caption in the Train tab for me. Was it removed or am i missing something ? |
Beta Was this translation helpful? Give feedback.
-
The lastest repo has a new option "Batch size" on training. Anyone know what dose it do? |
Beta Was this translation helpful? Give feedback.
-
Just wanted to ask I noticed something, if I train a network for 2000, steps look through that and find the best one for instance 1500steps, and then use that as the base to keep training to let's say 10000 steps, should it continue from 1500 to then go to 10000 - because I have noticed that happening, instead of starting from 0 again on the hyper network, is this default behaviour? |
Beta Was this translation helpful? Give feedback.
-
Just a question Is training at 5e-5 using 1000 steps the same with training at 5e-6 using 10000 steps? Will the trained data from 5e-5 lose more resolution, though the two can add up? |
Beta Was this translation helpful? Give feedback.
-
Hi, https://github.com/danielalcalde/stable-diffusion-webui Issue: #2740 (comment) The main idea is to add weight normalization to stop the model weights from exploding so quickly. It would be cool if someone could give it a try. If I get good feedback I will launch a pull request to add it to the maser branch :) |
Beta Was this translation helpful? Give feedback.
-
The training cannot be carried out due to the above error. plz help.. |
Beta Was this translation helpful? Give feedback.
-
In the last few days I've been thinking about how to get img2img to change only the style. If I give it a big denoising, it's completely different from the original image, if I give it a little, nothing changes. |
Beta Was this translation helpful? Give feedback.
-
Somebody please be so kind telling me what's the bottom input in Learning Rate is for? |
Beta Was this translation helpful? Give feedback.
-
Hey I have a problem where the training always cuts off early and I can't see any info about that. I set it to 2000 and maybe it will run for 299 and then stop. I just set this up yesterday so i'm on the latest versions of everything, i'm on osx 12.5.1. hypernetwork learning rate 0.00005. save image and save a copy of embedding each 100. The message just reads loss:nan followed by the number of steps. Hope someone can help! |
Beta Was this translation helpful? Give feedback.
-
Hi folks, If a hypernetwork is just an additional layer for UNet (have I caught it right?), why can't we train on 1024x1024px pics? A similar trick has been done with inpainting, hasn't it?.. |
Beta Was this translation helpful? Give feedback.
-
In order of complexity: TI, HN, LoRA, DB, FT. LoRA has many new subsets now that train more layers but my heart always belongs to HN as I found it the best for what I did. Since SDXL we can't train it, but so many things we can't with XL as far as training related. |
Beta Was this translation helpful? Give feedback.
-
Starting a discussion where we can exchange on hypernetwork training tips and tricks.
One tip I can give is to use a learning rate of 0.00005 for training hypernetwork... if you use the default 0.005 you will get to NaN very quick.
Discoveries:
Questions:
Suggestions/Requests:
Beta Was this translation helpful? Give feedback.
All reactions