[New Model] The finished tile model is released #1142
Replies: 69 comments 30 replies
-
If you have any problem, make sure that you are using "control_v11f1e_sd15_tile", not the old "control_v11u_sd15_tile". Pay attention to the name "v11f1e" |
Beta Was this translation helpful? Give feedback.
-
Beside official ways to use it, for a1111 users, we provide a special functionality: the tile model can be combined with "my prompt is more important" to do arbitary prompt-based transfer |
Beta Was this translation helpful? Give feedback.
-
some other up scaling tests should be trivial. just try your self |
Beta Was this translation helpful? Give feedback.
-
when the detail replacement is large (using "my prompt is more important"), it can interprete the contents and replace with promts while maintaining the overall structure. This method does not rely on image depth or edges Below is "balanced" |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
OK perhaps I need to give an upscale example so that it can be really called "tile" and prove that it is not off topic. Below is protogen without using any external upscaler (except the native a1111 Lanczos, which is not a super resolution method, just for resampling) This is from 192x256 to 1536x2048 meta: beautiful woman in sweater, high-quality, extremely detailed Note that image content may be changed because I used prompts to specify "in sweater" screen (note that I am using the more challenging t2i, which is weaker than i2i in upscaling): non cherry picked, random seed "1": non cherry picked, random seed "2": Note that I am using the worst method to do upscaling to test the worst performance. I am not using any XXX upscaler. This is just a basic t2i diffusion. Users may find much better ways. This is just the "naked" results without the help of special things. I am not even using esrgan this example just show it can draw detailed hair at high resolution. To really get best results, perhaps you should use I2I + ControlNet or something like high res fix + esrgan, or even other tile extensions. |
Beta Was this translation helpful? Give feedback.
-
Hi. I have a problem. Seems like controlNet tile doesn't work for me. It doesn't affect an image at all. Image generated same with and without control net I am using this versions: ControlNet v1.1.111 Model Preprocessor:
python: 3.10.6 • torch: 2.0.0+cu118 • xformers: N/A • gradio: 3.23.0 • commit: [22bcc7be] What can be wrong? Maybe I do something wrong? Do I need to setup big resolution? Use inpaint? Upscaler? Highres fix? Thanks |
Beta Was this translation helpful? Give feedback.
-
@torealise I dont know if it's your intention to use no preprocessor or not, but does it work with tile_resample preprocessor? |
Beta Was this translation helpful? Give feedback.
-
my bad. i forget to rename yaml. now fixed. update and restart completely |
Beta Was this translation helpful? Give feedback.
-
Thank you. Could I please ask about the chance of contacting you and having a talk? We have been using your ControlNet in our research, and it would be fascinating to have you as our adviser. If it is possible, would you be available for a conversation? |
Beta Was this translation helpful? Give feedback.
-
now all 1.1 models and usage codes are finished. We will not release more things unless there is bug fix. The next update will be ControlNet 1.2. But this may depend on what Stability is going to release. |
Beta Was this translation helpful? Give feedback.
-
I downloaded control_v11f1e_sd15_tile.pth, and updated control_v11f1e_sd15_tile.yaml, then even restarted pc, but still control net tile doesn't make any difference to generated image. Maybe I use it wrong way?
but result is the same with control net turned on and off |
Beta Was this translation helpful? Give feedback.
-
share logs please |
Beta Was this translation helpful? Give feedback.
-
is this a log that you asked me? Unfortunately I don't know what is log exactly Loaded state_dict from [C:\stable-diffusion-portable-main\extensions\sd-webui-controlnet\models\control_v11f1e_sd15_tile.pth] 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:00<00:00, 27.43it/s] |
Beta Was this translation helpful? Give feedback.
-
On that test no. However I did test with ControlNet is more important. Even at denoise strength 0.2 , it changes the face drastically unlike Dave's May give Dave's settings a try but I think the VAE did this to me. I have bad habit using 84000 mse vae for everything. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
it seems that Realistic_vision is a preferred model for tile. Protogen is relatively meh in my tests, not sure why |
Beta Was this translation helpful? Give feedback.
-
Thanks! The tiling upscaler pushed my image to AAA level. |
Beta Was this translation helpful? Give feedback.
-
"Great job! Could you tell me how you trained this model and what is the underlying principle behind it?" |
Beta Was this translation helpful? Give feedback.
-
Could you provide some details on how this is actually working? I read through as much as I could find on this and I found a lot of examples but without knowing what the controlnet is supposed to be doing it was difficult to figure out what those examples were trying to demonstrate. I understand that its emphasizing/deemphasizing prompts based on what was in the original tiles, but how is it choosing tile sizes and partitions? Are the tiles arbitrary sizes or is it just something like a 32x32 grid across the whole image? Is there a way for me to control the tile sizes? Or am I misunderstanding the term "tile" here entirely? Similarly, if the tile_resample preprocessor "down sampling rate" is set to 1, which is its default, is it doing anything at all? That setting seems like its not resampling at all, right? So is that essentially the same as not having a preprocessor? Does a "down sampling rate" of 4 mean that it is sampling the image at 1/4 resolution? And what is the use case where downsampling before passing it to the controlnet is useful? |
Beta Was this translation helpful? Give feedback.
-
Any tips on how to avoid this strong depth of field? I start with nice images, but the last ones geet this strong DOF and ruined everything |
Beta Was this translation helpful? Give feedback.
-
Putting DOF, blur, blurred in the negative prompt helped a lot! The wings get a new detail too. Thanks for such amazing job on this! |
Beta Was this translation helpful? Give feedback.
-
Anyone knows what causes huge color distortion like this? I compare metadata and and only difference is BREAK keyword ( which non-BREAK is the one cause issue). Is this webUI/Multidiffusion's bug or tile model? Never had this problem using Multidiffusion extension alone so this must be combination of these. |
Beta Was this translation helpful? Give feedback.
-
Is just me or after the last updates Tile is completely ruined? I cant run more the denoiser slide than 0.1 and the final image gets completely different and ruined? :( |
Beta Was this translation helpful? Give feedback.
-
OP says that the model also supports 768 and 1024 resolutions, but I cannot seem to find how to customize the tile size bigger from 512 pixels for the preprocessor resolution... To be absolutely clear: Any ideas how I could use bigger tiles? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
is there something wrong with recent update? Base tile_resampling seems to be overly detail the image @lllyasviel
|
Beta Was this translation helpful? Give feedback.
-
I did a test here, I used the tile resample along with the ultimate sd upscale, tested with down sampling 1 and 8, exactly the same settings for both, including the seed, the two images are identical, or practically identical. So, what's down sampling supposed to do? I understand that down sampling is to reduce an image and use the extra pixels to increase the details, but apparently, this option in CN isn't doing anything at all. |
Beta Was this translation helpful? Give feedback.
-
i think the tile_colorfix+sharp is very useful, but i can't find it in comfy,how can i use tile_colorfix+sharp in comfy? A1111 run slowly on my laptop. |
Beta Was this translation helpful? Give feedback.
-
hello everyone, we released the finished tile model. The name is
'control_v11f1e_sd15_tile' (The old "control_v11u_sd15_tile" will be removed)
The desc is at https://github.com/lllyasviel/ControlNet-v1-1-nightly#controlnet-11-tile
Download model here https://huggingface.co/lllyasviel/ControlNet-v1-1/tree/main
You have 5 ways to use it:
The performance of this model will be between production-ready and experimental. Final quality flag is [e].
You need ControlNet at least v1.1.107
The preprocessor can be none or tile_resample (tile_resample is just a resizing to make your image smaller if you want more variations)
Hint: do not forget to try different control modes - it can produce interesting effects
Hint2: this model supports resolution 768 or 1024 well. You can try it if you find SD's 512 is too small and blur.
Beta Was this translation helpful? Give feedback.
All reactions