CLIP Interrogator extension! #7968
Replies: 6 comments 3 replies
-
nice extension! do you have any info on which ones take the least vram? Since the model has loaded into vram at startup, I unloaded it with https://github.com/hako-mikan/sd-webui-supermerger, an extension that has an "unload model" option. "fast" and "classic" mode seem to use the least. |
Beta Was this translation helpful? Give feedback.
-
Updated now with batch mode so you can point at folder of images and it will create prompt for each and store in either txt file for each image, big txt file with all the prompts, or csv file. I haven't sorted out how to display progress bar nicely. I've been looking at how it's done for the textual inversion training and tried to replicate but run into Gradio errors so people still have to watch the console to see how it is progressing. Does anyone know more info on how to display progress bars in your extension nicely? |
Beta Was this translation helpful? Give feedback.
-
Hello, |
Beta Was this translation helpful? Give feedback.
-
This is VERY useful! There are error messages when offline that indicate it is possible to set up to work locally. As usual, they are very cryptic and go off talking about installing transformers, etc. So far I have not beemn able to set everything working. Is there a simpl;e guide to local automatic1111 use that even I can follow?? Thanks. |
Beta Was this translation helpful? Give feedback.
-
I love this extension. Lately it seems to be running really slow. I see that it has switched to low VRAM because I have 10 GB of VRAM. How much VRAM would keep it from doing that? |
Beta Was this translation helpful? Give feedback.
-
How to install additional models to this extension? |
Beta Was this translation helpful? Give feedback.
-
Hello friends!
I've created an extension so the full CLIP Interrogator can be used in the Web UI now. Give it an image and it will create a prompt to give similar results with Stable Diffusion v1 and v2. It can give you a nice starting point and ideas for your prompts.
https://github.com/pharmapsychotic/clip-interrogator-ext
It honors the low/med vram option of the web UI and does it own detection to switch into a low VRAM mode so should hopefully work across a wide range of GPUs. I've tested with nvidia 1070 8GB. Precomputed text embeddings are downloaded in safetensors format from huggingface and stored in
models/clip-interrogator
to speed up processing.The first time you run it the progress doesn't display so check the console to see how it's doing.
It's my first attempt at an extension so let me know if there's anything I've done wrong or could improve!
Beta Was this translation helpful? Give feedback.
All reactions