Pix-to-Pix as Viton? #6624
Replies: 2 comments 2 replies
-
I think it's better to use IP Adapters. I responded the same thing in the issue you had created. |
Beta Was this translation helpful? Give feedback.
-
Have you given existing VTON papers a try btw? I haven't played around much with training but I feel directly training Pix2Pix on cloth-model pairs may not work. A few current approaches involve learning separate embeddings for clothes and applying a warping module for cloth placement.
They're mostly unable to handle cloth consistency from what I've seen, but it can do a decent job at times with the right prompting + img2img and correct parameters. |
Beta Was this translation helpful? Give feedback.
-
Is it a good idea to train pix-to-pix-instruct to act like a viton (virtual try-on)? The training data would include images of cloth and prompts as input and a model wearing that dress as ground truth.
Beta Was this translation helpful? Give feedback.
All reactions