I’ve been using this for the past week. Game changer is absolutely right.
I spent hours trying to get a specific pose, hundreds of generations and seed changes, trying to dial in aspects of one vs another, taking aspects I like and clumsily patching together a Krita mash up then passing that back through img2img… only to get something kinda close.
One shot with canny or depth on Controlnet and I had exactly what I wanted.
Amazing how fast and easily it works. It’s been a WILD 6 months that this tech has been available.
You've convinced me to go for it. I'm downloading the models from HuggingFace and the source from Github.
I'm not sure how it all fits together but I'll try to figure it out. I have experience with Python development, git LFS, and I've been following research papers for years but this is my first attempt.
Is this all self-contained? Should I start elsewhere? I've been pretty intimidated by this stuff TBH, relying on hosted Midjourney / Dall-E for my artwork.
I notice the two repos have the same name. Do I merge the file-trees? How do they work together?
https://github.com/AUTOMATIC1111/stable-diffusion-webui
And then this https://www.reddit.com/r/StableDiffusion/comments/1167j0a/a_...
Step by step video tutorial https://youtu.be/vhqqmkTBMlU