How to use the new "discriminating" Stable Diffusion Img2Img algorithm

Key moments:
0:57 – Human Faces
4:19 – Oak Tree
6:54 – How to Install

We walk through how to use a new, highly discriminating stable diffusion img2img model variant on your local computer with a “webui” (Web UI), and actually achieve decent results.

Webui Repo: https://github.com/AUTOMATIC1111/stable-diffusion-webui
Reddit Thread: https://www.reddit.com/r/StableDiffusion/comments/xboy90/a_better_way_of_doing_img2img_by_finding_the/
A solid tutorial on installing another webui: https://www.youtube.com/watch?v=bu2xwx-6JGk&ab_channel=Aitrepreneur
***NOTE***, k installs webui from this repo: https://github.com/sd-webui/stable-diffusion-webui
you should use this repo instead: https://github.com/AUTOMATIC1111/stable-diffusion-webui to get the new version of img2img

Why are there two repos you ask???? Smart move, So did I!!! the nerds just started arguing about IP law: https://github.com/sd-webui/stable-diffusion-webui/discussions/686

Technical Support on discord: https://discord.gg/kD94kVnvE4

#stablediffusion #aiart #ai #art

63 Comments

  1. to get the closest prompt to your image you can interrogate, the clip interrogator code will spit out the closest prompt and you should have less margin of error generating photos very different from the original

  2. As an artist, right autumn tree is very similar. Just compare the shapes of tree crown, it has the same shapes and structure. Of course trunk is different, but that’s is a result of higher CFG scale

  3. Thanks for the information. Very good to know! Would anyone be able to suggest a way to use the Web UI to connect to a locally installed version of Stable Diffusion but do so without being on your local network? I'm afraid I've become a bit addicted to using SD and sometimes during my morning walk at the park I'm thinking it would be great to be able to generate some images while I'm walking. I know there are collabs and there is the paid dream studio. But that gets a bit pricey for me. Since the local install is essentially free, I'd like to make the most of that (if possible) while I'm on the road.

  4. how did it take you a long time to figure out the parameters? I used the parameters in the readme and got similar results to you (you also used those same parameters)

  5. This is a good introductory video, but you kinda did do the oak tree and this feature dirty when testing stuff with that image. I managed to get great results with both autumn and winter themed prompt changes with the same image as you used. The three biggest points are:
    1. Putting the CFG scale higher usually does not have a positive effect. Which value is "right" for each image varies, but usually putting it between 0.3 and 0.6 gives better results in my experience with this.
    2. Higher decode step counts can help dramatically with getting you more impactful (yet more accurate to the original image) changes at lower CFG scales and also helps a lot with the "compression artifacts"
    3: High sampling steps are a major contributor to the compression artifacts. Around 15-30 is usually much better in the majority of cases, preventing the majority of the artifacts!

    Finally prompt engineering can help quite a lot with getting the changes that you want. Just adding or changing a single keyword isn't always enough, like what you did with summer/winter/autumn in this case. Here's a demonstration of my results using these points: https://imgur.com/a/tTrkUFb.
    Hope this is helpful and doesn't come off too critical, I don't actually know the underlying code that makes any of this work any better than you do. I just want to showcase that this algorithm is more universally powerful than you show in the video!

  6. Is it possible to take a photo of a real person and make them into a 2-d flat color cartoon (not like those hideous cartoon filters for phones or PS) but actually in the style of cartoons like Hilda, Over the Garden Wall, or Adventure time. I have been experimenting with lots of different AI and all of them seem to have trouble doing this. Stable Diffusion seems to be the only one to actually attempt an artists style effectively, but there are still lots of gaps and issues.

    My project requires a lot of 2-d cartoon backgrounds in styles similar to the cartoons mentioned above and I was hoping to use IMG to IMG as a way to turn real backgrounds into cartoon versions in those styles. So far the version of colab I'm working with is only up to 1.4, hope 1.5 will come soon, but do you think it's possible?

  7. I've been studying and using Stable Diffusion UI V2 img2img and it's converting my background sketches into something good. As I'm a comic book artist, I was trying to find a way to speed up my gigs. Even with the right seeds and the right prompts, it still needs a lot of retouches. AI is not good to keep consistency, mainly for sequential art. Still, it's faster than my previous method (modeling 3D for backgrounds). Also,, it's decent to generate a character for illustrations, but when I add two characters or more, AI seems "on drugs", creating a lot of nightmarish images XD haha
    So, img2img works like a charm for one character on the cover (Right now, I'm using it the create a lot of content for my advertising campaign for my products). I'm still studying it. Great for illos, not so good for sequential art. The problem is that we can't be creative when we're working for a publisher, they want consistency throughout the project. But, for personal projects, I think it's valid to experiment and find creative ways to solve it… for now. I hope AI is more advanced in the next five years.

  8. I did everything you said everything the site said and still didn’t work for me! Not only did it not add what I wanted, it ruined the picture and oversaturated it. Changed around all the settings and nothing helped

  9. Was not able to replicate your results with the new AUTOMATIC1111 updates….don't know if they updated the Img2Img Alternative Script…..😂

  10. It is pronounced "Oiler" not "U-Lar", the name of the famous mathematician who contributed heavily to the branchs of Maths like Graphy Theory and Topology.

  11. Img2Img doesnt work for me, Txt2Img does, but if I use anything with an image in it I get the error – RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.cuda.HalfTensor) should be the same – I am using a 512×512 image, any idea how to fix this error?

Leave a Reply

Your email address will not be published. Required fields are marked *

© 2024 AI Art Video Tutorials