Helpful Resources

I’ll add more here as I remember them. Feel free to add more in the comments.

Models

Models are basically the brains of Stable Diffusion. They are the data SD uses to learn what your prompts mean.
The built-in models that come with Stable Diffusion are really bad for porn. Don’t use them. In fact don’t use them at all unless you’re training your own models, there are better SFW models.

Here are some of my personal favourites:

Anime

  • MeinaHentai is a great model to start with. Compared to other models it’s really easy to prompt
  • AOM3 also does really well, though it might be a little more difficult to guide

For all of those, I recommend installing https://github.com/DominikDoom/a1111-sd-webui-tagcomplete, as they heavily rely on danbooru tags.

  • Berry Mix (Pre-mixed version here) can also work pretty well, depending on what you want to do. AFAIK it uses rule34 tags instead of danbooru, so it probably won’t work all too well with prompts used for the above ones

Realistic

  • Uber Realistic Porn Merge is the only realistic model I know of that does hardcore stuff. It’s unfortunate problem is that it’s REALLY DAMN HARD TO USE

VAEs

VAEs are mostly used for finetuning colors, sharpness, what have you. Some models come with a VAE builtin, but for ones that don’t, it’s recommended to have one on hand.

  • “Anything VAE”, “Orangemix VAE”, and “NAI Leak VAE” are the same exact thing under different names. If you already have one on hand, don’t bother with the others. Most VAEs are renamed versions or modifications of this one.
  • Waifu Diffusion’s kl-f8-anime2 is also a pretty good one. It doesn’t require Waifu Diffusion.
  • The one that comes with Stable Diffusion is the only one that seems to work for realistic stuff.

LoRAs

LoRAs teach models about concepts (characters, clothing, environments, style, …) they might not know about. There are a LOT of them, so feel free to browse Civitai to find ones you might want.

LoRAs tend to be specific for families of models, or at the very least styles (using anime LoRAs on realistic models tend to be a bad idea), but there are a fair few that will work across the board.

Locon and LyCORIS are newer formats of LoRAs. Not sure on the technical differences between them, but they will not work out of the box and need an extension such as https://github.com/KohakuBlueleaf/a1111-sd-webui-lycoris to get working

Textual Inversions / Embeddings and Hypernetworks

These are mostly obsoleted by LoRAs. There are a few embeddings such as Deep Negative and EasyNegative that are still quite useful, but in most cases you’ll want to use LoRAs instead.

  • @RestfulD34m0N
    link
    English
    11
    edit-2
    2 years ago

    Are these softwares all free to download/use? Also, how does one start doing this? do i just need the WebUi or do i need extra files to feed into it and stuff?

    • @FreezingInFuckingHell
      link
      English
      52 years ago

      Yea the webui and stable diffusion models/checkpoints are generally free to use.

    • Triple UnderscoreOPM
      link
      English
      5
      edit-2
      2 years ago

      It’s all free, yes.

      For the how: Aside from the webui (or whatever else you’re running the stable diffusion code on), you’ll need models that instruct it what to create. The builtin models are terrible for porn.

    • @echoct
      link
      English
      32 years ago

      I just figured it out tonight playing around with the links and readmes available above. If you get stuck I can try to answer more specific questions.

      • @RestfulD34m0N
        link
        English
        22 years ago

        Hmm ill probably have mroe questions but for now im curious:

        • how much space was the download(s)?
        • how confusing is the software to use?
        • what kind of limitations does the software have? can i do multiple people? monsters? futa? etc.

        Thanks for your help! :)

        • @echoct
          link
          English
          72 years ago

          Sure,

          1: The initial download was pretty small ~10GB. But with the models, lora(s), extensions, ect. I’m up to ~60GB.

          2: The guide at the top was pretty easy to follow. Install the dependencies, then install the UI. Launch and run. There is a bit of a learning curve with all of the options but so far it hasn’t been too confusing.

          3: That’s where the extra models/lora(s) come in. Various models are trained in different styles, poses, actions, ect. The lora files are smaller things, like poses. IE: Cowgirl is it’s own lora file that tells the model how to use the prompts you give.

        • Triple UnderscoreOPM
          link
          English
          6
          edit-2
          2 years ago

          how much space was the download(s)?

          On my end, it’s sitting at ~64GB (with btrfs compression shenanigans), though 60 of those are from all the models I have installed. The download would probably be ~2GB, even less if you disable downloading the “default” models with --no-download-sd-model and instead pick models off of Civit or wherever manually.

          Edit: Should have mentioned. Most full models are between 2-4 GBs each. Some can be 5+ but they tend to be “full” versions intended for merging & such. LoRAs are generally smaller. Depending on how much they’re pruned they’ll be anywhere between 10-100 MBs each.

          how confusing is the software to use?

          There’s definitely a learning curve, yes. But there’s plenty of resources (and more importantly, examples) out there.

          what kind of limitations does the software have? can i do multiple people? monsters? futa? etc.

          As long as you have the correct models set up it can generate basically anything. At least with anime models, monsters and futa are a given. Your main issue will probably be multiple people, although there are solutions to that. (See the multidiffusion upscaler GitHub repo on the main post)

          • @RestfulD34m0N
            link
            English
            22 years ago

            After extracting the WebUI files and doing the git clone, ive tried to double click webui-user.bat but the terminal that opens up says “python not found” but i’ve got python downloaded…

            • Triple UnderscoreOPM
              link
              English
              4
              edit-2
              2 years ago

              Not on Windows right now so I can’t confirm, but you probably forgot to pick “Add Python to PATH” or whatever the option is in the installer. Try running the Python installer again, maybe it’ll let you add it without needing to uninstall & reinstall

              Edit: If you’re on Nvidia, there seems to be a simpler install method now: https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Install-and-Run-on-NVidia-GPUs (Method 1)

              • @RestfulD34m0N
                link
                English
                12 years ago

                Hmm i uninstalled and reinstaqlled and i’m not seeing an option for “Add Python to PATH” Could there be an alternate name?

                • Triple UnderscoreOPM
                  link
                  English
                  62 years ago

                  It’s probably “Add python to environment variables.” They must have changed the wording on that at some point.

          • @RestfulD34m0N
            link
            English
            22 years ago

            Thanks! this is very helpful :). Also, and maaybe more importantly, idk how to download things off of github. I ‘Go To File’ and theres no download button. I’m sure im missing something simple lol

    • @NSFW_Enthusiast
      link
      English
      22 years ago

      I recommend following a setup tutorial on YouTube. They usually link to models and stuff youll need. Look for a recent one, this stuff evolves a lot.

  • @blackmayne110
    link
    English
    72 years ago

    I’ve been meaning to figure this out. So the commonly used ones are labeled WebUI, but just how much of the content goes to the web itself?

    If I wanted to train on images that I don’t want going online, will they? Or will the products that I create end up online, or does all this stay local?

    • Triple UnderscoreOPM
      link
      English
      172 years ago

      By Web UI it means that the graphical part of it – where you write your prompt and hit generate – is running inside your browser and not as a separate window or command line. Everything is kept on your own computer unless you explicitly tell it to open up remote access.

      • @blackmayne110
        link
        English
        62 years ago

        It was probably a dumb question, but I wanted to clarify just the same, thanks. When ya get these kinds of pics under strict confidence, you don’t wanna risk breaking that in any way.

        • @NSFW_Enthusiast
          link
          English
          4
          edit-2
          2 years ago

          It was probably a dumb question

          No such thing as a dumb question. but yes the WebUI only exists so you dont have to type into a command prompt. It looks nice and pretty for humans, and then translates all those settings into a command that is then sent to stable diffusion. All of it stays 100% local, unless you go into the WebUI settings and tinker with remote access. But it is off by default

  • @taladar
    link
    English
    51 year ago

    Is it just me or do all the AI generated porn images have something in common that makes them immediately recognizable? I am not quite sure what exactly it is but they do all have some quality in common.

    • @pazzedqM
      link
      English
      31 year ago

      Its the uncanny valley effect I think. Your brain recognizes that something is wrong in the image.

      • @taladar
        link
        English
        11 year ago

        That might be part of it but I think there is also a sort of sameness, certain aspects that are always the same in the generated images and more varied in real ones. Possibly related to the averaging of all the input data.

  • @casnsfw
    link
    English
    41 year ago

    Did anyone manage to find out what model sites like anydream.xyz might be using? I have hard time emulating the style…

  • @yore
    link
    English
    31 year ago

    Forgot to add, when experimenting with prompts. Use a fixed seed number so you can see how your prompt changes effect the image between each generation

  • @dense_crane44
    link
    English
    31 year ago

    It would be amazing to have a list of remote web hosts that just take a prompt and spit out an image. Some I know:

    I remember a quick and uncomplicated site that gave access to some models and took a prompt. It wasn’t used too much and pretty fast if you selected a non default model. Sadly, I don’t remember the name but I miss it. It didn’t have a whole lot of bloated J’s framework in the frontend like the ones mentioned above.

  • @yore
    link
    English
    31 year ago

    This is a great guide and was really helpful when I decided to experiment to see how this works.

    A couple of things that confused me when trying this out that might save the next person some time:

    1. where to put models etc *.ckpt and *.safetensors files live in stanle-diffusion-webui/models/stable-diffudion These will automatically be loaded when you new start wrbui-usr.bat

    2. how to change models This took me waaay longer to figure out than I’d like to admit. There’s a drop down top left of the webui to select the model after you restart

    3. I find the range of models, loras, checkpoints, extensions etc overwhelming. Im still not sure exactly what each of these do and which ones I’d need. Eg: Whats a checkpoint for?

    4. prompt writing is clearly a fine art and can drive you mad. For both 3&4 I found civitai.com/images to be a fantastic resource. Browse through the images for styles or images you like and most of them will have the resources used and generation data there to recreate it. I found this to be a great starting point, particularly for negative prompts.

    5. deformity Deformed faces have mostly gone away for me by changing this webui setting: settings> face restoration > code former weight = 0 Just need to figure out hands and phantom limbs now…

  • gremlinwife
    link
    English
    32 years ago

    Why did mine get downvoted to oblivion

    • @NSFW69
      link
      English
      12 years ago

      deleted by creator

  • Porn Account
    link
    English
    21 year ago

    So what settings do y’all use when playing around with prompts, before you generate a few really nice HQ ones?

    • Cavendish
      link
      English
      51 year ago

      I use 600x800 (or 800x600) and 50 steps when experimenting with prompts. Then when i get one i like, I lock the seed, maybe do a few .01-.03 variations. Then lock the variation seed and turn on 2x HighRes Fix. That outputs a 1200x1600 that very closely looks like what I expected. (I’m using a 3090 with 24GB vram)

      In my experiments, i found that doing quick and smaller tests with a low number of steps, then increasing it for the highres would change the output too much. I settled on this so that theres less options to toggle back and forth.

  • @slimelord
    link
    English
    22 months ago

    I find Automatic1111’s WebUI to be pretty confusing and difficult to work with. I’ve switched to Easy Diffusion and I love it waaay more. It’s also a 1-click install program, all dependencies are already included

    • Triple UnderscoreOPM
      link
      English
      7
      edit-2
      2 years ago

      ROCm is flaky in regular consumer GPUs at the best of times. I’m surprised you could even get that far on a steam deck.

      Try the command line arg --opt-sdp-attention. You might also want to try out --medvram or --lowvram (4GBs is considered low when it comes to AI). although I have a feeling it’s just because of the custom nature of the deck’s APU.

      Your best bet would be to search for builds of ROCm, PyTorch and Torchvision that are specifically made for the deck, if such things even exist.

      • @HoldingMyDick
        link
        English
        11 year ago

        Can confirm on ROCm sucking donkey balls. Even with directml it’s a struggle.

        On 5700xt, tried both forks by isqqytiger or whatever his name is and vladmandic for directml on Windows and a1111 for ROcm on Linux.

  • @DudeWTF
    link
    English
    2
    edit-2
    1 year ago

    Is anyone here using SD with Blender or for stuff other than NSFW too? Looking to see if a laptop embedded RTX 4050 or 4060 is viable? What is the real limit here? Like can anyone tell me something like “8gbv just can’t do (XYZ)” or the iteration time becomes so long the workflow is not practical/sw crashes. Anyone running Linux on a 4050/4060? What kind of impact is there with storage speeds, what size of DDRx is used in practice?

  • Sine Nomen
    link
    English
    27 months ago

    I recently came across Krita’s AI plugin. It’s pretty easy to install (especially if you have nVidia or ComfyUI already set up). And lets you easily fine tune your images.

    • @pppromotions
      link
      English
      11 month ago

      I haven’t tried it, but Krita’s plug-in is very impressive to watch.

  • @Dugnuts93
    link
    English
    11 year ago

    Anyone else use the OpenPose control net? I’m a little confused on the head portion and which control node is doing what. Every thing else about the editor makes sense and is intuitive. Do the outer most nodes on the head affect head tilt and rotation? And are the inner nodes eye placement? this seems to be the case but I’d rather not end up confused and mad when I’m inevitably wrong and their heads are backwards and eldritch abominations. lol