/sdg/ - Stable Diffusion General

Old Thread Mech Bulborb Edition

>user Guides
Training guide: rentry.org/informal-training-guide
voldy guitard: rentry.org/voldy
voldy guide 2: rentry.org/zfawb
inpainting: rentry.org/drfar
textual inversion: rentry.org/aikgx
AMD (docker): rentry.org/sdamd
AMD (onnx): rentry.org/ayymd-stable-diffustion-v1_4-guide

>Web UIs
voldy: github.com/AUTOMATIC1111/stable-diffusion-webui
hlky fork: github.com/sd-webui/stable-diffusion-webui

>Krita Plug-ins
github.com/sddebz/stable-diffusion-krita-plugin
github.com/nousr/koi

>Colabs
voldy: colab.research.google.com/drive/1Iy-xW9t1-OQWhb0hNxueGij8phCyluOh
hlky: colab.research.google.com/drive/1AfAmwLMd_Vx33O9IwY2TmO9wKZ8ABRRa

>In-Depth Usage Info
Any Forums wiki: wiki.installgentoo.com/wiki/Stable_Diffusion
SD Akashic Records: github.com/Maks-s/sd-akashic

>Tools & Resources
big index: pharmapsychotic.com/tools.html
artist styles: sgreens.notion.site/sgreens/4ca6f4e229e24da6845b6d49e6b08ae7 | rentry.org/artists_sd-v1-4
prompt sharing: lexica.art/ | krea.ai/
dataset: rom1504.github.io/clip-retrieval/ | laion-aesthetic.datasette.io/laion-aesthetic-6pls/images
interrogrator: replicate.com/methexis-inc/img2prompt

>SaaS Sites
txt2img: huggingface.co/spaces/stabilityai/stable-diffusion
img2img: huggingface.co/spaces/huggingface/diffuse-the-rest
txt2img, no account needed: dezgo.com
txt2img, NSFW: pornpen.ai

>Finetuned Models
huggingface.co/hakurei/waifu-diffusion
huggingface.co/naclbit/trinart_stable_diffusion_v2

>Related
/h/ /sdhb/: /vg/ /aids/: /mlp/ /ppp/: Official: discord.gg/stablediffusion
NSFW: discord.gg/2ZhBzDGf8u
Training: discord.gg/xVsyrmhQWS
Stability.ai updates: twitter.com/EMostaque

Attached: 1662950199848-5.png (512x512, 240.24K)

Attached: chex3.png (1024x1408, 1.63M)

Attached: 3148.png (512x512, 525.8K)

do we like sweater dresses here?

Attached: grid-0003-219749782.png (1024x1536, 2.01M)

nice picture. reminds me of megaman enemies. Which would be great for SD maybe.

Stop spamming this

God save the King

Attached: 1662916816441694.png (448x640, 392.32K)

grid. It's not as interesting as you think it is.

Attached: 00102-2611717610-cat_girl,_medium_shot,_anime,_extremely_detailed,_intricate,_((clear_focus)),_((sharp_focus)),_perfect_face,_very_deep_eyes,_((r.png (512x512, 308.03K)

I guess setting CFG from 7 to 9 gave more yarniness

Attached: 00013-2491312614-landscape_with_flowers,(((yarn_art))),photograph,cinematic_lighting,realistic.png (512x512, 441.9K)

I like this one more than euler_a

Mountain river valley, by Ivan Kramskoi and Makoto Shinkai
Steps: 2500, Sampler: Heun, CFG scale: 7, Seed: 232122911, Size: 2048x512

Time taken: 4752.52s

are steps not linear? I was kindof under the impression they were, but it seems more steps stays blurry for the longest time, and only gets fine detail at the very end. This one didnt start to really clear up until 2400 steps.

Attached: 00715-232122911-Mountain_river_valley,_by_Ivan_Kramskoi_and_Makoto_Shinkai.png (2048x512, 1.75M)

one of ya tits is fucked up bruv

Okay, so here is a problem I have.
I generated pic related.
Originally it had a fucked up face.
So I cropped out a square around the face, dirty upscaled it in paint.net to 512x512 and popped it into img2img, it came out with the face you see. I then placed that back on the image roughly where I wanted it, and at the scale I wanted it.
I took the whole thing as you see it, into inpainting. I masked off the face and like, her necklace basically because I didn't want it to touch that.
I planned to use a low denoise level, and a similar prompt to what I generated the image with, but then it goes totally nuts and doesn't seem to follow the image at all.

Have any of you done this type of thing before?

Attached: 1662861896770-9-input.png (512x512, 382.33K)

Any advice on img2img based on real people? I have a willing model and a few photos but it's super hit or miss. I'm just masking the face and some hair and letting it fill everything else based on original, if I clear the background it just poops out but 99% of the time it just ends up with a face floating randomly in the image.

on voldy's colab, it links to a very nice webui that i open in a new tab
problem is that i tend to time out when using it because i'm not interacting with the colab notebook
any workarounds for this? or do i just have to remember to switch tabs every once in a while

this might be the best one it has made for me in a minute

Attached: 02663-365099941-landscape_with_flowers,_trees_and_a_river_and_a_bridge,_(made_of_cotton),_photograph,_cinematic_lighting,_realistic.png (576x640, 752.19K)

..why are you using 2500 steps with an ancestral sampler?

Attached: image.png (362x116, 20.84K)

Can you make her wear a micro bikini?

looks like shit

nice style, which artist did you use in the prompt

no artist, the full prompt is in the filename

Got it. Had to use masked content - Original bubble.

Nice. Looks mostly OK. Might hit a couple more seeds to get it better but this is roughly what I wanted.

Attached: 1662954049288.png (512x512, 369.5K)

by the way, (sultry) crazy eyes, detailed pupils, Is fucking amazing

Attached: grid-0006-1718787222.png (1024x1536, 1.97M)

Carne Griffiths gives such insane results.

Attached: 00064-2201299011-painting_of_an_eldritch_wheel_of_eyes_with_many_enormous_feathered_wings_at_midnight,_in_the_sky,_atmospheric_haze,_by_Carne_Gri.png (1152x896, 2.1M)

Attached: -1662951549.2978423-1805847468-a_(busty)_((anthro))_furry_female_lynx_((fursona))_in_a_space_station,_((wearing_an_open_jacket)),_wide_angle,_nikon_fisheye,_60.png (512x512, 381.53K)

I have 8GB VRAM, but I would like to use one of the optimization settings from voldy's repository just to normally use the rest of my programs. Using medvram and opt-split-attention both drop VRAM usage to ~45%, which is way too low and slow, I am ready to give it some more to get more performance. Which one should I leave enabled?

Attached: 00002-1613753114.png (640x448, 640.57K)

fantastic news friend

I find to get anything reasonably interesting, I am pretty often going between img2img and some basic editing in gimp, or photoshop if you have it, whatever. Treating it like a one-stop black box solution is kind of fun for exploring or playing around, but it's better to give it a little guidance with outside tools. Even basic bitch copypasting and smudging in gimp works a lot of the time. Photoshop would make this part exceptionally good. There's also the krita plugin which people seem to love, haven't tried yet.
But basically, get involved in the process. Erase those extra faces. Also by playing around, you'll get better at knowing what CFG/denoise/steps should be, when to inpaint, blah blah. Make a batch of 4, take the best. Or a batch of 10, take the best parts from each, and repeat.

heun isnt ancestral? I'm testing shit man.

thanks

I know, but I did a git pull since I made that Brit-waifu and it changed the random seeds, so now I can't make corrective variants of her ;_;

Huh. Anons this is a wild one. I trained a textual inversion model on a woman's face to see if it was a viable means of referencing specific people. I ran my next inversion model in the same Colab window and apparently it didn't overwrite the old output file, so I ended up with a duplicate of that first model, which I then downloaded and named vintage-scifi.bin, thinking it was my new model. What's wild is, I never ran that model on it's own, so I didn't realize the mistake. I swear to you both though, I have been running that inversion model as part of my prompts which is rather hilarious. I'm running a batch right now, but as soon as it's done, I'm going to check that same prompt with the faulty model removed and see what kind of difference that makes lmao
Interesting, I noticed the opposite when I tested, but I could also be wrong given what I said above regarding the incorrect nature of my one inversion model

Attached: 00440-2874377963.png (1088x512, 837.19K)

khajiit has wares if you have scrip

ah I misread the post my bad it's almost 6am

I'm not confident in this, but I believe opt-split-attention needs to be run alongside a VRAM specification, so if you're going to run one on it's own, I'd recommend running just --medvram. Conversely, if I'm wrong then --medvram is absolutely what's tanking your usage, and you should run with just opt-split-attention. I'd experiment with both and see which works better.

Can you sure the actually correct SF model .bin then?

alright. is gimp the free one? I used to have cs3 but that computer caught on fire.

Attached: -1662953284.1530843-710044974-a_busty_anthro_furry_female_big_cat_in_a_space_station,_(wearing_an_open_jacket),_wide_angle,_hasselblad,_60s.png (512x512, 406.39K)

Lol I thought it looked pretty good

nta. GIMP is free, may be a bit unusual at first, but there is a settings preset for those used to Photoshop. I use it, it's good enough to make asset bash for img2img and make a mask. But CS3 is based, last good version basically.

gimp and krita are free. krita is probably better, I'm just used to gimp and don't do anything very complicated with it.
Some people were really raving about the SD plugin for krita, it probably is worth checking out.

Attached: 02686-1396843903-landscape_with_flowers,_trees_and_a_river_and_a_bridge,_(made_of_wool),_photograph,_cinematic_lighting,_realistic.png (576x640, 813.73K)

Keep in mind, doing an image at 100 steps vs how an image looks at 100 steps when doing an image at 200 steps are not the same thing. It tries to make it converge at the step count you give it.

Also, with an ancestral sampler, 20 steps is often more than enough and over 100 steps is probably meaningless. Ancestral samplers quickly get to the goal and steps after that just change the image. Other samplers move toward a specific goal so more steps means more refined, but ancestral samplers just keep changing randomly through all the steps it goes through. Doing 2500 steps is completely pointless and wasteful. Try the same image at 20 steps or so.

>I have been running that inversion model as part of my prompts which is rather hilarious

Yeah no this isn't true, because if you had been using it you would have noticed almost instantly. That model inserts her face into at least half the images you generate, even with totally unrelated prompts

I don't know what the point of this ruse is, you wasted like 3 minutes of my time at best and I assure you my time is worthless

Bros how do you cope with the fact that big corporations like google, apple, and ibm have already achieved perfection as far as generative AI while we're still soaking up the scraps because they want to profit off it.

user that's the problem, I literally don't have a model that was trained on those vintage sci-fi pictures. Somehow that model of the woman's face is tying into my prompts and not completely fucking them up. That being said, I'm working on retraining a model with the original photos and as soon as I have that, I will absolutely post it here, but it won't allow you to replicate my images since again, I'm producing them using the face model, which is absolutely hilarious

Attached: 00813-3820325772.png (1088x512, 754.23K)