I Don't Agree with your Disagreement Edition
Discussion of Free and Open Source Text-to-Image/Video Models
Prev:
>>105758237https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Models, LoRAs, & Upscalershttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
>Cookhttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanX (video)Guide: https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1
>ChromaTraining: https://rentry.org/mvu52t46
>Illustrious1girl and beyond: https://rentry.org/comfyui_guide_1girl
Tag explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage | https://rentry.org/ldgtemplate
>Neighborshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/celeb+ai>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>>105761419 (OP)>>>/vp/naptis your neighbor
update your neighbor lists
>>105761336noted kek
>yet another horrible autistic collage where the baker intentionally skips people
FUCK YOU
>>105761429this is better.
blessed thread of shitfits
>posts one image per thread
>always makes it into the collage
kino
>>105761454what a coincidence
The baker is literally a schizophrenic
imagine hating illustrious
imagine saltmineposting daily
imagine being this fucking unhinged\drunk
its not fucking worth it, it really just fucking isn't
Is Kontext capable of outpainting? Or is it only edits?
>>105761479It's a direction just like anything else. It can zoom in and out.
>>105761479it works like any other dit model for generating txt2img so maybe
>>105761429LITTTLE SHIT!!!!!!
>>105761479its mostly only editing. Like a local handy photoshop to make small adjustments.
>>105761430Cool style anon
lol even ani doesn't like the spam
Is lumina2 anime good yet
the anime girl is sitting in a chair, watching TV in her living room. A window in the background shows a sunny beach.
kita aura
>>105761561the anime girl is sitting in a chair, watching TV in her living room. She is wearing a white tshirt that says "LDG" in black text, blue jeans, and white sneakers. A window in the background shows a sunny beach.
peak comfy
>>105761429So announcement #20 to leave "forever" hold for like half an hour
Does anyone have any techniques to add a bit more polish to a sex scene?
I'm getting better at going through a set of positions, adding things like sweat or motion lines though the scene, changing the facial expressions, but what do you anons do to add that little bit of polish?
should I try out a few loras like - https://civitai.com/models/1541642/implication-off-screen
and
https://civitai.com/models/797890/offscreen-sex-nai-vpred-or-pony-or-illustrious
into the mix? How do you guys to consistent camera zooms? any seldom used but good tags?
I don't wanna use flux or flux kontext because I know max versions exist and my results are sub par
file
md5: 0d9fd285f23d61f3f42bcfb8c7a5073b
🔍
>yes you can zoom out with kontext>>105761595sexo. SEX. SEX! SEX!!!!!!!!!!!
ROUWEI TESTING (https://civitai.com/models/950531?modelVersionId=1882934)
I can confirm that the text capabilities of RouWei are significantly better than noob. However, I'm doing something wrong and it's frying my images. anyone know what I need to change?
https://files.catbox.moe/2zelol.png
I haven't found any instructions or workflows online for using RouWei in comfy.
>>105761595>add that little bit of polish?use a cool traditional media style
>>105761605Is this a nsfw model or is WAI still the go-to goon checkpoint?
>>105761605My first guess would be the prompt considering how precise the author seems to be about that. Looks neat and vpred, too. I'll download it and try it out thanks anon.
>conditioning zero out to the negatives Is that the same as having an empty negative prompt?
>>105761611I've been doing colored pencil, traditional media, but I should try the other mediums like paint
>>105761617I like to mix in floating hair, bouncing breasts, motion lines, once the action starts
Can you do txt2img with Kontext, and if so is it any different/better at it than Flux Dev?
>>105761575same prompt except rei as input:
>>105761630>Is this a nsfw modelyeah, https://files.catbox.moe/0e5ki8.png
if I can get this un-fried it may be worth some deeper testing for NSFW purposes
>WAI still the go-to goon checkpointThat's ΣΙΗ_illu_noob_vpred. Wai is OK but is lacking in styles.
>>105761646>My first guess would be the promptyeah, I think I need to install an extension that enables BREAK, but that doesn't explain how fried my outputs are.
>>conditioning zero out to the negatives>Is that the same as having an empty negative promptI'm not sure, IIRC there was a difference in my testing
>>105761595>any seldom used but good tags?Bookmark this, it's very useful for finding tags:
https://danbooru.donmai.us/related_tag?commit=Search&search%5Bcategory%5D=General&search%5Border%5D=Cosine&search%5Bquery%5D=breasts
also i categorized 59,000 tags here:
https://github.com/rainlizard/ComfyUI-Raffle/blob/main/lists/categorized_tags.txt
but that's more probably useful for tools rather than people
>>105761676diff image, better result:
>>105761684>Wai is OK but is lacking in styles.Maybe but I don't like Noob's retarded prompting. When I feel like writing an essay, I just stick with Chroma
>>105761713ΣΙΗ_illu_noob_vpred is the easiest noob model I've ever used. illus prompts are completely portable to ΣΙΗ.
>>105761741Can it eat illustrious loras? All my goonshit is for WAI.
Is it possible to change the size of Kontext output images? Or is it always going to be the same size as the input image?
>>105761713if you ever throw one of the WAI example prompts into Noob you'll see you don't really /need/ a lot of tags to generate something good with it. Noobs outputs surpassed it desu.
>before getting into ai
>played video games, had hobbies
>spend 99.9% time in comfyui gen'ing porn now
>cant play games because gpu constantly in use
>no time for hobbies
>discord buddies think im dead
ai has ruined me.
>>105761750I'm a proompter and don't use LORAs a lot, but illus LORAs frequently work on noob so you should test some and find out
>>105761767specified clothes, now it's better
the green cartoon frog is on a fishing boat in the ocean. a cooler nearby is filled with beers. the frog is using a fishing rod and sitting on a chair. the frog is wearing a red tshirt and blue shorts.
>>105761765>Can easily ask for girls phone numbers at social events.>Talk and text to them.>Realize they talk and text like literal shallow NPCs.>Rather chat to LLMs instead.It's over....
>>105761765>no time for hobbies AI is the hobby now
with a ksampler what would happen if the seed was randomized inbetween each step?
i have a theory that it might be good for upscaling
>>105761685thanks for the info
>>105761671I just tried giving it a totally black 1344x768 image and said "replace this blank image with a landscape painting" and it worked. That might as well be txt2img.
>>105761765I think the gacha element of AI makes it worse in terms of it being addictive. Like one is trying to fine tune a slot-machine. I can only imagine the hell it could be if, say, a 30 sec video model is released and people try to constantly hammer away the tiny imperfections.
weebsip2
md5: 63d52ff206e21d5f33f05047a51ee26e
🔍
>>105761765>tfw uninstalled most of my games to make room for models
>>105761605try one of the following, i can't be arsed to open comfyui rn
>use model sampling discrete with vprediction sampling and zsnr enabled>use the built-in checkpoint vae
>>105761798setup 50x ksamplers, set them all to 1 step, then pass the output to the next ksampler with the different seed and post the final result
the green cartoon frog is wearing a NASA spacesuit with no helmet, and is on the surface of the moon. A cartoonish looking Earth is in the background. The frog is planting a flag with a similar looking frog on it that looks like his appearance.
>>105761809I was only ever keeping up with one game(genshin), but i've all so far behind now that it feels impossible to get caught back up considering how much content they drop. i want to get back into it but then i remember all the shit i could be gen'ing or experimenting with. i just can't do both.
>>105761801although the result is trash aesthetically, even more slopped than flux dev without loras. looks more like a render than a painting.
>>105761807this. comfyui isn't exactly a good choice for just finishing something. I think workflow is a bad word to describe the graphs since you still have work to do most of the time. you can get it out but then you spend nine hours automating a niche gen into a convoluted spaghetti nightmare and have to change it for every particular solution to get something done
>>105761684>ΣΙΗ_illu_noob_vpredsorry, but what model is that? I don't understand the first three symbols and search didn't bring up anything.
>>105761798you sample 1/steps amount of data points from different Gaussian distributions with a mean of 0 and 1 sd. nothing magical happens, you simply get a quite uninspiring mix of various distributions.
>>105761833if you make it convoluted, it'll be like that.
>>105761807Imagine waiting hours for a 30 second gen only to have it go off the rails and fuck up in the last 10 seconds.
I don't look forward to that.
>>105761854https://civitai.com/models/1217645/sih
>>105761811seeing some improvement but it's still pretty fucked up
pale
md5: bf7effb6d862306034b3b95ab87acbf9
🔍
repost. I have a sfw gen in the queue to post shortly.
>>105760700 (You)
>>105760700 (You)
anyone? Like way too much noise left in the image.
Anyway, I loaded the basic wf and fixed it back up, it's working fine...
....I must have changed something, any idea what in the Kontext wf would do that?
Man kontext is some weird magic. Like sometimes you can get it to generate similar faces.
I haven't seen a discussion of not-quite-a-likeness as a feature. But it's got me excited. You can generate a SIMILAR person.
>50/50 NoobAI vPred/Rouwei vPred
>>105761713chroma? that rarely needs an essay, but you can prompt fairly extensively
>>105761753most workflows I've seen explicitly pick the size
>>105761671> is it any different/better at it than Flux Dev?to me it seems on average no but I haven't really tried it much yet. it might have its niches. it's also annoyingly censored I think.
>>105761941>NoobAI vPred/Rouwei vPredI lost track because I kind of switched to Illustrious. What even is vpred? Black backgrounds?
>>105761901I just remembered why.
Teacache is not fully compatible. It can work.
>>105761949Better colors. Something like that
I'm the guy that had an issue with flashing looped videos and apparently it's this thing here. Even with the 'fun' model loaded it seems better to have this off.
No idea what it does apart from make the loop freak out in the last few frames. It still happens with one frame, slightly, but it's better than a few frames of blindness.
the anime girl waves hello and smiles.
>>105761973"better colors" is so vague. There's a thousand different techniques you can use to improve colors aside from using an entirely different model.
>LORA on civit.ai
>none of the replies in the image gallery are using it
>>105761995now with no floating drink
>>105761999More accurate colors compared to the greyness of regular XL
Change the text "DOOM" to "SAAR". Replace the man in green armor with an Indian man holding a sign saying "REDEEM".
Trying to into video, the WAN guide on /gif/ I can figure out most of but am using Linux not windows so cant run the bat file so I've no idea what plugins I'm missing or how to set it up.
Does a more general/linux specific guide for WAN exist?
>LORA on civit
>completely redundant to using a tag
>>105762055it works basically the same way, pull via git, install dependencies and stuff via [uv] pip, drop the models in the correct folders, install the rest when comfyui is running via the manager?
stability matrix also works on linux if you want to make the install of comfyui and other uis easier
>new snake oil drops
>no comfy node
>>105762054i had to try the meme on kontext
Gigant
md5: a2910eea5c3af16241d1af8b4bd96163
🔍
Found this image online in the wild. Kontext API version is absolutely wicked bros. Shame about local
f
md5: fc1bd1a5bd0e75dc0c7d594c4e4655e3
🔍
her skin has suffered enough
>>105762357Pretty damn smooth
its like you guys can't go an entire hour without seething
comfyui tells me "please install xformers". I was on the nightly pytorch, cu128 & couldn't get it working. so I downgraded to the latest stable pytorch 2.7.1, still no cigar.
"pytorch version: 2.7.1+cu128
WARNING[XFORMERS]: Need to compile C++ extensions to use all xFormers features.
Please install xformers properly" bla
is there a way to make this work? does xformers need pytorch 2.7.0?
so is detail calibrated just shit or what?
>>105762466I used this guy's .bat because I wanted to install sage attention and the previous dozen attempts didn't work. https://www.reddit.com/r/StableDiffusion/comments/1jdfs6e/automatic_installation_of_pytorch_28_nightly/
But I had to change line 78 to ..\python_embeded\python.exe -s -m pip install --pre torch==2.8.0.dev20250619+cu128 torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cu%CLEAN_CUDA%
and line 93 to echo ..\python_embeded\python.exe -s -m pip install --upgrade --pre torch==2.8.0.dev20250619+cu128 torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/nightly/cu%CLEAN_CUDA% -r ../ComfyUI/requirements.txt pygit2 because the active nightly of pytorch was broken.
MASSIVE fucking headache to get this stuff working.
>>105762459shut the fuck up
So Kontext is nice. But the details get messed up for anime style. So close to making LoRas redundant.
>>105762556You can use the version with Enhanced Style Transfer right from ComfyUI! Here's a workflow to get started: https://docs.comfy.org/tutorials/api-nodes/black-forest-labs/flux-1-kontext
>>105761463I too downloaded the graphis torrent for training data, before I realized they badly photoshop their eyes bigger in 50% of images and put it in the trash. What an absolute waste.
>>105762556too much <safety> to be close
>>105761890Got a link for the merge?
>>105762524thank you very much. headache indeed. I don't even remember what I need xformers for anymore, my brain is all mush. thanks again, copied and saved!
>>105761684Hmm, try res_ samplers (I used res_2M_ode) and sgm_uniform sched. I've been hovering around 3-3,5CFG and 20-25 steps and it seems less baked.
>>105761419 (OP)>>Cook>training resources Don't models "bake" not "cook"?
>>105762357booba https://files.catbox.moe/1o6v4w.png
>race race race
>race to the bump limit!
garbage
>>105761809Same bro, every month I see what's remaining in my Steam folder on my 1TB SSD to be uninstalled
>>105762466>xformersuse sage attention 2, nigga
xformers is only good for sd1.5
>WARNING[XFORMERS]: Need to compile C++ extensions to use all xFormers features.or install 0.0.30 it doesn't have this warning
>>105758343reminds me of tf2 insta-gibbing, i wonder if that's in the training data
>>105762748I am running sage attention (wan rentry setup from op basically) but something wants xformers. I was messing with general object detection stuff. mesh graphormer, segment anything/grounding dino, florence2, that shit. & thanks for tip
does nunchaku give your different image each time? Same settings & seed
>nvidia, github, and runpod supported the recent comfy event
dang
>>105762843that's the power of selling out to saas. comfyui is now the world's most powerful API model provider
>>105762824no should be reproducible
>>105762860he's simply the only one who got and continously had his shit gogether.
>>105762948>he's simply the only one who got and continously had his shit gogether.it's more like he had a smart idea to make a software that is easy to make custom nodes on, that way it means people do the work for him with custom nodes, Kijai is doing it for FREE
>>105762961>Comfy Photographing Kijai
>>105762843nobody tell them
>>105762906me after eating entire pitza
>>105762948>no should be reproduciblewhich version you are using 0.3.1 or 0.3.2.dev?
>>105762961what's this about? did kijai upset someone?
>>105761605I don't know what's wrong with your workflow and it produced black images for me so I didn't try to debug, but you can check mine. It's heavily recommended that you put style and quality tags inside an isolated CLIP chunk.
My current impressions on RouWei is that it has better prompt adherence than IL, Noob, and the merges I've tried; it might have wider artist support (or at least "different"); it seems more inflexible with tag weights on artist styles; and it has a strong bias for a very "RouWei" look that I'm not a fan of (orange tint, depth of field, and other qualities I can't describe off the cuff).
>>105762990>they called him all sort of names, but never a liar
test
md5: 9f34556ec726d498ba1dfb02d1118290
🔍
>>1057629860.3.2. ok I get a pretty large variance now. but the left one was done with pytorch 2.8.n & --fast, the right one with 2.7.1 & w/o --fast
>>105762961yeah (and we pay the price for that wild west ecosystem every day)
the evaluation system is so dumb on civitai, everything has 100% positive reviews because people review only if they post something
>>105763089I agree. You use this stupid binary system for social media slop, not to review something. There should be a scoring system for various criteria on a 1-5 scale.
>>105760637>>105761061This would only bug me if Max used as much VRAM as Dev, but I have a feeling it's probably closer to 80GB.
>>105762086>using the activation words without the actual LoRa gives better results
>>105762999>strong bias for a very "RouWei" lookindeed. im not sure if i like it mixed with noob, maybe for coom. dunno why the second pass brings the style out like this
>>105762999>strong bias for a very "RouWei" lookindeed. im not sure if i like it mixed with noob, maybe for coom.
>>105763171could be fail merge
>>105761605res_multistep
sgm_uniform
20 steps
CFG 3
Also idk what VAE are you using because you haven't hooked up the vae included in the model.
And simple scheduler removes the blue tint altogether.
>>105763239No nevermind, this was a fluke.
>>105763239Yeah, on this topic, I've noticed that karras with vpred models will have serious malfunctions unless you use custom sigmas, so I stick to simple scheduler.
sche
md5: 458467a7310e158265fddf8380093afb
🔍
What the fuck is with the creepshot/pov vignette that Rouwai does? Also It gave me several outputs with blurred pussy and black bar censors lol. This model is cooked.
>>105763472Holy shit, my brain is so accustomed to censors I forgot that more than half of my booru favorites have it. That must be a nightmare for datasets.
Meme
md5: b85e33b69667ac81802681b234c600dd
🔍
Is there a magic trick to fix Comfy when it randomly decides to hang and not display random nodes, without just restarting it completely? The program isn't completely crashed, it still processes the queue in the background but I can't interface.
>>105763610refresh the page?
>>105763610whens the last time you updated, boss
Not that most of you care, but:
lodestones
Upload chroma-unlocked-v41-detail-calibrated.safetensors with huggingface_hub
this
md5: 056b3d4d44b29f151f8b94a396609277
🔍
>>105763619>Not that most of you careyep, I really don't care, Kontext Dev is my new friend
>>105763616Closing the entire thing sort of works but I have to wait for my prompt to finish to avoid other problems. But I'm mistaken, it's not a random glitch, my workflow consistently implodes after a single prompt.
>>105763618This started after I updated 2 days ago, never happened before that.
>>105763594there's some artist with this style just cant remember who
>>105763619gguf waiting room
>>105763610the interface can be a bit glitchy but I've never had something like that happen. (using comfy since 2023)
>>105763619this is a very nice gen anon (goes for the others you posted as well, good shit)
>>105763619>>105763640>this is a very nice gen anon (goes for the others you posted as well, good shit)
>>105763636Ok if I press it once, close the tab, and reload it, I can then keep pressing the queue button until the first once finishes, at which point it breaks again.
skull
md5: 1d23d75fe850eb87988185e8988184a5
🔍
>>105763619>>105763640>>105763648https://www.nimh.nih.gov/health/topics/schizophrenia
>>105763639I can't remember it either
>>105763640Thanks man.
>>105763693I'm just posting gens because I think Chroma's neat. I don't care about the drama surrounding it, I just like that I can get output I like without using LoRAs. But perhaps my use case is a bit specific since the styles I like are replicated pretty well and I don't really gen photo realistic stuff.
Any vace chads here know the best wan lora slop combination to get my gen to match my input image? It always changes the faces. I have;
>causvid
>accvid
>self forcing (kinda keeps the input image but moves poorly)
>lightx2v (changes the input image completely into flux face)
The closes I've gotten is combining causvid and accvid
Is the Mayli anon back?
did he finish sorting his folder?
To the anons that offered advice regarding "dictionary state" errors on kontext yestday, it turns out my env was fucked for Kontext (it works fine for everything else though???) so i made a fresh env and 2nd install of comfy and the default comfyui workflow worked.
>>105763755you replied to the same anon lol. yeah I understand, chroma has a way with shapes and forms and certain styles that's just beautiful, it just flows. and the upscales almost always come out really nice as well. not sure if you know this dude, pretty dope stuff https://civitai.com/user/TijuanaSlumlord
and the drama thing, w/e. it's just one nutjob whose horizon doesn't seem to reach past "asian, feet"
>>105762531https://huggingface.co/rocca/chroma-nunchaku-test
Though should wait for official one to release soon
https://github.com/mit-han-lab/nunchaku/issues/167
>>105763876What settings do you use for chroma upscale?
>>105763679Looks like one other guy has reported the same issue, and somehow it's apparently Flux+Firefox specific, which is my config as well. I tried an older Illustrious workflow and no issues.
https://github.com/comfyanonymous/ComfyUI/issues/4235
>>105761992Last time I had this problem it was bc a few nodes were fucking everything up, so try to see wich part of your workflow is doing that, then which nodes
>>105761992Are you looping with the fun model? instead of the base one nor vace?
>>105763911like this but it's always WIP. between 0.3 and 0.4 denoise, steps 10-14, bit of noise manipulation via detail daemon, either x1.5 or x2 size. some flux loras might work ok for the initial gens but I found that removing them or lowering the strength for the upscale seemed to yield better results.
sampler/scheduler, I need to run various XY plots to figure that out - not done it yet.
wtf
md5: 60b6a8e0c29440990a8d70d50178bd26
🔍
this "load image from outputs" node is a good idea but it's slow as shit
You can get rid of the manlet effect by decreasing the FluxGuidance value
>>105764051let's fucking goooooo (I'm gonna try to compile it)
>>105764020could be sweat
3060
12gb
16ram
do I have any hope of genning decent i2vids?
I don't normally use diffusion models, but I had a dream about a new type of diffusion model that could generate perfect text.
It generated long and detailed texts by segmenting the image and separately generating each letter one after another, akin to inpainting each letter.
Now I'm curious if there's any model that actually works like this.
>>105764108It will either take forever or OOM
>>105764061Ok, after some test, I got a 13.8% speed improvement overall (RTX 3090), I'll take it, the quality is the exact same too
>>105764167Why is AI so shit at text anyway?
Sometimes kontext has visible progress in the vae, then it just undoes everything and wastes the rest of the steps changing nothing.
>>105764051>>105764061>>105764170>the chinks didn't betray us after allI NEVER DOUBTED THEM
>>105764170Happy for you anon, my gcc is too new to compile it and i have to fuck around with other stuff first before trying a workaround.
I only noticed the release as i am just getting around to setting up sageattn on a new install.
>>105764051Cool, now I'm waiting for SA3 (unless it only got its speed improvement from the 5090 in which case I won't give a damn lol)
>>105764185You have to use a good LLM as the text encoder for it to do text properly.
T5 and CLIP are neither good, nor LLMs.
>>105764217>Happy for you anon, my gcc is too new to compile it and i have to fuck around with other stuff first before trying a workaround.lucky for you KJ God saved the day and added some 2.2.0 wheels here
https://huggingface.co/Kijai/PrecompiledWheels/tree/main
>>105764222The downside with this is it significantly increases VRAM requirements so it's usually an API model thing. I know Lumina uses Gemma but it still fails at text most of the time, though it's understanding of the prompt is significantly enhanced by using Gemma as a decoder instead of clip/t5.
>>105764167>I had a dream about a new type of diffusion model that could generate perfect text.4o imagegen is really close to that
>>105764222No, I mean drawing the actual letters, not text comprehension.
>>105763974Ok I debugged it apparently, the command window didn't indicate any error but deleting the U-NAI Get Text node (which wasn't essential but useful for saving the chosen wildcard) fixes it.
>>105764233Thanks for that anon, i'm a bit dim but i'll take a look at it and figure out how to use it to my advantage somehow.
>>105764061For those who said the SageAttention guys would gatekeep their code, APOLOGIZE
file
md5: 54fb1100d8545348efcf4ea677d96cb4
🔍
he can't get away with this!
>>105764061Is this good only for Wan and kontext?
>>105763876Damn, sorry. Didn't know that dude but I'll check him out. I see he's futzing around with LoRAs, which I never really did with Chroma since the results haven't been too great for me. Thanks man.
>>105764358Sage works on everything
>>105764061https://github.com/thu-ml/SageAttention/blob/main/sageattention/core.py#L140
>only works for sm80 or more (it means the 30xx series or more)VRAMLETS BFTO
I can't get her to heil, but that's okay
file
md5: b6eb4d57988f8a289104f8d71faa1e68
🔍
>>105764366So I can just directly git clone it and comfy won't explode when doing Chroma?
>>105764410you have to install the wheels, use kijai's one
>>105764233
>>105764325is it possible, nay probable, that this is in direct response to that very backlash?
So adetailer detects the faces and hands and makes them better quality using a custom model and the trained LoRa?
Is there any way I can also upscale the rest of the image? Clothes look low quality compared with the adetailed faces.
Normal Upscaling models don't make the cut
>>105764498Upscale first, and run detailer on the finished image.
>>105764061God bless the chinks
>>105764498>So adetailer detects the faces and hands and makes them better quality using a custom model and the trained LoRa?No, the only ADetailer-specific models are the detector models. It'll sample using whatever model you give it, usually the same as the model you used for your base gen.
>Is there any way I can also upscale the rest of the image? Clothes look low quality compared with the adetailed faces.Is there any way I can also upscale the rest of the image? Clothes look low quality compared with the adetailed faces.Hiresfix (before ADetail), or detail the clothing with manual masking. See the "Hiresfix", "Face and hand detailing", and "Inpainting" sections of the guide: https://rentry.org/comfyui_guide_1girl
>>105764051>>105764061if you want to see an improvement you have to upgrade cuda to 12.8 or more
>>105764582you'll get more improvement if your gpu is a sm89 (4090 or more)
https://github.com/thu-ml/SageAttention/pull/196/files
>>105764233this one isn't compiled with cuda?
https://github.com/woct0rdho/SageAttention/releases
does have wheels with cuda, but seems to say it only improves speed on 40XX and 50XX?
>>105764597>but seems to say it only improves speed on 40XX and 50XX?there seem to be 2 optimisations there, and the 40xx and 50xx will get the both of them
Soon wansisters...
https://github.com/Yaofang-Liu/Pusa-VidGen
>Extended video generation
>Frame interpolation
>Video transitions
>Seamless looping
https://github.com/mit-han-lab/radial-attention
>4× longer videos
>3.7× speedup
>>105764658can't wait to use those on the future Wan 2.2 ;D
https://xcancel.com/bdsqlsz/status/1939574417144869146?t=UeQG__F9wkspcRgpmFEiEg
https://www.reddit.com/r/StableDiffusion/comments/1loyav1/important_psa_you_are_all_using_fluxdev_loras/
Someone managed to make flux dev loras work on kontext
Can I leave the negative prompt in NAG empty or do I have to enter something?
>>105764704you can leave it empty, it works like cfg
>>105764189>Sometimes kontext has visible progress in the vae, then it just undoes everything and wastes the rest of the steps changing nothing.that's because the filter has been triggered during the process, can't wait for someone to do some abliteration and uncuck this shit
https://huggingface.co/blog/mlabonne/abliteration
>install cutting edge acceleration tech from bright chinese minds (actually got it running lol, thanks Kijai!!)
>back to genning SDXL smut
>>105764040but they got extra arms and shit
>>105764108check out framepack studio and wan2gp
>>105764233Does this autoinstall into wherever it needs to be if I just CD into comfy root?
do you need to upgrade trition for sage_attn 2.2?
I'm still at v3.0.0
has anybody tried new sage attention?
>>105764803[..]\python_embeded>python.exe -m pip install [drop your *correct* wheel here]
>>105764831oh yes
>>105764882It's impressive that even ampere got speedup when it seemingly targeted optimizations for 40 and 50 series
>>105764897s-stupid flatchested little brat...
>>105764893yeah, that update has 2 optimisations, and one of them is for all card
>>105764640
>>1057648933090 former chads not dead yet, sage 3 will probably fix that. 5090 prices coming down tho, 2300 now here for the cheapest one
>>105764882how to install or should i just wait till its properly released?
>>105764939>how to install- you need to upgrade to cuda 12.8
then you can inspire from 4) from that post to make it work
https://www.reddit.com/r/StableDiffusion/comments/1h7hunp/how_to_run_hunyuanvideo_on_a_single_24gb_vram_card/
the wheels are here
https://github.com/woct0rdho/SageAttention/releases
just made a pic of me and my elementary school (20-21 years ago) oneitis with kontext (nothing sexual) using pics of us from when the last time she blocked me when we were 19 (10 years ago)
God I need help
>sageattention-2.2.0-cp312-cp312-win_amd64.whl not a supported wheel on this platform
huh
same with sageattention-2.2.0-cp312-cp312-linux_x86_64.whl
nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2024 NVIDIA Corporation
Built on Tue_Feb_27_16:28:36_Pacific_Standard_Time_2024
Cuda compilation tools, release 12.4, V12.4.99
Build cuda_12.4.r12.4/compiler.33961263_0
ah, need 12.8
is it worth it though? does it break anything?
>>105765008>is it worth it though?it is
>does it break anything?no
>>105764966that's the cuda with torch and not the toolkit right?
seems slightly slower to me, torch 2.7.1 on 3090 using the normal rentry i2v workflow
>>105764857The cuck who was claiming they were gatekeeping Sage2++ must be seething right now.
>chroma-unlocked-v41-detail-calibrated-Q8_0.gguf
>>105765019if you want to create the wheels by yourself, you need the toolkit to be cuda 12.8, if you just want to use someone else's wheels, it'll be just cuda+torch
>>105764061NTA. Surprisingly painless. VS 2022 build tools and had to set DISTUTILS_USE_SDK=1 and it worked.
Maybe there was some bullshit I dealt with the previous time that carried over.
do I need the 12.8 toolkit to install these wheels?
>>105764061I forgot to enable --use-sage-attention with the previous version. I had it installed unused for months.
>>105765026>oh look another distilled version of chroma
>>105765057>do I need the 12.8 toolkit to install these wheels?yes, and you need cuda+torch to be on 12.8 aswell
>>105765026>oh look another version of chromaThanks anon!
>>105765079>another version*distilled version
>>105764386How is that okay ?
>>105764969>God I need help I recently celebrated re-discovering some toddler butt photos that I thought were lost forever on an archive site with a triple goon session. It could always be worse, anon.
But having a oneitis in elementary school at all is really cute. I didn't care about girls at all until middle school
>>105765026>mfw my favorite Chroma-fp8-scaled version is no longer being updated by Clybius...
>>105765026wake me up when i reach v69
https://www.reddit.com/r/StableDiffusion/comments/1lozgre/kontextdev_id_card_faking/
hello saar we can now easily falsify ID saar, Kontext is the best model saar
>>105765026>yaay number go upbut is it improving though?
>>105765138This could maybe save 1 euro 20% of the time at a museum by lying about having a student card and showing them a fake image but you could already do that with photoshop
>>105765161Nice I did this with shotas and Persian mommies
It's crazy how well WAN is able to generalize similar concepts it's seen before into never before seen kino
>>105765176>It's crazy how well WAN is able to generalize similar concepts it's seen before into never before seen kinonot only that but the apache is also kino, everything in Wan is pure kino, and we'll even get an improvement soonTM
>>105764667
>>105764667It will be GLORIOUS! I almost forgot Jenga https://github.com/dvlab-research/Jenga
>up to 6.12x boost>>105765020Good catch, might give it a couple of days before trying to install this
>>105765198>>up to 6.12x boostthe quality will suffer though?
>>105765186I already wasn't dooming for video for all of 2025 because of FusionX and lightx2v but we just keep getting improvements too. 720p at home in 2026 doesn't seem far fetched at all
>>105765185Nice wax museum bro
>>105765244Not that's some nigga begging in issues to add poorfag support
>>105765248oh I thought it was a PR and he found a way to make it work kek
So with pusa video extension (doesnt specify duration), radial attention with 4x the duration and a 3.7x speed up + sage2++ a 30 second video would take about 1 minute to slop up? I'm just guessing, my math sucks.
Currently my 4070tis, does around 1 min 30 sec for 5 sec clip (NAG, not new sage attention, VACE and lightx2v lora, 4 step lcm)
>>105765212Not a clue, we'll have to just wait and see
>>105765259>>105765300I gave up on sunny side up eggs a long time ago
It's not worth the rubberyness of fucking it up and also the extra work cleaning the yolk from the plate in the sink
I just stick to scrambled eggs and if I want that yolky flavor I make an egg sauce pasta like carbonara
>>105765242You can tell forrens on github, they phrase requests as demands, like children.
>>105764995>install it just fine>realize my python is too old>update it>not a supported wheel on this platform
>>105765329Can you share a workflow? OP wan rentry doesn't have a good NAG preset and sage attention update should just be drop in
>>105765331nigga what? I'm just posting my gens before i hit the gym.
Anon who shared the json for two images kontext, on which basis does it decide which is image 1 and which is image 2?
>>105765360Post one of those videos as a cat box then
>gymDon't forget to do wrist exercises
>>105765369image 1 is considered the main one by the model because it's the first on the reference conditioning cascade (and the workflow also fits the resolution as the same as image 1), but you shouldn't use "image 1" and "image 2", it won't understand, just reference the image 1 normally, and for image 2 you go for "other", like this
NAGGER
md5: d3306d30c872b2048ae4e1145e33c91d
🔍
>>105765355Its just this workflow https://docs.comfy.org/tutorials/video/wan/vace with kijai's version of NAG https://github.com/kijai/ComfyUI-KJNodes
You can leave negative prompt empty
does the op have sageattention 2 yet?
>>105765429Oh alright thanks. In my one test I did NAG doesn't play well with either lightx2v + vanilla wan or FusionX merge or FusionX merge + lightx2v lora. Faces get fucked up on t2v but thank you I am excited for more prompt adherence
>>105765186>>105765230feelin positive for videogen. wan is truly magical
>>105765387How does it work if you have two character in your first image, and want to swap one of them for the character in the second image? I tried going for "swap the white haired character by the blonde one" but it only swap the first character's hair
>>105765477separate the character, and put them on each "load image", that's better than stitching the images together
why does load image node have a very old image by default, how do I clear that/cache or make it blank
>>105765518>how do I clear that/cache or make it blankyou remove the images from ComfyUI\input folder
>>105765525ty, why is it saving cached images though?
>tfw you're not the goat anymore
>>105765560I guess people want to keep their input images to use them for another day
So this is it then
The last local diffusion we ever made
>>105765300those sausages are extremely penicular
>>105765154I noticed improves in furry subjects but not anime or realistic
Can't get two image kontext to work
someone share a .png catbox with prompt
Thanks xoxo
>ERROR: Could not find a version that satisfies the requirement sageattention-2.2.0-cp312-cp312-win_amd64 (from versions: none)
Does the desktop comfy not have it already in? I have to get the old version too?
>>105765791Can't get two image kontext to work
someone share a .png catbox with prompt
Thanks xoxo
>>105765800try this workflow and set it to 2 images (bypass the third)
https://openart.ai/workflows/amadeusxr/change-any-image-to-anything/5tUBzmIH69TT0oqzY751
the anime girl is standing beside the green cartoon frog. change the location to a sunny beach. the frog is holding a beach ball.
>>105765791>>105765800there's a full example here
https://www.reddit.com/r/StableDiffusion/comments/1lo4lwx/here_are_some_tricks_you_can_use_to_unlock_the/
do you need 2.8 torch to get sageattn2.2 working? i installed it fine and there is zero speed change
be
md5: ff4fb32bf042bd1eecd849077436a290
🔍
>ERROR: sageattention-2.2.0-cp312-cp312-win_amd64.whl is not a supported wheel on this platform
erm
are we not going to gen more?
>>105765842you can still fit another 10 images in this thread
>>105765778Thought same.
>>105765259>>105765300Phalic phrankfurter
>>105765841Use those wheels instead
https://github.com/woct0rdho/SageAttention/releases
sage_attention_2++ is the same as 2.20?
I've updated 2.20 from 2 and I don't get any speed boost on 3090 TI
Installed Comfy, tested a bit sd.3.5 and flux dev. Any other models I should be using for realistic, anime, stylistic styles? Does any of the default template/models do porn or I have to download another one? I tried Biglust, is not very good, though I just used the sd template to run it
kontext is surprisingly good at the stalin effect:
remove the man on the right.
>>105765850What's the cp39 and so?
>>105765882replace the man on the right with miku hatsune.
>>105765889it's the python version
cp 39 = python 3.9
>>105765861what cuda and pytorch version are you on? i am on 2.7.0dev and cuda 12.8 and i also do not see any speed increase on a 3090
>>105765861>>105765896>i also do not see any speed increase on a 3090I do, my wan gens take me 4 mn insteas of 5 now, maybe it's because I built the wheels by myself Idk
>>105765904kek, you probably have an old ass gpu
>ERROR: Could not install packages due to an OSError: [Errno 2] No such file or directory: 'C:\\Users\\Gaming\\Documents\\Downloads\\sageattention-2.2.0+cu128torch2.7.1-cp313-cp313-win_amd64.whl'
NIGGA IT'S LITERALLY THERE
>>105765879you can do cool shit with biglust but takes time to master those sdxl models. no idea what you are into tho, "model I should be using", kinda vague.
>>105765902I can do that with ms visual studio installed, correct? how long does it take?
>>105765929>I can do that with ms visual studio installed, correct? how long does it take?it takes ~10 mn, you can use this tutorial to see how it can be done proprelly
>>105764966
>>105765945he went on vacation
>>105765922I never put anything like that in the users folder. It's just cursed. C:\\downloads or something else.
How do I set desktop to use sage? There's no bat file to edit. Do I put it as a launch attribute of the exe?
>>105765837Yes, you need torch+cu128
anons, is there a trusted mirror of full size bf16 Kontext?
I'm not giving BFL my details
>>105766057Desktop. Not portable.
>>105766057I mean standalone, fuck.
https://huggingface.co/ostris/kontext_big_head_lora
>Kontext has the manlet effect
>he makes a lora that makes the manlet effect even worse
what does he mean by this? can we use negative strength to remove the manlet effect?
>>105766052if you're using comfyui, you already are
>>105766085You can try, any lora loader lets you set a negative value, I've seen anons using it like that in the past (with various success). Put the trigger phrase into the negs too
>>105766086haven't run Kontext on comfy yet
also, where in the code are they sending telemetry?
I can just comment it out
>>105766138>where in the code are they sending telemetry?there is non he's trolling
wow it's over
no more local diffusions
>>105766179it's only page 7, plenty of time for one of the bakers to make another (mediocre) collage.
>>10576410816gb ram is gonna be tuff, but u could try gguf q4_k_s with self forcing. unironically install linux because windows is bloated, and every drop of ram matters. you should probably go for a light-weight desktop environment like lxqt or a window manager (i3/dwm..)
t. 3060 12gb vram 64gb ram enjoyer
>>105764378fuck i feel bad, we're gonna be on the chopping block soon too, i wish someone could backport it to the 2000 series for my bros