← Home ← Back to /g/

Thread 106362104

320 posts 182 images /g/
Anonymous No.106362104 >>106362128 >>106362942 >>106366548
/ldg/ - Local Diffusion General
Cozy Weekend Edition

Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106358470

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://github.com/Wan-Video
2.1: https://rentry.org/wan21kjguide
2.2: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-HD/tree/main
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.106362122 >>106362148
I am white.
change my mind
Anonymous No.106362124 >>106362653
Blessed thread of frenship
Anonymous No.106362128 >>106362321 >>106363471
>>106362104 (OP)
How do i make these ?
Native Backend only please.
Anonymous No.106362148
>>106362122
Eyy holmes, don't worry about these gringos, I'm white too ese
Anonymous No.106362169
>>106362114
Lower your resolution significantly while you're getting a feel for the prompting. The prompts follow a general format of: Subect -> Scene -> Motion. You can read the official prompt guide here if you're interested: https://alidocs(dot)dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y
Anonymous No.106362187 >>106362225
comfy should be dragged out on the street and shot
Anonymous No.106362225
>>106362187
true. fuck this whole gay ecosystem
Anonymous No.106362231
>comfyui nunchaku still not update
AIEEEEEEEEEEEEEE
Anonymous No.106362250 >>106362299 >>106362311 >>106362858
Hehe, she spin.
blueCube No.106362276
How would one do those QR Code things with newer models? Like Illustrious, Wan, Flux, Qwen? This is a bad example as it doesn't add elements where the code is, it just makes it darker.
Anonymous No.106362299
>>106362250
kawaii
blueCube No.106362311
>>106362250
really nice gen. lovely.
Anonymous No.106362321 >>106362340
>>106362128
i can tell this is using the "everything is cake" lora on civit
blueCube No.106362340 >>106362404
>>106362321
I wonder if it was done with this one https://civitai.com/models/1337473/cakify-cake-everything-wan-14b-i2v
Anonymous No.106362404 >>106362420
>>106362340
yeah if thats the most popular one then probably

and if you can understand how genning crushing content like that saves the animals, you should be able to understand how genning cuties saves the children
blueCube No.106362420 >>106362615
>>106362404
I don't get the second part of your comment. lol
Anonymous No.106362467 >>106362489 >>106362543
>>>/g/sdg
Anonymous No.106362489
>>106362467
based namefag filterer
Anonymous No.106362503
shoot comfy in drag
Anonymous No.106362543 >>106362550
>>106362467
What's with this place and namefags? Seems like a new one pops up every week.
Anonymous No.106362549
Anonymous No.106362550 >>106362656
>>106362543
chan culture is dead. it's all tourist-tourists now
the only ones who care about anonymity anymore on this website have something to hide. the rest are all discord zoomers that dont care about their privacy at all
Anonymous No.106362615 >>106363024
>>106362420
>I don't get the second part of your comment. lol
here is a visual explanation
Anonymous No.106362620
Are there any guides on how to quantize models into FP8 scaled on your own?
Anonymous No.106362630 >>106362668
https://youtube.com/shorts/fFGZMZKFD1E?si=AHfuQOWMk5qcG2wy

Can local make something of this quality yet?
STATLER N WALD0RF No.106362653
>>106362124
BEAHAGHAHAH
blueCube No.106362656 >>106362695
>>106362550
I don't disagree. You're right. 95% of my posts are anon though.
Anonymous No.106362668
>>106362630
maybe months ago it was that bad. this is absolutely trash
Anonymous No.106362695 >>106362718 >>106362719 >>106362888
>>106362656
well of course you don't disagree you're a namefag lol

honestly i was a little inaccurate. everyone still on this website has something to hide. i think the argument that removing usernames results in higher quality discussion by focusing on the subject matter of the thread over thread personalities has been proven false many times over by the fact that there is a resident schizo in EVERY general (where did the deebster go btw? can we remove him from the OP? he only showed up like once in the last 2 weeks. I'm more of a schizo troll than him at this point)
blueCube No.106362718 >>106362734 >>106362942
>>106362695
I leave the name in because I post a lot in degen and respected that some might want to hide my content.

But I'll take the hint and remove it from now on.
Anonymous No.106362719 >>106362733 >>106362734 >>106362942
>>106362695
if we're all schizo trolls, then none of us are schizo trolls
Anonymous No.106362733 >>106363110
>>106362719
Oooh I like that art style is it a specific artist?
Anonymous No.106362734 >>106362763 >>106363110
>>106362718
oh right you're a /b/ namefag. well you shouldn't respect that imo, people can either learn to ignore or filter you through something else like file name or dimensions

>>106362719
cool gen thank you for sharing it. did you get your muscle girls fetish from a muscle girl sitting on you in elementary school or something
Anonymous No.106362763 >>106362772
>>106362734
I mean the /b/ thread is literally the only one that isn't consistently full of anal-retentive autistic faggots arguing about utter bullshit
Anonymous No.106362772 >>106362780
>>106362763
there was literally. like, LITERALLY an anal-retentive faggot with trauma about anuses in the last bake who thought rimjobs = poop kek

there's plenty of arguments and grossness but with the general decline in the interest in AI it's gotten better just through obscurity
Anonymous No.106362780 >>106362793 >>106362794
>>106362772
The real mental illness is thinking that assholes DONT mean poop. I’m not saying don’t eat ass, it’s fine, I’ve done it, just don’t delude yourselves.
Anonymous No.106362793 >>106362808 >>106362813
>>106362780
only if you think that vaginas and penis = pee as well. otherwise the logic is lacking which was the whole reason people were shitting (no pun intended) on the schizo
Anonymous No.106362794
>>106362780
>I’m not saying don’t eat ass, it’s fine, I’ve done it
real and gay (it was the ass of a male)
Anonymous No.106362796
imagine actually licking the stinker
Anonymous No.106362808
>>106362793
> only if you think that vaginas and penis = pee as well.
Indeed I do. Which is fine, there is nothing more patrician than enjoying the warm golden flow of a fair maiden upon your face
Anonymous No.106362813 >>106362820
>>106362793
>only if you think that vaginas and penis = pee as well.
but like, you DO know that women have a separate tiny urethral opening solely dedicated to peeing that isn't connected to their main "pussy hole" at all, right?
Anonymous No.106362820 >>106362835
>>106362813
i already covered this in the other thread. if you're licking pussy you will lick both. also when they wipe they wipe it all around their pussy anyways lmao and it dries
Anonymous No.106362833 >>106362861
>40 / 9 / 1
Anonymous No.106362835 >>106362861
>>106362820
yeah ok, i've just literally encountered many people who straight up didn't know there were two actual holes down there in front, so figured i'd check
Anonymous No.106362858 >>106362866 >>106363016
>>106362250
nice, how are you getting so much movement, all my lightx2v wan gens are in slow motion :(
Anonymous No.106362861
>>106362835
pee is stored in the balls

koalas have two vaginas

>>106362833
i have an hour free. what would you like to see anon? i have no limits i like girls younger than you do 1000%
Anonymous No.106362866
>>106362858
i suggest mixing the 2.1 and 2.2 loras in a 60/40 split and trying with that. great quality and good movement, not a lot of slow motion often
Anonymous No.106362882 >>106362933 >>106362942 >>106363175
Anonymous No.106362888 >>106362911
>>106362695
the reason to use this site is simply that there is no alternative for discussing this tech that is not either filled with shills/hustlers (eg reddit) or even worse schizos (discord).
Anonymous No.106362911 >>106363039
>>106362888
there's plenty of schizos, and we have our own shitters (myself included). you can't actually ban people on """anonymous""" platforms

i think a bigger reason is because pretty much everyone here is interested in some kind of "wrongthink" with AI, which is why we gravitate here. otherwise we could just ask questions on discords, and post our stuff to normie social media

of course, everyone has their own wrongthink which is why its interesting to keep coming back to these threads even if there's not a lot to talk about in between model releases
Anonymous No.106362933 >>106363066
>>106362882
Holy sloppa, I wanna make my own. What model?
STATLER & WALD0RF No.106362942 >>106362959
>>106362882
BWAHAHAH
>>106362718
>catering to the schizos here
>not outmaneuvering them at every step instead
N G F M i
bewahghahahah
>>106362719
**beaghahahs nervously**
>>106362104 (OP)
>collagebait
Anonymous No.106362959 >>106363000
>>106362942
how much of the muppets have you actually watched mr smith and wesson
STATLER & WALD No.106363000
>>106362959
NONE
>beahgahahaaha
just like this thread
i read NONE of it, cant bear to LOOK
BEHAAHAHFGHGAHAHAH
Anonymous No.106363016 >>106363051
>>106362858
I do 2 steps on the high model without the LoRA, that already provides a ton of motion. The rest of the samplers can use the LoRA again. Takes a bit longer to gen but gets better results in my opinion.
elf-hugger No.106363024
>>106362615
based art understander
Anonymous No.106363039
>>106362911
It's impractical to ask even benign questions on "normie" platforms. Reddit and Civitai are populated by a bunch of wannabe startup monetizers attempting to exploit people too clueless to know what git pull means, Github is a mixed bag but not really an appropriate place for discussion, and Discord is where the truly insane ones are. Sometimes Discord does get you useful answers if you're able to pass the gate, but even then the whole platorm should be avoided on principle for the harm it's done to the free exchange of information.
Anonymous No.106363051 >>106363141
>>106363016
Nice, are you using the 3 samplers chain then? I saw a similar wf on civitai
Anonymous No.106363060
Anonymous No.106363066 >>106363075 >>106363077
>>106362933
chroma
Anonymous No.106363075 >>106363251
>>106363066
What are your Chroma weights, settings, and lora?
Or a catbox to help?
Anonymous No.106363077
>>106363066
me when i use flash chroma
Anonymous No.106363105 >>106363146
Well, I was going to demonstrate a 4x wan 2.2 14b i2v gen, but catbox is either broken or something changed about uploads.
Long story short, 1 original plus three screenshot continuations of more or less the last frame of each subsequent video is as far as I could go. Things ended up very fried looking on the fifth gen.
Anonymous No.106363110 >>106363272
>>106362733
No artist or lora used, main thing style tags were

>[(ligne_claire, fisheye, Surreal, abstract:0.7):0.40], colored_pencil_\(medium\)

with a noobai vpred model at a low cfg


>>106362734
i got it from street fighter
Anonymous No.106363141 >>106363272
Tensorrt upscale works surprisingly quick for Wan, but... kinda pointless to upscale 720p, I guess.
https://file.garden/aIdN6xfH0QVghCy0/._00092.mp4
>>106363051
I started with that but added another sampler. High no LoRA -> High with LoRA -> Low no LoRA -> Low with LoRA
I don't think it matters too much, though, 3 should be enough.
Anonymous No.106363146
>>106363105
use litterbox or register for an account silly
Anonymous No.106363175
>>106362882
I fed this to Gippity 5 for a big chungus NLP prompt (so long that it exceeds the 4chan comment character limit by 600 lmao)
then prompted Flux Krea and Qwen with it on the same seed at the same resolution
both decent takes IMO (and relatively similar to each other overall)
Anonymous No.106363233 >>106363347
Left my pc running comparisons between HD, Base, T5, T5Flan, and Gner at a lower resolution (768). Seems to confirm the other anon's observation that it tends to blur the image at lower resolutions. T5 also seems to provide the best realism results somewhat consistently. I'll post some examples when I have time.
Anonymous No.106363251 >>106363261
>>106363075
>to help?
help with what?
Anonymous No.106363259 >>106363273 >>106363288 >>106363296 >>106363319 >>106363338 >>106368173
What do you think if Comfy take the bloatpill and upload his UI to steam make an UI like those Dota games?

pic related could be the main comfyUI menu, at the right of the pic it is a open chat to Comfy discord server
Anonymous No.106363261 >>106363268
>>106363251
So we can see the settings for ourselves
Anonymous No.106363268 >>106363284 >>106363322
>>106363261
for what?
Anonymous No.106363272 >>106363379
>>106363110
amazing gen, are you the same anon that used to post those Athena Asamiya gens?

>>106363141
Thank you so much anon
Anonymous No.106363273 >>106363288 >>106363296 >>106363319
>>106363259
pic related ultrahd 3d node map
Anonymous No.106363284 >>106363444
>>106363268
So we can reproduce the results
Anonymous No.106363288 >>106363296 >>106363319
>>106363273
>>106363259
chekpoint selector
Anonymous No.106363296
>>106363259
>>106363273
>>106363288
lora selector
Anonymous No.106363319
>>106363288
>>106363273
>>106363259
And with an API connection directed to GPT where an assistant comments you of your workflow and suggest things to do

all this very bloated, you will have an assistant and at the same time youwill have the Comfy discord open while de 3d rendering nodes runs
Anonymous No.106363322 >>106363444
>>106363268
I want to make images like it for myself anon help a nigga out damn
Anonymous No.106363338
>>106363259
AniStudio is a lot more suited for that
Anonymous No.106363341 >>106363528 >>106363936
Is there like some secret sauce workflow to fix faces on a video?
Anonymous No.106363347 >>106364732
>>106363233
Flan has been blatantly misaligned with the original embeddings every time I ever tried it versus normal T5 XXL
Anonymous No.106363360
Chunners got stood up
Anonymous No.106363379 >>106363414
>>106363272
>amazing gen, are you the same anon that used to post those Athena Asamiya gens

Thanks, I am not that anon. I change my shit up fairly regularly
Anonymous No.106363414
>>106363379
Love kimono stuff with patterns very nice
Anonymous No.106363444 >>106363676
>>106363284
>>106363322
why? its nothing special
Anonymous No.106363471
>>106362128
delete this
Anonymous No.106363480
Anonymous No.106363528 >>106363605
>>106363341
meaning?
Anonymous No.106363579 >>106363711
What's the current best cfg solution for models that get weird or fried when you use high values?
I remember trying skimmed cfg in the past and not liking it too much. Maybe I just couldn't get it working I dunno.
I am using V-pred Noob and the recommended 3.5-5.5 range is a bit too low at times.
Anonymous No.106363605 >>106363936
>>106363528
When a video gen fucks up the face is there a way to fix it?
Anonymous No.106363611
any prompting tips for getting rid of the bokeh on chroma?
Anonymous No.106363676
>>106363444
Jesus Christ if you don’t want to share your prompt and settings just say so don’t be such a faggot about it
Anonymous No.106363711 >>106363728
>>106363579
just use low cfg for base pass and then 4-5 euler for upscale pass for better detail while keeping things non-shiny.
Anonymous No.106363728 >>106363905
>>106363711
Do you mean 4-5 step euler here or 4-5 cfg euler? Also what denoising value do you recommend for this?
Anonymous No.106363824 >>106363839 >>106363892
i really hate it, when creators release 15gb+ models, and never make gguf versions. i don't want your stupid 20gb model!
Anonymous No.106363839
>>106363824
Skill issue.
Make your own quant.
Anonymous No.106363892
>>106363824
OK, I'll make sure to upcast any future releases to FP64.
Anonymous No.106363905 >>106363963 >>106363969
>>106363728
I mean 4-5 cfg euler for upscale pass. For my main pass I do uni-pc_bh2 at 3 to 4 cfg (using nag/pag/seg so really like 7 actual cfg) then upscale I use euler or dpm2m at 4 or 5 cfg. gets prompt adherence, clean details, and then the upscale pass locks in details but keeps it from too much ai sheen. sometimes. shit will always look ai on some level.
Anonymous No.106363908 >>106363974
Are quantized models slower in ComfyUI than the base models? Or are ggufs slower than non-ggufs? I was trying a gguf version of chroma and getting about 1.6s/it at 1 megapixel generation, then I downloaded the original checkpoint and ran it and got 1.0s/it. I always thought smaller checkpoints were faster for some reason.
Anonymous No.106363936 >>106364047
>>106363341
>>106363605
I figured out you can run GFPGAN on video in comfy. It does work somewhat but mostly on realistic content and with varying results. Maybe the order of operations should be upscale, facefix and then interpolate so it's not trying to interpolate the fucked up faces. Or even run it before and after interpolate.

Are there any other nodes like this? Maybe something that works on 2d?
Anonymous No.106363944 >>106363964 >>106364143 >>106364172 >>106364341
I managed to make a lora that even the sloppiest images are soulful

I would gladly train a Qwen-Image version of this if someone points me to a flawless training tutorial specifically for Qwen
Anonymous No.106363963 >>106363973
>>106363905
>using nag/pag/seg
Do you mean this?
https://github.com/pamparamm/sd-perturbed-attention
From the resolution you seem to be doing 1.5x upscale. What denoising do you use with that?
Anonymous No.106363964
>>106363944
just use default ai-toolkit
Anonymous No.106363969 >>106363978
>>106363905
Your gens are mid Asukafag imagine thinking you can give advice when your output are slop tier nornal
Anonymous No.106363973 >>106363990
>>106363963
yes. nag 2, pag/seg 1.5. you can roughly think of each value as "adding" to cfg. main pass is 40 steps (uni-pc loves high steps, fuck the docs), upscale pass is 30 steps, .25 denoise.
Anonymous No.106363974 >>106363987
>>106363908
I think on higher end hardware the quant speedup becomes less and less noticeable, especially if you already have the memory for the full model. But I'm really not sure.
What quant did you try?
Anonymous No.106363978 >>106364048
>>106363969
lets hear your setup then, always looking to improve.
Anonymous No.106363987 >>106364020 >>106365201
>>106363974
The gguf was Q8_0. I have a 4090 and can just run the base one, my VRAM usage is > 20GB, higher on upscale.
Anonymous No.106363990
>>106363973
Thanks, will try this.
Anonymous No.106364016
Anonymous No.106364020
>>106363987
kino
Anonymous No.106364047
>>106363936
Posted some nsfw examples in the /gif/ thread
>>>/gif/29345317
>>>/gif/29345327
Anonymous No.106364048 >>106364101
>>106363978
Haha stay mad faggot you are not a mentor or role model to anyone, you like your stuff because you have a fixation with Asuka. Your gens look like slop to everyone else and nobody cares about your waifufagging
Anonymous No.106364101 >>106364221
>>106364048
when you insult anons on the internet you are only exposing your own insecurities, jungian shadow and all that
Anonymous No.106364136 >>106364196
Anonymous No.106364143
>>106363944
I don't think anyone has one yet
I'm using musubi with default settings except all 16-bit. ai-toolkit wouldn't run and debugging it was a waste of time.
Anonymous No.106364172 >>106364206
>>106363944
diffusion-pipe? what optimizer are you using anon? looks very nice
Anonymous No.106364196
>>106364136
>knock off asuka
I mean 98 is 0-2, its so perfect
Anonymous No.106364200
Cozy Weekend
Anonymous No.106364206 >>106364413
>>106364172
Yes, diffusion-pipe.
I am using basically the same settings as the rentry, training for the latest Chroma HD, but for inference I am using a custom version of Chroma Flash that is a merge of the latest Chroma1 HD with the delta for Flash
My loras are completely fucked/busted on the base model and I have no idea why, but they work fine on Chroma Flash for whatever the reason
Anonymous No.106364221 >>106364295
>>106364101
Thanks anon, yeah I'm insecure and all that. BUT about your gens and giving advice on image gen, your Asuka has a VERY massive chin. Wouldn't matter if canon Asuka had one, also her calm expression doesn't capture her personality at all and Asuka has a slim body with medium breasts.
This Asuka, your Asuka with a very massive chin, hag body and calm expression, it's your subjective version of her, subjective and sloppy, unworthy of giving advice to others.
Anonymous No.106364295 >>106364508 >>106364511 >>106364588 >>106365942
>>106364221
lol, umad?
Anonymous No.106364341
>>106363944
How do you set up diffusion-pipe?
Anonymous No.106364413 >>106364427
>>106364206
Thank you anon, I found the rentry, what do you mean by this
>Eval datasets are the same format as training. Copy 1-4 image/text pairs from your training into eval. Normally you don't poison eval with training data, but we're intentionally biasing the model here.

it means copy 1 to 4? sorry im a retard
Anonymous No.106364427
>>106364413
Just disable the eval, you won't need that shit
Anonymous No.106364446 >>106364464 >>106365293
I TOLD IT TO BITE THE FINGER NOT DEEP THROAT IT AIIIEEEEEEEE
Anonymous No.106364459 >>106364677 >>106365036 >>106365095
Tip for those training a Chroma lora for the first time: do not train on HD, train on Base. HD is fucked ATM it will output blurry images even if you train a lora on highres images. So just train on Base (which is the same weights as v48) to get the hang of it
Anonymous No.106364464
>>106364446
>tail wag
LEWD
Anonymous No.106364508
>>106364295
Umad Mostaque
Anonymous No.106364511
>>106364295
Yes, your Asuka sucks. You need to own that.
Anonymous No.106364534
Anonymous No.106364574
Anonymous No.106364588 >>106364636
>>106364295
smoking is bad
mmm'kay
Anonymous No.106364636 >>106364772
>>106364588
you are after me in every time, ldg bro lol. i'm still drinkin, still smokin.
Anonymous No.106364677 >>106365036 >>106365095
>>106364459

Confirming this as well. It unfrtounately seems like some of the fuckery from v50 still hasn't been fixed yet with the re-train. It's a shame since anatomy *does* seem more consistent, but the vaseline effect and prompting issues are back.
Anonymous No.106364711
Oh look, the chroma high comes down again and every realizes the model is shit until the next time loadofshitstones drops a post on is discord.
Anonymous No.106364720
Anonymous No.106364732
>>106363347

This lines up with my tests as well. It just seems to produce worse versions of t5.
Anonymous No.106364748
Anonymous No.106364772 >>106364858
>>106364636
fixed
Anonymous No.106364775
i hear we're posting our shitty Asuka's in this thread
Anonymous No.106364787
Pizza time
Anonymous No.106364804
Noobai is all you need
Anonymous No.106364814
>do comparison between debian 12 chroot running inside debian 13 and debian 13 itself
>debian 13 outright kills comfyui on the 4th step
lol, lmao even!
Anonymous No.106364846
Anonymous No.106364858
>>106364772
thx bby, smooch
Anonymous No.106364859
Purple prose natural language prompts feel vile
Anonymous No.106365019 >>106365105
Look what goth girl deliveries brought you
Anonymous No.106365030
debian 13 is slower than debian 12.. sigh
>WHY 580.65.06
dont ask, it's also slower than the 570.133 driver
Anonymous No.106365036
>>106364677
>>106364459
Samefag, come on, start posting photos of Asian women with ugly feet.
Anonymous No.106365095 >>106365219
>>106364459
>>106364677
make a post on huggingface discussions about it
Anonymous No.106365105 >>106365234
>>106365019
>sass gen
>in my ldg
Anonymous No.106365125 >>106365229 >>106365328
is there anything stopping me from using diffusers to train loras and forgetting about guis forever
Anonymous No.106365157
https://github.com/nunchaku-tech/nunchaku/pull/624
HAPPENING!!! CPU OFFLOADING!!! HAPPENING!!!!
(qwen image)
Anonymous No.106365178
Anonymous No.106365201
>>106363987
> those lines
hello chroma
also, have a link to that lora? is it a flux lora or a retrained chroma one?
Anonymous No.106365219
>>106365095
I'll get a comparison ready.
Anonymous No.106365229
>>106365125
Their training scripts are highly unoptimized and inference with diffusers is much lower quality.
Anonymous No.106365234 >>106365239
>>106365105
Those are nice
Anonymous No.106365239
>>106365234
what do you not get? LOCAL DIFFUSION GENERAL YOU FUCKING NIGGER GO KILL YOURSELF FUCKING NIGGER
Anonymous No.106365256
uh oh, looks like the localcreature lost his marbles after chroma flopped and qwen slopped!
Anonymous No.106365259
where is shotabro?
Anonymous No.106365269 >>106365749 >>106365764
>lightning
>teacache
>nunchaku
Just buy a better GPU
Anonymous No.106365271 >>106365281 >>106365303
>>106360525
>I LOADED UP FLUX KREA SVDQ WITH --reserve-vram 8 FOR THE LULZ
> 2315MiB / 12288MiB
wait does this mean I could run it on a GPoorU with only 8 gb?
Anonymous No.106365281 >>106365295
>>106365271
Why'd you want to run krea
Anonymous No.106365293
>>106364446
hell yeah
Anonymous No.106365295
>>106365281
it would be nice to try it out if it doesn't take a year per gen
Anonymous No.106365303 >>106365387
>>106365271
yea if ur gpu is RTX (2000 and later)
Anonymous No.106365308
Why does he hate me?
Anonymous No.106365309 >>106365321 >>106365329 >>106365397 >>106366468
>5090 32gb
>4090 48gb
can't decide
Anonymous No.106365311
You guys think LTX will ever be anything more than just a meme video gen model? They have minute long generations and fast generations but it's just ass compared to WAN
Anonymous No.106365321 >>106365331
>>106365309
4090 48gb is a frankengpu and will be so expensive after shipping that it's not worth it
Anonymous No.106365328
>>106365125
Use kohya musubi tuner
Anonymous No.106365329 >>106365340
>>106365309
only third worlders (who can't afford it) want your hacked up card chen. ada 6000 on sass less then $1 btw
Anonymous No.106365331 >>106365338 >>106365342 >>106365347
>>106365321
a regular 4090 is itself "not worth it" compared to a 3090 if you care about efficiently spending money
Anonymous No.106365338 >>106365346
>>106365331
i literally have a 3090 in a waterblock on my desk because it is worthless. Get with the times.
Anonymous No.106365340
>>106365329
but I can't gen illegal pixels on the could, can you?
Anonymous No.106365342 >>106365346 >>106365367 >>106365404
>>106365331
Isn't 30XX series alerady incompatible with the newest toys?
Anonymous No.106365346 >>106365455
>>106365338
You being retarded doesn't debunk my argument of perf/$
>>106365342
No
Anonymous No.106365347
>>106365331
a regular 4090 wasn't hacked apart and stitched back up by a third party in an attempt to make it better. you can do what you want but you're spending a lot of money for something even more dodgy than normal
Anonymous No.106365367 >>106365391
>>106365342
Yes, newer cards have hardware support for some novel quant formats, besides being generally faster overall.
Anonymous No.106365387
>>106365303
yeah I got one guess I gotta try it out then
Anonymous No.106365391 >>106365424
>>106365367
Anything except Q8 is retard quant territory that noticably changes the image away from the full model. So paying 2000-8000$ for a single gpu to have noticably worse quality gens than a 500-700$ 3090 is absolutely retarded.
>besides being generally faster overall
No shit newest top card will be faster than previous top card in the gen, the problem comes when you have to pay 3+ times the price for performance that's not even 2x while still getting fucked with vram capacity.
Anonymous No.106365397
>>106365309
wait for the 5090 super with 64gb
Anonymous No.106365404
>>106365342
wow I literally can't fucking breathe it's over. buy the new product right now
Anonymous No.106365424 >>106365466
>>106365391
sub-Q8 may be trash for images but not if you want to use LLMs as well
Anonymous No.106365455
>>106365346
thinking of this hobby in terms of "value" is cope for poor people
Anonymous No.106365463
>4090 48gb
>comes with warranty
scam?
Anonymous No.106365466
>>106365424
I mean sure but if then if you want to use non toy LLMs you need more than 128gb of VRAM which is gonna cost you dozens of thousands of dollars vs buying a single gpu and fast ram for 4-8+ t/s with top MoE models.
Anonymous No.106365625 >>106365641 >>106365665 >>106365714
>https://github.com/nunchaku-tech/ComfyUI-nunchaku/commit/10c9947159123414cc3cbc347accc08f1b8076f5
b-bros?????? QWEN NUNCHAKU COMFYUI JUST DROPPED?!!??!!?!
Anonymous No.106365641 >>106365647 >>106365719
>>106365625
what does this mean?
t. stupid
Anonymous No.106365643
Anonymous No.106365647 >>106365719
>>106365641
it means jeets can now generate their ghibli images faster
Anonymous No.106365665 >>106365719
>>106365625
how much vram do I need
Anonymous No.106365695 >>106365711 >>106365895 >>106365921 >>106365974
new to this, what tag gets you small milf eyes like this, not neccessarily small pupils per say just small milf eyes as opposed to bigger shojo eyes
Anonymous No.106365711 >>106365728
>>106365695
I think it's more style dependent but slanted eyes?

I get that style all the time when I don't even ask for it which is kinda annoying because WAN then wants to open up the eyes full shojo style and ruins the look.
Anonymous No.106365714 >>106365719
>>106365625
>not supporting loras
ok two more months
Anonymous No.106365719 >>106365743
>>106365641
>>106365647
>>106365665
>>106365714
loras are next, the retards also didnt update shit I had to manually install the needed wheels and update the comfyui node manually along with its install, but FAST
Anonymous No.106365728 >>106365744
>>106365711
yeah slanted eyes don't work, it's surprisingly hard to get those consistently. but thank you for your reply anon
Anonymous No.106365743 >>106365760
>>106365719
Anonymous No.106365744
>>106365728
You could try asian woman, korean woman etc. or find a lora that's specific to that style of eye
Anonymous No.106365749
>>106365269
>Just buy a better GPU
At a certain point the speed upgrades between better GPUs diminish and the fact remains that some models are slow as fuck.
Anonymous No.106365760
>>106365743
>56s on 16gb VRAM 20steps 2.5cfg
mikubros... we won!!!
Anonymous No.106365764
>>106365269
or you could shit out slop faster on your 5090
Anonymous No.106365765
Qwedit soon then?
Anonymous No.106365823
VRAMlet life is rough. Can't even watch youtube without slowing down gens.
Anonymous No.106365831 >>106365836
I did one run with wan2.2 using kijai's workflow and it worked out fine and now I get this error, help
Anonymous No.106365836 >>106365851
>>106365831
are you running the torch compile node? try bypassing
Anonymous No.106365851 >>106365917 >>106365928 >>106365968
>>106365836
Anonymous No.106365895
>>106365695
half-lidded eyes
bedroom eyes
Anonymous No.106365917 >>106365956
>>106365851
Are u using the wan2.2 vae? the 2.2 model needs the 2.1 vae
Anonymous No.106365921
>>106365695
narrowed_eyes
Anonymous No.106365928 >>106365956
>>106365851
It could also be that the resolution of the inputs between the latent going from the first pass to the second pass are the wrong size.
This error seems to cover a lot of possible issues.
Anonymous No.106365942
>>106364295
Nta, and I'm going to make it harder for you, anon. But I like your asukas, and it's not about the settings, since 1girl has been solved for a while. But you have artistic vision, keep going.
Anonymous No.106365948 >>106366366
so, Qwen Image, rank 32 or rank 128?
Anonymous No.106365956 >>106365967
>>106365917
I'm using 2.1
>>106365928
How do I check and fix that
Anonymous No.106365967
>>106365956
Did you change the resolution between samplers? If no. Then it's fine.
Anonymous No.106365968
>>106365851
Post a screenshot of your workflow.
Anonymous No.106365974
>>106365695
If in doubt, consult danbooru. Half-closed eyes is a danbooru tag with 110k images, seems plenty enough.
Anonymous No.106365981 >>106366011 >>106366319
qwen image edit 2.0 at some point too probably
Anonymous No.106365989
>ayymd
Anonymous No.106366001 >>106366018 >>106366254
I see comfyui memory management is still fucked
Anonymous No.106366011 >>106366037
>>106365981
idk about you guys, but I've never seen a labubu in person.
Anonymous No.106366018
>>106366001
ramlet issue
Anonymous No.106366031
still four more years until we get cards with higher vram
Anonymous No.106366037
>>106366011
I saw them in China. about as cringe and ugly as you could imagine. truly zoomies' beany babies just much uglier
Anonymous No.106366053 >>106366068 >>106366090 >>106366150 >>106367335
i'm doing fine with low end cards like the 4070, what's with the fuss about getting 5k+ cards itt, do you guys really need to generate a million 4k images in a minute that bad
Anonymous No.106366068
>>106366053
Anonymous No.106366087 >>106366170
spritesheet came out kinda wonky
Anonymous No.106366090 >>106366110
>>106366053
I want faster video gen
Anonymous No.106366110
>>106366090
believe in nunchaku, they'll COOK
Anonymous No.106366150 >>106366159 >>106366165
>>106366053
>4070
>low end
Anonymous No.106366159
>>106366150
cope
Anonymous No.106366165 >>106366189
>>106366150
for AI gen it absolutely is low end
Anonymous No.106366170
>>106366087
I assume this is for giggle right? Nobody would do the ENTIRE sprite sheet in one go.
Anonymous No.106366189 >>106366198 >>106366203 >>106366215 >>106366267
>>106366165
what's a 3060 then?
Anonymous No.106366198
>>106366189
poorfag cope
Anonymous No.106366203
>>106366189
If the 4070 is undead burg
3060 is blight town
Anonymous No.106366207 >>106366235
i am using Draw Things on mac and using Chroma v48. i want to add a controlnet to add poses and i haven't been ablt to get it to work via experimenting. does anyone know what controlnet models work with Chroma v48?
i tried a Flux controlnet because i thought they were somehow related but chatgpt says this isn't true.
can you pose and scribble etc for chroma ?
Anonymous No.106366215
>>106366189
it's all slightly different variations of low end shit that's barely capable of video generation
Anonymous No.106366217 >>106366230 >>106366234
qwen edit keeps zooming out my fucking pictures (check bottom right of this image) and this is by doing the reference latent conditioning outside of the shitty comfy qwen text encoder. If I do the vae directly in it, it zooms out even further.
FUCK
Anonymous No.106366221
Anything less then a 4090 is low end
Anonymous No.106366228
Anonymous No.106366230
>>106366217
Use the wan 2.1 vae
Anonymous No.106366234 >>106366242
>>106366217
is the source image one of its default aspect ratios?
Anonymous No.106366235
>>106366207
chroma is a never-finished-soon-will-be-abandonded project
Anonymous No.106366242
>>106366234
AR is 1:1, both source and target
Anonymous No.106366251 >>106366254 >>106366256
Did something ACTUALLY fuck up with comfy this time? My ram keeps capping out when it definitely did not before.
Anonymous No.106366254 >>106366258
>>106366251
yes >>106366001
Anonymous No.106366256 >>106366261 >>106366265
>>106366251
which version?
Anonymous No.106366258
>>106366254
its just I can't find anything about it on the github. Only my own personal experience and the memey screenshot.
Anonymous No.106366261
>>106366256
all ver since qie p2 update
Anonymous No.106366265 >>106366322
>>106366256
>ComfyUI version: 0.3.52
Anonymous No.106366267 >>106366271 >>106366281
>>106366189
Basically bare minimum to even bother playing around with AI.

That said, the 3060 Ti is the best bang for the buck of all xx60 cards, from then on Nvidia went full jew.
Anonymous No.106366271
>>106366267
>the 3060 Ti is the best bang for the buck of all xx60 cards
and it's way too late to buy one now
elf hugger No.106366278 >>106366300 >>106366323 >>106366377 >>106366439
Has anyone compared windows vs linux gen times? How much does the OS overhead actually effect your efficiency?
Anonymous No.106366281 >>106366284
>>106366267
>from then on Nvidia went full jew
wouldn't you? it's free money
Anonymous No.106366284 >>106366287
>>106366281
No, I'm not jewish.
Anonymous No.106366285 >>106366308
wan 2.2 has problems with hands
Anonymous No.106366287
>>106366284
prove it
Anonymous No.106366293 >>106366355
Anonymous No.106366300 >>106366318
>>106366278
>gen times
It's more about how much mem you can squeeze out with a WM or headless.
Anonymous No.106366308
>>106366285
You need to elaborate.
Anonymous No.106366311
confy should be draged and shod
elf hugger No.106366318 >>106366323
>>106366300
are you tunneling into a server to prompt and fetch your gens? what do you run on, tinycore or something?
Anonymous No.106366319
>>106365981
Now that we have our first iteration if inage editing weights, let's hope they use it as tool for good: taking real photos/art and deconstructing it into input pieces, and then training it as pairs of synthetic inputs/real output, the way lora makers were doing. Because those horrible photoshopped compositions must go.
Anonymous No.106366322
>>106366265
I'm still on 0.3.50
guess I won't be updating soon then
Anonymous No.106366323 >>106366347
>>106366278
>>106366318
box?
elf hugger No.106366347
>>106366323
the model is fluxmania_kreamania
Anonymous No.106366355 >>106366404
>>106366293
i this rank 32? how are you finding it?
Anonymous No.106366366 >>106366371
>>106365948
i usually go big rank and resize lower with kohya's sd-scripts, they work for loras trained with musubi, haven't tried 128 on qwen yet though, only 64
Anonymous No.106366371
>>106366366
ah, I meant the nunchanku quants
Anonymous No.106366377 >>106366409 >>106366439
>>106366278
For starters, Linux is a lot less vram hungry than Windows, even if you are running a shitty DE like Gnome or KDE (why would you even ?)

For example, I'm using Sway (a tiling WM), which makes the system take up ~80mb of vram, that's a LOT of vram left for AI.

Then you have the fact that AI is 100% optimized for Linux, since ALL research and ALL professional usage of it is on Linux, from the cloud to the research labs, Linux as a kernel / OS is also truly optimized for heavy workloads, and the NVidia Linux driver is also optimized for AI / GPU accelerated workloads, since that's where all the money from their Linux customers come from, it's NOT coming from gaming.
Anonymous No.106366404 >>106366425
>>106366355
128
i like the higher base res but definitely needs loras to consistently stick with styles
elf hugger No.106366409
>>106366377
Thank you, that makes a lot of sense. I switched to CachyOS today for Blackwell architecture related reasons but defaulted to KDE. I'll look into swapping to tiling windows. I figured ai was probably linux first in development but I didn't realise it was quite so extreme, that's cool.
Anonymous No.106366425
>>106366404
have you tried rank 32 to compare the speed difference?
Anonymous No.106366435 >>106366444 >>106366500
I've tried to train a lora for QIE and it seems to have failed completely. I kept all of the default settings in musubi tuner other than changing to bf16, as I wasn't able to find a known working recipe anywhere.

The dataset was about a thousand pictures from a particular artist dumped straight from danbooru, along with the original danbooru tags. I know the tags aren't optimal for Qwen but I thought it would be a good start for a test. The goal is to allow QIE to change an image to the new style.

The lora has a very small effect, and it's hard to tell if it's only adding noise. It definitely isn't obviously working, but there might be a little bit more of an effect on from the later epoch checkpoints.

It's also likely that I made mistakes with musubi tuner. The documentation from it is pretty terrible auto-translated Japanese and a lot of parameters have no explanation. The doc indicated that --edit should be used for QIE training, but it seemed like the only effect of that was that it would generate test images from references along the way so I ended up omitting that. Maybe it was actually necessary, but I'm not sure because I couldn't find a description.

More likely the basic training settings were bad, particularly with the number of images I fed it. network_dim 16 may be too small and learning rate 5e-5 may be too weak as well. The tagging could be bad too, but I think it would still have a more visible effect.
Anonymous No.106366439 >>106366447
>>106366278
>>106366377
there is no reason to use linux over windows just to save vram from the desktop rendering. in either OS you should be plugging your monitor into the iGPU to avoid loading the dGPU at all.
Anonymous No.106366444 >>106366500
>>106366435
>The dataset was about a thousand pictures
Whoo boi.

Did you even train on A/B pairs?
elf hugger No.106366447
>>106366439
Oh right, of course, this is like when gpus with no outputs on them were made during the crypto boom.
Anonymous No.106366456 >>106366466
Anyone has the Chroma radiance workflow?
Anonymous No.106366466
>>106366456
Go ask in the discord.
Anonymous No.106366468 >>106366472
>>106365309
it's all about vram
so buying 5090 is retarded when you can buy 4090 48gb
Anonymous No.106366472 >>106366474
>>106366468
>so buying 5090 is retarded when you can buy 4090 48gb
There are other consideration to make when the object you're buying was soldered together in some Chinese dude's basement with zero warranty.
Anonymous No.106366474 >>106366623
>>106366472
they come with warranty
stfu if you don't know any better
Anonymous No.106366500
>>106366444
No, because I couldn't find a reference to explain how to configure that and it was easier to hit go and see what would happen than to try to reconstruct something from Qwen's own docs.

Joke's on me though, while I was typing the wall of text at >>106366435 some more test generations run on photos actually worked pretty well. So now seems that it works great on photos but not on anime-style images.
Anonymous No.106366513
Anonymous No.106366530
alright fuck it i'm gonna try and train a huge lips lora because i don't trust a civitai indian to ever care about that. the only loras that are close to that is a lora for blonde bimbos specifically

people with lora training experience how many images/videos do i realistically need? at least 30 right? and for a body part specifically i'm assuming i need multiple races etc

can anyone recommend some women with disgustingly huge lips? better to go overboard for the training data so i can lower the weight from 1.00 later right?
Anonymous No.106366536 >>106366563
Wake up sweaty, new snakeoil mix dropped.
Anonymous No.106366548
>>106362104 (OP)
oh neat someone animated that painting of the two boys carrying the angel

that's like considered the national painting of finland or something like that. i have no idea where i saw/learned about it. hacker news maybe?
Anonymous No.106366563 >>106366574 >>106366617
>>106366536
What the hell is this
Anonymous No.106366574 >>106366606 >>106366926
>>106366563
https://huggingface.co/silveroxides/Chroma-GGUF/tree/main/Chroma1-HD/Chroma1-HD-dc-super-mix
Anonymous No.106366581 >>106366633 >>106366664
Hey, I'm a dumbass, is there a particular reason xformers broke when I upgraded to my GPU to a 5080? Using Reforge, if that has anything to do with it.
Anonymous No.106366606 >>106366616
>>106366574

So like every other variant, it just appears on HF with zero information lmao. I'm starting to think the high resolution epochs are just cursed at this point.
Anonymous No.106366616 >>106366627
>>106366606
I've been trying the supermix and the outputs are weirdly dumbed down like cfg1 model. I'll try the normal DC later.
Anonymous No.106366617
>>106366563

Oh and thank (you) for the link by the way
Anonymous No.106366623 >>106366645
>>106366474
>stfu if you don't know any better
Alright, sure. I'm sure the random Chinese guy who loots 4090s from the mailbox for their cores will honor the warranty.
Anonymous No.106366627
>>106366616

Thanks for the link. I'll play around with it tomorrow.
Anonymous No.106366633 >>106366641
>>106366581
Did you change your pytorch version?
Anonymous No.106366641 >>106366646
>>106366633
Not that I know of, all I did was update my drivers, and I even did a fresh install of Reforge when I noticed the issue.
Anonymous No.106366645 >>106366649
>>106366623
by your logic neither will AIBs because they will say it's user error
Anonymous No.106366646 >>106366661
>>106366641
What error does it give you when X-formers breaks? It's highly dependent on having the right cuda and pytorch version for the distribution you're using.
Anonymous No.106366649 >>106366654
>>106366645
All I'm saying is 4090s sourced through dubious means will likely have a dubious warranty. You might not like that fact but it's true.
Anonymous No.106366654 >>106366666
>>106366649
>trust me bro
I rather hearing from facts
Anonymous No.106366661 >>106366671
>>106366646
Anonymous No.106366664 >>106366671 >>106366674
>>106366581
Why are you using xformers ? Pytorch has its own built-in attention mechanism that is practically identical (same quality) and also a bit faster (SDP), if you use this then it will always work.
Anonymous No.106366666 >>106366673 >>106366688
>>106366654
Your source is also just "Trust me bro". I'm just arguing from a pragmatic standpoint where yours is arguing from a point where the vendor is operating in good faith.

I don't doubt the products work, but I do doubt their ability and intent to protect the customer once they get their money.
Anonymous No.106366671 >>106366674
>>106366661
idk bro. I've never seen this before.
But >>106366664 is right. X-formers is like SD-1.5 era technology.
Anonymous No.106366673 >>106366696
>>106366666
>t. spoonfeed me where can I buy a 48gb 4090 with warranty
also, nice slippery slope anon
Anonymous No.106366674 >>106366719
>>106366664
>>106366671
Been using them since the NAI leak, haven't really looked into them until there was an issue. Heard they helped speed up gens and just threw it in my args. Guess I'll just keep them turned off. Does the alternative need a commandline arg, or is it automatic?
Anonymous No.106366688
>>106366666
checked
Anonymous No.106366693
>>106366692
>>106366692
>>106366692
>>106366692
Anonymous No.106366696 >>106366761
>>106366673
>spoonfeed me where can I buy a 48gb 4090 with warranty
What? Where did I imply that? If I was going to upgrade to a GPU with larger VRAM I'd buy a 5090 or an rtx6000.
I'm simply strongly implying that the warranty from a vendor that sells a modified 4090 will not honor a warranty as they have no incentive to.
I don't understand why this is a controversial point.
Anonymous No.106366719
>>106366674
>Does the alternative need a commandline arg, or is it automatic?
Never used Reforge, but typically SDP is the default attention mechanism (since it's just as good and built-in) unless you explicitly specify something else, like xformers
Anonymous No.106366761
>>106366696
>trust me bro
yep
Anonymous No.106366926
>>106366574
So it has this merged in? https://huggingface.co/DFloat11/Chroma-DF11
Anonymous No.106367335
>>106366053
5 seconds for 1024 square 1girls already feels way too slow for me, after upgrading to a 5070ti, yes. It took me 25 before so it’s much better, but I want more.
Anonymous No.106368173
>>106363259
unironically love this idea, especially the checkpoint selector and assistant guy