← Home ← Back to /g/

Thread 106244742

315 posts 220 images /g/
Anonymous No.106244742 >>106244939
/ldg/ - Local Diffusion General
Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>106242211

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://github.com/Wan-Video
2.1: https://rentry.org/wan21kjguide
2.2: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-HD/tree/main
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.106244755
Left or right, white man.
Anonymous No.106244761 >>106246423
This collage is an outrage.
Anonymous No.106244781
local dumpers general
hats off to the lora maker this is an impressive level of zoom in
Anonymous No.106244791
Blessed thread of frenship
Anonymous No.106244802 >>106244814 >>106244860
>>106244741
>>106244757

I'm using a dynamic split between High and Low through Wan MoE KSampler, which uses the same formula as the one used by wan team.
And the result in my tests is always skewed towards less steps in High vs Low.
Anonymous No.106244814
>>106244802
show me a gen nerd also i was wrong its 23 spells 17 lands i guess no one here right now plays magic the gathering
Anonymous No.106244860
>>106244802
I just feel like you don't need a million low noise steps.
Anonymous No.106244876
So is this just a better sd1.5? https://huggingface.co/opendiffusionai/sd-flow-alpha

Also this might be of interest https://github.com/OPPO-Mente-Lab/X2I

Both work in comfyui apparently
Anonymous No.106244911
Anonymous No.106244939 >>106244944 >>106244965 >>106244996 >>106245010 >>106245753
>>106244742 (OP)
Remember pony? That faggot wants to get some downloads on his shitty sfw app, "fictional", before releasing v7.
Anonymous No.106244944 >>106244984
>>106244939
How can one man make so many wrong moves after capturing lightning in a bottle and holding onto it for the better part of a year?
Anonymous No.106244965
>>106244939
You can't be serious ?
Anonymous No.106244968
Is there any job more fit for beasts of burden than python environment management?
I'm going to put a bullet between my teeth.
Anonymous No.106244975 >>106246376
Anonymous No.106244980
Anonymous No.106244984 >>106245010 >>106245020
>>106244944
What wrong moves? As far as I understand, he:
>picked auraflow over flux because flux is not properly trainable
>released an app that lets people generate their fetish pics and have cringe ai chats while avoiding being branded as a porn app
Doesn't sound that bad.
Anonymous No.106244996
>>106244939
if the money he gets from that funds the training, sure why not
Anonymous No.106245010 >>106245027 >>106245114
>>106244939
>Remember pony?
I want to forget.
>>106244984
>>picked auraflow
And how'd that work out for her?
Anonymous No.106245020 >>106245114
>>106244984
Except it turns out auraflow wasn't properly trainable, what is it now, 7-8 months and the few images I've seen from it are 'meh'

All that time and money for a model that will already be obsolete when it releases

I mean it's shitty luck, he obviously thought he could get auraflow to work great for his huge finetune, but unfortunately it didn't
Anonymous No.106245027
>>106245010
>Her
OMG WHERE ONLYFANS?!
Anonymous No.106245053
Maybe this is just anime being anime, but I'm having a hard time prompting good movement with anime gens.
Anonymous No.106245073
Anonymous No.106245084 >>106246423
Who do I lodge a complain to about none of my gens getting into the collage this time despite there being plenty of room?
Anonymous No.106245111
Anonymous No.106245114 >>106245149 >>106245156 >>106245763 >>106246410
>>106245010
>I want to forget.
v6 is great, nigga

>>106245020
Idk about that. I've seen examples with great prompt adherence, such as multi character interactions. And the style, which is at times deep fried is no different from what happens on v6 when retards pick a high cfg. At least that's what I'd like to believe. It seems to suck at small details though. Pic related is a collage I made from some pics I found in the app, probably more to reassure myself than anything else.

I'm not a shill btw.
Anonymous No.106245125 >>106245128
I'm having fun learning
Can't wait to make slop for Indians
Anonymous No.106245128
>>106245125
Usually it's the reverse.
Anonymous No.106245149
>>106245114
Well, time will tell, I hope you're correct, more local quality models is always better.
Anonymous No.106245156 >>106245169
>>106245114
You've been spending too much time in their coomcord, anon.
Anonymous No.106245159
This one went over 4mb so I had to convert it to a shitty webm
Anonymous No.106245160 >>106248381
surprised this worked 1st try
Anonymous No.106245169 >>106245179
>>106245156
I only occasionally check their coomcord for v7 news
Anonymous No.106245179 >>106245210
>>106245169
Leave the discord and never look back. You're only padding his metrics for when he inevitably sells out to someone looking to buy his eternally never to launch product.
Anonymous No.106245185 >>106245205 >>106246295
Anonymous No.106245205 >>106245216
>>106245185
Not bad, would be better without the midget
Anonymous No.106245210 >>106245238
>>106245179
He is a jap, they usually get shit done. Plus it's not like we had lots of similar cases.
Anonymous No.106245216
>>106245205
She was supposed to grab it by the hair and slam it into the tree, pick up her sword and turn to the viewer. Complete failure.
Anonymous No.106245238
>>106245210
That goes a huge way to explaining their irrational fear of unenforceable licenses.
Anonymous No.106245250
Anonymous No.106245257 >>106245271 >>106246395
how do I monetize gens on twitter?
Anonymous No.106245260
Anonymous No.106245271
>>106245257
You always gotta find a niche in a poorly serviced fetish and lock your good shit (actually just random shit) behind a paywall, with the promise to random coomers that there are even higher highs if you just fork out 5 bucks.
Anonymous No.106245325 >>106246888
Oh man this turned out worse that I ever could have hoped.
Anonymous No.106245337 >>106245345 >>106245353 >>106245381 >>106245406
how do I make sure the workflow isn't saved in the image?
Anonymous No.106245339
Anonymous No.106245345 >>106245355
>>106245337
There is no reason to hide the workflow in an image.
Anonymous No.106245353
>>106245337
save as a jpg
Anonymous No.106245355 >>106245360
>>106245345
how do I make sure the workflow is embedded in the image?
Anonymous No.106245360
>>106245355
There is no reason to confirm this.
Anonymous No.106245381
>>106245337
delete metadata
Anonymous No.106245387
Anonymous No.106245406
>>106245337
found a node in WAS suite that lets you choose to embed workflow or not
nice
Anonymous No.106245435 >>106245464
In case you niggas want more info the 3 sampler setup that other Anon is trying out, here's some experiments
https://huggingface.co/lightx2v/Wan2.2-Lightning/discussions/20
Anonymous No.106245451 >>106245481 >>106246970
>All the cats trot like horses.
Anonymous No.106245464
>>106245435
Don't forget to try upscaling the latent on the low pass for free upscale.
Anonymous No.106245467 >>106245483 >>106245499
https://github.com/kijai/ComfyUI-WanVideoWrapper/blob/987e92cc43a77013f941ec50c7b2789969c620cc/example_workflows/wanvideo_Stand-In_reference_example_01.json

https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Stand-In

new image source to video workflow to try out (stand in)
Anonymous No.106245481
>>106245451
kawaii
Anonymous No.106245483 >>106245499 >>106245511
>>106245467
Should probably explain what it's about
https://www.stand-in.tech/
portraits, basically
Anonymous No.106245499 >>106245510
>>106245483
>>106245467

So this could be used to replace one character with another in a video?
Anonymous No.106245510 >>106245524
>>106245499
what I understand so far is you plug in an image and it uses that as guidance for your t2v gen, so a pepe image would result in a frog that looks like pepe in your output.

which is interesting cause i2v is good but you need a starting image with a pepe. this would generate a video of a pepe from nothing but the source.
Anonymous No.106245511 >>106245523
>>106245483
Kind of?
It seems to do two things; facial deepfakes and pose-to-pose. You could replace one character with another using pose-to-pose. I'm not entirely sure about the quality of the results though. Pose-to-pose seems to throw away everything but the pose.

Here's a summary of the history of a tangential topic: video stylization
https://www.youtube.com/watch?v=TZ0W5UQ5mrQ
These guys are pretty much one of the only ones in the field doing it.
Anonymous No.106245523
>>106245511
wrong video
https://www.youtube.com/watch?v=gbgJhk138W8
Anonymous No.106245524
>>106245510
I can see the appeal. I'll wait to actually see the results before I go and download it myself though.
Anonymous No.106245548
I tried chroma and no, high frequency noise isn't detail. it's noise.
Anonymous No.106245551 >>106245593 >>106245605
We have night elves at home
Anonymous No.106245593
>>106245551
It is quite insane that you can gen this locally on consumer hardware

This could be a scene from a network / streaming children show that costs $500k-1 million per episode to produce
Anonymous No.106245605 >>106245641 >>106245646
>>106245551
What clash of clans clone is this?
Anonymous No.106245612 >>106246145 >>106246271
Anonymous No.106245641
>>106245605
Clans of War World
Anonymous No.106245646
>>106245605
Age of Heroes
Anonymous No.106245698
Just testing various levels on denoise on the LoRA pass.
Anonymous No.106245753
>>106244939
didnt v7 already fail or something like that? why are we still talking about this guy?
Anonymous No.106245763
>>106245114
>v6 is great
I never really cared for pony
Anonymous No.106245776 >>106245791
What the fuck happen here? She eat all the owls or something?
Anonymous No.106245791 >>106245844
>>106245776
Now we're talking... still no jiggle though, a crime I'd say
Anonymous No.106245844 >>106245900
>>106245791
If you prompt the word "voluptuous" with the cfg cranked up, it takes that word very seriously it seems.
Anonymous No.106245900
>>106245844
Based chinks, they know what they're doing
Anonymous No.106246008 >>106246023
Seeing what lowering the light LoRA weight on the second pass does.
Anonymous No.106246023 >>106246031
>>106246008
Faster movement, as expected I suppose
Anonymous No.106246029 >>106246050 >>106246053
testing a lora and t2v
Anonymous No.106246031
>>106246023
Yeah gonna see how low I can get it before it goes back to being not worth it.
Anonymous No.106246050 >>106246053
>>106246029
Anonymous No.106246053 >>106246057 >>106246076
>>106246029
>>106246050
I wonder what LoRA this could be.
Anonymous No.106246057
>>106246053
I mostly use i2v but for t2v the most obvious test is the ghibli meme lora
Anonymous No.106246062
Anonymous No.106246076
>>106246053
I think it's Metallica
Anonymous No.106246106 >>106246130
Bro what is my sampler doing.
Anonymous No.106246107 >>106246143
Here is a strength of 1.5
Gonna try 1 then cut it off there. I can't imagine it get any better past that.
Anonymous No.106246130
>>106246106
Anonymous No.106246143 >>106246157
>>106246107
She keeps becoming less stylized, why are you doing this, stop!
Anonymous No.106246144
but, I still think i2v is the most fun. if I need a base image I can use reforge or google whatever.
Anonymous No.106246145 >>106246169
>>106245612
Is she gonna be okay? It looks like she's having trouble swallowing something.
Anonymous No.106246156
So I have a question about I2V, sometimes an object in the background has a similar color to hair/outfits and get absorbed into it. Is there a way to prevent that with prompts, or any specific node?
There's no way something like "hair is a singular color" would work right?
Anonymous No.106246157
>>106246143
I forgot I deleted the 3D cgi tag. She's irl now. Sorry.
Anonymous No.106246167 >>106246199 >>106247169
Is there a site that tracks nvidia driver performance across versions?
Anonymous No.106246169
>>106246145
She's used to it.
Anonymous No.106246199
>>106246167
Probably for gaming, but not for AI
Anonymous No.106246205 >>106246212 >>106246308 >>106246541
ah nice, with kijai's workflow and the block swap stuff I can do 120 frames and no oom and im on 16gb.
Anonymous No.106246212 >>106246254
>>106246205
*actually it's cause I matched the img size which was small. larger image = more vram needed
Anonymous No.106246213
Here it is at strength 1. Definitely moves better, but I'm done with fake nelfs.
Anonymous No.106246244
Anonymous No.106246249 >>106246260
from the Wan research paper:
>Shaky camera footage: Amateur recordings with pronounced camera shake, often causing motion blur and ambiguous foreground-background differentiation. This category is systematically excluded from training consideration.
i was failing to gen something like this earlier and this explains why
i suppose a lora would help
Anonymous No.106246254
>>106246212
>larger image = more vram needed
https://www.youtube.com/watch?v=REU7-64-vCM
Anonymous No.106246260
>>106246249
I can kind of forgive them for it. Think about how awful the hands and feet would look.
You can try handheld footage. it works a bit.
Anonymous No.106246271
>>106245612
goddamn she toofless
Anonymous No.106246288 >>106246318
swapped text with kontext, made it move with wan 2.2:

still want the gun to fire, maybe next gen.
Anonymous No.106246295
>>106245185
Trying this one again with my revised settings. It definitely followed the prompt better.
Anonymous No.106246308
>>106246205
If you don't use the Torch Compile node on KJ's workflow, set the "rope_function" on the WanVideo Sampler node to "comfy_chunked" for less VRAM usage.
Anonymous No.106246318 >>106246326 >>106246331
>>106246288
there, that's better.
Anonymous No.106246326
>>106246318
*121 seconds with the 2.2 kijai lightning loras, seem to work well
Anonymous No.106246331 >>106246340
>>106246318
Looks like there were a few rounds missing from that revolver.
Anonymous No.106246340
>>106246331
miku roulette I suppose
Anonymous No.106246350 >>106246355 >>106246358 >>106246406 >>106246473
Chroma status?
Anonymous No.106246355 >>106246359
>>106246350
Eternal v48 for now.
Anonymous No.106246358
>>106246350
>ctrl+f chr*ma
>4 hits

It's over.
Anonymous No.106246359 >>106246374 >>106246473
>>106246355
What happened to v49 and v50?
Anonymous No.106246367 >>106246436 >>106246562
perhaps he's wondering why he would fist bump a gun outside a plane?

I just said pointed gun at the man on the right:
Anonymous No.106246374
>>106246359
It's like each of them has a flaw that makes it suck at a specific thing.
Anonymous No.106246376
>>106244975
With new settings
Anonymous No.106246380 >>106246389
Chroma showed great promise, we needed a new NsFW local model.
Anonymous No.106246389
>>106246380
Chroma could never really be the thing everyone truly wanted it to be. It does a lot of things, sort of okay, but always falls short in some way that makes the whole image kind flawed, and not in a good way.
Anonymous No.106246390
Anonymous No.106246395
>>106245257
git gud
Anonymous No.106246406 >>106246428
>>106246350
anon is too busy training loras for 49. The new hd version lodestone is baking is getting sharper too. My favorite photo sets? Now I have endless variations
statler/waldorf No.106246410
>>106245114
>topleft
beahagaha
Anonymous No.106246413
Anonymous No.106246420 >>106246475
Is Chroma a one person project or does it have a team? Maybe he needs help from people who know what they're doing.
statler/waldorf No.106246423 >>106246463
>>106245084
contact your local schizo
>>106244761
outrageous? Outrageously BAD beahagaha
Anonymous No.106246428
>>106246406
Ok that right leg is totaly shit
Anonymous No.106246436 >>106246557 >>106246562
>>106246367
better

now he respects the big guy
Anonymous No.106246463
>>106246423
Are you rocketshizo? Yes you are.
Anonymous No.106246473 >>106246499
>>106246350
>>106246359

V49 was an attempt to stuff all higher resolution training into one epoch. It slopped old prompts and acts strangely the closer your resolution is to 1024x1024. V50 was lode merging v49 with undisclosed "experiments" that made prompts even more unruly and put a fun vaseline filter on everything. In conclusion just use v48.
Anonymous No.106246475 >>106246501
>>106246420
Him and his discord yes men. But let's not float the idea of sinking more cost into a model has had the same issues since its inception.
Anonymous No.106246486
Anonymous No.106246499 >>106246509 >>106246535
>>106246473
Why recommend 48 over 49 if it got less training?
Anonymous No.106246501
>>106246475
Lets spend more money!!
Anonymous No.106246509
>>106246499
Because it has better quality
Anonymous No.106246523
Using img2img gives me better results than using openpose, am I the only one?
Anonymous No.106246535
>>106246499

It consistently gives better results. More training or not, consistency is key, specially for chroma.
Anonymous No.106246536
Huh, Krea is pretty good.
Anonymous No.106246541
>>106246205
Native can too and I'm also on 16gb. I used to be able to push it to 200 frames (640x480) but with all the updates, changes, etc I OOM around 152, fucking bullshit. Even with VACE low steps, used to be able to go beyond and really slop the gen past 200 but again, it OOMs around 152 for me.

I can only assume some kind of memory nerf has been put on these custom nodes or something based on your card size
Anonymous No.106246542 >>106246555 >>106246623
Well, well, Anons, let's look ahead.
Thank you, Chroma, for everything you gave us.
We learned a lot from you, and we won't make the same mistake of marrying the first model we see.
Now the future is bright with WAN, Qwen, and future other models.
Flux itself is a crappy base model, sorry Iodestones for feeling slopp at first glance and marrying the worst line of the worst models from the worst image generation company.
Anonymous No.106246555 >>106246579
>>106246542
Why do you hate him so much?
Anonymous No.106246557
>>106246436
suddenly...
Anonymous No.106246562 >>106246580
>>106246367
>>106246436
Love the focus pull, wan is so good.
Anonymous No.106246579 >>106246623
>>106246555
He's right though. Blowing a load of cash on a model that has basically done everything it can to resist training was a bad idea.
Anonymous No.106246580
>>106246562
yeah the focus change is pretty neat. when I first tried wan I used a low res haruhi image and it actually made it HD, pretty cool.
Anonymous No.106246607 >>106246627 >>106246639 >>106247093
...im batman
Anonymous No.106246623 >>106246646
>>106246542
>>106246579
Samefag, stop being so pathetic
Anonymous No.106246627 >>106246633
>>106246607
bruh good one
Anonymous No.106246633
>>106246627
for once
Anonymous No.106246639
>>106246607
also I just noticed there is someone doing something in one of the plane windows

wan is pretty cool
Anonymous No.106246646 >>106246660
>>106246623
Okay, you got me. Want a prize?
Anonymous No.106246660
>>106246646
post bussy
Anonymous No.106246676 >>106246691
how do you make something like this
>>>/bant/23102013
Anonymous No.106246678 >>106246934
Anonymous No.106246691 >>106246708
>>106246676
You gotta install the chat gpt phone app and use the DAN jailbreak for it to make sora videos like that.
Anonymous No.106246699 >>106246972
alternate timeline cia:
Anonymous No.106246708 >>106246715
>>106246691
so i'm new to this and haven't been following. Chat gpt can edit videos?
Anonymous No.106246715 >>106246737
>>106246708
Yeah sure, why not? Don't forget you can also do it locally with A1111 and stable diffusion 1.5 img2img.
Anonymous No.106246737
>>106246715
so for example if i wanted to make videos of a model from IG (for private use lol)
How would i go about doing that, what is the best way with the best results?
Anonymous No.106246742
Anonymous No.106246758 >>106246773
thoughts on skyreels a3?
Anonymous No.106246771 >>106246958
Anonymous No.106246773 >>106246791
>>106246758
?
Anonymous No.106246779
Anonymous No.106246782 >>106246834 >>106247118
Anonymous No.106246791
>>106246773
nta, but I think a trailer dropped. Looked like a bunch of uncanny talking heads with voice.
Anonymous No.106246834 >>106246838
>>106246782
looks krea
Anonymous No.106246838 >>106247118
>>106246834
it is
Anonymous No.106246888
>>106245325
average video game mount
Anonymous No.106246890
Anonymous No.106246902 >>106246941 >>106246972
CIA...why did you do it...

the men on an airplane dive out the door on the right.
Anonymous No.106246917 >>106246924 >>106246972 >>106247298
I've been out of the game for a couple years in imagegen, what's the current local State of the Art these days.
I see Flux is still around, as well as various finetunes of SD 1.5, but I was curious if anything else has cropped up, particularly something that would benefit from 48GB+ VRAM?
All the models I've seen are <10GB, is there really no race for SOTA models like there are for LLMs?
Anonymous No.106246924
>>106246917
qwen for image generation. wan 2.2 for video generation
Anonymous No.106246934 >>106246962
>>106246678
starting to look more natural
Anonymous No.106246941
>>106246902
diff angle
Anonymous No.106246958
>>106246771
this feels like a meme template
just put something unexpected on the horizon
Anonymous No.106246962 >>106246985
>>106246934
I don't think there's much you can do about the 3D rotoscope looking movement, but I wonder what settings you'd need to finally get rid of the blurring between frames on high motion scenes.
Anonymous No.106246970
>>106245451
Try using CFG on the high noise model, no lightning lora. I got one running properly, though it's a low res, low step gen.
Anonymous No.106246972
>>106246699
>>106246902
Now make the whole film

>>106246917
Wan2.2, it's video focused but does great images too. other models are just to assist Wan.
Anonymous No.106246985
>>106246962
probably some kind of lora, just like how you need one for proper boundaries on pixel art in img gen
but desu i dont mind the rotoscoped look in this, as long as the movement itself looks good and doesnt feel weightless
Anonymous No.106246992
Ok so if I:
1- generate a 1girl image with crea
2-WAN 2.2 with NSFW lora img2vid
3- good quality nsfw dataset that overkills Chroma?
Anonymous No.106247026 >>106247110
so how does wan t2i work? low noise model only?
Anonymous No.106247093
>>106246607
>I am the big black guy now
Anonymous No.106247100
Anonymous No.106247110 >>106247203
>>106247026
i don't know why you assume that you'd use only one of the models
Anonymous No.106247118 >>106247129
>>106246782
>>106246838
great gens, mundane subjects
Anonymous No.106247129 >>106247139
>>106247118
Thanks, that's what I was going for.
Anonymous No.106247139 >>106247159
>>106247129
fair enough, they're good like what you'd see on a model card
Anonymous No.106247159 >>106247167
>>106247139
I was pretty impressed with Krea, still limited in all the same ways that Flux is, but there's a lot to like.
Anonymous No.106247167 >>106247178
>>106247159
I quite like it as an upscaler
Anonymous No.106247169 >>106247171 >>106247174 >>106247336
>>106246167
I actually just tried updating my drivers today and my gen times went up by like 30%, rolled it back and it went back to normal lmao.
Anonymous No.106247171 >>106247208
>>106247169
30 series?
Anonymous No.106247174 >>106247208
>>106247169
OS, driver versions, gpu, cuda versions
Anonymous No.106247178 >>106247191
>>106247167
I could see it being pretty good for that.
Anonymous No.106247191
>>106247178
Not that guy but cuda seems fine across several versions. It's updating the cuda itself to 13 is what breaks things.
Anonymous No.106247197 >>106247294 >>106247502 >>106247688
neat, wan t2i (1 frame) works pretty good
Anonymous No.106247200 >>106247210 >>106247213 >>106247332 >>106247879
Anonymous No.106247203
>>106247110
is there a good workflow for the 2.2 t2v ones? any time I try non 2.1 for making a still image it's blurry or noisy.
Anonymous No.106247208 >>106247237
>>106247171
yes
>>106247174
win 10, 576.02 studio, 3060 12gb, cu128.
The driver I updated to is the latest one up on nvidia right now.
Anonymous No.106247210 >>106247299
>>106247200
one of the most disgusting gens ever conceived
Anonymous No.106247213
>>106247200
Anonymous No.106247237 >>106247257
>>106247208
>yes
yeah i have a 30 series too and i'm still sitting on 566.26. last time i tried updating the fans went haywire. i don't know wtf nvidia is doing
Anonymous No.106247257
>>106247237
yeah after that and the fucking slow download speed on their driver archive, i think i'm sticking with 576.02.
Anonymous No.106247294 >>106247319 >>106247342
>>106247197
share catbox for workflow please
Anonymous No.106247296
teacache for qwen image when
Anonymous No.106247298
>>106246917
>but I was curious if anything else has cropped up, particularly something that would benefit from 48GB+ VRAM?

Not that people talk about here. Sure you can load all your models in vram at once with 48GB but generation software will put models you aren't using into ram and move it back to vram when it needs it, the speed difference isn't big. Also nobody wants to release open source models for VRAM quantities hardly anyone has. I don't feel pressured to go from 24 to 32GB.
Anonymous No.106247299
>>106247210
actually its not disgusting enough because she needs to be covered in more sauces to really sell it that she works there

should probably try the upskirt lora with her walking away/from behind in the fast food restaurant setting too
Anonymous No.106247319
>>106247294
nnnnnnnnnnyo :3
Anonymous No.106247332 >>106247731
>>106247200
>no grimace bikini

Im going back to burger king
Anonymous No.106247336
>>106247169
wait, it wasn't 30%. It was much MUCH worse. These are the sampler times after i rolled back. They were 200 seconds each on the latest nvidia driver, LMAO
Anonymous No.106247342
>>106247294
thats what you get for begging a tranny for an image thats here just to spam his agp fetish 1girl basic, same type of trash image over and over again, lmao
Anonymous No.106247346 >>106247389 >>106247469
does gguf unet loader automatically offload model to system ram? I have 16gb vram but I can run qwen q8, which is 21gb?
Anonymous No.106247389
>>106247346
yes
Anonymous No.106247469 >>106247485
>>106247346
All the models are automatically offloaded to system ram in Comfy as needed, it's not a gguf thing
Anonymous No.106247485 >>106247564
>>106247469
but why do I get oom when I run with fp8?
Anonymous No.106247501
Anonymous No.106247502 >>106247512 >>106247688
>>106247197
and this is with 2.2, first one was 2.1

2.1 t2v seems better? idk why
Anonymous No.106247512
>>106247502
other workflow had shift at 1, idk im gonna mess with stuff and see how it changes
Anonymous No.106247546 >>106247575 >>106247583
first attempt at a chroma lora (v50). not happy about this one but the dataset was only 9 images (uncaptioned) at 512 and fp8
Anonymous No.106247564 >>106247666
>>106247485
because you're retarded ?
Anonymous No.106247565
Anonymous No.106247575 >>106247611
>>106247546
can I try it? you could get better result by lowering lora str and doing second pass
Anonymous No.106247583 >>106247871
>>106247546
512 is enough to get good likeness, but 9 images are very few, and training at fp8 does eat into quality, good for testing though since it's faster
Anonymous No.106247609 >>106247677
Anonymous No.106247611 >>106247924
>>106247575
https://litter.catbox.moe/zrkc9msmsamww6eo.safetensors
Anonymous No.106247666
>>106247564
kek
just admit you have reading comprehension already
it's partially offloading, not fully offliading. is it so hard to infer from 16-21gb?
Anonymous No.106247677 >>106247706
>>106247609
Can WAN do NSFW like Chroma?
Anonymous No.106247688
>>106247502
>>106247197
please share workflows, for me you are not a trans person but a very brave woman!
Anonymous No.106247706 >>106247730
>>106247677
Not out of the box, but as always, loras will be the best way to get NSFW, and this includes Chroma, which although uncensored, is still trained on tons of different things with no particular focus on NSFW, which a lora or further finetuning can effectively provide.
Anonymous No.106247730
>>106247706
>Not out of the box
I wouldn't say that. It's can do some super interesting things out of the box just not hardcore close ups
Anonymous No.106247731
>>106247332
>grimace bikini
Even the McDonald's bikini is inconsistent but I'm asking for patterns not just a logo on each boob

What would that even look like just a dark purple bikini with cartoon eyes on it?
Anonymous No.106247754 >>106247775
the man in the blue shirt jumps on top of the plane behind him as it takes off into the air.

now that's some action.
Anonymous No.106247775 >>106247806
>>106247754
better jump:
Anonymous No.106247806
>>106247775
Make him reveal himself as bane
Anonymous No.106247808 >>106247811
Wan images can never be as sharp as Chroma, plus it really only has 1 style and it's quite noticeable. Chroma is better.
Anonymous No.106247811
>>106247808
buddy we have eyes to see
Anonymous No.106247825
>512x in 2025
kekkycope! localpoors are absolutely devoid of compute, no wonder saas is winning so hard
Anonymous No.106247834
Anonymous No.106247846 >>106247856
is there a way to save the samples from the 2.2 high noise model so you can gen a bunch with the high noise model, pick the best one, then feed it to the low noise model and reroll? or do you just decode it and save it as a video?
Anonymous No.106247854
what do you do if some guy dms you that he wants you to help him destroy onlyfans
Anonymous No.106247856
>>106247846
I tried decoding the latent and using video combine with high noise but always got garbage output. i have no idea why. can someone else confirm if it works for them?
Anonymous No.106247870 >>106247916
the man in the blue shirt shoots a fireball with his hands at the plane behind him, making the plane explode.

SORCERY
Anonymous No.106247871
>>106247583
you can make a good lora with at minimum 8 images if they're all very high quality and provide all the important features it needs(full body, facial close up, multiple expressions, multiple angles)
Anonymous No.106247879 >>106247910 >>106247978
>>106247200
How come Timmy gets TWO drinks and bobby gets NONE?
Anonymous No.106247884 >>106247957 >>106247969
Ai toolkit is best to train chroma?

How long to train a chroma person lora with ~30 image dataset on 3090? Any good guides?

I want to train the same lora on v48 and v50 to see the difference
Anonymous No.106247910
>>106247879
bobby got caught in the bathroom with a cucumber
Anonymous No.106247916
>>106247870
wasnt specific enough

the man in the blue shirt turns and shoots a fireball with his hands at the plane behind him, making the plane explode.

now he turned and shoots the fire.
Anonymous No.106247924 >>106247946 >>106247980
>>106247611
it's decent with 0.9 str. small dataset and lack of tagging probably hurts general use
Anonymous No.106247926 >>106247995
Anonymous No.106247946
>>106247924
FIX IT NOOOOOOOOW
Anonymous No.106247957 >>106247991
>>106247884
>Ai toolkit is best to train chroma?
No, not only does it have huge vram requirements but there seems to be a bug in the way it trains, causing artifacts, unless it's been fixed by now.

Best current option is Diffusion-Pipe, OneTrainer support is imminent as it is being worked on right now and is seemingly training fine, will likely be faster than Diffusion-Pipe since it has a more efficient offloading implementation.

A ~30 image lora at 512 (which is enough for good portrait quality of a person) would probably take ~30-40 minutes on a 3090 if doing 100 epochs
Anonymous No.106247969
>>106247884
depends on the amount of steps/resolution
Anonymous No.106247978
>>106247879
Timmy said he wants to marry the senorita when he grows up
Anonymous No.106247980 >>106248007
>>106247924
that looks more like him than anything i've genned, guess I'm a promptlet
Anonymous No.106247991 >>106247999 >>106248020
>>106247957
>No, not only does it have huge vram requirements but there seems to be a bug in the way it trains, causing artifacts, unless it's been fixed by now.
I've trained multiple Chroma loras just fine with AI Toolkit and I had no artifacts whatsoever.
Anonymous No.106247995 >>106248008
>>106247926
are you doing t2v or i2v?
Anonymous No.106247999
>>106247991
I also trained a few XL loras with it, and it turned out perfectly fine.
Anonymous No.106248007
>>106247980
That's with a lora trained on him, still, just 9 images at 512 resolution, not bad
Anonymous No.106248008 >>106248022
>>106247995
T2V
Anonymous No.106248020 >>106248154
>>106247991
People have reported a lot of horizontal lines with loras trained on AI Toolkit, but that was a while back, maybe it's been fixed
Anonymous No.106248022 >>106248069
>>106248008
2.2? which workflow, also are you using the light2x loras or not, outputs look good
Anonymous No.106248069
>>106248022
Yes and no.
Sampler 1 (advanced) generates 5 steps out of 20 WITHOUT the 2.1 LoRA. (the way the noise denoises needs to be set to 20 because it's different based on how many steps it will think it takes).

After that it is fed to a sampler with a LoRA strength of 1.

Then denoising on low for 3 steps with Heun LoRA at strength .8
Anonymous No.106248077
whoosh.
Anonymous No.106248151
Three more years of Illustrious supremacy
Anonymous No.106248154 >>106248163 >>106248256
>>106248020
I've generated thousands of images from multiple loras. No horizontal lines. If anything diffusion-pipe is the worse choice for Windows since you have to use a linux subsystem(WSL) to make it work.
Anonymous No.106248163
>>106248154
It's the only choice if you have more than one gpu.
Anonymous No.106248177 >>106248203 >>106248214 >>106248254 >>106248383
>make chroma loli
>100% likeness
>use wan to animate
>100% perfect animated sex
I AM IN HEAVEN YOU DONT UNDERSTAND THE PLEASURE I AM EXPERIENCING MY GOD
Anonymous No.106248198
Anonymous No.106248203 >>106248215
>>106248177
post proofs
Anonymous No.106248214
>>106248177
Nice try fed
Anonymous No.106248215 >>106248271
>>106248203
you know i cant do that. im probably already on a list for hinting at it.
Anonymous No.106248254 >>106248274
>>106248177
Shouldn't Chroma be illegal to have on your computer since it can generate illegal content?
Anonymous No.106248256 >>106248269 >>106248335
>>106248154
AI Toolkit hardly works unless you have 24gb, you can train full precision Chroma loras with Diffusion-Pipe on as little as 12gb, less if you use fp8
Anonymous No.106248269 >>106248290
>>106248256
>Diffusion-Pipe
I can't use that because the linux in WSL doesn't have gui
Anonymous No.106248271 >>106248314 >>106248639
>>106248215
do you wanna slide in my dms? got matrix/element?
Anonymous No.106248274
>>106248254
who said anything about illegal? I'm talking about 2d cartoon loli officer :)
Anonymous No.106248290 >>106248320
>>106248269
Why would you need a gui to train ? Only real caveat with Diffusion-Pipe is that it doesn't have sample generation during training, that does suck.
Anonymous No.106248314
>>106248271
i hate "loli" (pedo) shit but i'll gen some loli kino and not share it just to spite you
Anonymous No.106248318 >>106248327 >>106248372
>Kids... roll up the windows.
Anonymous No.106248320 >>106248347 >>106248486
>>106248290
Because I can only use ms-dos-like cli navigation. I literally cannot navigate linux commandline.
Anonymous No.106248327
>>106248318
>haha....oh no ladies! please... you cant just have your way with me...haha...oh noooooo!!!
Anonymous No.106248335
>>106248256
I see. Well, I trained a 128 rank chroma lora @ 1024x1024, 4k steps just fine on my 3090. I tried diffusion-pipe but for whatever reason the lora didn't work. it completed just fine but the output was black squares. not sure what happened.
Anonymous No.106248347
>>106248320
Navigate what ?

Just run a command with a config file
Anonymous No.106248372
>>106248318
kek
Anonymous No.106248381 >>106248416 >>106248440
>>106245160
just for funsies didnt put a whole lot of effort into prompting or anything
Anonymous No.106248383 >>106248439
the M in mcdonalds was designed to remind you of large breasts

>>106248177
unfortunately you'll get over the honeymoon phase just like everything else
Anonymous No.106248396
https://openart.ai/workflows/dowhatyouwantcuzapirateisfree/wan-22-t2v-for-high-end-systems-speed-and-quality-focused/97QzdiAgLDihbeoSKHIt

this seems to work decent, has interpolation also
Anonymous No.106248407
putting togethert qwen and wan 2.2 gens for a Migu and Rin short
Anonymous No.106248416
>>106248381
Sir, we are going to space
Anonymous No.106248423 >>106248441
matrix is pozzed btw, it came to me in a meme
Anonymous No.106248430 >>106248715
Retard here, can you train loras on lightning/dmd models? There's the model ponyhofdmd4_v10 which gives me extremely good outputs especially regarding anatomy for an XL model, seems like it'd be a good base. But so far my attempts at trainign a lora either resulted in black pictures or blurry outputs with datasets that worked on other models. Do you need special settings, or do you need a version without the dmd lora baked in for training?
Anonymous No.106248439
>>106248383
She waxes, nice
Anonymous No.106248440 >>106248465
>>106248381
Tried again
>The man makes a shocked expression and grabs onto the toilet as he passes a particularly large piece of poop.
it's puzzling how literal prompting can be sometimes
Anonymous No.106248441
>>106248423
truthbomb, wanna hmu on tox instead?
Anonymous No.106248458 >>106248472 >>106248506
A question for weebs itt: which japanese artist consistently does art like this (the artstyle)?
Anonymous No.106248465
>>106248440
I keked
Anonymous No.106248472
>>106248458
When I wanna know these things, I go to r/art or something and ask them. But only because if they knew why I wanted to know they'd seethe.
Anonymous No.106248476 >>106248484
the choppy framerate of anons videos makes me nauseous
Anonymous No.106248479
Anonymous No.106248484 >>106248502
>>106248476
sorry im new to this and im not sure how to do the interpolating to 32fps or whatever yet. i followed the 2.2 guide and use its workflow. even though it said it would happen, it definitely does not.
Anonymous No.106248486
>>106248320
>I literally cannot navigate linux commandline.
If i were in your predicament I would simply kill myself
Anonymous No.106248502 >>106248524
>>106248484
film vfi node before last savevideo node and change savevideo node to 32fps
Anonymous No.106248506
>>106248458
The closest I could find so far is Range Murata
I might play around with this image on Flux Kontext a bit to see if it can create variations of this without changing the artstyle
Anonymous No.106248524 >>106248541
>>106248502
not sure where to put it because none of the nodes from the workflow the guide provided are called that. i tried changing videocombine to 32fps and it worked, but it cut the length of my video by half.
Anonymous No.106248531 >>106248550 >>106248566
is rife vfi best for interpolation?
Anonymous No.106248541
>>106248524
film vfi node before the last node to save the video https://github.com/Fannovel16/ComfyUI-Frame-Interpolation
Anonymous No.106248550
>>106248531
film vfi is best aside from cracked topaz chronos
Anonymous No.106248562
>clip encode two sentences once
>clip encode each sentence and then combine them with condition combine node
difference?
Anonymous No.106248565
>>106248560
>>106248560
>>106248560
>>106248560
Anonymous No.106248566
>>106248531
bang
Anonymous No.106248639
>>106248271
been awhile since i used elements. i'll have to create a new user
Anonymous No.106248715
>>106248430
don't use pony, it is really bad you just can't see yet. Skip the part where you generate piles of shit images you will hate later once you see (and can never unsee) pony eyes and just go to Flux/Chroma/Qwen.
don't train on slop mixes, only train on base model or a finetune.