← Home ← Back to /g/

Thread 106277108

319 posts 188 images /g/
Anonymous No.106277108 >>106282955
/ldg/ - Local Diffusion General
Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>106273830

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://github.com/Wan-Video
2.1: https://rentry.org/wan21kjguide
2.2: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-HD/tree/main
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.106277114
the two girls hug:
Anonymous No.106277116
>*braps*
Anonymous No.106277122
Anonymous No.106277165
Anonymous No.106277191
Anonymous No.106277196
Anonymous No.106277205 >>106277219
>106277084
>Why are so many (nsfw) wan loras in fucking SLOW MOTION?

Anything plugged into the high noise model that fucks with the motion will make it slow motion if the training data isn't perfect. You need to make some motion with one sampler without a LoRA then move it into the next one with a LoRA.
Anonymous No.106277219 >>106277223 >>106277257
>>106277205
So basically 2 samplers back to back? One to set the proper "speed" and the second with the loras you want?
I never thought of that, how many frames do you go with with the first sampler?
Anonymous No.106277223 >>106277226
>>106277219
5 at most. 3 at least.
Anonymous No.106277226
>>106277223
OK, I will try.
Anonymous No.106277253 >>106277372
Anonymous No.106277257 >>106277266 >>106277295 >>106277427
>>106277219
5 steps is a good ballpark

however this is 5 steps that you'll have the cfg at 3.5 or something because no lightning lora. So it's slow, but you get proper motion and some prompt adherence. I scale my empty latent down 50% so it does the first high noise pass at half res which is way,way faster and this first pass ends in a blurry mess anyway so I dont think the lower res compromises anything, but I really don't know.
Anonymous No.106277266 >>106277278 >>106277427
>>106277257
>I scale my empty latent down 50% so it does the first high noise pass at half res
Can you share your wf anon?
Anonymous No.106277278 >>106277341
>>106277266
its a mess because I tried to make it do both i2v and t2v and I just went from 2 samplers to 3 with remnants of shit I don't use, let me get the important parts in a sfw screenshot after this gen in done
Anonymous No.106277285 >>106277298 >>106277417 >>106279301 >>106279322 >>106283753
What are the best models based on either sd 1.5 or 2.1?
I have 8 gigs of ram and it's not enough to run SDXL.
I'm looking for models that are great at processing long/detailed prompts and can understand context.
The issue with the models I have is that they clump it all together or make a mandala or makes it into an object or it just lacks consistency or something like that.
Anonymous No.106277295 >>106283762
>>106277257
I tried cheating my way with Flux the similar way with the 1st ksampler at half resolution and I always ended up getting a total mess that couldn't be fixed no matter how many steps you'd set in the next ksampler with proper resolution. Then again, completely different model and that was a static image so go figure
Anonymous No.106277296
Anonymous No.106277298
>>106277285
Anime: https://civitai.com/models/118495/exquisite-details
Realistic: (this one is better than the examples lead on) https://civitai.com/models/550737/1girl-printer
>I'm looking for models that are great at processing long/detailed prompts and can understand context.
Not with 1.5 or 2.1 you're not.
Anonymous No.106277311 >>106277546 >>106281449
Anonymous No.106277341 >>106277427
>>106277278
Thank you anon, I mostly want to see how you connected stuff anyway.
Anonymous No.106277349 >>106277365 >>106277368 >>106277392 >>106277835 >>106280675
Anonymous No.106277365 >>106277368
>>106277349
>mememe dance
why is there no other hot dances loras, why is it always the same one, it's kind of crazy
Anonymous No.106277368 >>106277387
>>106277349
>>106277365
this, i want my twerk loras.
Anonymous No.106277372
>>106277253
lul
Anonymous No.106277374
Anonymous No.106277387 >>106277502
>>106277368
there like dozens of sexy dances with easy to find material, from kpop to latin, but it's always the same mememe instead
Anonymous No.106277392 >>106277405 >>106277502 >>106280675
>>106277349
can you make him dance too please?
Anonymous No.106277405 >>106277568
>>106277392
>figure collecting fag
i regret not buying this when i had the chance. its sold out, now.
Anonymous No.106277417
>>106277285
>https://rentry.org/5exa3
check this out, it's well trained furry model combined with really well trained anime lora which made it so good that in some cases it's still better than sdxl illustrious
Anonymous No.106277427 >>106277434 >>106277473 >>106277508 >>106277595
>>106277257
>>106277266
>>106277341
This is a rip off of another anon that definitely knows more than me I'll post his screenshot after

>I tried cheating my way with Flux the similar way with the 1st ksampler at half resolution and I always ended up getting a total mess that couldn't be fixed no matter how many steps you'd set in the next ksampler with proper resolution. Then again, completely different model and that was a static image so go figure

I tried to use 3 ksampler advanced but when I upscaled the latent, the noise of the latent became basically christmas lights on the image. Tried adding noise but that degraded it. But the normal ksampler denoise fixes it.
Anonymous No.106277434 >>106277453 >>106277508
>>106277427
original workflow
Anonymous No.106277453 >>106277498
>>106277434
Oh hey it's my schizo workflow. I'm trying stuff with the kij workflow now but idk. I think as long as you have the motion and the high noise pass looks good enough you can upscale it and feed it to the low noise model at are .65~ denoise and it will turn out fine. I feel like the finer motions get lost in the process though, I'm trying to figure out how to fix it.
Anonymous No.106277473 >>106277553
>>106277427
Yeah it looks like latent size to dumber doesn't work. The point was it's better to get your width and height from wherever you set it so you don't have to manually change it with every aspect ratio. You can just use another 'upscale latent by' but if you decide on different scales than 50% or 75% the math changes to get it back to 100%.
Anonymous No.106277476 >>106277489 >>106277790
Anonymous No.106277489
>>106277476
I don't remember this xena warrior princess episode
Anonymous No.106277498
>>106277453
its still way better than anything I had with 2 samplers and lightning on both. Slow motion is fixed, prompt adherence is better and I can make darker scenes, and sometimes things just wouldn't move at all or just ignore me with lightning
Anonymous No.106277502 >>106277554 >>106277558
>>106277392
>>106277387
idk i'm just using a lora
Anonymous No.106277508 >>106277522 >>106277543
>>106277427
>>106277434
Thanks anon, I guess screenshots are better than nothing, I'll try to implement that.
I don't use lightx2v because I think it fucks up outcome too much (it's nsfw loras that make things slowmotion for me), so maybe I'll go something like that :
- first sampler on high only 5 steps at 50% size + no lora
- upscale latent to get back to 100%
- feed latent to second sampler with 40 steps using a dynamic cutoff between high and low (Wan MoE KSampler)

- same seed
- same scheduler/sampler

Would that work or am I missing something?
Anonymous No.106277513 >>106277802 >>106279524
Anonymous No.106277522 >>106277553 >>106277564
>>106277508
if you don't use loras at the start, wouldn't that defeat the purpose of what you want even if it's able to be normal speed? especially for nsfw
Anonymous No.106277543 >>106277553 >>106277564
>>106277508
Yeah it's mainly to fix lightx2v while still benefiting from the speed by just adding a little gen time at the beginning, but if your loras are giving you slowmo problems you can use it for that too.
Statler&(maybe)Waldorf No.106277546
>>106277311
beahagahah
Anonymous No.106277553 >>106277574
>>106277473
>latent size to dumber doesn't work
So resizing doesn't work?

>>106277522
>>106277543
Yeah that's the thing I want to test. If I'm able to retain the motion speed, but not impacting too much the actual scene itself and what's going on, since that's what I want added from the lora.
Statler&(maybe)Waldorf No.106277554 >>106277581
>>106277502
rocketnon this is just sad
Anonymous No.106277558
>>106277502
thanks anon <3
looks really cute
Anonymous No.106277562 >>106277795
Anonymous No.106277564 >>106277580 >>106277598
>>106277543
actually fuck what I said this anon >>106277522 is right for probably 99% of nsfw loras because they are usually about content/composition and that needs to be in the latent from the start
Anonymous No.106277568
>>106277405
there is a 40cm one up for preorder.
go for it same price as beach queen
Anonymous No.106277574 >>106277598
>>106277553
>So resizing doesn't work?

No that particular node that I was trying to use to fetch the starting latent size doesn't work like I thought. You just gotta do something else to make it more user friendly for yourself. The resize works fine.
Anonymous No.106277580
>>106277564
Yeah well fuck you enjoy your slow motion video. You can't have your cake and eat it too.
Anonymous No.106277581
>>106277554
who?
Anonymous No.106277595
>>106277427
forgot to mention you can OOM on the 2nd sampler if the upscale is too high. My 24GB was barely able to do 75% of 720p. There's probably some way to fix this and my workflow isn't dumping stuff from vram like the third sampler that switches to low noise does.
Anonymous No.106277598
>>106277564
I will still try it, maybe with enough noise and in i2v it could work.
Man, I wish there was an easy way to control speed.

>>106277574
I see.
Anonymous No.106277601
Anonymous No.106277636 >>106277682
Yeah so trying the upscale in the middle of the kij workflow without using the low model from a fresh set of noise doesn't really work well. Gonna try again just doing the high noise on its own, upscale it and then do low noise and see if that is cleaner.
Anonymous No.106277662
Has anybody tested the difference between t5xxl, flan, and gner encoders for Chroma? Gner is the newest, but it was apparently trained with flan.
Anonymous No.106277682 >>106277801
>>106277636
And here it is without upscaling the denoise pass as well.
Anonymous No.106277686
Anonymous No.106277722
Anonymous No.106277753
Anonymous No.106277759 >>106277763 >>106277792 >>106277812
3060 12gb + 16gb ram
can I run any t2vid? Would it be extremely painful?
Anonymous No.106277763
>>106277759
>16gb ram
legitimately might be an issue.
Anonymous No.106277790
>>106277476
>Get the fuck out gaijin bitch, or I'll slice you!
>Xenophobe Warrior Princess
Based Japan!
Anonymous No.106277792
>>106277759
you probably want to increase the size of your swap file so it offloads more onto your ssd which will make everything ultra slow but atleast it works that way.
Anonymous No.106277795 >>106277899
>>106277562
>That clenching
More of this japanese masterpiece please!
Anonymous No.106277798
Anonymous No.106277801 >>106278378
>>106277682
Back when your PC had sovl, no shitty bling
Anonymous No.106277802
>>106277513
Anonymous No.106277812 >>106277869
>>106277759
16gb system ram is the big issue, 32gb is like the minimum, for a good experience you'd want 64gb

A lot of vram to ram offloading is needed for these huge models
Anonymous No.106277835 >>106278059
>>106277349
Anonymous No.106277837 >>106277854 >>106278388
Anonymous No.106277854
>>106277837
lol
Anonymous No.106277869 >>106277900
>>106277812
let's say I buy 2 5060 ti 16gb + stick the 3060 12gb on the new system. Does comfy let you spread models across multiple cards? It's overkill but I'd be using it for llms too
Anonymous No.106277899 >>106277915 >>106278020
>>106277795
Anonymous No.106277900
>>106277869
I think you need some extension for that in Comfy, but I seem to recall people reporting it working

But two 5060 i 16gb ? Why not start with one 5060 ti and the 3060 ? The 5060 ti is about 2x the speed of the 3060 for inference.
Anonymous No.106277915 >>106277959
>>106277899
S-surely there are some scenes where they make out, like just for silly fun ?
Anonymous No.106277959 >>106277972
>>106277915
How exactly would that help them defeat the shogun?
Anonymous No.106277972
>>106277959
Morale building! Stress release!
Anonymous No.106277979
Anonymous No.106278020 >>106278034
>>106277899
catbox? surprisingly good hair
Anonymous No.106278034
>>106278020
https://files.catbox.moe/mo0qda.mp4

Beware the ancient Chinese secrets in this workflow.
Anonymous No.106278059
>>106277835
cute
Anonymous No.106278113 >>106278505
Anonymous No.106278183 >>106278305
how does lightx2v lora work and why can it speed up wan steps?
Anonymous No.106278244
>just tried nunchaku kontext
>8s~ to do 8stp 1cfg
BROS, when we get comfy nodes for QWEN and later WAN, it's gonna be a game changer, this multiplied my shitposting potential to 5x
Anonymous No.106278253
{
"nodes": [
{"type": "LoadImage", "id": "person_image"},
{"type": "LoadImage", "id": "suit_reference"},
{"type": "SegmentAnything", "id": "clothes_mask", "inputs": ["person_image"]},
{"type": "Flux_Kontext", "id": "clothing_transfer", "inputs": ["person_image", "suit_reference", "clothes_mask"]},
{"type": "Prompt", "id": "conditioning", "inputs": {"text": "a person wearing a formal business suit, realistic, professional corporate attire"}},
{"type": "NegativePrompt", "id": "neg_conditioning", "inputs": {"text": "extra arms, distorted, blurry"}},
{"type": "FluxSampler", "id": "sampler", "inputs": ["clothing_transfer", "conditioning", "neg_conditioning"]},
{"type": "SaveImage", "id": "final_output", "inputs": ["sampler"]}
]
}
Anonymous No.106278289
THESE NIGGAS TRYNA KILL ME
Anonymous No.106278298 >>106278424
>>106276887
New midwit upscale cope just dropped:

Upscale only the first frame of a smaller video then use the latent motion of that video fill in the gaps during an img2vid pass
Anonymous No.106278305
>>106278183
I dont know shit about shit but it doesn't speed up steps, it makes it so very few steps gives acceptable quality
Anonymous No.106278315 >>106278323 >>106278342
Is there a reason to use wan2gp over comfy?
Anonymous No.106278323
>>106278315
Yeah to get (You)s on 4chan by saying you use it.
Anonymous No.106278327 >>106278421 >>106278599 >>106278640
Is there are any way to remove background for pixel art that i was generated for
i try to used Inspyrenet, but it give out anti-aliasing, while Transparency Background Remover always remove line
Anonymous No.106278342
>>106278315
it avoids spawning the deranged anticomfy schizo
Anonymous No.106278378
>>106277801
It wasn't the PC. It was the world that was different.
Anonymous No.106278388 >>106278400
>>106277837
>hands clipping through the bars
they're definitely using shitty CGI to train these models
Anonymous No.106278400
>>106278388
there's probably game footage in there where NPCs clip through shit like that without the clipping being tagged
Anonymous No.106278413
Anonymous No.106278416
Anonymous No.106278421 >>106278599 >>106278843
>>106278327
The problem is that it's generating discoloration outside of what should be edge of the character. For the most part, the bg can be removed with a simple color selection. Unfortunately the extra garbage around the outside matches colors on the inside of the character, so it's more difficult to remove. I'm sure there is some algo you could write to select only those colors that are touching transparent pixels or something.
Which model/LoRA are you using?
Anonymous No.106278424 >>106278681
>>106278298
Actually, if the upscale isn't immaculate the whole world warps and shifts.
Anonymous No.106278425
Anonymous No.106278456 >>106278469
has someone trained a LoRA on Vixen media group movies?
Anonymous No.106278469 >>106278480
>>106278456
Sorry I'm not a coomer what is that?
Anonymous No.106278480 >>106278503
>>106278469
>Vixen Media Group owns and operates nine online adult film sites: Vixen, Tushy, Blacked, Blacked Raw, Tushy Raw, Deeper, Slayed, Wifey and Milfy.
Anonymous No.106278503
>>106278480
I'd rather see loras on classy porn, almost never done.
Anonymous No.106278505 >>106278581 >>106279323
>>106278113
Just testing my i2v upscale idea.
Anonymous No.106278516
this beauty took me only 3 hours to gen.
Anonymous No.106278525
Another day without svdquant for wan.
Anonymous No.106278581 >>106278593
>>106278505
we need something like multidiffusion but for video, would make it possible for vramlets to upscale
Anonymous No.106278593
>>106278581
Personally I would not mind an iterative upscaler for the low noise pass.
Anonymous No.106278599 >>106278675 >>106278843 >>106279030
>>106278327
>>106278421

The way I cut out backgrounds from my pictures is by using Flux Kontext then a paint program. On windows you can use paint.net with the color grim reaper extension (I think they renamed it?). On Linux I use Krita (as shown below in workflow).

Probably not so good with green characters, but usually bright green isn't found anywhere else on the image. When prompting you can pick a color which isn't found anywhere on your character. If you don't have such an option, then pick a color which isn't found near the edges of your character and you can either select around them, then use alpha, or use the magic wand tool.

The advantage with this strategy is that it allows you to get smaller details like flowing hair with some transparency. It is the current year of 2025 and I have yet to find a better solution. All the background-removal models I have tried are total dogshit. I have to wait over a minute to let Kontext do its thing before I can post-process it to have a transparent background.

I've attached an example, my workflow was exactly like this:
>comfyUI with Flux Kontext
>"replace the yellow background with solid bright green, keep the characters exactly the same" (into a 1024x1024 image, your input is still 128x128 though)
>open the output in Krita
>image -> scale image to new size -> 128x128 (your original size)
>"Filter -> Colors -> Color to Alpha" then there's a bar of white with the title "Color Sampler" -> use the color picker in the new menu to select the background green
>tweak the threshold until it looks right (I made part of a new background layer black and part of it blue so I can see how much of the character's other colors I'm losing and how neat of a cut I'm getting, and I go somewhere that looks good)
>export image as PNG with Alpha channel for transparency and no HDR

Nice sprites, how did you generate the sprite map? I can never get consistent sprites.
Anonymous No.106278638
Anonymous No.106278640
>>106278327
Pirate photoshop and use background remover.
Also is the character sheet actually consistent enough?
Anonymous No.106278670
Anonymous No.106278675
>>106278599
you're getting a lot of anti-aliasing with that method, which he said he doesn't want
Anonymous No.106278681 >>106278696
>>106278424
this is what it actually looks like when you eat mushrooms
Anonymous No.106278684
Anonymous No.106278692
Anonymous No.106278696 >>106278705
>>106278681
acid too.

Try another upscale idea with i2v. here is no upscale.
Anonymous No.106278705
>>106278696
And here is upscale. Not too bad desu.
Anonymous No.106278742 >>106278751 >>106278926
what happens if you give this to wan
Anonymous No.106278744
anyone here have tips for training wan loras? my only lora experience is with SDXL
Anonymous No.106278748
Anonymous No.106278751
>>106278742
or this
Anonymous No.106278762
Anonymous No.106278800 >>106278808 >>106278984
No i2v upscale
Anonymous No.106278806
>the plan
push /ldg/ somewhere else
(into the trashcan & or dumpster)
Anonymous No.106278808 >>106278844 >>106278984
>>106278800

with i2v upscale. There seems to be some ghosting?
Anonymous No.106278815 >>106278817 >>106278845 >>106278987 >>106279104
should i get amd or nvidia for a linux server ?
Anonymous No.106278817
>>106278815
AMD all the way for image gen.
Anonymous No.106278843 >>106279030
>>106278421
>>106278599
lmao no the model only work once in blue moon, and if it need to draw creature or fantasy-like humanoid, than it would break down and cry like like crazy
But thank for helping
Anonymous No.106278844 >>106278984 >>106278989
>>106278808
upped the denoise slightly.
Anonymous No.106278845
>>106278815
don't listen to him
get Nvidia
Anonymous No.106278926 >>106278998
>>106278742
Where do I find a gf like this?
Anonymous No.106278974 >>106278980
Another one. No i2v upscale.
Anonymous No.106278980
>>106278974
with i2v upscale.
Anonymous No.106278984 >>106279005
>>106278800
>>106278808
>>106278844
You're trying to get that book to say "FUTANARI HENTAI" aren't you?
Anonymous No.106278987 >>106279079
>>106278815
Nvidia

Sadly there's no other choice for AI at the moment, in a year or so AMD should be competitive, not now though
Anonymous No.106278989 >>106279005
>>106278844
You might want to manually calculate denoise or use Ksampler Advanced, that one automatically sets denoise like this: (total steps - steps on the 2nd sampler)/total steps
If you use 4 steps in the 1st sampler and 4 steps in the 2nd sampler then denoise would be 4/8=0.5
If you have 2 steps in the 1st sampler and 6 in the 2nd sampler then denoise would be 6/8=0.75
Anonymous No.106278998
>>106278926
In the caves below the japanese subway system
Anonymous No.106279005
>>106278984
Yes but it really resists it.

>>106278989
I'm using kij nodes right now so I'm kind of fucked in that regard. I can always switch.
Anonymous No.106279030
>>106278599
>>106278843

Here's my best shot to fix the anti-aliasing issue:
>start with Krita and open your image (I used your original provided in the thread)
>image -> scale image to new size (128x128 -> 1024x1024) + filter type is set to "nearest neighbor"
>put it into flux and remove any garbage outlines (I had to tweak the prompt a bunch but I got this to work: "keep the characters exactly the same, clean up the outlines so they are only black, there should be no peach outlines, remove any excess incorrect outlines by turning them into background")
>I run the output image through flux a second time, this time to convert the background to solid green
>Krita -> image -> scale image to new size ->256x256 then ->128x128 (the 256 in the middle helped a little with the outlines)
>"Filter -> Colors -> Color to Alpha" (as in my previous post)
>export image as PNG with Alpha channel for transparency and no HDR

The outlines on this one are a little thicker but that was flux trying to clean up any unwanted outlines like the peachy color outside the characters model. The character is a lot clearer than my last attempt and a lot "less" anti-aliased (at least on the inside) but still has blurry outlines. Honestly, the best strategy might just be using magic wand tool to select the outside, then using a straight edge cutter to clean up the outlines. It will be more consistent and probably faster too (at least on pixel art) than my method. I tried a workflow where I simply added a bunch of yellow background to the original image until it was 1024x1024 in size, but the characters remained the same, however Kontext kept blurring the characters.

This is the best I can do with my crappy workflow. The only time I can see this helping is if you have to do very large sprite sheets. I think the best strategy is just magic wand tool with a little clean up by hand. Hope you find a good solution that works for you, good luck.
Anonymous No.106279042
Anonymous No.106279079 >>106279107
>>106278987
>year or so AMD should be competitive
they said this last year
Anonymous No.106279094 >>106279105 >>106279673
>in a year or so AMD should be competitive
How much longer people gonna say that?
Anonymous No.106279104 >>106279126
>>106278815
linux support is better for AMD cards, and they are starting to sell 9700 pro cards with 32GB VRAM for affordable pricing. speed will be slower than nvidia though
Anonymous No.106279105
>>106279094
they'll keep saying it until china suddenly reveals something for AI that's competitive. it wont be intel or amd
Anonymous No.106279107
>>106279079
Well, you gotta have hope. If there's no real competition to Nvidia, hardware improvements will continue to be heavily throttled, while local models just become larger and more demanding.
Anonymous No.106279125
Anonymous No.106279126
>>106279104
For overall usage, yes, probably.

For AI usage, no.

When it comes to AI usage, the linux NVidia driver is the best there is, because Linux is where all the AI research and production is being done.
Anonymous No.106279132
Anonymous No.106279211 >>106279245
Anonymous No.106279245
>>106279211
Heartwarming bullying, my new fetish
Anonymous No.106279270 >>106279284 >>106279288 >>106279312 >>106279378 >>106279393
qwen nunchaku bros... WE WON!
Anonymous No.106279284 >>106279289 >>106279298
>>106279270
>not a single qwen image in the thread
Anonymous No.106279287
Bros, can I run qwen on a 12gb4070 and 32gb RAM?
Anonymous No.106279288 >>106279293
>>106279270
How much of an upgrade it is to Q4?
Anonymous No.106279289 >>106279753
>>106279284
I was playing around with kontext, I thought I could use normal flux loras with it but.. you cant!!!! MEGA SAD!! cant even undress the anime ladies FUCK KONTEXT!!!!!!!!
Anonymous No.106279293
>>106279288
it's better than FP8
Anonymous No.106279298 >>106279306
>>106279284
Is there even a official workflow yet, I thought it was to be done by Monday ?
Anonymous No.106279300
Talking about inpaiting, is chroma good at it?
Anonymous No.106279301
>>106277285
>I have 8 gigs of ram and it's not enough to run SDXL
wtf are you talking about. you gen with sdxl no problem
Anonymous No.106279305 >>106279461 >>106279735
Anonymous No.106279306
>>106279298
the official nunchaku nodes arent even updated yet and yes, like you mentioned next week they'll provide updated nodes + wf. Right now you have to use diffusers to gen
Anonymous No.106279312 >>106279314
>>106279270
wansister, our response?
Anonymous No.106279314 >>106279335
>>106279312
wan is next on the list of nunchaku. Imagine genning video slop in 30 secs in stead of 3 minutes! AGI BROSSS
Anonymous No.106279320
But have they fixed sage attn working with Qwen?
Anonymous No.106279322
>>106277285
>I have 8 gigs of ram and it's not enough to run SDXL.
lol
skill issue
I run flux on 6 GB of VRAM just fine
Anonymous No.106279323 >>106279334
>>106278505
cool, is this using WAN to upscale? Or a separate upscaling model?
Anonymous No.106279334 >>106281858
>>106279323
Upscale the first image and latent of a video and then running it through i2v. I don't know if it's actually worth it though.
Anonymous No.106279335
>>106279314
As long as I can get rid of the awful (but useful) lightx2v and still be able to gen under like 10min, I'll be happy.
Anonymous No.106279378
>>106279270
>no lora support
LOL!
Anonymous No.106279387 >>106279541
But fr, why do they all look like Xena?
Anonymous No.106279393
>>106279270
no comfy support yet
Anonymous No.106279404 >>106279422 >>106279460
>hyping up 4bit slop
new low for this general
Anonymous No.106279422 >>106279435 >>106279460
>>106279404
I gotta agree. I don't understand the excitement. Qwen as an image model has been nothing but bland. (this is not a message in support of chroma.)
Anonymous No.106279435 >>106279460
>>106279422
it was just a new flavor, the honeymoon period wore off almost instantly
Anonymous No.106279456 >>106279637
Anonymous No.106279460 >>106279466
>>106279404
>>106279422
>>106279435
>>>/sdg/
is that way. Imagine playing down the ability to generate slop faster.
Anonymous No.106279461 >>106279517 >>106279755
>>106279305
comfy. why are they in their underwear though?
Anonymous No.106279466
>>106279460
>NOOO YOU CANT CRITICIZE MODELS
Anonymous No.106279469 >>106279478 >>106279548 >>106279655 >>106279823
is 64gb ddr4 ram enough?
Anonymous No.106279478 >>106279555 >>106279823
>>106279469
if you want to gen at 720p, not really
Anonymous No.106279517
>>106279461
it's comfier
Anonymous No.106279524
>>106277513
nice
Anonymous No.106279541
>>106279387
>Oh, my hands just slipped, sorry!
Anonymous No.106279548 >>106279555 >>106279561 >>106279641 >>106279823
>>106279469
Yes, but barely, which is kinda sad
Anonymous No.106279555 >>106279585
>>106279478
>>106279548
what about DDR5?
Anonymous No.106279561 >>106279641
>>106279548
Anonymous No.106279568 >>106279589 >>106279597
>wan2.2
>requires 2 fucking models
>still 16 fps 5 seconds
>still worse than online t2v
holy fucking shit. xi failed locally
Anonymous No.106279585
>>106279555
There's no noticeable difference between using ddr4 or ddr5 when it comes to vram offloading, so just buy more ddr4 if you already have it and just make sure the timings are the same to maximize stability / performance
Anonymous No.106279589
>>106279568
It's a pretty god damn amazing tool to have on your own machine. Nothing to doom about.
Anonymous No.106279597
>>106279568
go back to /saasdg/
Anonymous No.106279637
>>106279456
kek
Anonymous No.106279638 >>106279712 >>106279747
>>106270744
>>106271257
>>106271326
https://www.reddit.com/r/StableDiffusion/comments/1mraxv0/sdxl_with_native_flux_vae_possible/

Why is /ldg/ always wrong?
Anonymous No.106279641 >>106279678
>>106279548
>>106279561
is 96gb enough?
Anonymous No.106279655 >>106279823
>>106279469
On Linux yes.
Source: me running a 3090+64GB RAM on a headless ubuntu server and generating wan videos in 720p.
Now I have 128GB but also 2x3090.
Anonymous No.106279673
>>106279094
it's competitive right now in the 3090 range.
Anonymous No.106279678
>>106279641
64gb is enough atm moment, so yes, 96gb is certainly enough

The more free ram you have, the more greedy the offloading algorithm seems to be, which speeds things up, if it speeds things up to be worth 96gb, not sure, I only have 64gb ram in my machines
Anonymous No.106279692
I was maybe too enthusiastic (and horny) and planned too many wan video gens (each takes 20min).
Is there a way to save the queue?
Anonymous No.106279712
>>106279638
i think like 60% of the way may be easy, the hard part is the rest of getting it actually usable
Anonymous No.106279733
I have had my workflows chug multiple times with 64gb, but I'm loading a fair bit of unnecessary snake oil too.
Anonymous No.106279735
>>106279305
was that vramlet
Anonymous No.106279747
>>106279638
I assume this is a troll post, because nobody can be this retarded. The posts you link are 100% correct. Nobody ever said you can't retrain SDXL with a Flux VAE, just that nobody will.

It would be incompatible with all existing loras, and the SDXL finetunes would gain nothing from this unless they too were retrained from scratch.

Nobody will spend the huge amounts of money and time to retrain such an old model architecture for something 100% incompatible.
Anonymous No.106279749
Is there any blower style 5090? Aka sending most of its hot air to the back of the case and not recirculating it in the case?
Anonymous No.106279753
>>106279289
>I thought I could use normal flux loras with it but.. you cant!
You can but might require upping the strength
Anonymous No.106279755 >>106279772
>>106279461
I can't help myself, everyone needs bikini armor.
Anonymous No.106279760 >>106279791 >>106279809 >>106279816
can fragments in vram accumulate over time?
it's weird but comfy just took an age to do what it had been doing sharpish for days, or weeks, or months, I never turn off my PC, and whatever was jammed in the pipes cleared out after a reboot
Anonymous No.106279772
>>106279755
>everyone needs bikini armor.
This
Anonymous No.106279791 >>106279802 >>106279803
>>106279760
>can fragments in vram accumulate over time
lmao dude, do you know how pcs work? can you articulate this in a non-retarded manner? fucking retard
Anonymous No.106279802 >>106280061
>>106279791
fragment = small thing
vram = video random access memory
you're welcome
Anonymous No.106279803 >>106280061
>>106279791
someone didn't have their fruity pebbles this morning
Anonymous No.106279809
>>106279760
it's a great way to farm latent crystals actually
Anonymous No.106279816
>>106279760
Yes, not only can, it does.

CUDA has special environment variables to deal with this, like:

expandable_segments

max_split_size_mb

garbage_collection_threshold

which you place in PYTORCH_CUDA_ALLOC_CONF=""

These help minimize fragmentation, HOWEVER, they don't do anything in Comfy because he uses his own homemade CUDA memory allocator, which seems to bleed vram profusely at times.
Anonymous No.106279823 >>106279851 >>106279955
>>106279469
>>106279478
>>106279548
>>106279655
I have 24 & 32 and I gen at 720p, no virtual vram. My ram does hit max at some points and I see my swap get used, not necessarily at the same time which is confusing. But I haven't noticed poor gen times compared to anyone else around here. There's extra time when genning the first time, which should be true for everyone because it has to pull the models off your drive first, and there's extra time added when I change the prompt, but I think the actual gen when it's showing the progress bar goes at a normal speed. On wan 2.1 my gens were somewhere over 5 minutes. My typical gen now is 3 minutes for the first two samplers and 4 1/2 minutes for the last one, which is 4 steps at 720p. Prompts executed in 10 minutes. Though I think it was something like 14 minutes when I was on two samplers at full res. I'm going to upgrade ram just because I don't like seeing it topped out and I'm curious what improvement I get. But I'm definitely not getting it because I'm suffering.
Anonymous No.106279851 >>106279955
>>106279823
Well if you have 24gb vram then there's less need for offloading, meaning you don't need as much system ram.
Anonymous No.106279928 >>106279947
today I learned at that 2.2 lightning was absolutely trashing my attempts at dark scenes and made them impossible, and probably crippled environments too

candlelit in a dark mansion with lightning was a dumpster fire, with lightning off it's as good as flux.
Anonymous No.106279947 >>106280027
>>106279928
You should subscribe to me posts more, Anon. I've been screaming this from the rooftops for the last three days.
Anonymous No.106279955 >>106279981
>>106279823
>>106279851
also I wonder if its like raid 0 where if you have more memory it can split the model to different ranks/sticks/channels/idontfuckingknow and read it twice or four times as fast
Anonymous No.106279975 >>106280003
I want to buy a pre-built PC because the cost of individual parts will exceed the pre-built one
the tower PC I want to buy says
>Maximum Memory Supported 32 GB
it means that the maximum a single RAM chip memory it can support is 32 GB, right?
It has four slots so I can go 32 x 4, right?
Anonymous No.106279981
>>106279955
I very much doubt it has any impact.

The bottleneck is the GPU bus, you can have fast DDR7 both on the GPU and system ram, it likely won't make any noticeable difference since both sides are mainly waiting while they squeeze data through the bus back and forth.
Anonymous No.106280003
>>106279975
Find out what motherboard it uses and you can check directly.

32x4 for a max of 128 is a common figure.

Saying 32GB is the max supported is weird phrasing and there's no need to speculate what they meant anyway. If they dont say what motherboard it uses you should buy from someone else
Anonymous No.106280027 >>106280125
>>106279947
It probably has a lot to do with cfg 1 too. But with lightning, turning the cfg up much more than 2 burns the output even if you are willing to wait
Anonymous No.106280048 >>106280297 >>106281827
I am making a card game about exotic tropical flora and fauna, what is the chance the AI will know how draw a "buriti palm" or "mico leΓ£o dourado"?
Anonymous No.106280061 >>106280070
>>106279802
>>106279803
there's actual technical terms instead of this normie computer illiterate garbage you're spewing, but I guess this is /g/ after all, can't expect any level of competency here.
Anonymous No.106280070 >>106280103
>>106280061
by all means, go ahead
Anonymous No.106280103 >>106280124
>>106280070
sure, it's called memory leak, retard, mostly due to shitty drivers or software that cant manage vram memory allocation properly.
Anonymous No.106280124
>>106280103
Fragmentation is not the same as memory leak though

Memory leak is unused memory not being returned to the system

Memory fragmentation becomes a problem when you end up with a lack of large continous memory blocks available for allocation
Anonymous No.106280125
>>106280027
I wonder if this could work; https://github.com/Extraltodeus/Skimmed_CFG
Anonymous No.106280185 >>106284077
Anonymous No.106280246 >>106280372
Anonymous No.106280279 >>106280796
Anonymous No.106280297
>>106280048
Very unlikely. Actually, zero chance. You'll need to train a lora on plant species.
Anonymous No.106280336 >>106280377
Anonymous No.106280372
>>106280246
>suddenly, blood appears!
That aside, it's insane how far we have gotten and how it's accelerating.

With minor fixing, this could be straight out of a seasonal anime.
Anonymous No.106280377
>>106280336
kek
Anonymous No.106280446 >>106280475 >>106280522
Anonymous No.106280475
>>106280446
>Dungeon twerking for amateurs
Anonymous No.106280522 >>106280759
>>106280446
lusty argonian maid when
Anonymous No.106280573
Anonymous No.106280675
>>106277392
>>106277349
cursed
Anonymous No.106280688 >>106280773 >>106280799
how do I get over the fact that the big companies and govt have 10x better than what you have in local now?
Anonymous No.106280759 >>106280790 >>106280797
>>106280522
I should never have prompted this.
Anonymous No.106280767 >>106280831
So I want to do some large batch of images while I am away. I want to mix different CFG strengths, samplers, etc for this purpose.
I know that technically I can just add to the queue, but is there a nice, convenient comfyUI extension that can randomize gens across a given batch number within the specified parameters?
Anonymous No.106280773
>>106280688
how do I get over the fact that very rich people have 10x better houses, cars, clothes, women, than what you have right now ?

indeed, sir, how do you ?
Anonymous No.106280790
>>106280759
You sacrified your sanity for science, a noble effort
Anonymous No.106280796
>>106280279
you should do a music vid like that one anon did using these
Anonymous No.106280797
>>106280759
SEXO
Anonymous No.106280799 >>106280884
>>106280688
can they coom to it though?
Anonymous No.106280831 >>106281005
>>106280767
There's probably a hundred of those. Would be better if you could make one yourself using native nodes in Comfy and tailor it exactly to your needs, but Comfy doesn't have the necessary nodes natively, instead you get a worthless HUD display.

Enjoy!
Anonymous No.106280877 >>106282457
Anonymous No.106280884
>>106280799
they must have a nsfw version.
no reason trump doesn't have one
Anonymous No.106280928 >>106281027
Anonymous No.106281005 >>106281059
>>106280831
I mean I am not a good programmer at all but I can probably make a variant of the sampler node allowing randomization within the user controller parameters if I take my time with it.
There is probably already some node like this out there though, just gotta find it.
Or maybe good excuse to not be a lazy piece of shit and brush up some basic python skills.
Anonymous No.106281027 >>106281517
>>106280928
nice
Anonymous No.106281042
Finally got WanGP to work with the poor setup or whatever. Now I just need to get the proper 14B setup to work...
Anonymous No.106281059 >>106281125
>>106281005
I doubt you can, you would need nodes to pick random words/sentences from lists and then combine them and pass them off to the prompt

I do not think the necessary nodes exist natively, I'd be happy if you proved me wrong
Anonymous No.106281080 >>106281200
retard here, I used to train loras with prodigy/prodigy plus with learn rate 1 getting some solid results, but some people ITT(or maybe it was the same guy doing it multiple times) kept recommending me Adamw_8bitat 1e-4 + cosine, and learning here seems much slower. At 1000 steps, the results with the prodigy trainer wer pretty solid already, but with the mentioned Adamw_8bit setting it seems to only be starting to pick up the concepts at the same number of steps. Is it supposed to be lie this or did I get memed?
Anonymous No.106281125
>>106281059
That's not what I described at all but sure.
Anonymous No.106281200 >>106281263
>>106281080
Well, Prodigy is a 'self-learning' optimizer, as in you set the LR to 1, but the optimizer will figure out the (in its estimation) optimal LR and use that.

I typically use adamw with a constant scheduler, meaning it will keep ~the same LR throughout the training (adamw is a bit adaptive, but not like Prodigy).

Using cosine scheduler like you did, means it will start at 1e-4 and then lower the learning rate as the training goes on. For cosine 1e-4 sounds low, but I don't know what model you're using or how many images, batch size etc.
Anonymous No.106281263 >>106281307
>>106281200
model is biglust_v16, sample size 60, batch size 4, same as when doing it with prodigy. From what I got from your post as a noob it means that the LR is much slower with the Adamw settings used compared to the previous prodigy settings so it needs more steps to learn right?
Anonymous No.106281307 >>106281338
>>106281263
Well, I don't know what Prodigy decided was the optimal LR, but if the learning is much slower with adamw 1e-4 then yeah, it's a safe bet that prodigy decided on a higher LR.

Now lower is not necessarily better even if you have infinite time, if it's too low it will simply fail to learn certain things, sadly there's no exact science here, just baseline estimates.
Anonymous No.106281338
>>106281307
So I guess this is where you just have to try out things to see how they work out. Thanks.
Anonymous No.106281345 >>106282331
It came up a couple days ago, there was a ComfyUI node that had resolution presets you could pick from. What was the name of that?
Anonymous No.106281378 >>106281430
Anonymous No.106281406
GOMFYUI
PUT BENIS IN NODE
EBIN :-D
Anonymous No.106281430
>>106281378
added too much interpolation on this one, with less it looks better
Anonymous No.106281435 >>106282161
comfy should be dragged out on the street and shot
Anonymous No.106281449
>>106277311
Made me throb
more?
Anonymous No.106281454
I want to try out making some qwen loras

What is the best way to mass tag some images at the moment, I think I need natural language not booru tags for qwen right?
Anonymous No.106281491
I want nano-banana.
Anonymous No.106281517 >>106281618
>>106281027
It's not too shabby, though the panties are all very modest.
Anonymous No.106281519 >>106281578 >>106281634
Anonymous No.106281578
>>106281519
kino
Anonymous No.106281618
>>106281517
I like them
I prefer white ones but these are also nice.
Anonymous No.106281634
>>106281519
Cyclist was at fault.
Anonymous No.106281827
>>106280048
maybe you can image to image them. I also saw a thing recently where you can gen entirely new images that will maintain a consistent subject that you provide an image of
Anonymous No.106281858
>>106279334
Have you tried this on real not genned videos?
Anonymous No.106282059 >>106282478
Anonymous No.106282083 >>106282373
For Wan2.2 Kijai's i2V workflow. It says that I can easily adapt to 720p. Is the whole part about switching to the 720p models specific to Wan2.1?

Can i just up the resolution on that Wan2.2 workflow and it should be fine? Maybe remove the lightx2v node?

(also i got a 5070ti with 16gb of VRAM and 64gb of RAM, maybe I should give up now on 720p if im too much of a vramlet?)
Anonymous No.106282110 >>106282154 >>106282672
Anonymous No.106282154 >>106282178
>>106282110
really cool
Anonymous No.106282161 >>106282382
>>106281435
*trani
Anonymous No.106282178 >>106282672
>>106282154
Nice water physics, but I wanted her to become completely submerged and it just refused to do it. This is the closest it got in 3 attempts.
Anonymous No.106282271 >>106282292
Anonymous No.106282292 >>106282321 >>106282452
>>106282271
we're barrelling full force towards climate collapse just so you can make pure garbage like this
Anonymous No.106282321
>>106282292
Maybe people from the old testament wouldn't have been so moody about continental floods if they had this kind of tech to pass the time
Anonymous No.106282331 >>106282691
>>106281345
It was D2 size selector (you can change the presets by editing the "config.yaml" file in the node folder); https://github.com/da2el-ai/ComfyUI-d2-size-selector
If you gen with FLUX/Chroma I recommend Flux Resolution Calculator; https://github.com/gseth/ControlAltAI-Nodes
Anonymous No.106282373
>>106282083
for wan 2.2 you can just change the resolution. it was wan 2.1 that had seperate models for 480p/720p.
Anonymous No.106282376
please anon for the love of god interpolate your videos
Anonymous No.106282382 >>106282697
>>106282161
nah, ani made a faster frontend instead of heaped POOP bullshit that locks you into fascist nodes
Anonymous No.106282452
>>106282292
Greta, it's over, it was all a sham
Anonymous No.106282457
>>106280877
saar!
Anonymous No.106282478
>>106282059
>Dear fellow scholars...
Anonymous No.106282607
FUCK OFF /LDG/ SCHIZO
>>>/vp/58122207
Anonymous No.106282672
>>106282110
>>106282178
Seeing this made me think you could do the GITS making of cyborg sequence
Anonymous No.106282691
>>106282331
I was being a dumbass and didn't hit up google. When I did I found this Olm Resolution Picker which has a text file you can edit which works perfect for my use case.
Anonymous No.106282697 >>106282900
>>106282382
>faster
Anonymous No.106282712 >>106282768 >>106282914
besides comfyui what's the go to ui today? sdnext? forge? reforge? something else?
Anonymous No.106282768
>>106282712
bash scripts calling python calling cuda.
Anonymous No.106282772 >>106282892
Anonymous No.106282892 >>106282915
>>106282772
Is that the retro videogame Chroma lora someone posted here ?

Really nice.
Anonymous No.106282900
>>106282697
>can't separate backend from frontend
forgot this was /g/
Anonymous No.106282914 >>106282958
>>106282712
Forge/reforge
Anonymous No.106282915 >>106282957
>>106282892
It's a retraining experiment of that on Qwen-Image. It's actually undertrained; right now, I am training something else
Anonymous No.106282943 >>106282955 >>106282958
Yeah. I guess i'm going for that.
What's the difference between the different forge forks? I don't know which one to choose.
Anonymous No.106282955 >>106283023
>>106282943
>>106277108 (OP)
>re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
Anonymous No.106282957 >>106283173
>>106282915
What were the ~results vs Chroma ?

I gave Qwen a try on 640 resolution, but the results were so-so, that said I haven't experimented with settings at all. If I have to use the native 1328 resolution for Qwen training, my old dusty 3090 will be too slow.
Anonymous No.106282958
>>106282943
>>106282914
Anonymous No.106282975
>>106282974
>>106282974
>>106282974
>>106282974
Anonymous No.106282978 >>106283104
Eternal SDXL VRAMlet here
>>106282861

Can anyone of you, VRAMGODS, test this NSFW anime model. The person who made this model is a "renowned" anime checkpoint maker.

Thanks
Anonymous No.106283023 >>106283150
>>106282955
which fork?
Anonymous No.106283104
>>106282978
I would but I'm also a VRAMlet
Anonymous No.106283150
>>106283023
whichever has the features you want, fren
Anonymous No.106283173
>>106282957
>What were the ~results vs Chroma
Just undertrained / less effective
Anonymous No.106283753
>>106277285
Dreamshaper
Anonymous No.106283762
>>106277295
Flux is garbage.
Anonymous No.106284077
>>106280185
exceptional