← Home ← Back to /g/

Thread 105669256

323 posts 166 images /g/
Anonymous No.105669256 >>105669353 >>105670099 >>105670118 >>105673344
/ldg/ - Local Diffusion General
Good Luck with the Experimenting Edition

Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>105667276

https://rentry.org/ldg-lazy-getting-started-guide

>UI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Models, LoRAs, & Upscalers
https://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info

>Cook
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>Chroma
Training: https://rentry.org/mvu52t46

>WanX (video)
https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1

>Misc
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Archive: https://rentry.org/sdg-link
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Bakery: https://rentry.org/ldgcollage | https://rentry.org/ldgtemplate
Local Model Meta: https://rentry.org/localmodelsmeta

>Neighbors
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/celeb+ai
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.105669291 >>105669329
You're not entitled to my frenship.
Anonymous No.105669329
>>105669291
:(
Anonymous No.105669350 >>105669361
Anonymous No.105669353
>>105669256 (OP)
neighbors: >>>/vp/napt/

:3
Anonymous No.105669361 >>105669386
>>105669350
>High CFG - The animation
Anonymous No.105669373 >>105669420 >>105669535 >>105669761
Is a trigger word tag necessary for a lora? Isn't SD supposed to automatically apply the Lora if you... use the Lora?

I don't get why you have to train a lora with one extra tag.
Anonymous No.105669386
>>105669361
No, it's the shader noise sampler. Actual mental illness generator.
Anonymous No.105669420 >>105669429 >>105669441
>>105669373
Shit like "pu55ytw3rk1n" is not necessary and utterly retarded, yes. Models learn by association and tag frequency in the set. If you're training a person LoRA, you always want their name in the caption for every image they're in, and then it learns to associate said person in the images with that name/tag. For styles, you literally just describe the scene in tags/natural language, depending on the model. You can add stuff like "pixar style, 3d" if it's a Pixar LoRA to reinforce the style though, but you still want to caption it just like you'd actually prompt it
Anonymous No.105669429
>>105669420
>not necessary
It's my lucky charm, okay? I don't feel good unless I put it in.
Anonymous No.105669431
>>105669270
/LDG/ ladies & gentlemen
Anonymous No.105669441 >>105669447 >>105669493 >>105669923
>>105669420
But isn't stupid to use the same tag for all of the images if you are copying an artstyle? You are going to use the Lora for that artstyle, why would you tag it as if you wanted it out of the Lora if you would not tag it.
Anonymous No.105669447
>>105669441
The Lora learns what you don't tag as a core trait for the Lora. Or so I'm told.
Anonymous No.105669466
>>105669315
gooning is vital daily exercise.
Anonymous No.105669493 >>105669923
>>105669441
I made a 3D style LoRA based on a videogame, then trained it for illustrious. I only used normal tags, no same tag for the style (game cg, 3d). The end result was weak. Then I trained it again with those two tags, and it was a LOT stronger. For styles, I'd say train two loras, one with 'trigger' style tags and one without and see what works best. Most image models don't take long to train on
Anonymous No.105669535 >>105669599
>>105669373
I remember reading that you want to attach and associate styles to tags and words and concepts that the model already knows. So if you're training an oil painting lora, you wouldn't use the artists name, you'd use 'oil painting', and it'd sort of replace the generalized knowledge and images it was trained on for 'oil painting' with the new style you teach it, making the effect stronger than if you used no trigger word at all

Is it true? I have no idea. There's a lot of misinfo and misinterpretations when it comes to training loras

Either way, I don't see how it could hurt to have them though, unless someone has evidence to the contrary
Anonymous No.105669557
Anonymous No.105669564 >>105669577
cursed thread of hostility
Anonymous No.105669577
>>105669564
Fear not friend, I'm here now!
Anonymous No.105669599 >>105669605
>>105669535
yeah that's how it works. if you're training a lora and 3d stuff without the 3d tag you're training the 3d style from scratch. with the 3d tag you're training what the model already knows to be like what you want. you're basically telling it "this is what 1girl, 3d, hatsune miku, huge breasts, sagging breasts, futanari, huge cock, veiny cock, huge testicles, sagging testicles, excessive pubic hair, armpit hair, armpit hair peek, projectile cum, projectile lactation very sweaty, shiny skin, smell, steaming body, scat, cat" looks like now
Anonymous No.105669605
>>105669599
Anonymous No.105669612 >>105673373
Anonymous No.105669706 >>105669764
so ermmmm what happened to that flux release comfykek was hype-raising about?
Anonymous No.105669707 >>105669760 >>105669781
I've seen this slapped as a note in several workflow. Do multiples of 2 of these also work fine?
Anonymous No.105669735 >>105669852
Blessed thread of frenship
Anonymous No.105669755
it cant be blessed if my beautiful creations are never in the collage but these disgusting pics are
Anonymous No.105669760
>>105669707
The goal is get an image with roughly 1 million pixels total in it, because that's what sdxl was trained on. With some optimizations and side nodes like kohya's upscale you could get away with 1.5M or even 2M pixel base resolution images. Flux also can handle 1.5M just fine on its own too, not sure about going higher than that.
Anonymous No.105669761
>>105669373
It's not strictly necessary, but it helps a lot with training. The model needs to know what it's learning. Without a main keyword, the model won't have a main concept to associate its new knowledge with and will take a longer time to learn or might not even learn it at all. Some people compensate for this by using very high learning rates, which would easily lead to overcooking and poor generalization. However, if you use the same main keyword too often, it will also lead to overcooking.
Anonymous No.105669764
>>105669706
cumfy org has an deal for early exclusive access to the pro model through api nodes once the dev model is out
that's why they have a financial interest in shilling it
Anonymous No.105669772
why no Loras to make video game HUDs? all i find is a shitty dark souls health bar. sad
Anonymous No.105669781 >>105669791 >>105670185
>>105669707
Do you mean like, say, 2048x2048? If so, no, you want to do an upscaling process where you decode the image, upscale it with an upscale model, then encode it and use it as the latent for a separate KSampler process with low denoise (Ultimate SD Upscale consolidates this and has tiling options, though tiling is best avoided if you have VRAM)

Optionally use a Controlnet with the upscaled image into the second KSampler, this will make the output more consistent to the original since inference can change details

The reason those resolutions are listed is because those are the resolutions trained at. It can handle a little variation, but if you go too high you'll start seeing weird aberrations, humans merged into second humans with excessive features and limbs, things like that.
Anonymous No.105669791
>>105669781
*i mean you want to use the output of the first KSampler this way, to be clear
Anonymous No.105669834
why does batch generating fucks with regional prompter in Forge? such a strange phenomenon
Anonymous No.105669852
>>105669735
My dude!
Anonymous No.105669855 >>105669896
I'm doing the /ldg/ Comfy I2V 480p workflow from the wan rentry. Is it normal, that without offloading (unless there is offloading by default) it takes up 43 GB of my RAM?
I got the same error as this guy https://www.reddit.com/r/comfyui/comments/1lgxzym/last_dimension_must_be_contiguous/ and had to select CPU for the CLIPLoader. Could this be why so much RAM is used? I thought, or hoped, after the initial CLIPLoader node only the GPU would be used for the rest
Anonymous No.105669886
>>105668302
noice
Anonymous No.105669896 >>105669937
>>105669855
offloading is enabled by default
if you have 24gb vram just replace the distorch node with plain unet load gguf
Anonymous No.105669909
>>105668302
Do Kat too please
Anonymous No.105669923
>>105669493
>>105669441
for art styles i usually leave the captions completely empty lmao.
granted this is semi retarded because it takes on certain aspects like eye colors if for instance you have an overwhelming amount of red eyes in the dataset. best to tag "red eyes" then, so it doesn't learn that as part of the lora.
Anonymous No.105669937
>>105669896
Nah I'm a 16GBlet, so I guess I need offloading. I mean I was prepared for it but I didn't think it would need that much
Anonymous No.105670099 >>105670126 >>105670226 >>105670402 >>105670424 >>105670608 >>105670870 >>105671323 >>105671804 >>105672062
>>105669256 (OP)
I made a tag explorer with Illustrious gens for all face/hair/style/composition tags as well as 18000+ artist tags. Could be useful.

https://tagexplorer.github.io/
Anonymous No.105670118 >>105670870
>>105669256 (OP)
anyone able to prompt this without img2img i tried for like half hour but couldnt kek
Anonymous No.105670126
>>105670099
thats neat bro, I was looking for something like this.
Anonymous No.105670161 >>105670238 >>105671142 >>105671200
what are the next things to get excited for?
-chroma v50+ unfurried fixed anatomy version
-kontext dev finetuned unslopped
-some new sage attention extra turbo meme edition
anything i'm missing?
Anonymous No.105670185
>>105669781
>Ultimate SD Upscale consolidates this and has tiling options, though tiling is best avoided if you have VRAM)
Oh found that. Also do I run SEGS for hands and faces before or after upscale?
Anonymous No.105670226
>>105670099
Cool, bookmarked
Anonymous No.105670238 >>105670992
>>105670161
netani lumina
illust vpred 3.5
illust lumina
Wan 3
Anonymous No.105670402
>>105670099
that is awesome... thanks and never delete pls
Anonymous No.105670414 >>105670431 >>105670447
What do you use your gens for?
Are you working on some project or generating just for the heck of it
Anonymous No.105670424
>>105670099
Based beyond belief, fucking hate having to lurk through extracted booru tags without image examples
Anonymous No.105670431
>>105670414
Both. For work and for shits and giggles
Anonymous No.105670447
So I guess it is segs after upscale. I don't want to prompt so it doesn't change their expressions randomly baka.
>>105670414
Prompt goonstuff for pixiv>goon>use postnut clarity to work on sfw art
20Loras No.105670481 >>105670532
I want to give image2video a try.
If I already have forge up and running without issues, do I have to mess about with pytorch? I don't want to brick my usual genning.
Anonymous No.105670532 >>105670545
>>105670481
install comfyUI
Anonymous No.105670540 >>105670561 >>105670621 >>105673337
20Loras No.105670545 >>105670552 >>105670726
>>105670532
Sure, I know WAN requires that. But does the pytorch version being higher than what forge needs by default, ruin it for me? Right now it's at 2.3 iirc. You can't have multiple pytorch versions running I'm guessing.
Anonymous No.105670552 >>105670682
>>105670545
why bother? just use comfyUI for both, whatever the forge crap can, comfyUI can do better.
Anonymous No.105670561
>>105670540
The beer starts spilling way after the beer reaches the edge looks kinda silly
Anonymous No.105670608 >>105673351 >>105673364
>>105670099
very nice
Anonymous No.105670609
Anonymous No.105670621
>>105670540
She has a drinking problem
Anonymous No.105670624
20Loras No.105670682 >>105670793 >>105671304
VROOM

>>105670552
Because forge is comfy, comfyui is not.
Doesn't comfyui have a really convoluted way to tile upscale? Everything is just extra steps, not comfy.
Anonymous No.105670692 >>105670696 >>105670705 >>105670794 >>105670870
How do I improve SEGS detection of hands that are fists?
Anonymous No.105670696
>>105670692
wash your mouth out
Anonymous No.105670705
>>105670692
Test other ultralytics detectors from Civitai trained on anime, test a SAM2 setup, test a Florence2 setup.
Anonymous No.105670726 >>105670793
>>105670545
i am seconding this question. i also use forge and don't want to break anything as i'm technologically inept. if you find any good answer that doesn't involve quitting forge, i'm interested
Anonymous No.105670793 >>105671349 >>105672544
>>105670682
it's just a different way. a bit ironic that having a zoomable UI on PC is what's giving zoomers conniptions. that, and a total unfamiliarity with package management.
>>105670726
just so you know, once you get comfortable with this part, you'll be past the hurdle that filters 90% of comfy complainers. setup conda/venv, install dependencies, start service. it's basically the same for every manual install section of every python project across github.
Anonymous No.105670794
>>105670692
Just mask them manually, it takes no time at all and you don't need to be precise.
Anonymous No.105670802 >>105670809
I'm getting ~12-15% speed improvement on Wan when using sage attention (in Comfy), is that what you'd expect ?

I'm testing it on a 3060 ti.
Anonymous No.105670809 >>105670819
>>105670802
I don't remember, it's been a while and it was a different GPU but I recall it being significant
Anonymous No.105670815 >>105670962
is there a way to use ai to edit a video? i want to get rid of weird artifacts but i dont want to edit every frame. i want to select the area with the artifacts, have it automatically track it and fix it
Anonymous No.105670819 >>105670827
>>105670809
Hmmm... I've just installed sage attention and launched comfy with --use-sage-attention, do I need to change something in my workflow ?
Anonymous No.105670827 >>105670873
>>105670819
on native no, on kijai's wrapper it won't work if using the flag
Anonymous No.105670870
>>105670118
throw it into joy caption and go from there.
>>105670099
neat.
>>105670692
manual is the way. no detection is foolproof
Anonymous No.105670873 >>105670898
>>105670827
Ok, thanks, maybe that's all you get on a 3060, I am testing at 512x512, perhaps it gives larger gains at higher resolutions ?
Anonymous No.105670885 >>105670891
what was the name of that song generator model lads?
Anonymous No.105670891
>>105670885
Ace-Step ?
Anonymous No.105670898 >>105670906
>>105670873
stretching my memory, I think I was looking at 15 minutes for 40 frames of 480 on fp8
and that was the best I could get out of it, anything higher was exponentially longer, like extra 16 frames an extra 10 minutes
this was just with teacache, if you don't have that get it
Anonymous No.105670906
>>105670898
I was avoiding teacache because I heard it had a noticeable degradation on the results, but I should probably give it a go
Anonymous No.105670937 >>105671014
Did anyone find a way around the lightx2v making the motion slower in gens? I know there's the RIFLEx workaround, but it makes my gen times longer and I'm inpatient.

I've tried the :2 strength trick in the prompt, and it seems to burn the frames.
Anonymous No.105670962
>>105670815
You mean inpainting a video? I don't think it is possible. Or at least I've not seen any discussion of it around here.
Anonymous No.105670992
>>105670238
Wtf is netani?
Anonymous No.105670997 >>105671021 >>105671025 >>105671039 >>105671043 >>105671046 >>105671129
What's should I use if I literally just want to creat porn videos featuring my coworkers?
Anonymous No.105671014
>>105670937
you can increase the frame rate but that makes the video shorter obviously. maybe we need a motion lora or some shit.
Anonymous No.105671021
>>105670997
Use a camera.
Anonymous No.105671025
>>105670997
A good lawyer
Anonymous No.105671039 >>105671373
>>105670997

Hypothetically if anyone wanted to do this, there are two routes. Keep in mind this is HYPOTHETICAL because it is EEL LEGAL!!!!

Rookie:
>Get I2V setup
>DL average titty drop/cumshot loras
>Gen and Goon


Level 99 Mafia Boss Gooner:
>Get a bunch of images of this person together
>Crop and Caption them
>Train a lora
>Repeat 1-3 in the rookie setup
Anonymous No.105671043
>>105670997
>What's should I use if I literally just want to creat porn videos featuring my coworkers?
Anonymous No.105671046
>>105670997
At least 20-30 images of each person, train a lora of them, generate a start image, use Wan img2video with a lora which is trained on the sexual activity you want them to perform.

Do not post the results online, it is illegal unless you have their permission.
Anonymous No.105671070 >>105671085 >>105671144
Where do I get celebrity LORAs now?
Anonymous No.105671085
>>105671070
duckduckgo and musubi
Anonymous No.105671121 >>105671178 >>105671233 >>105671326
AI is just like humans sometimes, can't draw a straight line if it goes behind a character
Anonymous No.105671123 >>105671128 >>105671153
retard here,

I'm trying to use image2image in Forge to get a desired basic outline of the image (a person standing behind a desk). The outputs always look way too much like my reference image though (more like what I remember controlnet canny stuff doing). What setting should I be tweaking so that the image2image weighs the prompt more and the reference image less?
Anonymous No.105671128 >>105671170
>>105671123

Have you tried adjusting the denoise?
Anonymous No.105671129
>>105670997
VACE
Anonymous No.105671142 >>105671173 >>105671190
>>105670161
>things to get excited for
AMD catching up to Nvidia
Anonymous No.105671144
>>105671070
Train your own or use the MANY already existing, most should be available at places like https://civitaiarchive.com
Anonymous No.105671153 >>105671170
>>105671123
try using something like a scribble controlnet instead if you want more variety
Anonymous No.105671170
>>105671128
I have and it kind of works but not to the extent I will need.
>>105671153
Noted. I'll look into that.
Anonymous No.105671173
>>105671142
I wish
Anonymous No.105671178
>>105671121
sdxl does that. it's fun!
Anonymous No.105671190
>>105671142
Just 2 more weeks
Anonymous No.105671200 >>105671229
>>105670161
>anything i'm missing?
krea I guess, but at this point it's really unlikely it'll happen
Anonymous No.105671229
>>105671200
I think it will release, I doubt the post would have been made unless the decision was at least already 90% made.
Anonymous No.105671233
>>105671121
Yes background discontinuities frequently trip up the models. Though I found flux is noticeably better at that (not perfect).
Anonymous No.105671240 >>105671262 >>105671307
why am i getting OOM with chroma safetensors on 3090? it's only 17.4gb
Anonymous No.105671262
>>105671240
check nvtop/nvidiasmi and look at the output of the service.
Anonymous No.105671304 >>105671357
>>105670682
https://github.com/deepbeepmeep/Wan2GP
people are retarded and don't just give people this
Anonymous No.105671307
>>105671240
try:
>offloading t5 to cpu
>tiled VAE decode
Anonymous No.105671323 >>105671789
>>105670099
>no bedroom eyes/half-lidded eyes
ngmi
Anonymous No.105671326
>>105671121
That's actually one of the more inhuman of the common AI mistakes. One of those little things that reminds you the image was made by something that doesn't "see" it as a 3D scene like you do, because even though the problem is objectively subtle, human visual reasoning immediately notices something wrong. A human artist of that skill level could freehand both ends of the staff better than that, but they wouldn't, they'd draw the full staff in the sketch, or they'd use a ruler.
Anonymous No.105671349 >>105672520
>>105670793
it's bad software design to expect people to fix their own app all the time which is why comfy needs to die
Anonymous No.105671357 >>105671365
>>105671304
cuz it wasnt using good optimizations and looked like shit, does it even use SLG now?
no reason to use it, if you got 24gb vram, you shouldnt use it and should go for quality, if you got anything below, you shouldnt use it since with comfyui ldg workflow you can finetune optimizations a lot better to get the most out of everything

you gotta learn the basics of what the optimization do either way, might as well just do it through the main workflow then instead of cope
Anonymous No.105671365 >>105671400
>>105671357
you are actually fucking retarded
Anonymous No.105671373 >>105671407 >>105671420 >>105671694
>>105671039
Why the fuck can't I just make a video to jack off?
What's next? Needing their permission to have a wank?
Anonymous No.105671400 >>105671459
>>105671365
fp16 accumulation?
sageattention?
torch compile?
latest pytorch?
can you easily swap quants of wan and clip?
virtual vram settings?
does it support all the different things you can do with vace? easy looping videos?
will it implement sageattention2++ when it comes out soon as fast as comfy which will give a big boost?
just no point, its not like current video tools are like images where you need to mask things all the time which is better done with a mouse in a big UI, you just quickly set up ldg wan workflow and thats it, then just select lora and prompt away until something big drops when you update the workflow
Anonymous No.105671407
>>105671373
>Why the fuck can't I just make a video to jack off?
You have to ask that to your Commander-in-chief Sergeant Johnson.
Anonymous No.105671420 >>105671477 >>105671484
>>105671373
Because to get the resources you have to basically stalk them, retard
Anonymous No.105671447
Anonymous No.105671459 >>105671536
>>105671400
fp16 acc is in, safe attention is a requirement retard, pytorch version matters very little, torch compile memory leaks, it automatically selects the quant, don't really care about virtual vram, yes there is a whole vace interface, looping videos don't really loop properly so no UI can really do it unless you are talking about ping pong which looks like shit. why not ask the author about sage++. just no point playing around with noodle garbage instead of just having what you want to see right in front of you. it's actually easier to extend the clips with new input images than comfy. you can be a noodle faggot but it doesn't make you better than anybody. stop sniffing your own farts. not to mention I think a lot of people are comfy fatigued
Anonymous No.105671461 >>105671471
>still no sage2 update
>i weep
Anonymous No.105671471
>>105671461
They sayd "around 20th of june" they didn't specify the year
Anonymous No.105671477 >>105671484
>>105671420

If you don't cyberstalk your crush, can you even say you truly love her?
Anonymous No.105671484
>>105671420
If you post pictures of yourself on social media they are free to use for everyone.

>>105671477
based
Anonymous No.105671506
Anonymous No.105671536 >>105671632 >>105671633
>>105671459
>it automatically selects the quant, don't really care about virtual vram
The problem is there is a big difference when going lower than Q8, and for some things surely people want to actually have a high quality output despite needing to wait an hour or two, so offloading like this is a requirement.
>looping videos don't really loop properly so no UI can really do it unless you are talking about ping pong which looks like shit
Picrel
>>105571625

There will definitely be better UI at some point that will dominate but it's hard to compete with the flexibility of comfy given the speed of developments.
Anonymous No.105671560 >>105671601 >>105671607
>>>105663104
>chroma svdquant
where do i find this though?
Anonymous No.105671574
>>105668859
great gen anon, can you please post the catbox?
Anonymous No.105671601 >>105671607
>>105671560
https://huggingface.co/rocca/chroma-nunchaku-test/tree/main
So far only experimental v29, with v38 coming in 2 more hours. Or days. Or weeks https://huggingface.co/rocca/chroma-nunchaku-test/discussions/1#68557c81961b7e57afe5f902
Anonymous No.105671607
>>105671560
https://huggingface.co/rocca/chroma-nunchaku-test
>>105671601
*kicks you in the ass*
Anonymous No.105671632 >>105671679
>>105671536
>offloading like this is a requirement.
I make vids in ~70 seconds. you are dramatically overthinking things. if speed is a concern for making videos, just buy a better GPU instead of gobbling snake oil
Anonymous No.105671633 >>105671748
>>105671536
>big titty girl, boob physics like shes on the moon. sloshing bags of water. remove safeties. execute
Anonymous No.105671647 >>105671658 >>105671670 >>105671714
>finally figured out the workflow
>been gooning for 16 hours straight
help when do i get bored of infinite fully customized pornography
Anonymous No.105671658
>>105671647
>fully customized pornography
doesn't exist yet. the model is great but it's still limited what concept motions it knows. like seriously, why is it so obsessed with blowjobs?
Anonymous No.105671670 >>105671680
>>105671647
when you realize it's limited by your imagination and tastes, and all your niche fetishes rapidly become dull and boring when you have an unlimited supply
Anonymous No.105671676 >>105671716 >>105672217 >>105673118
Chinkmodded 4090D 48GB... yes, no? I'm really sick of losing the 5090 FE lottery, and honestly, 32GB isn't enough to run the best local video models. It's $3K. What else is there? Drop $4K on a M4 Max 40-core GPU 128GB Mac? That's going to be slow at imagegen and utterly shit at video, right? DGX Spark? Seems like a shitty, expensive cloud-service upsell box to me.
Anonymous No.105671679 >>105671765
>>105671632
>I make vids in ~70 seconds.
Do post those 70s made videos that blow out the colors, have stiff motion and warping the fuck out of anything moving.
>if speed is a concern for making videos, just buy a better GPU instead of gobbling snake oil
You are the one using cope projects instead of having a 24gb card to just go for max quality workflows.
Anonymous No.105671680
>>105671670
reduction is the key. I never go all out and that's a strict rule.
Anonymous No.105671694
>>105671373
>Why the fuck can't I just make a video to jack off?
You literally can, the law only has any effect if you generate porn (sexually explicit) images or video of REAL PEOPLE and then POST THEM ONLINE without their permission.

If you don't do either of these things, you can wank off as much as you want.
Anonymous No.105671714
>>105671647
post workflow
Anonymous No.105671716 >>105671811 >>105671925 >>105671933 >>105671954
>>105671676
https://www.newegg.com/amd-100-300000075-radeon-pro-w7800-32gb-graphics-card/p/N82E16814105115?Item=9SIA24GKAP2001
this has drawbacks but it's a better deal on VRAM and can actually be purchased
Anonymous No.105671748
>>105671633
checked
also
>pic related
Anonymous No.105671765 >>105671780
>>105671679
>max quality workflows.
that would be without the snake oils. pretty much raw fp16. do you not pay attention to what people say around here?
Anonymous No.105671780 >>105671850
>>105671765
>no 70s made video posted
yawn
Anonymous No.105671789 >>105671850
>>105671323
Thanks for the heads-up. I added two new groups for eyes and pupils tags.
https://tagexplorer.github.io/#/?tagGroupFilter=eyes
Anonymous No.105671804 >>105671851 >>105671942 >>105672042
>>105670099
>github
really cool project thankyou anon
Anonymous No.105671811 >>105671969
>>105671716
Drawbacks like: If you even get it to run it will run much slower than the NVidia equivalent.

There is hope now that AMD actually (after 3 fucking years) are pulling their heads out of their asses and is starting to work on strong AI support, but it's not even close yet, so you'd be an idiot to buy an AMD GPU for AI use today.
Anonymous No.105671850 >>105671870 >>105671879 >>105671923 >>105671930
>>105671780
i have one here that took 73 seconds. I just run a script in anistudio to send requests to the wan2gp backend. no quants since I have a 4090. dunno if he added magcache yet but I've been too busy.

>>105671789
this is really helpful anon! thanks!
Anonymous No.105671851 >>105671861 >>105673466
>>105671804
F U C K O F F
Anonymous No.105671861 >>105671879
>>105671851
someones moody
Anonymous No.105671870 >>105671884
>>105671850
>mostly static sketch of 10 different colors in total of a simply sketched cartoony anime girl that barely moves and when she does picrel happens
ah, so this is the power of 70s generated videos... i now truly see
Anonymous No.105671875
based ani proving my point
Anonymous No.105671879 >>105671884
>>105671861
I do not accept the pedo creep. end of story.
>>105671850
vnice!
Anonymous No.105671884 >>105671929
>>105671870
that is teacache doing that and it's been in pretty much every video that moves a little quickly. wan really is made for 3dpd too

>>105671879
tyvm
Anonymous No.105671923
>>105671850
Pedo
Anonymous No.105671924
SNAKE OILED
Anonymous No.105671925 >>105671969
>>105671716
32GB is not enough, most big video models give 40/48GB as the minimum, plus whatever small discount there is on the hardware you pay for it in wasted time trying to get shit to work with AMD.
No thank you.
Anonymous No.105671929
>>105671884
the cope trannies have to tell themselves, lol
Anonymous No.105671930 >>105671944
>>105671850
does your implementation support loras? does it support sage attention? does it support torch compile?
genuinely curious.
how much space does the venv take up in total?
does it support kijai models?
Anonymous No.105671933 >>105671969 >>105671975
>>105671716
>2300 dollars for a 32gb vram card
might aswell buy 2x3090 with that money and get 48 gb of vram total
Anonymous No.105671942 >>105671950 >>105672042 >>105672063
>>105671804
Anonymous No.105671944
>>105671930
>does your implementation support loras
it's literally a script that just runs the backend. just read what the author has to say. I use the kijai distill yeah

https://github.com/deepbeepmeep/Wan2GP
Anonymous No.105671950 >>105671968
>>105671942
was rife used for interpolation here instead of film vfi?
Anonymous No.105671954
>>105671716
ok but how big is it..
tell me the size, i need to know
my comfyui venv takes up 6.2GiB
Anonymous No.105671957
Wan trolling me with nudity:
https://files.catbox.moe/le11vb.mp4

scrubbing nipples from your dataset vs tagging "rating:expicit"... which to choose...
Anonymous No.105671968 >>105671972 >>105672042
>>105671950
there was no interpolation at all, just self forcing and movement reward lora
Anonymous No.105671969 >>105671986 >>105671989
>>105671811
>it's not even close yet, so you'd be an idiot to buy an AMD GPU for AI use today
it works for me ¯\_(ツ)_/¯

the AMD cards are a good VRAM value if you're on Linux. I use AMD for this reason and because AMD's Linux GPU drivers are open source and much more stable than Nvidia's.

currently, the 7900 XTX is around a 3090 in AI performance using chroma as a benchmark. If we're lucky and FineWine kicks in with newer ROCM versions, this card could get to 4090 perf.

>>105671925
>>105671933
totally fair. just mentioning we have options.
Anonymous No.105671972 >>105671989 >>105672011
>>105671968
so thats why its bad
Anonymous No.105671975 >>105671999 >>105672008
>>105671933
>might aswell buy 2x3090 with that money and get 48 gb of vram total
Well that's my setup right now. I even have the nvlink. The problem is diffuser stuff needs to have the whole model in memory, you really need more memory on a single GPU at the moment.
Anonymous No.105671986 >>105672023 >>105672081
>>105671969
>the 7900 XTX is around a 3090 in AI performance using chroma as a benchmark
don't know if I can believe this
Anonymous No.105671989 >>105672023
>>105671969
have you tried wan? what performance are you getting? do you have a single 7900 xtx?
>>105671972
resolution too, yeah
Anonymous No.105671999 >>105672017
>>105671975
isnt there a multigpu node
Anonymous No.105672008 >>105672039
>>105671975
>The problem is diffuser stuff needs to have the whole model in memory,
you can split the model into multiple gpus if that's a gguf though
https://rentry.org/wan21kjguide
Anonymous No.105672011
>>105671972
Wan was trained at 16 FPS. Either you minterpolate the output in ffmpeg or use cosmos and deal with limbs falling off and characters growing a second head.
Plus, anime shit is like 6 fps, so there's that too. Maybe i2v in wan and then v2v back into cosmos to make it smoother?
Anonymous No.105672017 >>105672037
>>105671999
I've seen someone say previously that multigpu setups are mostly for loading encoders separately, and that "use other vram" doesn't work properly but no proofs were posted
Anonymous No.105672023 >>105672090 >>105672139
>>105671986
We were discussing this in an old thread, a 3090 user said he gets ~60 to 70 s per standard chroma gen at 20 steps. I get the same result.

>>105671989
Haven't tried Wan, because local video just doesn't look worth the effort to me yet. Single card. Would be interested to know if any AMD users have been using Wan.
Anonymous No.105672024 >>105672289
What's this whole deal with "detected dubious ownership in repository at"?
Is it because it's installed on an external drive?
Anonymous No.105672037
>>105672017
>"use other vram" doesn't work properly but no proofs were posted
it works, I'm currently splitting the model into 2 of my nvdia gpus
Anonymous No.105672039
>>105672008
>https://rentry.org/wan21kjguide
Ah so "To manage VRAM limitations, offload to RAM/CPU using the virtual_vram_gb setting in the UnetLoaderGGUFDisTorchMultiGPU node, though this slows generation and you can only offload so much." OK. Eh, at the moment I just run two comfy instances on different ports and pinned to separate GPUs, and just tandem gen. It's like 30-50% success rate on gens, so more gens is preferable I think.
Anonymous No.105672042 >>105673466
>>105671804
>>105671942
>>105671968
kys
Anonymous No.105672062 >>105672083 >>105672112
>>105670099
nice but
>Illustrious 1.1.
why not 2.0 which used the largest dataset out of all versions currently available?
Anonymous No.105672063 >>105672074 >>105672123 >>105672140 >>105672158 >>105672180
>>105671942
setting her on fire isnt very nice
Anonymous No.105672074 >>105673466
>>105672063
remember, kys
Anonymous No.105672081
>>105671986
>don't know if I can believe this
I know I don't believe this
Anonymous No.105672083
>>105672062
it's all the same booru shit. basically nothing changed
Anonymous No.105672090 >>105672207
>>105672023
weird, i get the same speed, albeit with SVDQuant (sadly still at v29 but its good enough)
100%|...| 20/20 [00:48<00:00, 2.44s/it] @100W
100%|...| 20/20 [00:37<00:00, 1.89s/it] @ 170W
t. 3060 vramlet
Anonymous No.105672112 >>105672145
>>105672062
Honestly, I didn't think about it too hard, I just got the general impression that the image gen community overall didn't fully accept 2.0 as the rightful successor to 1.0. For example, WAI v14 chose 1.0 over 2.0 as a base. It probably makes little difference for the site's purpose.
Anonymous No.105672120 >>105672151
Does controlnet, openpose, work with illustrious, noobai? I can't get it to work.
Anonymous No.105672123 >>105673466
>>105672063
fuck off you disgusting piece of shit, you're not welcome here
Anonymous No.105672139 >>105672207
>>105672023
it's a bit faster. Prompt executed in 59.40 seconds. 3090@80%, 26 steps, 2.25s/it, sage attention & --fast, the comfy wan build from the rentry basically. sniff.
Anonymous No.105672140
>>105672063
Back to trooncord, sis
Anonymous No.105672145
>>105672112
>I just got the general impression that the image gen community overall didn't fully accept 2.0 as the rightful successor to 1.0. For example, WAI v14 chose 1.0 over 2.0 as a base
unfortunate considering its ability to handle larger initial latent sizes and NLP. i thought anon would be all over its NLP ability at least. maybe that wont be fully realized until/if/when 3.5vpred drops...
>It probably makes little difference for the site's purpose.
fair point
Anonymous No.105672151 >>105672163
>>105672120
It does, use xinsir's promax ControlNet model. However, the OpenPose preprocessor works poorly with anime input images. I'd recommend using Depth or Canny if your input image is anime-style. If you set it up like in my guide or download the workflows it should work without much fuss. https://rentry.org/comfyui_guide_1girl#controlnet-pose-transfer
Anonymous No.105672158
>>105672063
Uncanny
Anonymous No.105672163
>>105672151
Thank you kindly. Haven't used it in a while since the checkpoints are so good at poses these days.
Anonymous No.105672180 >>105672291
>>105672063
cute, would
Anonymous No.105672207 >>105672284 >>105672309
>>105672139
I am not using sage attention or flash attention. Some other AMD anon figured out how to install these, but I haven't figured it out yet. this kind of issue is obviously the big drawback of AMD for the time being.

>>105672090
that makes sense because SVDQuant is a lot faster.
Anonymous No.105672217 >>105672257 >>105672330
>>105671676
>Chinkmodded 4090D 48GB... yes, no?
Up to you but if you're wondering where to get one, I bought mine from
https://www.c2-computer.com/products/new-parallel-nvidia-rtx-4090d-48gb-gddr6-256-bit-gpu-blower-edition
Anonymous No.105672257 >>105672395
>>105672217
Yep that's where I'm intending to purchase mine. Did they pack it well? I'm always nervous of them sending an item like that in a "speed pack" and having it get trashed.
Anonymous No.105672269
I wish I could run Lumina 2 at a reasonable speed :(
Anonymous No.105672284 >>105672352
>>105672207
60s/gen just kills the vibe. sick shit tho
Anonymous No.105672289
>>105672024
> Is it because it's installed on an external drive?
I usually get that when I try to access a repo initially checked out by another user or checked out on another computer and accessed via a samba share. I would think an external drive would be ok as long as it isn't a network drive administered by another machine.
Anonymous No.105672291 >>105672350
>>105672180
regenned on newer model
Anonymous No.105672304 >>105672324
Anonymous No.105672309 >>105672352
>>105672207
did they say how they got sage working i didnt have any luck with that flash attention is part of rocm now and uses triton but pytorch also has a triton flash attention thing built in i think they might be the same thing you can enable the torch one with TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1
Anonymous No.105672324 >>105672334
>>105672304
>be me
>middle school teacher
>tell class the next assignment is to make a political cartoon since our current topic is world politics
>kid in the back asks if he can use "ai" to make his
>other kids chuckle
>"sure i dont care"
>ff to end of the week
>this is his submission
Anonymous No.105672330 >>105672395
>>105672217
>https://www.c2-computer.com/products/new-parallel-nvidia-rtx-4090d-48gb-gddr6-256-bit-gpu-blower-edition

Interesting, slightly cheaper than a 5090 and 48gb? Is this the best place to buy these? I can only find ebay and aliexpress listings
Anonymous No.105672334
>>105672324
>>this is his submission
he deserves an A+, he nailed that shit
Anonymous No.105672350 >>105672379
>>105672291
the new gen is less cute, it looks more slopped and its close to the uncanny valley, the old gen looks cartoonish in comparison and has cute hearts and the old gen has a lewder outfit
Anonymous No.105672352 >>105672416 >>105672586
>>105672284
for complex prompts and styles, the ROI is worth it. chroma can do in a single 60s gen what SDXL and SD3.5 models can't do in six 10s gens.

>>105672309
no, sadly they didn't explain how.

>you can enable the torch one with TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1
I use this command:
>TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1 python main.py --use-pytorch-cross-attention --bf16-vae
I have no idea if I'm getting a perf boost from this.
Anonymous No.105672379
>>105672350
its a wip im trying to get analog photography style but getting it working nice requires merging in more realism stuff which is hard to balance kek also the hearts werent prompted so theyre completely random
Anonymous No.105672395
>>105672257
It comes with a 12VHPWR adaptor and I forget if it was styrofoam, bubble wrap or something else but there was cushioning between the inner and outer box, I think.
https://litter.catbox.moe/14v4hucayh4elbjh.jpg

>>105672330
>Is this the best place to buy these?
I'm not sure. Some other anon on /lmg/ bought one from there so I did the same thing. There are some listings on ebay that I can see but they ship from china too and they cost slightly more than what I paid. Depends on where you live probably.
Anonymous No.105672397 >>105672513
Anonymous No.105672416 >>105672586
>>105672352
might be worth trying this to see if you get an perf differences between the rocm implementation and the torch one i think in general theyll both be slower than nvidia flash attention because theyre both using triton https://www.reddit.com/r/LocalLLaMA/comments/1jh0n3q/psa_get_flash_attention_v2_on_amd_7900_gfx1100/
Anonymous No.105672433 >>105672437
Anonymous No.105672437
>>105672433
nice
Anonymous No.105672448
"very pale skin", t. chroma
Anonymous No.105672450 >>105673027
>tfw Bethesda is too stupid to make Fallout 5 with NCR as the main antagonist and them having their own frank Horrigan that enforces their fascist end stage capitalism with corrupt oligarchs onto the populace of the wasteland.

feels bad
Anonymous No.105672470 >>105672521 >>105672542
should i use
https://github.com/TTPlanetPig/Comfyui_JC2
https://github.com/StartHua/Comfyui_CXH_joy_caption
or something else for joycaption in comfy?
Anonymous No.105672477
Anonymous No.105672498 >>105672531
Could you anons post your illustrious training settings? I can't manage to get a goof safetensor. Do you have any tips?
Anonymous No.105672513 >>105672581
>>105672397
heh
Anonymous No.105672520
>>105671349
It's incredible how many g tards don't see it
Anonymous No.105672521 >>105672539
>>105672470
I use this as per recommendation of another friendly anon, works well
https://github.com/silveroxides/joycaption_comfyui
it's a big boy tho. there is a quant floating around on huggingface of the joycaption model but I have no idea how one would use that.
Anonymous No.105672531
>>105672498
I don't know if anyone has finetuned it. I'm waiting for 3.5 before trying.
Anonymous No.105672539
>>105672521
thank you anon, but why are you using that repo and not the main repo, that repo is a fork of a fork of the main repo, up to date with the main repo branch
it would probably be safer to use the main repo.. Thank You!!
Anonymous No.105672542 >>105672566
>>105672470
https://github.com/EvilBT/ComfyUI_SLK_joy_caption_two/blob/main/readme_us.md works fine with nf4 quant but its requirements.txt doesn't have one additional requirement: pip install timm==1.0.13
Anonymous No.105672544
>>105670793
>having a zoomable UI on PC is what's giving zoomers conniptions.
zoomies have no frame of reference, they dont care in the least
Anonymous No.105672548 >>105672603 >>105672902
Anonymous No.105672566
>>105672542
Thank You anon!!
Anonymous No.105672581 >>105672592
>>105672513
lol nice
Anonymous No.105672586
>>105672352
>>105672416
from a quick test, there is zero or negligible difference from using TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1 and/or main.py --use-pytorch-cross-attention

this kind of thing is where AMD really needs to get their shit together. instead the retards in their AI department are busy releasing "optimized" SDXL and SD35 versions that nobody will fucking use. they should instead be contributing to comfyui and underlying libraries to make sure consumer AI runs fast.
Anonymous No.105672592
>>105672581
Anonymous No.105672603
>>105672548
you could fool every poster in >>>/b/irl
Anonymous No.105672646 >>105672660 >>105672738 >>105672787 >>105672813 >>105672825
NetaAniLumina_Alpha examples
Anonymous No.105672649 >>105672700 >>105672713 >>105672739
I often get anime characters when i prompt random gibberish like w0e9mfv-w0me9vf-0awm9v-0awmve with Chroma.
This makes me think Astracuck might have meddled in the dataset, and that the artist names are all like that.
Anonymous No.105672654 >>105672729
This is what I get with the prompt
"w0e9mfv-w0me9vf-0awm9v-0awmve-a0wmve9-aw0mef9v-m0wae9vf-wa0me9v-e00-e33333333335555555555555"
Anonymous No.105672660 >>105672685 >>105672738
>>105672646
nice.. ungated download?
Anonymous No.105672685 >>105672705
>>105672660
Only the HF repo is gated for some reason. On Civitai it's open https://civitai.com/models/1612109
Anonymous No.105672700
>>105672649
does the ponycuck not have a page where he posts training logs and the captions?
Anonymous No.105672705
>>105672685
this is why, the civitai model is over a month old
Anonymous No.105672713 >>105672739
>>105672649
it's more like the chroma dataset has way more anime images than real photos, so the model is more biased to render an image image if it has no idea what your prompt even mean
Anonymous No.105672729 >>105672746 >>105672761 >>105672874 >>105672929
>>105672654
"98A0S9sudv0 8s9dv09ns8 oiuxouv 0s987dv7": https://files.catbox.moe/u0gi64.png
"sdfsajhgk 79873f 9x8v90 xvxdrve23": https://files.catbox.moe/7tu5jr.png
"ef9dgue9rue 948t9ut as898fg urgu9874g4" (nsfw): https://files.catbox.moe/dxu4am.png

Just mashing my keyboard and it always makes anime esque art. This doesn't make sense unless he hashed all the artist tokens and it's picking up the tokens. And thus explains why the model "is learning them slowly" after 38 epochs.
Anonymous No.105672738 >>105672746
>>105672646
>>105672660
you have to request access and then beg for acceptance on their d server KEK
Anonymous No.105672739
>>105672713
>>105672649
certain keywords also trigger anime in chroma. For example "Japanese"
Anonymous No.105672741
AI killing real porn industry is the best part about it
Anonymous No.105672746 >>105672792
>>105672729
>1
sovl
>2
cute
>3
would
>>105672738
thats my point, someone ITT should leak their new models on hf
DO IT!!!!
Anonymous No.105672761 >>105672840
>>105672729
>This doesn't make sense unless he hashed all the artist tokens and it's picking up the tokens
FUCK

let's test. do any substrings of these random characters consistently reproduce a single style?
Anonymous No.105672787
>>105672646
So is this the obsolete version from Civitai or the latest one from HF?
Anonymous No.105672792 >>105672807
>>105672746
I have access to them and they aren't that different compared to the public version on CivitAI, they tend to have more artifacting. I'd just wait for a newer version to show up, i'm guessing they're trying different training settings (roundnnnn versions)
Anonymous No.105672807
>>105672792
post them on hf or wherever and ill post them on hf for you
Anonymous No.105672813
>>105672646
looks way more promising than chroma unironically desu
Anonymous No.105672825 >>105672840 >>105672873
>>105672646
why did they go for lumina? what's so special about that model?
Anonymous No.105672840 >>105672873 >>105672874 >>105672907 >>105672911 >>105672939
>>105672761
Yeah, it stays consitent for each string. Basically any sequence of alphanumerical characters that isn't a word will always generate anime with Chroma in my testing. This is exactly what happened with Pony, the artist tokens were turned into random strings.

Here's "dfvdfh8bh7df".

>>105672825
It's a DiT with 2.6b parameters so not too heavy, has a better text decoder and a 16 channel vae. It has way nicer backgrounds compared to sdxl, just better to work with, it's already this good with four epochs.
Anonymous No.105672852
>astra actually cucked the model
Anonymous No.105672873 >>105673087
>>105672825
Apache-2.0 license? The tooling https://github.com/Alpha-VLLM/Lumina-Accessory ?
>>105672840
>compared to sdxl
Compare it to SD3.5m but I'll take what I can get honestly with a 16ch and further enhanced prompt understanding.
Anonymous No.105672874 >>105672915
>>105672729
>>105672840
Would you mind posting your findings to some shitty chink spreadsheet website pls
Anonymous No.105672902
>>105672548
prompt?
Anonymous No.105672907
>>105672840
It would make sense that there's something since they have "aesthetic 1" to "aesthetic 11" thing
Anonymous No.105672911 >>105672939
>>105672840
>>Here's "dfvdfh8bh7df".
you have to test this with different prompts though. eg 1boy playing baseball, 1girl in kitchen, whatever.
so far I am not seeing conclusive evidence that random strings produce a consistent and distinct style when paired with a real prompt.
Anonymous No.105672915
>>105672874
Maybe I'll do more testing and see if I recognize any.
Sometimes it's not anime, but it's definitely "stylized" enough to be some artist.

"ahsdjkvhs0d9v09"
Anonymous No.105672929
>>105672729
>prompt '4n0nd03sntkn0wwh4th4lluc1nat10n1s'
wow, i think you're really onto something here
Anonymous No.105672937 >>105672971 >>105672984 >>105672987 >>105673017 >>105673029 >>105673095
I'm a total newfag to this, but tested comfy ui on a 4060 Ti 8gb for a bit, and loved what I got. I'm planning on buying a new pc, what should I focus on in my GPU? Should I look at relative performance, or try to get as much VRAM as possible?
Also a poorfag, so budget is really important.
Asking this because there's cases like 3060 12gb apparently having worse performance than 4060 ti 8gb. Would this also hold true for image generation?
https://www.techpowerup.com/gpu-specs/geforce-rtx-4060-ti-8-gb.c3890
https://www.techpowerup.com/gpu-specs/geforce-rtx-3060-12-gb.c3682
I don't care about gaming on that pc btw, I only want the best performance and speed on image generation.
Anonymous No.105672939
>>105672840
oof

>>105672911
I agree. a grid would be better suited for this
Anonymous No.105672971
>>105672937
vram vram vram
Anonymous No.105672984
>>105672937
Anything less than 24gb and you will regret it, so 3090/4090 or a 5090 even better
if you are a poorfag, then just save more for one of those three, rather than getting a 12/16gb now and complaining later about shit results and being out of memory when you will do video stuff too
Anonymous No.105672987 >>105673187
>>105672937
3060 12gb is infinitely better than any <12gb vram gpu
but maybe you should get a 4060 ti 16gb, maybe 5060 ti 16gb
depends on the prices
if you're feeling lucky check out intel and amd too they're cheaper and have fairer amounts of vram but keep in mind many optimizations are nvidia only
Anonymous No.105672988 >>105673155
Can you give me feedback for my training settings?
https://pastebin.com/pdnQG4fj
Anonymous No.105673012
Anonymous No.105673017 >>105673187
>>105672937
if you're not going for fast WAN or big LLMs, you'd still need around ex- 20GB for chroma. if you can get a 40/50 series 16GB used or for MSRP, you could run quants at full speed quickly, but I'd personally prefer a 3090 over either.
also, prompt?
Anonymous No.105673027 >>105673086
>>105672450
looks like a gay man drawed this
Anonymous No.105673029 >>105673187
>>105672937
>cases like 3060 12gb apparently having worse performance than 4060 ti 8gb
and which are those cases exactly? for ai you need 1. vram size 2. memory bandwidth

used 3090
Anonymous No.105673058 >>105673094 >>105673188
Been a long time since I've proompted anything, has speed on 4gb lol vramlet machines improved? last time i messed around with this stuff auto1111 was still king
Anonymous No.105673071
8273987298735
Anonymous No.105673074 >>105673094
is chroma worth the time in a 3060?
Can it do more not realistic styles? Not anime but something more 3d-2.5d
Anonymous No.105673086 >>105673129
>>105673027
>drawed
Anonymous No.105673087 >>105673167
>>105672873
Lumina's stated Apache 2.0 license shouldn't be taken seriously given
1) requires Gemma 2 which has a very obviously incompatible license that makes reference to a Google URL for the terms and has extensive use restrictions, including use for generating sexually explicit content
2) the model weights were necessarily trained on Gemma which makes the model a "Model Derivative". The terms related to this are not compatible with Apache 2.0. Technically given the license and the fact they're violating some of the terms already, Google could actually tell them to delete the model at any time, if you take the terms at face value... that goes for any derivative model thereafter

Whether all that matters you to personally, as someone who can download the weights and give Google a big middle finger as you sail off into a sunset comprising of explicit pornographic material, is another matter. But I wouldn't take Lumina 2 license seriously and it surprises me anyone does. If you used it for commercial purposes you'd be taking a huge risk
Anonymous No.105673094 >>105673134 >>105673172
>>105673074
with svdquant 20steps is 48 seconds on a 3060 @100W pl
>>105673058
gpu? rest of the specs?
Anonymous No.105673095 >>105673147 >>105673187
>>105672937
Grab a 5060Ti 16gb, don't throw away your old gpu, you can use them both by partially offloading larger models to the second card. Not nearly as good as a single gpu with 24gb but many times faster than offloading to system RAM
Anonymous No.105673118
>>105671676
What? With a $3k budget anything other than 5090 is asking for trouble. Chinks are very crafty scammers, don't do it anon.
Anonymous No.105673129 >>105673144
>>105673086
am i wrong
Anonymous No.105673134 >>105673163
>>105673094
is there any guide on making it work?
Anonymous No.105673144
>>105673129
No, you're Indian.
Anonymous No.105673147
>>105673095
>many times faster than offloading to system RAM
Don't most of the consumer hardware requires you to pass the data to ram first before going into another gpu aside from old nvlink tech? the main benefit from multigpu systems is if you can permanently keep layers in another gpu and pcie isnt the bottleneck
Anonymous No.105673149 >>105673220
Anonymous No.105673155 >>105673194
>>105672988
Please, it's for an illustrious training
Anonymous No.105673163 >>105673241
>>105673134
https://files.catbox.moe/x4ev0o.png
heres a workflow, super ez to set it up on linux, the guide is on the nunchaku comfyui github repo
if you want 20 steps change res_multistep to euler
Anonymous No.105673167
>>105673087
Interesting point, thanks for the explanation. I'm not open source brained enough yet to make sure to check the upstream licensing. I'll also add this to my list of pointless arguments why someone should do an SD3.5m animetune.

On that note, any news about Animaestro?
Anonymous No.105673172 >>105673190
>>105673094
>gpu? rest of the specs?
It's an old 1650 super, not sure what gpu specs are even relevant besides vram
But I was more asking for a general direction in terms of whether low vram gens have gotten significantly faster for other anons
Anonymous No.105673187 >>105673207 >>105673217 >>105673261
>>105673017
What's WAN, quants and chroma?

prompt was something like Ghislaine, simple style, flat colours, loli, very chibi, huge breasts, etc, etc
>>105672987
>if you're feeling lucky check out intel and amd too they're cheaper and have fairer amounts of vram but keep in mind many optimizations are nvidia only
I'm definitely not feeling lucky after reading billion people saying "Nvidia or suffer".
>>105673029
>and which are those cases exactly?
I linked these 2 sites that compare relative performance.
>>105673095
>Grab a 5060Ti 16gb, don't throw away your old gpu
That's the funny part, I have no GPU right now. Would a 3090 work faster than a 5060 Ti? Anyone has experience with upgrading VRAM but using an older card?
Anonymous No.105673188
>>105673058
comfy or reforge are the quickest for what youd be doing
you be able to at least run XL but it would be slow
Anonymous No.105673190 >>105673281
>>105673172
i meant rest of the rig
theres ggufs now and svdquant
Anonymous No.105673194 >>105673248
>>105673155
better do a test run and see how it goes. You could use some language model to interrogate those settings
Anonymous No.105673198 >>105674637
Anonymous No.105673207
>>105673187
>Would a 3090 work faster than a 5060 Ti?
yes
wan is a video model (image=>video/text=>video)
quants are 4bit,8bit...
chroma is a new text to image model, very good
Anonymous No.105673217
Is anything new?

>>105673187
>That's the funny part, I have no GPU right now. Would a 3090 work faster than a 5060 Ti? Anyone has experience with upgrading VRAM but using an older card?
loads of people use a 3090.
Anonymous No.105673220
>>105673149
cool shoes
Anonymous No.105673241 >>105673255
>>105673163
>quantization process, known as SVDQuant, involves compressing the model weights and activations to 4 bits, significantly reducing the memory footprint and computational load. This is achieved by absorbing outliers in the data using low-rank components, which helps maintain the model's performance and visual quality. The extension integrates seamlessly with ComfyUI, allowing users to set up and execute workflows that take advantage of these optimizations.
Wow am I getting filtered. It what?
Anonymous No.105673246
Anonymous No.105673248 >>105673259
>>105673194
Already did, that's why I'm asking for some smart anons.
The generated images don't pass the bar, any ideas?

LLM invent stuff many times and sometimes they don't know what they are talking about.
Anonymous No.105673255 >>105673327
>>105673241
reduces vram for flux or chroma by 4 times while keeping quality super good
Anonymous No.105673259 >>105673503
>>105673248
>they don't know what they are talking about.
What?
Anonymous No.105673261
>>105673187
3060Ti was the only good 60 series RTX card, easily being on par or even beating 2080. 4060Ti could barely hold out against 3070 and 5060Ti actually loses to 4070 in terms of speed. 16gb is nice of course but you'd still take your sweet time genning stuff, unless the only thing you're interested in is smaller models like sdxl
Anonymous No.105673281 >>105673308
>>105673190
Thanks I'll check these out
the rest of the rig is a ryzen 5600x and 32gb ram
Anonymous No.105673308 >>105673544
>>105673281
https://huggingface.co/city96/stable-diffusion-3.5-medium-gguf/blob/main/sd3.5_medium-Q4_K_S.gguf
try this with comfyui gguf node
https://huggingface.co/city96/stable-diffusion-3.5-medium-gguf
report back with results, try with 512x512 and make sure to use linux to reduce vram usage, you could also disable browser hardware acceleration to reduce vram usage on linux too
Anonymous No.105673316 >>105673344
I've been out of the loop for a while.

Is Hunyuan not even worth considering as a video model or something? No guides in the op? was Wan equally uncensored?
Anonymous No.105673318
Anonymous No.105673327 >>105673343 >>105673360
>>105673255
Tradeoffs?
Anonymous No.105673337
>>105670540

I did not manage to make beer spilled.

How do you prompt it?
Anonymous No.105673343 >>105673360
>>105673327
none
it's black magic
Anonymous No.105673344
>>105673316
Wan is the only good video model right now >>105669256 (OP)
>>WanX (video)
>https://rentry.org/wan21kjguide
Anonymous No.105673351
>>105670608

kino
Anonymous No.105673360
>>105673327
>Tradeoffs?
>>105673343
>none
of course there's a tradeoff, the quality isn't close to bf16 or Q8
Anonymous No.105673364
>>105670608
yjk
Anonymous No.105673365
Fresh

>>105673353
>>105673353
>>105673353

Fresh
Anonymous No.105673373
>>105669612
Every day man, every fucking day!

SMASH
Anonymous No.105673466
>>105672042
>>105672123
>>105672074
>>105671851
>please post more, i-i mean that is disgusting!
imagine trying this hard to convince others?
kys
Anonymous No.105673503
>>105673259
Sometimes LLM makes stuff on the fly. It happened just to me today.

If you really know about a topic and the AI does not, sometimes instead of saying "I don't know" it invents stuff. You shouldn't trust everything a LLM model says.
Anonymous No.105673544
>>105673308
I'll give it a try later today. Thanks anon!
Anonymous No.105673723
This was the most drastic hires fix I've ever seen and the denoise is only at 0.3. What the fuck lol
Anonymous No.105674637
>>105673198

I'm trying the nietaani lumina found on civitai too
Anonymous No.105674754
Netaanilumina is really promising, very satisfying in terms of prompt adherence and character knowledge(recent characters tho)
Can't find more artist tags other than the one declared in the example image found on civitai.
Really wonky anatomy ,but all in all this alpha is usable