← Home ← Back to /g/

Thread 106162717

317 posts 148 images /g/
Anonymous No.106162717 [Report] >>106162832
/ldg/ - Local Diffusion General
Discussion of Free and Open Source Text-to-Image/Video Models

GenJam3: https://forms.gle/hWs19H4vTGTdwARq8

Prev: >>106157414

https://rentry.org/ldg-lazy-getting-started-guide

>UI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://github.com/Wan-Video
2.2 Guide: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-Base/tree/main
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.106162736 [Report] >>106162777
Blessed thread of frenship
Anonymous No.106162754 [Report] >>106162804 >>106163397
wincucks can't into qwen-image
Anonymous No.106162759 [Report] >>106162774
Anonymous No.106162774 [Report] >>106162792
>>106162759
wan 2.2?
Anonymous No.106162777 [Report] >>106162915 >>106162964
>>106162736
Aww, Anon-kun...
Anonymous No.106162781 [Report] >>106162789 >>106162962
is diffusion-pipe the most retarded training script ever?
useless deepspeed bloat, nonsensical cryptic output, cannot create samples during training, doesn't work on windows
Anonymous No.106162789 [Report]
>>106162781
yes
Anonymous No.106162792 [Report]
>>106162774
I have no idea, I didn't create it.
Anonymous No.106162797 [Report]
InvokeAI?
Anonymous No.106162804 [Report] >>106162982
>>106162754
chroma is number 1 anyway
Anonymous No.106162832 [Report] >>106162996
>>106162717 (OP)
>including ani in the collage
Anonymous No.106162838 [Report]
Joolien
Anonymous No.106162840 [Report] >>106165697
Send the bastards back!
Anonymous No.106162855 [Report]
Why do normoids like ghibli style gens so much?
Anonymous No.106162864 [Report] >>106164095
>DeepMind Genie3 architecture speculation
Gonna be fun playing with Hunyuan-World2.0 next year or whichever chinese company does open source
Anonymous No.106162878 [Report] >>106162947 >>106163009
Why does Comfy seethe about Ani so much? They used to be friends ;_;
Anonymous No.106162915 [Report]
>>106162777
>I call this piece 'autistic japanese girls on a beach'
Anonymous No.106162937 [Report]
>>106158221
is there a HF space for this? I don't want to have to load an LLM every time I want to rice one of my prompts
Anonymous No.106162947 [Report] >>106163022
>>106162878
comfy said C was too hard but ani is a gooner idiot and can do it proving comfy is just a lazy asshole phoning it in
Anonymous No.106162948 [Report]
>julien
Anonymous No.106162962 [Report]
>>106162781
The only valid complaint there was that it can't create samples, that's a clear negative.

Works fine on Windows using WSL2, it's kind of hilarious since Deepspeed is made by Microsoft, and it doesn't run on native Windows. Then again, all AI research and development is done on Linux.

Deepspeed is used since it allows training with multiple GPUs.
Anonymous No.106162964 [Report]
>>106162777
check'd
Anonymous No.106162974 [Report] >>106163118
>>106158052
Sombrabox please.
Anonymous No.106162982 [Report]
>>106162804
This man speaks the truth
Anonymous No.106162996 [Report]
>>106162832
ani-taur is a cute!!!! OK?!?!?!
Anonymous No.106163009 [Report]
>>106162878
i think it was something about cumfarts and pedos
Anonymous No.106163013 [Report] >>106163026 >>106163874 >>106167943
How do you guys feel about Statler/Waldorf?
Anonymous No.106163022 [Report] >>106163072 >>106163095 >>106163716
>>106162947
wrong
comfy said that it makes no difference because once everything is on the gpu it doesn't matter if the wrapper is python/c/whatever
Also TraniStudio is somehow 10x slower than the same workflow in comfy
Anonymous No.106163023 [Report]
I don't know what I'm doing.

Why doesn't this change the image output?
Anonymous No.106163026 [Report]
>>106163013
forced meme. would be better in moderation
Anonymous No.106163054 [Report] >>106163884
Now trying neta lumina again,and neta yume lumina
Anonymous No.106163066 [Report]
Ladies(males)&Gentlemen:
Please
'''
git pull
'''
your ReForge repository.
Panchovix made important changes!
Yes, we are back!
Anonymous No.106163072 [Report] >>106163172
>>106163022
comfy also said learn C but doesn't write C himself. ani doesn't have access to a lot of the optimizations since they were written in python first. also there is no vulkan option for comfyui. if people actually care about cleaning up the huge amounts of waste from deps they should be making stuff that works with C/C++ first then provide python bindings not abuse the python wrapper for everything. that's how a disk becomes a landfill fast
Anonymous No.106163085 [Report]
>>106161237
I have 3 datasets that i got from the internet. Each zip has ~250 pics + captions. Is that too much? Should I thin it out? They are former sdxl datasets.
Anonymous No.106163095 [Report] >>106163362
>>106163022
>wahhhh the inference speed is everything!!!! don't think about the bloating or the shitty frontend runtime!!!!
Anonymous No.106163118 [Report] >>106163248
>>106162974
I'm sure you can figure it out on your own anon, i'm rooting for you
Anonymous No.106163158 [Report]
Holy shit, bong_tangent acts like it's got a built-in controlnet. If you use it in an upscaler and set denoise to 1.00 it sticks really close to the original image.

You can't push it too far (you start getting minor body horror at 3x or 4x resolution), but it's certainly different than the other schedulers that start sprouting multiple heads at 2k resolution.
Anonymous No.106163172 [Report] >>106163200
>>106163072
>ani doesn't have access to a lot of the optimizations since they were written in python first.
This sounds like you saying it's the tools fault for being written in python instead of her fault for choosing to not use python
Anonymous No.106163192 [Report]
waaaaa my shitUI is better than your shitUI waaaaa
Anonymous No.106163200 [Report] >>106163322
>>106163172
it has python interop anyways so just plug in whatever. it's good ani is opening up new paths for people to reverse engineer optimizations that work and trashing dep bloat
Anonymous No.106163231 [Report]
Remember:
VRAMlets hobby
VRAMlets general
VRAMlets website
Anonymous No.106163248 [Report]
>>106163118
No, I can't. You have tried that with the Mercy gen already and it didn't work. Please don't be like that.
Anonymous No.106163313 [Report]
Anonymous No.106163322 [Report]
>>106163200
ok I see how anistudio can destroy comfy now. ani has to work harder though it's taking forever
Anonymous No.106163331 [Report] >>106163344 >>106163589
will the release of the old chatgpt models 3 and 4 have any impact on pic genning?
or is this just for chatting?
Anonymous No.106163344 [Report]
>>106163331
anon, those don't have any impact at all. it would just outright refuse nsfw
Anonymous No.106163362 [Report] >>106163381
>>106163095
Are you argueing for using a super slow tool that's missing 90% of expected features because... people should just use it mkay?
Anonymous No.106163381 [Report] >>106163462
>>106163362
it's only as slow as comfy inference if you just use the interop but it's hard for you to comprehend an actual modular design instead of forced nodeshit
Anonymous No.106163397 [Report]
>>106162754
Just use the Linux replacement, I mean WSL.
Anonymous No.106163462 [Report]
>>106163381
it would be nice if it could mix cpp and python nodes so we can watch comfy get deprecated over time as each node gets replaced
Anonymous No.106163463 [Report]
Qwen tagged all the GPT slop so you can neg it out... right?
Anonymous No.106163510 [Report] >>106163530 >>106163603 >>106163856
so qwen is the new hot thing to use?

Can you train loras with 24gb vram
Anonymous No.106163514 [Report]
Anonymous No.106163530 [Report] >>106163551 >>106163577 >>106164584
>>106163510
>20b
you can't even run fp8 with less than 32GB of vram. I think you would need 48GB to train.
Anonymous No.106163551 [Report] >>106163884
>>106163530
I have 12gb Vram and 32gb Ram and I can run fp8 fine.
Anonymous No.106163555 [Report]
>open thread
>he's shilling his shitty wrapper again
>close thread
Anonymous No.106163564 [Report] >>106163690 >>106163889
in his defence i actually got it to run the other day but it crashed after a few gens :d
Anonymous No.106163577 [Report]
>>106163530
I am running it just fine it seems to use 23gb vram with the default comfy setup.

I do have two gpus but multiGPU just doesn't seem to work, that node is worthless
Anonymous No.106163589 [Report]
>>106163331
those models are getting dunked on as worse than existing open source models
Anonymous No.106163603 [Report]
>>106163510
With system ram offloading, yes, and of course with quantization you will be able to train without offloading, but quality will suffer

The training will be slow though, it's a big model
Anonymous No.106163690 [Report]
>>106163564
I wish the metadata/project layout load button functioned so I don't have to set everything up again everytime it crashes
Anonymous No.106163716 [Report]
>>106163022
>it makes no difference because once everything is on the gpu it doesn't matter if the wrapper is python/c/whatever
GIL and other Python issues do affect the speed at which everything is on the GPU. The difference using the same code but with libtorch would be marginal, maybe more of a difference on slower/older CPUs but any real performance gains come from optimizing the operators and custom kernels.
>aniStudio is somehow 10x slower
That's because sd-cpp has shit kernels
Anonymous No.106163790 [Report] >>106163887 >>106164773
I have a question:
Let's say I have generated an anime-style video involving 1-2 characters in a room. Let's say for example it's a school classroom.
What if I want to gen a continuation of this video in a new cut? For example, a different angle and composition, but the character(s) are in the same pose. wan cannot cut mid-gen so obviously I would need to generate an image that somehow believably captures what's going on in exactly the same environment.

What are my best options for this? Any good workflows? I know openAI's sora is pretty good at this, but the output resolutions are non-standard and I don't want to deal with censorship.
Anonymous No.106163802 [Report] >>106163862
I added Qwen-Image support to diffusion-pipe

The model is large, but with fp8 base model plus a bit of block swapping you can train loras on a single 24GB GPU. I've added instructions and an example config file to do this.

Ideally, you would have something like 2x3090 and use pipeline parallelism to split the layers across two GPUs, which works very well for this model.
Anonymous No.106163850 [Report]
If the 5070TiS gets 24GB is it gonna be a better purchase over a used 4090?
Anonymous No.106163856 [Report] >>106163875 >>106165212 >>106165918 >>106166197
>>106163510
i'm sorry anon
Anonymous No.106163862 [Report]
>>106163802
Nice, not that I think I'll torture my rusty old 3090 with this model either way, it will likely still feel too slow to be worth it.
Anonymous No.106163872 [Report] >>106163901
I have a 5070Ti, and I am not a vramlet. This is all just petty elitism from embarrassing manchildren. 16GB vram is more than capable for good AI generation in a timely manner.
Anonymous No.106163874 [Report] >>106163910 >>106164598
>>106163013
Anonymous No.106163875 [Report]
>>106163856
Well played, chuckles were had
Anonymous No.106163884 [Report]
>>106163551
What fun do you see in running that model at low quality without Loras? That's why I remain loyal to SDXL. Also, for Vramlets and SFW this model is free on LMarena now.
>>106163054 good consistency,
Anonymous No.106163887 [Report] >>106163899
>>106163790

Hmm... Probably Design Doll or Cascadeur with controlnet to maintain posing. Those are posing software.
Anonymous No.106163889 [Report]
>>106163564
Yeah also crashed all the time for me
Anonymous No.106163899 [Report] >>106164022
>>106163887
>Cascadeur
Cascadeur? I recognize this as a physics based animation software.
Anonymous No.106163901 [Report]
>>106163872
> I have a 5070Ti, and I am not a vramlet
Sayd the vramlet
Anonymous No.106163910 [Report]
>>106163874
This could actually be a scene directly from The Muppets

We could have those Bert and Ernie nazi memes in 720p video, what a time to be alive.
Anonymous No.106163914 [Report]
If you don't have at least 48VRAM itt, you are not prepared for this hobby and you a lesser human being.
Anonymous No.106163918 [Report] >>106163944 >>106164445
High vram cannot compete with a creative mind.
Anonymous No.106163919 [Report] >>106164000 >>106164119 >>106164328
>spend 2k to generate a 5s hit or miss slop video
KEK
Anonymous No.106163944 [Report]
>>106163918
unironically many do not understand this
better card =/= better gens
Anonymous No.106163945 [Report]
Do not forget we have a genjam:

https://forms.gle/DekqEyMcMpjtyxjy9
Anonymous No.106163979 [Report] >>106164019
Is there no I2V light2x 2.2 lora?
Anonymous No.106164000 [Report]
>>106163919
>"Your hobby is stupid because I can't afford it!"
Anonymous No.106164019 [Report]
>>106163979
i think kijai released one but it works worse than the 2.1 lora
Anonymous No.106164022 [Report] >>106164047
>>106163899

Yeah. Animations.
Anonymous No.106164047 [Report] >>106164800 >>106164812
>>106164022
Not really sure what this can do that Blender can't.
Also still not sure how I can take a still frame from a wan gen, then somehow transition to a 3D workflow to change the camera angle while maintaining the same environment and character poses.
Anonymous No.106164095 [Report]
>>106162864
>Hunyuan-World2.0 next year or whichever chinese company does open source
Tencent are not big fans of open source anymore
Pray alibaba looks into it
Anonymous No.106164119 [Report]
>>106163919
my 3090 was 500$ and i can use it for a crap ton of things anyway, sis
Anonymous No.106164136 [Report] >>106164149
>>106161112
>if one attosecond passed for every single possible i2v prompt, how long would it take?
Let’s break down what you're asking and work through it carefully.
Goal:

You want to calculate how long it would take if we spent 1 attosecond (1×10-18 seconds) per image generated by an image-to-visual (i2v) model like Stable Diffusion or DALL·E, for every possible image that could be generated, considering:

Every possible text prompt

Every negative prompt

Every resolution (within reason)

Every random seed

Every possible combination of pixel values that a consumer PC can realistically process

Step-by-step Breakdown

We’ll estimate the total number of possible outputs, then multiply that by 1 attosecond.
1. Pixel-Level Possibilities

Let’s use a single resolution as an example — say 512×512 (a common resolution for models like Stable Diffusion). We'll generalize later.
Bits per pixel

Assume 24-bit RGB images:
Each pixel = 3 color channels × 8 bits = 24 bits = 1 byte per channel × 3 = 3 bytes

So,

512×512 = 262,144 pixels

Total possible pixel combinations:
256(512×512×3)=256786,432=2(8×786,432)=26,291,456
256(512×512×3)=256786,432=2(8×786,432)=26,291,456

That's ~10189404 possible pixel combinations — just for 512×512 RGB images.

2. Resolution Variants

A "consumer PC" can reasonably handle resolutions from around 64×64 up to maybe 2048×2048 (though that's pushing memory limits). That gives a range of image sizes.

There are perhaps a few thousand practical resolution combinations (square, portrait, landscape) within that range.

Let’s be generous and say 104 resolutions.

3. Random Seeds

Models like Stable Diffusion often use 32-bit seeds. That’s:

232 = ~4.3 billion seeds = ~4.3×109
Anonymous No.106164143 [Report] >>106164192
so radian attention was a snake oil
Anonymous No.106164149 [Report]
>>106164136
4. Prompt and Negative Prompt Space

This is the trickiest part because it's infinite in theory. But you limited it to “everything a consumer PC can handle.” Let’s assume:

A model can take a max 77 tokens (common in transformer models like CLIP).

A vocabulary size of ~50,000 tokens.

So, the number of unique prompts is:
∑k=17750,000k
k=1∑7750,000k

This grows extremely fast. To get a rough upper bound, we can approximate with:
50,00077=10(log10(50,000)×77)≈10(4.7×77)≈10362
50,00077=10(log10(50,000)×77)≈10(4.7×77)≈10362

Let’s assume the same for negative prompts.

So total prompt combinations ≈
10362×10362=10724
10362×10362=10724
Total Possible Unique Generations

Multiply all components:

Pixel combinations (for 512×512): ~10189404

Resolutions: ~104

Seeds: ~109

Prompt + negative prompt combos: ~10724

Total images≈10189404×104×109×10724=10189404+4+9+724=10190141
Total images≈10189404×104×109×10724=10189404+4+9+724=10190141
Time in Attoseconds

1 attosecond = 1×10-18 seconds

So:
Total time=10190141×10−18=10190123 seconds
Total time=10190141×10−18=10190123 seconds
Convert to Years

There are ≈ 31.56 million seconds in a year:
10190123÷3.156×107≈10190123−7.5≈10190115.5 years
10190123÷3.156×107≈10190123−7.5≈10190115.5 years
Final Answer:

If 1 attosecond passed for every single possible image under your constraints:

It would take approximately 10^190115.5 years.

For comparison, the age of the universe is ~13.8 billion years ≈ 1.38 × 1010 years — a blip compared to this number.
Anonymous No.106164192 [Report]
>>106164143
too annoying to get running and wastes memory
Anonymous No.106164304 [Report] >>106164389 >>106165361
High-quality digital art in an anime style from pixiv. depicting a fantastical scene of a busy medieval tavern in the fantasy genre. The central figure is a girl wearing fantasy inspired armor resembling a metal bikini and has a toned, athletic physique, she is leaning back in a simple wooden chair with a stylised speech bubble saying "another successful commission!" . Surrounding her is an average looking adventuring party including one wizard with a hat and a paladin with a shield next to his chair she is drinking beer and talking with. The girl is holding a beer glass and waving it around enthusiastically. The atmosphere is celebratory and the people are all looking at each other. The tavern has an aged look, the wooded tables and barstools looking worn. The character's expression is intense, reflecting determination and focus.
Anonymous No.106164318 [Report] >>106164325
is 16gb vram and 32gb of ram enough to generate videos?
Anonymous No.106164325 [Report]
>>106164318
8gb is enough so yeah
Anonymous No.106164328 [Report] >>106164846 >>106165318
>>106163919
well worth it
Anonymous No.106164370 [Report]
Anonymous No.106164389 [Report] >>106164411
>>106164304
This is a good SDXL replacement for anime
Anonymous No.106164411 [Report] >>106164555
>>106164389
it's probably going to have similar problems to flux in refusing to change style much or properly learn characters/artists
Anonymous No.106164445 [Report] >>106164539 >>106164652
>>106163918
well if you generate 100 videos or pics and get a good one by chance, it can be more productive than a creative mind. there might also be some AI agent to pick the best ones for you. that could be the future of art
Anonymous No.106164456 [Report]
does qwen do nsfw
Anonymous No.106164462 [Report] >>106164486
is there a guide for retards on how to actually generate videos once you actually get this shit installed? ive never messed with it, but id like to tinker with it while i work from home.
Anonymous No.106164486 [Report] >>106164516 >>106164560
>>106164462
>2.2 Guide: https://rentry.org/wan22ldgguide
Anonymous No.106164516 [Report]
>>106164486
thats what i used for install. i was looking more for a guide on how to properly use workflows and what each section does, etc.
Anonymous No.106164539 [Report]
>>106164445
You keep relying on machines my guy.
Me? I have more than enough faith in my vision.
Anonymous No.106164555 [Report] >>106164571
>>106164411
Oh... so this will be forever SDXL?
Anonymous No.106164560 [Report]
>>106164486
the 2.2 t2v link is broken. also what is this flowmatch stuff?
>then delete "ModelSamplingSD3" from the workflow and replace "BasicScheduler" with the new "FlowMatchSigmas" node
those aren't in kijai's workflow
Anonymous No.106164571 [Report]
>>106164555
No, you must trust
Trust in RowWei
Trust in Neta Lumina
Trust in NovelAI when they release their open weights
Trust in illusitrous
Trust in Noob
Anonymous No.106164580 [Report] >>106164593 >>106164702 >>106166614
If only the outputs weren't completely slopped... (I used the same prompt about 2000s digital camera from the last thread)
Anonymous No.106164584 [Report]
>>106163530
Anonymous No.106164587 [Report]
I tire of these sidegrades when will we get a Real New Model
Anonymous No.106164593 [Report]
>>106164580
It's Qwen-Image (forgot to say)
Anonymous No.106164598 [Report] >>106164631
>>106163874
what is your height and width set to?
Anonymous No.106164623 [Report]
Anonymous No.106164627 [Report]
SDXL WINS AGAIN
Anonymous No.106164631 [Report] >>106164748
>>106164598
Anonymous No.106164635 [Report]
comfy should be dragged out on the street and shot
Anonymous No.106164652 [Report]
>>106164445
brownest post I've seen in a while
Anonymous No.106164674 [Report]
Anonymous No.106164702 [Report]
>>106164580
i kinda like the slop
Anonymous No.106164748 [Report]
>>106164631
F
Anonymous No.106164767 [Report] >>106164776
anything new since kontext nudify loras?
Anonymous No.106164773 [Report]
>>106163790
Seems like a LoRA could be created for hard cutting to different angles
Anonymous No.106164776 [Report]
>>106164767
i got laid (no longer have a use for nudify loras)
Anonymous No.106164790 [Report] >>106165950
Anonymous No.106164800 [Report]
>>106164047
have you tried flux kontext?
Anonymous No.106164812 [Report]
>>106164047
in wan
>still scene. time frozen. camera rotates around the room
Something like that and then find a frame for the angle that you want.
Anonymous No.106164813 [Report] >>106164858
Anonymous No.106164846 [Report]
>>106164328
kek
t. 3090ti
Anonymous No.106164858 [Report] >>106164906
>>106164813
I checked my vram situation, and as a dual gpu chad, it seems I still get plenty of free vram in one of the GPUs. It may be worth look into creating a Comfy flow where I can use Chroma as a second img2img pass on Qwenimg outputs, acting as a filter that unslops the image. The downside is that Chroma (at least the current version) would give bad fingers etc
Anonymous No.106164875 [Report] >>106164891
oh i understand now, it thinks penises are pink soft serve ice cream so thats why smooth chunks get licked off
Anonymous No.106164891 [Report]
>>106164875
how do i delete a post
Anonymous No.106164906 [Report] >>106165595
>>106164858
>The downside is that Chroma (at least the current version) would give bad fingers etc
for some reason Chroma works really nicely on i2i second pass, doesn't ruin fingers. I use res_multistep + beta
Anonymous No.106164931 [Report]
Anonymous No.106164961 [Report] >>106164970
That's it, I'll filter the word chroma, I don't have to see the name of a old and inferior model. Qwen has better composition, and SDXL it's the king of quick cooms.
Anonymous No.106164970 [Report]
>>106164961
Based
Anonymous No.106165002 [Report]
>>106161650
>fucking sad that I have to rely on advanced color grading in premiere pro to fix my gens
does it work well and is it automatic?
Anonymous No.106165028 [Report] >>106165179
Anonymous No.106165032 [Report] >>106165048 >>106165057
Really thinking we need /vdg/ now.
Anonymous No.106165048 [Report] >>106165060
>>106165032
>vdg
exists on /gif/
>>>/gif/29249725
Anonymous No.106165057 [Report] >>106165070
>>106165032
You don't post gens here though.
Anonymous No.106165060 [Report]
>>106165048
Yeah well we need an sfw version. Do not care for hardcore degeneracy.
Anonymous No.106165070 [Report] >>106165080
>>106165057
I certainly do.
I have zero interest in chroma or qwen and that's all I see here.
Anonymous No.106165080 [Report] >>106165082 >>106165110
>>106165070
ok go cope in /sdg/ with the other vramlets
Anonymous No.106165082 [Report]
>>106165080
/vdg/ is coming soon.
Anonymous No.106165084 [Report]
*yawn*
Anonymous No.106165110 [Report]
>>106165080 (me)
oh nvm i misinterpreted your post
Anonymous No.106165139 [Report] >>106165169 >>106165188
flux is megacensored and barely usable even with extreme lora autism

krea looks slightly better but might be even more censored and has no loras

chroma still has sd1.5 anatomy

wan image is less censored, but weirdly sensitive about settings, looks sloppy most of the time, very slow especially with dual model autism, early lora situation

qwen follows a little better than wan with a single model, but looks super sloppy and soft, even slower

who is going to save image gen?
Anonymous No.106165145 [Report] >>106165156 >>106165166
Is it a bad idea to go beyond 81 frames on 2.2?
Anonymous No.106165156 [Report]
>>106165145
don't do it anon
Anonymous No.106165166 [Report] >>106165181
>>106165145
i'm doing 161 frames no problem. well except pic related
Anonymous No.106165169 [Report]
>>106165139
praying for some richfag to do a proper chroma anatomy finetune
Anonymous No.106165179 [Report]
>>106165028
Anonymous No.106165181 [Report] >>106165501
>>106165166
What's the output like though? Does the video try to go back to the beginning? Any side effects like earthquakes or overbrightness?
Anonymous No.106165184 [Report]
Have a good day /ldg/.
Anonymous No.106165188 [Report]
>>106165139
Praying for some richfag to do a proper SDXL 2.0
Anonymous No.106165212 [Report]
>>106163856
kek
Anonymous No.106165269 [Report] >>106165279
I don't understand, you' guy are simping over 5 seconds of the worst animation possible?
Anonymous No.106165279 [Report]
>>106165269
What you don't understand could fill a library
Anonymous No.106165303 [Report]
Hey, what happened to landscape diffusion? Nobody baked anything?
Anonymous No.106165318 [Report]
>>106164328
>blonde
>denim shorts
those are some nice genes
Anonymous No.106165361 [Report]
>>106164304
Mmmm this is a good base bit is slopish as hell. Appart i supose that if SDXL its 1b and this model its 20b, a coomer would have to finetune the checkpoint or Lora x20 times
Anonymous No.106165446 [Report] >>106165490 >>106165817 >>106166413
Anonymous No.106165490 [Report] >>106165588
>>106165446
Betting $100 theres a black man in that suit
Anonymous No.106165501 [Report]
>>106165181
>Does the video try to go back to the beginning
with complex prompts it does tend to do that but you can kinda fight it with the prompt, like "the woman turns around to face away from the camera. blah blah blah. she continues to face away from the camera."
>Any side effects like earthquakes or overbrightness
no
Anonymous No.106165563 [Report] >>106165578 >>106165619
If I wanted to do image classification for tagging images with a long form description of what seems to be happening in an image and/or general descriptors tags?
Also is there one for short gifs?
ZNt2C No.106165575 [Report] >>106165745
oh wow, qwen->second pass chroma i2i is... really fucking good when it works. like might be all we need
Anonymous No.106165578 [Report]
>>106165563
sorry I meant to ask what model I might use...
Anonymous No.106165588 [Report]
>>106165490
he is at once in hell and also having the time of his life
Anonymous No.106165595 [Report]
>>106164906
A screenshot of your i2i workflow kind anon ?
Anonymous No.106165619 [Report]
>>106165563
Depends if it's nsfw or not. Joycaption for porn, Qwen2.5-VL-7B-Instruct for non-porn.
Anonymous No.106165639 [Report] >>106165776 >>106166413
Anonymous No.106165697 [Report]
>>106162840
>"do the fucking pushups"
>*throws ayyliums*
Anonymous No.106165710 [Report] >>106165723
what is this qwen shit everyone is talking about?
man I swear I'm gone for like a day or two and some new shit pops up instantly.
Anonymous No.106165711 [Report]
so we have qwen image with all this fancy stuff like editing and all we can use is t2i?
Anonymous No.106165723 [Report] >>106165740
>>106165710
Imagine Flux but scaled up and more Chinese.
Anonymous No.106165740 [Report] >>106165803
>>106165723
so its a new model like a better flux or what?
Anonymous No.106165743 [Report] >>106165808
Anonymous No.106165745 [Report] >>106165807
>>106165575
oh wow your pic looks amazing. worth it
Anonymous No.106165776 [Report] >>106165833 >>106166413
>>106165639
Anonymous No.106165782 [Report]
Anonymous No.106165803 [Report] >>106165826 >>106165907
>>106165740
Better Flux with less censorship. But it's also unreasonably big.
Anonymous No.106165807 [Report]
>>106165745
he's still genning, be patient
Anonymous No.106165808 [Report]
>>106165743
Billy has upgraded his Nintendo Super Scope, Jobst better run
Anonymous No.106165817 [Report]
>>106165446
https://youtu.be/3iLY-EzLefw
Anonymous No.106165826 [Report]
>>106165803
Since I don't like AI slop it's wan2.2 14b lownoise > flux krea > qwen image > flux
Anonymous No.106165833 [Report]
>>106165776
she hot
Anonymous No.106165859 [Report] >>106165897 >>106165912
I believe that in a street test, 100 out of 100 passers-by would choose wan without hesitation and say that it is a real photo.
Anonymous No.106165876 [Report] >>106166000
>radial attention and jenga back to getting regular updates
what a time to be alive!
Anonymous No.106165897 [Report]
>>106165859
Anonymous No.106165903 [Report] >>106166019
hey KARL, look what i'm doing with your money!
Anonymous No.106165904 [Report] >>106166018 >>106166049
Anonymous No.106165907 [Report] >>106165926
>>106165803
>But it's also unreasonably big.
you cant run it local?
Anonymous No.106165912 [Report]
>>106165859
To be fair, 90 of them would say dreamshaper looks real too
Anonymous No.106165918 [Report] >>106166027 >>106166769
>>106163856
howd you make it longer than 5sec?
Anonymous No.106165926 [Report]
>>106165907
A quanted version, but LoRA training is also important
Anonymous No.106165950 [Report]
>>106164790
kek'd
Anonymous No.106165997 [Report] >>106166007
more karl being super rich:
Anonymous No.106166000 [Report]
>>106165876
eww
Anonymous No.106166007 [Report]
>>106165997
*billy

just kidding karl is poor
Anonymous No.106166018 [Report]
>>106165904
nice style
Anonymous No.106166019 [Report]
>>106165903
Hot mother and daughter combo, is this what they mean when they say WINNING ?
Anonymous No.106166027 [Report] >>106166758
>>106165918
by just doing it
Anonymous No.106166042 [Report] >>106167440
Reminder that if you get OOM after a first successful generation, there is a node called unload-model that can solve the problem.
It did for me.
Just place before the last node, usually the one for saving the generation.
Anonymous No.106166049 [Report]
>>106165904
cool gen, my take on it
Anonymous No.106166089 [Report] >>106166144
I've put example images and videos in my lora folders.
Is there a way to explore lora files directly from comfy and have it show the video or image?
Anonymous No.106166107 [Report] >>106166140 >>106166154
Anonymous No.106166140 [Report]
>>106166107
its ok, an amer*can made the lora
Anonymous No.106166141 [Report] >>106166166
is it true wan 2.2 t2v is the best image gen

if so, why?
Anonymous No.106166144 [Report] >>106166434
>>106166089
loramanager extension
Anonymous No.106166154 [Report]
>>106166107
>2k downloads
Anonymous No.106166166 [Report]
>>106166141
Basically does porn out of the box, or at least with an any 8000 step LoRA.
Anonymous No.106166184 [Report]
can gwen do edit like kontex
is it better?
Anonymous No.106166197 [Report] >>106166217 >>106166433
>>106163856
First they came for the 12gbs, and I did not speak out—because I was not a POORFAG.

Then they came for the 16gbs, and I did not speak out—because I was not a VRAMLET.

Then they came for the 24gbs, and I did not speak out—because I was not a OBSOLETEFAG.

Then they came for me—and there was no one left to speak for me.
Anonymous No.106166207 [Report]
Anonymous No.106166217 [Report] >>106166433
>>106166197
<16gbs aren't even human
Anonymous No.106166301 [Report] >>106166306 >>106166345 >>106166413 >>106166491
Anonymous No.106166306 [Report]
>>106166301
>he knew
Anonymous No.106166308 [Report]
kek, wan 2.2 camera test

a man runs to the right very fast. the camera tracks his movement as he runs.
ZNt2C No.106166342 [Report] >>106166470
Do your qwen image, i2i it to chroma, 20 steps sfg 4.5, res multistep beta, denoise .35, second prompt is just (amateur photo:3) or whatever you want if you want to add in elements that qwen doesnt do well
Anonymous No.106166345 [Report] >>106166417
>>106166301
add some 90s camera filter and make ragebait post on xister
Anonymous No.106166413 [Report]
>>106166301
>>106165446
you cant move like that and still be fat

>>106165776
>>106165639
pure sexo
Anonymous No.106166417 [Report] >>106166671
>>106166345
lol
Anonymous No.106166433 [Report] >>106166445 >>106166456
>>106166197
>>106166217
People doing image / video gen don't even know how good they have it. Over in /lmg/, 72 GB (3x3090) is considered by many to be King of VRAMlets. Only at 96GB+ are you no longer VRAMlet.
Anonymous No.106166434 [Report] >>106166450
>>106166144
This?
https://github.com/willmiao/ComfyUI-Lora-Manager
With its own interface?
Can it interact locally? It looks like it's mostly about pulling stuff from civitai.
Anonymous No.106166445 [Report]
>>106166433
Yeah but you can pool memory with LLMs, but I don't think you can do that with image and videogen.
Anonymous No.106166450 [Report]
>>106166434
all these interface extensions all universally suck
Anonymous No.106166456 [Report]
>>106166433
at least you chain cards relatively cheap cards. we need the memory all on one card and the card has to be fast
Anonymous No.106166470 [Report] >>106166614
>>106166342
Sadly it still looks kinda slopped with those settings
Less slopped for sure, but still slopped
Anonymous No.106166491 [Report]
>>106166301
kek
Anonymous No.106166493 [Report] >>106167700
Anonymous No.106166516 [Report] >>106166525
wan img2vid never gets old
the little things always get me
Anonymous No.106166525 [Report]
>>106166516
this, like my first time trying SD all over again
Anonymous No.106166570 [Report] >>106166653
a man runs to the right extremely fast and flies into the sky. The sky is blue and cloudy.
Anonymous No.106166614 [Report] >>106166639
>>106166470
It definitely has some effect though.

It improves skin (compare with >>106164580
), but it doesn't get rid of the background blur

I used (candid amateur photo:3)
Anonymous No.106166639 [Report]
>>106166614
maybe try something like "textured wallpaper in background" so its not just a blurry color? Might need to give it something to force it into bootleg f/22 mode
Anonymous No.106166653 [Report]
>>106166570
You gotta hand it to the chinks, Wan can handle basically every concept you throw at it, and you can run it locally on consumer hardware
Anonymous No.106166671 [Report]
>>106166417
>Women on the highway
Always a recipe for disaster
Anonymous No.106166696 [Report]
made a node that gets all the loras from high noise and allows you to reuse them for low noise with adjustable lora strength. reduces boilerplate/clutter from duplicate loras. do you think this is helpful?
ZNt2C No.106166697 [Report]
that looks a lot less slopped actually
Anonymous No.106166758 [Report] >>106166776 >>106166803 >>106167403
>>106166027
Anonymous No.106166769 [Report] >>106166926 >>106167044
>>106165918
Wait until you find out 4fps/8fps LoRAs work.
Anonymous No.106166776 [Report] >>106166959
>>106166758
kinda amazing that wan can even interpret that
Anonymous No.106166803 [Report]
>>106166758
lmfao
Anonymous No.106166926 [Report] >>106166964
>>106166769
>4fps/8fps LoRAs work
the what/where
Anonymous No.106166959 [Report] >>106167403 >>106168137
>>106166776
china numba wan
Anonymous No.106166964 [Report]
>>106166926
If you make a LoRA at 4fps or 8fps you effectively double or quadruple your running length. You basically rescale a clip from 4fps to 16fps (which will appear to be fast forwarded) but trained like that Wan will produce longer length clips. 120 frames = 30 seconds at 4fps, and it works.
Anonymous No.106166993 [Report]
i am STILL waiting for anon to prove mmaudio can do nsfw
Anonymous No.106167044 [Report] >>106167068
>>106166769
isnt the data set standardized at 80 frames 16fps? wouldnt extending the number of frames mess up motion? i was thinking about taking the last frame and re-sending it through the workflow and splice them together in post, but seems like a lot of work
Anonymous No.106167068 [Report] >>106167077
>>106167044
4fps is basically fast forward or sped up footage which is in the dataset, so the concept actually works.
Anonymous No.106167069 [Report] >>106167080
>https://x.com/jkbr_ai/status/1953154961988305384

this shit is honestly crazy, cant wait until some chink opensources the thing with 90% quality
Anonymous No.106167077 [Report] >>106167095
>>106167068
oh interesting... ill have to play around with that, didnt think sped up footage was in there
Anonymous No.106167080 [Report] >>106167096
>>106167069
>actually works, the prompt bled into the sliding door showing the outside
It's a gimmick but as I suspected the results are way over blown, it's just a slightly refined version of the Minecraft hallucination model.
Anonymous No.106167095 [Report]
>>106167077
I mean of course it is, sped up footage is in many normal videos. All you're doing is rescaling the 16fps "sped up" footage to 4fps. And of course that's not talking about how Wan is a very smart model in general, it knows a lot of small things due to its size and complexity.
Anonymous No.106167096 [Report] >>106167129 >>106167181
>>106167080
wonder how they handled "memory" of actions made in the model at runtime since if you look away it doesnt make stuff disappear
Anonymous No.106167129 [Report] >>106167148 >>106167224
>>106167096
It's just stacked up frames in context. I don't think they ever showed a true memory challenge, only that it doesn't forget within 5 seconds of moving away. I've also noticed in their cherrypicked examples that the motion is generally slow with no forms of fast travel, so maybe if I saw a demo where they were exploring an environment like Myst where they travel in a circle on an island back to a wall with custom writing on it after 30 seconds I'd be more impressed.
Anonymous No.106167148 [Report]
>>106167129
For example, I think with Wan if you had 20 frames of "context" rather than a straight first frame, you'd have a similar memory solution.
Anonymous No.106167181 [Report]
>>106167096
might be flux chin, it's always there even if you don't want it to be
Anonymous No.106167224 [Report] >>106167254
>>106167129
interesting, this is the only one ive seen with decent interactive memory, but like you said, it never really leaves the frame for more than 5 seconds, so good eye on that

>https://x.com/_rockt/status/1952735159834325210
Anonymous No.106167232 [Report] >>106167246 >>106167258
Finally got FastWan working
Anonymous No.106167246 [Report]
>>106167232
solid camera shake on this
Anonymous No.106167248 [Report]
Anonymous No.106167254 [Report] >>106167264
>>106167224
seems like a good match with vr, in that eternal tech demo state vr exists in
Anonymous No.106167258 [Report]
>>106167232
>deepfried in corn oil
Anonymous No.106167263 [Report]
>my preview nodes stopped working and only the last node is actually saving
wtfff
Anonymous No.106167264 [Report]
>>106167254
I mean it's good enough to make VR porn, not that long memory is that important.
Anonymous No.106167275 [Report] >>106167298
pro tip: it's hard to forget something when it's right in front of you
Anonymous No.106167298 [Report] >>106167316
>>106167275
then how can i forget about all the trash around my monitor so easily?
Anonymous No.106167316 [Report] >>106167348
>>106167298
liar, you know it's there because you were trained on slop
Anonymous No.106167331 [Report] >>106167348
>tfw parents are a slop dataset
Anonymous No.106167348 [Report]
>>106167316
>>106167331
Anonymous No.106167403 [Report] >>106167507
>>106166758
>>106166959
The future of jaks is bright
Anonymous No.106167404 [Report]
Anonymous No.106167408 [Report]
still not jumping on the wan bandwagon but would 10 GB be enough, or should i not bother
Anonymous No.106167419 [Report]
Anonymous No.106167440 [Report]
>>106166042
is there a better node system to control model management? i want to decide when a model is loaded or unloaded :/
Anonymous No.106167507 [Report]
>>106167403
my brainlet lora paying dividends already
Anonymous No.106167559 [Report] >>106167569 >>106167637
a man flies into the sky to the right at extremely high speed. The sky is blue and cloudy.

holy shit, billy became a plane. wan never gets boring. and thanks to the lightx2 i2v loras, it's fast. (no more 15 min gens)
Anonymous No.106167569 [Report]
>>106167559
my guy became jay jay the jet plane
Anonymous No.106167593 [Report] >>106167625
for video gen, new gpu or more ram? at 64gb ddr5 3090 atm
Anonymous No.106167625 [Report]
>>106167593
more ram would be a fraction of the cost of a new gpu so why not both?
Anonymous No.106167637 [Report] >>106167652
>>106167559
there we go. now he's flying (as a person)
Anonymous No.106167643 [Report]
genjam status?
Anonymous No.106167648 [Report]
Anonymous No.106167652 [Report]
>>106167637
>i'm off to take karl's money again
Anonymous No.106167681 [Report] >>106167695 >>106167817
"good" end?
Anonymous No.106167695 [Report]
>>106167681
not for that pony
Anonymous No.106167697 [Report] >>106167709 >>106167749 >>106167753 >>106167843
guys, can you give us some better themes? i'm sick of your ugly males gens. i know this thread is blue, but come on
Anonymous No.106167700 [Report] >>106167864
>>106166493
Model? This is badass.
Anonymous No.106167709 [Report] >>106167780 >>106167812
>>106167697
what do you have in mind
Anonymous No.106167718 [Report] >>106167730 >>106167785 >>106167823
just stopping by to say this is the best fucking collage I can remember lmao
Anonymous No.106167730 [Report] >>106167785
>>106167718
Anon posts kinosoul ITT doesn't he
Anonymous No.106167746 [Report] >>106167757
Anonymous No.106167749 [Report]
>>106167697
too busy losing my sanity with spicy 2.2 gens
Anonymous No.106167753 [Report]
>>106167697
The male form is beautiful tho and I'm not gay for enjoying it
ポストカード !!FH+LSJVkIY9 No.106167755 [Report]
>70 images @ bump limit
ooooffff ;3
Anonymous No.106167757 [Report] >>106167825
>>106167746
i...now...want...to...die...for...the central banking usury debt slave system of made up fiat currency not tied to anything physical...
Anonymous No.106167771 [Report] >>106167785
>70 pieces of pure kino
Anonymous No.106167780 [Report] >>106167838
>>106167709
robots theme, medieval, starwars, etc
Anonymous No.106167785 [Report]
>>106167718
>>106167730
>>106167771
u summoned the schizo now, gg
Anonymous No.106167812 [Report]
>>106167709
anime
schizo
Anonymous No.106167817 [Report]
>>106167681
my ass hurts...
Anonymous No.106167823 [Report]
>>106167718
i like to look at sort of the progression of the collages with how the number of gens included seems to ebb and flow "naturally" i.e. there'll be a bunch in one collage, then the number of gens will slowly decrease over time, and then suddenly there'll be one that includes almost 20+ gens
Anonymous No.106167825 [Report]
>>106167757
kek, also have this Kino.
Anonymous No.106167838 [Report] >>106167924
>>106167780
Wow how original anon...
Anonymous No.106167839 [Report] >>106167850
Anon-sans, how do I obtain less dithered animations?
ポストカード !!FH+LSJVkIY9 No.106167843 [Report]
>>106167697
nope you must visit other generals or wait until your cooldown\throttle is removed ;3

behave !
Anonymous No.106167850 [Report]
>>106167839
increase resolution. tone down the strength of lightx2v maybe
Anonymous No.106167861 [Report]
>>106167855
>>106167855
>>106167855
>>106167855
Anonymous No.106167864 [Report] >>106168206
>>106167700
Noobai with pixelization script.
Here's a typical output without it.
Anonymous No.106167924 [Report]
>>106167838
but nobody made them,except some rare robots gens. medieval theme is rarely here
ポストカード !!FH+LSJVkIY9 No.106167943 [Report]
>>106163013
he targets me a LOT
but such are the ways of \g\ schizos ;c
Anonymous No.106168137 [Report] >>106168153
>>106166959
prompt? I have been trying to get someone to puke for months now
Anonymous No.106168153 [Report]
>>106168137
nta but 2.2 is way better at prompt adherence
Anonymous No.106168206 [Report]
>>106167864
Very cool. Thank you anon.
Anonymous No.106168479 [Report]
>omg, the cat is working at mcdonald's. this is epic!