← Home ← Back to /g/

Thread 107154826

200 posts 142 images /g/
Anonymous No.107154826 [Report] >>107156856
/ldg/ - Local Diffusion General
Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107145378

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Neta Yume (Lumina 2)
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd
https://gumgum10.github.io/gumgum.github.io/
https://neta-lumina-style.tz03.xyz/
https://huggingface.co/neta-art/Neta-Lumina

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.107154861 [Report] >>107154885
Anonymous No.107154883 [Report] >>107154891 >>107154975
From "Localsong" + a lora:

https://voca.ro/1cbIetpoY6Gv

I am telling ya, this shit has potential
Anonymous No.107154884 [Report]
Anonymous No.107154885 [Report] >>107154896
>>107154861
based
Anonymous No.107154886 [Report] >>107154920
blessed bred
Anonymous No.107154891 [Report] >>107154913
>>107154883
that's no language I've ever heard, sounds like gibberish
Anonymous No.107154896 [Report] >>107155543
>>107154885
thanks. any OCR or VLM anons want to see if their model can read these?
Anonymous No.107154904 [Report] >>107154943
https://files.catbox.moe/9egs1f.png
Anonymous No.107154908 [Report] >>107154937
What program/model do I use to gen cool landscape images?
Anonymous No.107154913 [Report]
>>107154891
Who cares when the melody sounds cool
Modern music is garbage precisely because artists try to give emphasis to the lyrics way too much
Anonymous No.107154915 [Report]
https://files.catbox.moe/cqg2n9.png
Anonymous No.107154918 [Report] >>107154939 >>107154958 >>107156131 >>107156269 >>107156312 >>107156462
For those who missed it:
https://github.com/Lakonik/ComfyUI-piFlow
https://huggingface.co/spaces/Lakonik/pi-Qwen
https://huggingface.co/Lakonik/pi-Qwen-Image
https://huggingface.co/Lakonik/pi-FLUX.1
>>107154174
>Ok this thing is kind of insane. I made a workflow to compare it with normal Qwen, and it's basically the same level of quality while taking less than 10% of the time. Works out of the box with loras also. In fact, with a custom lora on a mediocre quality dataset, the results are arguably better with this thing at 4 steps. It is partially counteracting the shitty quality of my dataset. Absolutely the new meta for using Qwen, it will be impossible to go back with how fast it is.
Anonymous No.107154920 [Report] >>107154956 >>107155072
>>107154886
Anonymous No.107154937 [Report] >>107154943 >>107154944
>>107154908
You can try regional prompting, so that in one region of the image it'll follow this prompt, then in another this other prompt. You can also try inpainting

----
https://files.catbox.moe/2zhb62.png
Anonymous No.107154939 [Report]
>>107154918
>20s qwen gen
not bad, i would still give a little denoise with something to tidy it up.

if you gen with qwen then do wan denoise, where do you even post that on civit?
Anonymous No.107154943 [Report] >>107154972
>>107154937
>>107154904
how long do your WAN gens take anon?
Anonymous No.107154944 [Report] >>107154972
>>107154937
I just wanna go:

>landscape, big castle, atmospheric, dark clouds, lightning, mountains

What does that?
Anonymous No.107154948 [Report]
https://files.catbox.moe/e3dk4s.png
Anonymous No.107154956 [Report] >>107158038
>>107154920
Anonymous No.107154958 [Report] >>107155437
>>107154918
>6s flux gen with 4steps
Anonymous No.107154972 [Report] >>107154981
>>107154943
Takes about 3 mins per generation. The workflow I use has an upscaler that basically generates the image twice

>>107154944
Hmm, I see. Any image generator can do that. I thought you were going for a specific compostion, etc

https://files.catbox.moe/tcgxrp.png
Anonymous No.107154975 [Report]
>>107154883
alright, i'm gonna give this a try with some instrumental tracks and see what happens. this was convincing, lyrics aside (which i know the page said it wasn't trained on lyrics)
Anonymous No.107154981 [Report] >>107154997
>>107154972
you got the patience for that? asking coz I dont. I can get a 540p WAN video at least twice. I know your gens are super good. its just too long I fee.
Anonymous No.107154989 [Report]
>ram prices skyrocketing
>rumors of 5000 series supers being delayed

bros... I'm about to give in. I'm tired of waiting. Should I buy a used 3090 or 5070ti? they are about the same price
Anonymous No.107154997 [Report] >>107155004
https://files.catbox.moe/sp4jkj.png

>>107154981
Yeah, I actually set up a bunch of them in a row then I got eat a snack or something, lol. Thanks for the compliment btw.
Also: You can cut generation time by half by skipping the upscaler/upres part of the workflow
Anonymous No.107155004 [Report] >>107155038
>>107154997
Ill give it a shot. I havnet your level of realism till now.
Anonymous No.107155038 [Report] >>107155045 >>107155050
>>107155004
WAN is perfect to recreate the "modern digital" photography style, that you see with most photojournalism and some photographers
Also, it has pretty much perfect anatomical precision, but adding loras (i.e porn loras) decrease this precision

https://files.catbox.moe/wbkfmb.png
Anonymous No.107155045 [Report]
>>107155038
oh yeah the military ones look damn good.
Anonymous No.107155048 [Report] >>107155234
https://files.catbox.moe/rc3h45.png
Anonymous No.107155050 [Report] >>107155066 >>107155069 >>107155117 >>107155166 >>107155178
>>107155038
can you do images like this but with bikini thighhighs girls?
Anonymous No.107155066 [Report]
>>107155050
I can, but I don't wanna get the banhammer. Also, I don't have access to the 5090 I use to generate the imgs rn.
I'll post some NSFW next post. I'll just post the catbox link, i won't up the img on the thread
https://files.catbox.moe/3jpm5w.png
Anonymous No.107155069 [Report] >>107155166 >>107155178 >>107155206
>>107155050
+1
Anonymous No.107155072 [Report] >>107155117
>>107154920
nta making the other wan gens
Anonymous No.107155117 [Report]
>>107155050
>>107155072
I don't have access to the 5090 I use to generate images rn, sorry. The porn images I've are mostly artsy-fartsy ones
Anonymous No.107155166 [Report] >>107155188 >>107155225 >>107157290
>>107155050
>>107155069
This gen is a rare one made in the "digital photojournalistic" style I've on hands rn

https://files.catbox.moe/lei0s5.png
Anonymous No.107155178 [Report] >>107155190 >>107155217 >>107155225
>>107155050
>>107155069
An example of my typical "artsy fartsy" gens.lmk if you guys want more

https://files.catbox.moe/y93k43.png
Anonymous No.107155187 [Report] >>107158212
Anonymous No.107155188 [Report] >>107155256
>>107155166
can you generate feminist protesting free nipples or something feminist but are actually hot babes with big tiddies in underwear and wearing thighhighs?
Anonymous No.107155190 [Report] >>107155256
>>107155178
Yess ofc definitely!
Anonymous No.107155195 [Report]
man, all these dit models kinda suck. was raping ram really worth having nlp? everything was just fine if not better when we used controlnets and ipadapter. edit models were a mistake
Anonymous No.107155204 [Report] >>107155425 >>107156771
Anonymous No.107155206 [Report]
>>107155069
Nice Redditor Gold there, kind stranger!
Anonymous No.107155217 [Report] >>107155256
>>107155178
i too would like more
Anonymous No.107155222 [Report] >>107155272
Anonymous No.107155225 [Report] >>107155272
>>107155166
>>107155178
These are great
Anonymous No.107155234 [Report] >>107155240 >>107155272
>>107155048
Anonymous No.107155240 [Report] >>107155245 >>107155251 >>107155272
>>107155234
What track is this?
Anonymous No.107155245 [Report]
>>107155240
le circuit de wan
Anonymous No.107155251 [Report]
>>107155240
this is going to be the first playable "world simulator" game. just an infinite race track. probably releasable by someone like deepmind right now
Anonymous No.107155256 [Report]
>>107155217
>>107155190
https://files.catbox.moe/32hb6v.png
>>107155188
can't, sorry. this machine can't gen imgs
Anonymous No.107155272 [Report]
>>107155225
Thanks a lot, fren!

>>107155234
>>107155222
Awesome gens, fren! Loved how the lead car went to the F-Zero shield recharge strip at the end there, lmao

>>107155240
Reminds me of the start/finish line from Imola, but it's not any particular track
Anonymous No.107155339 [Report]
Fencing duel gens, complete pic(s) in the catbox
https://files.catbox.moe/10dpcm.png
https://files.catbox.moe/9g7xb8.png
Anonymous No.107155364 [Report] >>107157290
TW: suifuel (contains happy couple)
https://files.catbox.moe/ngt115.png
Anonymous No.107155370 [Report]
last one for now, gtg work. another duel, this time to the death
https://files.catbox.moe/y7jlxy.png
Anonymous No.107155410 [Report]
Blessed thread of frenship
Anonymous No.107155425 [Report]
>>107155204
recipe for this bread?
Anonymous No.107155437 [Report] >>107155505 >>107155799
>>107154958
Does it work with Chroma since it supports Flux?
Anonymous No.107155505 [Report]
>>107155437
try it and find out
Anonymous No.107155543 [Report]
>>107154896
Anonymous No.107155614 [Report] >>107156195
Sega Genesis Sonic-style track on "LocalSong":

https://voca.ro/13U9LKll5na4

Things got a bit bad in the end, but overall pretty good
Anonymous No.107155799 [Report] >>107156167
>>107155437
>60s with (30s -> face detailer), 12steps using 8step lora. no dice on chroma, it has hardcoded qwen and flux in the loader
Anonymous No.107155852 [Report]
Need a wan lora from the Tylers poop festival video
Anonymous No.107155866 [Report] >>107155875
>happily gen some cute anime 1girls at the start of the year
>look away from the screen for a moment
>Huge fucking pile of optimizations happen
I feel like unless you're keeping up with this daily, you're just hopelessly left behind because its impossible to find information on whatever sage attention or these other -attention fixes are, how to use it, or what they're for because it gets buried under a sea of new or conflicting information.
Anonymous No.107155875 [Report]
>>107155866
that would be the case if anyone used said optimizations. unless it's merged into mainline comfyui, most of the good optimizations (both for speed and quality) just get ignored/forgotten.
Anonymous No.107155946 [Report]
>>107154100
>>107154342
Nope, doesn't build with downgraded toolkit:(
Yaps about nvvc not existing after idling for half an hour. I guess the other anon who warned about incompatibility was right.
Gonna wait TM for official support or make separate docker for it later.
Anonymous No.107155977 [Report]
Anonymous No.107156022 [Report] >>107156045 >>107157732
What do you want the most for a local model?

https://poal.me/7udx6s
https://poal.me/7udx6s
https://poal.me/7udx6s
https://poal.me/7udx6s
Anonymous No.107156045 [Report] >>107156054 >>107156056 >>107156072 >>107156110
>>107156022
anyone voting anything than video is retarded, images are already mostly there, the biggest thing we need is edit model without vae, video has a long way to go in comparison
Anonymous No.107156054 [Report]
>>107156045
>anyone voting anything than video is retarded
*or vramlet
Anonymous No.107156056 [Report]
>>107156045
yep this was my take too
Anonymous No.107156059 [Report]
Retards rise up
Anonymous No.107156072 [Report]
>>107156045
Video models are less suitable for prompt alignment for a single frame
Anonymous No.107156110 [Report]
>>107156045
I'm excited for video because I know video brings audio in with it immediately as well. Immediately ASMR and braps and sound effects and short dialogue sentences and memes and swears and so much more are solved before we even get a text-to-audio model that's good
Anonymous No.107156130 [Report] >>107156157
You know deep in your hearts that you will not be able to run Sora 2 grade stuff without 48gb vram and waiting 10+ minutes per video even with distillation and quants
Anonymous No.107156131 [Report]
>>107154918
>ctrl f "edit"
>zero results
does it work for qwen-e
Anonymous No.107156157 [Report] >>107156170 >>107156736
>>107156130
correct, we will have something much better than dogshit sora lol
Anonymous No.107156167 [Report]
>>107155799
Lame ty. Glanced at the code and it seems like there's a few places that would need adapting
Anonymous No.107156170 [Report] >>107156194
>>107156157
I am an openai hater as well, but come on anon, let's not cope that way
Anonymous No.107156194 [Report] >>107156310
>>107156170
toy model for memes whose only great thing is the fact that they trained on the entire youtube dataset, without that its literally worse than wan 2.2
Anonymous No.107156195 [Report]
>>107155614
well it got the genesis instruments right for sure
Anonymous No.107156269 [Report] >>107156279 >>107156291
>>107154918
Loaded this up and I'm getting 20 second Qwen gens even with my shitty setup, what sorcery is this
Anonymous No.107156279 [Report]
>>107156269
vram?
Anonymous No.107156282 [Report] >>107156458
What is the current meta lora for speeding up wan 2.2 14b i2v?
Anonymous No.107156291 [Report]
>>107156269
16GB, RX 9070 XT.
Anonymous No.107156310 [Report] >>107156335
>>107156194
It's still superior to any open video model in existence by a country mile, and that will remain true for a long time. To this day, there isn't a single local model that can pull some of the stuff that dalle3 could in 2023
If you cherrypick things, Wan does mangled outputs just as often
Anonymous No.107156312 [Report]
>>107154918
does it work with gguf?
Anonymous No.107156335 [Report] >>107156393 >>107156920
>>107156310
>If you cherrypick things, Wan does mangled outputs just as often
not by a mile
sadly for you, the apicuck model cant be tested 1:1 with local because its locked into a chastity cage, like all who shill for it
Anonymous No.107156393 [Report] >>107156479 >>107156485
>>107156335
>sadly for you, the apicuck model cant be tested 1:1 with local because its locked into a chastity cage, like all who shill for it
You do realize there are other possible prompts other than porn and politically incorrect stuff, right? So yes, they can be compared
Anonymous No.107156458 [Report] >>107156523
>>107156282
Let me be more clear.
Apperantly I am still using this from 3 months ago:
https://huggingface.co/lightx2v/Wan2.2-Lightning/blob/main/Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1/high_noise_model.safetensors
Is this:
https://huggingface.co/lightx2v/Wan2.2-Distill-Loras/tree/main
Or anything else better than it?
Anonymous No.107156462 [Report] >>107156855
>>107154918
>uses own ksampler
>uses own model loader
INTO THE TRASH IT GOES
Anonymous No.107156479 [Report]
>>107156393
NTA compared =/= 1:1
Anonymous No.107156485 [Report] >>107156509
>>107156393
Wow. I didn't know that. You're telling me now for the first time
Anonymous No.107156509 [Report] >>107156559
>>107156485
You're welcome anon. It's enlightening indeed to know there are more prompts other that "1girl big bobs and vagene", who would have guessed!
Anonymous No.107156523 [Report] >>107156717
>>107156458
There also seems to be a moe distill lora...
Anonymous No.107156559 [Report]
>>107156509
damn, gotta step my game up, i mean imagine a 1girl with smal bobs... it got my creative juices flowing
(and unretarding for a minute: curiosity in how to setup those matrix comparison graphs people post every now and then, since those can be programmed, i think?)
Anonymous No.107156717 [Report]
>>107156523
There also seems to be v1030 that got deleted
https://huggingface.co/Kijai/WanVideo_comfy/blob/main/LoRAs/Wan22_Lightx2v/Wan_2_2_I2V_A14B_HIGH_lightx2v_4step_lora_v1030_rank_64_bf16.safetensors
I don't expect a wall of text spoonfeeding me strengths and weaknesses of all but just what are anons here using in their daily gens?
Anonymous No.107156736 [Report] >>107156761
>>107156157
we still don't have DALL-E 3 at home, stop coping
Anonymous No.107156750 [Report]
what's a good free software for managing gens? preferably one that shows the metadata like prompts. I'm getting to have too many. bonus points if it does wan too, though idk if it actually has metadata yet. I only just started with that
Anonymous No.107156761 [Report] >>107156774 >>107156845
>>107156736
correct yet again, we have something much better than dalle 3, the possibility to train a lora on anything you want and generate with any parameters you want with no limits, including training a dalle 3 style lora itself like picrel
Anonymous No.107156771 [Report]
>>107155204
It upsets me that I can't reproduce this solid vectorized style.
Anonymous No.107156772 [Report] >>107156804 >>107156860
nano banana 2 is too good
its over for local
Anonymous No.107156774 [Report]
>>107156761
lora https://civitai.com/models/2093591
Anonymous No.107156778 [Report]
Anonymous No.107156804 [Report] >>107156840
>>107156772
The better proprietarycuck edit models are, the better outputs the new qwen image edit model can be easily trained on, thanks for spending millions for local to snatch it all up for free before training a clothes remover lora within a couple hours lol
Anonymous No.107156840 [Report]
>>107156804
based
Anonymous No.107156845 [Report] >>107156854
>>107156761
it's not about the style, or any specific thing object/concept, retard
that you thought it was tells me all I have to know about your intellectual level, you don't understand what dall-e 3 has that local still has not and you never will understand because you're a moron
Anonymous No.107156854 [Report] >>107157028
>>107156845
>no argument
oof, thanks for conceeding
Anonymous No.107156855 [Report]
>>107156462
This. I can't fucking use this in my workflow. I needs my snake oil!
Anonymous No.107156856 [Report]
>>107154826 (OP)
>not collaging the real braphog
Anonymous No.107156860 [Report]
>>107156772
It still can't do maps. (Courtesy of some plebbitor.)
But yes the whiteboard math equation stuff is impressive.
Anonymous No.107156920 [Report] >>107156940
>>107156335
>not by a mile
No local model can gen multiscene videos WITH audio at the same time, so yes, nothing local comes close to it currently

The closest thing to it is this Wan fine-tune for multiscene, which has no audio:

https://holo-cine.github.io/


(and I haven't seen any anon use this)

Apparently they will release the weights for an audio component later though, so we'll see (there is a HoloCine-audio in the roadmap as well as an I2V version)
Anonymous No.107156940 [Report] >>107157076
>>107156920
no proprietary model is gonna allow you lora creation for whatever you want nor to tweak every gen parameter, that is the thing that actually matters, everything else can already either be done locally or can be done locally but with more manual work worst case scenario, but proprietarycucks literally CANT do these things and wont ever be able to in any way.
Anonymous No.107156982 [Report] >>107157073 >>107157453
>a- aunt jemima... is that OK to wear in public?
Anonymous No.107157028 [Report] >>107157052
>>107156854
keep on coping, copeboy
Anonymous No.107157052 [Report] >>107157076
>>107157028
>no argument
already accepted your concession lil bro, keep crashing out
Anonymous No.107157073 [Report]
>>107156982
Very nice anon
Anonymous No.107157076 [Report] >>107157092
>>107157052
you do whatever it takes to keep the cope alive
is this you?>>107156940
>everything else can already either be done locally or can be done locally but with more manual work worst case scenario
lol, lmao even
Anonymous No.107157092 [Report] >>107157141
>>107157076
>no argument
this has to be a bot, right? lol
Anonymous No.107157098 [Report] >>107157114 >>107157199
Most important things for new pc if I wanna do decent video gens in a non absurd timeframe?
I don’t wanna reply to ever webm in here asking for pc specs but if someone wants to post some with their specs/how long it took I’d greatly appreciate it
Budget is about 2.5k for new pc
Anonymous No.107157114 [Report]
>>107157098
16gb vram is the single most important thing. more than that is better. less than that you're fucked.
Anonymous No.107157141 [Report] >>107157370
>>107157092
of course, anyone who laughs at your lack of intelligence is a bot
the argument is that you're a retard, you give more weight to what can be done locally just to poop on the things local can't do yet, that's moron behavior
>can be done locally but with more manual work worst case scenario
ANYTHING can be done locally but with more manual work, just grab a camera, hire actors, make a set, film it, pay jeets to VFX it and there you have it, no Sora 2 needed
it's an useless statement, you absolute shit for brains baboon
the whole point of AI is to have less manual work, if Sora 2 can do it without the manual work then it is (even if just for now) better
Anonymous No.107157199 [Report] >>107157556 >>107157642
>>107157098
nvidia gpu is the only thing that really matters. 16gb vram+. 24vram is practically required if you want top quality video gens. minimum 64gb ddr5 ram for offloading model cache if needed. cpu isnt important but you'll want something made within the past 10 years at least.
Anonymous No.107157280 [Report] >>107157311
Question to the anons using Wan2.2 text-to-video (not I2V), which lora are you using?
Anonymous No.107157290 [Report]
>>107155166
crazy workflow, nice
>>107155364
im so lonely bwos
Anonymous No.107157311 [Report] >>107157546
>>107157280
There was this released two days ago if you're talking about lightx2v
https://huggingface.co/lightx2v/Wan2.2-Lightning/tree/main/Wan2.2-T2V-A14B-4steps-lora-rank64-Seko-V2.0
Anonymous No.107157370 [Report] >>107157739
>>107157141
>be proprietarycuck

>you cant train a lora to add a style to the model
>you cant train a lora to add a character or a person to the model
>you cant train a lora to add a concept to the model
>you cant train a lora for anything at all
>you cant finetune the model
>no big company can finetune the model like many companies are doing right now with wan
>you cant have anyone research around the model at all to improve its architecture, find optimization avenues, fix issues, change specific layers, text encoders, vaes, learn how to make better models in the future and advancing the entire ai industry itself etc
>you cant generate gore
>you cant generate pornographic material
>you cant generate anything else someone else would deem "problematic", no matter how mundane it might be
>you cant generate anything they at any point in time say you cant generate in the future when they change their mind overnight
>you cant generate anything at all if their servers are overloaded, not online, or broken
>you cant generate anything without it being logged and all your data harvested and sold
>you cant control dozens of generation parameters that would allow you to have precise control over what you generate, no matter how specific
>you cant write nor test out new generation parameters like new specialized samplers and schedulers
>you cant do anything about it if they decide to lobotomize the model you are using or remove it completely overnight, never being able to truly recreate what you once did and liked
>you cant test out new papers coming out with new technologies like completely changing how an entire portion of inference works, like completely changing how cfg works, completely changing how negative prompting works (https://github.com/hako-mikan/sd-webui-negpip) etc etc
As a proprietarycuck you are paying to be in a limited and spied on cuck cage and you lash out when someone calls out your evil corpo master and your pathetic cuck predicament.
Anonymous No.107157453 [Report] >>107157470 >>107157641
>>107156982
how the fuck are you guys, like pancakechad for example, genning animateinanimate like this? fuck this is so good.
man i know my brain is rotted when i find pancake and syrup women hotter than any e-girl kek
Anonymous No.107157470 [Report] >>107157516
>>107157453
very carefully
Anonymous No.107157516 [Report]
>>107157470
i asked how you gen them, not how you fuck them!

but true.
Anonymous No.107157546 [Report]
>>107157311
No lora I found works well with Holocine (the multiscene fine-tune)
Anonymous No.107157556 [Report] >>107157565 >>107157572 >>107157592
WAN 2.2 anons: just bought a 5070ti and I've been playing around all weekend to get a good workflow for keyframing a longer animation
>Generate ~12 separate 'keyframes' in SD for character LORAs
>Inpaint poses/details - create depth masks to quickly delete background in photoshop to keep character in white void for WAN
>send color 'keyframes' 1 + 2, 2 + 3, to FFLF2V to get a crude timeline of 2-3 second clips (turning, raising, pointing, draining a pint glass, etc. )
>i2v Q_8 gguf in the comfy 'workaround' gets jarring "Flashes" on reaching last frame as it quickly tries to compensate for color degradation, but LORAs are made for i2v.
>Inpaint Q_8 gguf seems to go faster and solves the flashes, seems to take the LORAs but i'm still unsure how well it will work long term.

curious how to proceed here:
>finish all the 2-3 second clips in i2v and try to save it in premiere
>keep playing with the inp. to get it to follow styles so I only need to fix the front half in post or re-gens
>Learn how to use VACE and how to use the last and first 8 frames of each clip to preserve the motion
>Take the entire 24 second video with jank coloring and learn VACE v2v to depth mask the entire thing and regen.


>>107157199
minimum 64gb ddr5 ram for offloading model cache if needed
I have 32 and have been holding off because prices are gay. is it actually super necessary?
Anonymous No.107157565 [Report] >>107157705
>>107157556
vace
Anonymous No.107157572 [Report]
>>107157556
>is it actually super necessary?
No but excessive swap use you get with 32 gigs slow generation down considerably.
Anonymous No.107157584 [Report]
Why are the vue nodes so fucking huge? I want to use them, but this is ridiculous.
Anonymous No.107157592 [Report] >>107157705
>>107157556
It's much faster with 64gb+ ram
Anonymous No.107157641 [Report] >>107158021
>>107157453
prompt for the original one:
>professional 4k high resolution hyperrealistic 3d render by Disney Pixar of a beautiful nude curvy woman slime girl who is made entirely out of maple syrup. Her whole body and face are translucent and seethrough syrup. Her hair is made out of melting butter. She sits cross-legged on top of a huge stack of pancakes. Her body melts onto the pancakes. The pancakes are on a modest porcelain plate in a 50s American diner restaraunt.
>raytracing, beautiful lighting.

standard chroma WF
Anonymous No.107157642 [Report] >>107157707 >>107157712
>>107157199
What does offloading model cache mean and what do you mean by 16gb vram + .24vram?
Anonymous No.107157662 [Report]
Anonymous No.107157685 [Report] >>107157713
Easy Cache, Lazy Cache, Apply First Block Cache, Wan Video Tea Cache, Wan Video Mag Cache, Wan Video Tea Cache Native, Wan Video Easy Cache
Which cope cache node do you use and at what settings?
Anonymous No.107157705 [Report] >>107157881
>>107157592
would 96 make any difference or is that just pointless? the price ladder from 64 is a lot narrower than it used to be due to being a weirder size + slower clocks for XMP

>>107157565
>Vace
what's the point of the 3gb "Module" Vace FUNs at https://huggingface.co/Kijai/WanVideo_comfy_GGUF/tree/main/VACE
versus the large models at https://huggingface.co/QuantStack/Wan2.2-VACE-Fun-A14B-GGUF/tree/main/HighNoise?

Do you load the modules in the same chain as the regular i2v (or inp) model to save on disk space while achieving the same result?
Anonymous No.107157707 [Report]
>>107157642
For example, Wan2.2-I2V-A14B-LowNoise-Q8_0.gguf is 15.4gb. If you only have 16gb of vram on your gpu, that leaves you with 0.6gb of vram. Keep in mind, the text encoder + loras + vae also are stored in the vram. Since all that can't fit on a tiny 16gb card, you can set a specific amount of the model to be swapped to your system ram. IE, 10gb of the wan model off loaded to system ram. This will allow you to gen without running out of memory. Off loading to ram is much, much slower, but it works.

Optionally, you can use a lower quant version of the model, like Wan2.2-I2V-A14B-LowNoise-Q6_K.gguf which is 12gb, but lower quants = lower quality.
Anonymous No.107157712 [Report]
>>107157642
He's saying you should aim for 16gb vram minimum but 24 is preferable. Offloading is when you can't fit the entire model into vram so you use your system ram. wan 2.2 q8 is like 15 gigs(?) for one of the models
Anonymous No.107157713 [Report] >>107157740
>>107157685
For Wan2.2, you don't use any of them.
Anonymous No.107157732 [Report]
>>107156022
imgchad rein eternal
Anonymous No.107157739 [Report]
>>107157370
Anonymous No.107157740 [Report] >>107157947
>>107157713
Is there a reason why?
Anonymous No.107157881 [Report]
>>107157705
>would 96 make any difference or is that just pointless?
Hard to say really. Depends of the motherboard combo I guess.
Anonymous No.107157947 [Report]
>>107157740
Video generation is iterative.
Anonymous No.107157987 [Report] >>107158042 >>107158049 >>107158090 >>107158435
https://civitai.com/models/2114848/2000s-amateur-photography
As requested. Not perfect, but reduces vaginahorror and manfaces.
Anonymous No.107157997 [Report]
I tried Holocine and I could not get the same results as their demo even with 15 seconds lol
I used the same prompt
I obviously had to make some sacrifices like using distillation models with 5bit quants

"b-but local is better than saas, trust me bro!"
"results are shit? It's your fault you are poor and don't own an H100, the pinnacle of LOCAL gpus :^)"
Anonymous No.107158021 [Report]
>>107157641
thanks <3
Anonymous No.107158038 [Report]
>>107154956
this is AI?
Anonymous No.107158042 [Report] >>107158114
>>107157987
That looks like a zoomer idea of what 2000s photography looks like, and some of the photos in the showcase don't look "amateur" at all. At least search for photos that used popular cameras from that time like Sony Cybershot, Olympus, Canon PowerShot etc, or search for old myspace photos or older photos from Flickr.

t. Millennial
Anonymous No.107158049 [Report]
>>107157987
bruh moment, as the kids say. https://civitai.com/models/978314/ultrareal-fine-tune?modelVersionId=1413133
Anonymous No.107158090 [Report] >>107158114
>>107157987
wait regular chroma cant do vageen? wtaf
Anonymous No.107158114 [Report] >>107158162
>>107158042
Dataset is mostly from 2000-2010 era.

>>107158090
It can, but it gets confused.
Anonymous No.107158131 [Report]
>cold weather
>gpu 100% to warm room
Ohh shit it is GOON season
Anonymous No.107158135 [Report]
But for what shall i goon to?
Anonymous No.107158137 [Report]
Anonymous No.107158147 [Report] >>107158193
correct me if im wrong, but is there any reason to make a high noise of a character lora for wan? there's no motion, so what would be the point?
Anonymous No.107158162 [Report] >>107158639
>>107158114
>Dataset is mostly from 2000-2010 era.
I am a Millennial boomer who lived that era and at least the showcase images don't resemble the amateur pics from that era at all
Anonymous No.107158193 [Report] >>107158224
>>107158147
It's less about "motion" strictly but denoising strength.
You might be able to make do if your character looks like a normal human with just low denoising lora. But for something like say Kirby or Sonic, you probably want for both.
Anonymous No.107158212 [Report]
>>107155187
Anonymous No.107158224 [Report] >>107158274
>>107158193
I see, thanks. I've been experimenting with my character lora while using other NSFW loras, and I noticed that using the low rank of some loras forces my character(person) to look like whatever person that lora was trained on. How can I avoid that? Increase the strength of my character's LOW lora? remove the NSFW's low model? I've tried both but haven't found anything solid that works. I can't get rid of the low lora for some NSFW loras because wan needs that data to create for example, a penis or cumshot.

The twerk lora for example, always makes the ass bigger and i don't want that. its so annoying. lowering the strength of the nsfw lora helps but also reduces the motion
Anonymous No.107158244 [Report] >>107158258
n00n0
Anonymous No.107158258 [Report]
>>107158244
nani kore wa yameto my ramenu betta stoppa acting up i'm gonna nækædæshi my ramanu
Anonymous No.107158274 [Report]
>>107158224
>How can I avoid that?
I should note that I never trained a WAN lora, but this seems like a generic lora compatibility issue to me. Try lowering the strength of other lora?
>Increase the strength of my character's LOW lora?
Maybe just a bit if you are desperate.
>remove the NSFW's low model?
Probably not.
>The twerk lora for example, always makes the ass bigger and i don't want that.
This just means the person who trained it, trained on big asses.
Train your on with diverse dataset of asses of all sizes?
Anonymous No.107158280 [Report]
Anonymous No.107158330 [Report] >>107158350 >>107158378
flux/chromosome users, how do you handle your text encoders? do you use specific quants? i'm starting to wonder if my shit gens are a product of what i'm using, but i'm not sure. cumfartui is very confusing as well so that's a variable. the default flux krea workflow is 3 whole seconds slower than an old workflow i was using earlier this year..
Anonymous No.107158346 [Report]
Anonymous No.107158350 [Report] >>107158417
>>107158330
keep t5 at fp16 imo.
Anonymous No.107158362 [Report]
Anonymous No.107158378 [Report] >>107158417
>>107158330
q8 chroma, fp16 clip, 26-35 steps, euler simple/beta
try "aesthetic 1" in negative
Anonymous No.107158385 [Report] >>107158418
it doesn't understand left/right but far/near seem to work
Anonymous No.107158417 [Report]
>>107158350
>>107158378
thanks. i guess i was trying too hard to save on vram by lobotomizing the text models.
Anonymous No.107158418 [Report]
>>107158385
Why do text encoders struggle with directions? That's not an isolated incident.
Quick theory:
Is this because right/left can mean both viewer's right/left and character's right/left, which ends up confusing the UNET during training?
Anonymous No.107158435 [Report] >>107158639
>>107157987
thank you for your hard work
Anonymous No.107158443 [Report]
Anonymous No.107158461 [Report]
Anonymous No.107158475 [Report] >>107158586
Anonymous No.107158532 [Report]
Anonymous No.107158542 [Report] >>107158586
Anonymous No.107158554 [Report] >>107158637
So sounds like it’s worth going down a generation to the 4x cards if I want 24gb vram at a more reasonable cost
Anonymous No.107158586 [Report]
>>107158475
>>107158542
neat
Anonymous No.107158587 [Report]
Anonymous No.107158595 [Report]
Anonymous No.107158607 [Report] >>107158665
i literally gooned for 12 hours today
Anonymous No.107158637 [Report]
>>107158554
so a 4090 then? aren't they like 1500 dollars
Anonymous No.107158639 [Report]
>>107158162
I guess I could rename it, fair point

>>107158435
npnp
Anonymous No.107158665 [Report] >>107158723
>>107158607
Can you catbox your picrel or a similar gen?
Anonymous No.107158683 [Report]
Anonymous No.107158711 [Report]
Anonymous No.107158723 [Report] >>107158739
>>107158665
https://files.catbox.moe/0andv6.png
Anonymous No.107158726 [Report] >>107158784
Anonymous No.107158739 [Report]
>>107158723
Thanks!
Anonymous No.107158781 [Report] >>107158784
Anonymous No.107158784 [Report]
>>107158781
>>107158726
>no large breasts, wide hips
Anonymous No.107158829 [Report]
>he doesn't (large breasts, wide hips, thick thighs:1.5)
Anonymous No.107158850 [Report]