← Home ← Back to /g/

Thread 106206599

317 posts 142 images /g/
Anonymous No.106206599 >>106207104 >>106208096 >>106209600
/ldg/ - Local Diffusion General
Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>106201767

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://github.com/Wan-Video
2.1: https://rentry.org/wan21kjguide
2.2: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-HD/tree/main
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
Anonymous No.106206613
mfw
Anonymous No.106206617
First for Chroma kino
Anonymous No.106206633 >>106206646 >>106206647 >>106206659 >>106206674
is there a recommended workflow for chroma?
maybe something shared on civitai?
Anonymous No.106206642
Anonymous No.106206646 >>106206667 >>106209991
>>106206633
latent image, sampler, encoders, padding remover, save image node is all you need
Anonymous No.106206647 >>106206667
>>106206633
Like with all things. The default workflow is best.
Anonymous No.106206652
Can I not update Comfy if I installed sage attention?
Anonymous No.106206659
>>106206633
If you have to ask you shouldn't be doing this
Anonymous No.106206667
>>106206646
>>106206647
OK so nothing special then, got it
Anonymous No.106206674
>>106206633

Legitimately just use the default.
Anonymous No.106206689 >>106206754 >>106206761 >>106206852 >>106206861 >>106206866
>flux
>krea
>hidream
>chroma
>qwen
what are the strengths and weaknesses of each model when it comes to photoreal gens?
Anonymous No.106206692
Anonymous No.106206698
Anonymous No.106206714 >>106207104
is wan 2.2 actually better than 2.1?
Anonymous No.106206747 >>106206753
I can't get qwen to do distorted vhs style at all.
Anonymous No.106206753
>>106206747
kek
Anonymous No.106206754 >>106207930
>>106206689
only krea and chroma are good for legitimate photorealism. the rest have a fake AI look.
Anonymous No.106206761 >>106206777 >>106206810
>>106206689
Flux - censored slop, same poses, fake skin
Pro: svdquant i guess?
Krea - lack of NSFW, lack of flexibility, no feet
Pro: better than flux
Hidream- same as flux but even worse, bigger model
Chroma- best overall, best flexibility, variety of styles, photoreal look, nsfw prompting
Con: no svdquant yet
Qwen- chronic same pose unless you change prompt, lack of nsfw, fake Seedream 3.0 based skin
Pro: too big to have any over chroma
Anonymous No.106206777 >>106207245
>>106206761
The real con of Chroma is the SD1.5 quality output.
Anonymous No.106206787 >>106206807
Really disappointed with v49 and v50 dropping the ball so hard. Hoping that people gravitate to v48 for fine tuning.
Anonymous No.106206795
>Training & Ganging
Anonymous No.106206807 >>106206816 >>106206821
>>106206787
>v49 and v50 dropping the ball so hard.
Nah they were fine, just needed a higher res to gen.
Anonymous No.106206810
>>106206761
what about camera angles? I've been using flux but it's really hard to change the viewpoint when the prompt got really complicated. it was way easier back in the sd 1.5 times
Anonymous No.106206816
>>106206807
And if you try them at 1152x1152 they are legit better than previous versions. Like the difference is tangible.
Anonymous No.106206818 >>106206823 >>106206829
So many new models, and they are all shit when it comes to weeb stuff
There is not even an once of hope for anything better than Noob/Illus this year, isn't it.
Anonymous No.106206821 >>106206909 >>106206927
>>106206807

Which is needless since its easiest to gen small then upscale. 49 and 50 were supposed to be high res epochs but it couldn't have been done in a stupider way.
Anonymous No.106206823 >>106206851
>>106206818
luminaluminalumina
Anonymous No.106206829
>>106206818
Chinks are the only who can save us, and they are unfortunately incompetent so there's very little hope. Maybe in 3 years if we get lucky.
Anonymous No.106206851 >>106206955 >>106207100
>>106206127
>>106206098
>>106204935
>>106206020
I think I got somewhat close with prompting.

(by tsukumizu yuu:0.9),(by kaamin \(mariarose753\), by fjsmu:0.8),chibi, (painterly:1.3), thick lineart, low poly, abstract, (impressionism:1.2),cubism, (acrylic paint \(medium\), brushstrokes:1.2), absurdres, (colorful:0.6), shiny skin,shiny clothes,
1girl,white dress, big eyes, (no pupils,jitome:0.9), (big head:0.7), sundress, straw hat, holding cup, drinking straw, drinking straw in mouth, walking, looking at viewer, happy,
from side, upper body,
beach, ocean, wind, (cumulonimbus cloud:0.8)

>>106206823
lumina needs more training, the artist names aren't fully baked in and it can't handle tag-only prompts as well as illustrious or noob
Anonymous No.106206852
>>106206689
>Flux
Needs a LoRA or a finetune to get photorealistic images. Cons are that the models are distilled crippleware and have shitty licenses (minus Schnell) so there's only so much you can really do with them.
>Krea
Just a Flux finetune. I've heard mixed things, but try it and see what happens.
>HiDream
Flux except it never got any sort of ecosystem developed around it. Skip it.
>Chroma
Probably does the best photorealistic images out of any of these, but it can be inconsistent with small details like hands. Some anons in the past few threads were saying that setting to the latent res to 1152 instead of 1024 helps, but I haven't tested that myself. It just came out so there's no finetunes and very few LoRAs yet.
>Qwen
Just came out like Chroma. Fuckhueg and has better prompt adherence than anything else but has the plastic-y look like base Flux does. People are already starting to train LoRAs for it so maybe soon it'll be good for photorealism but I'd pass right now.
Anonymous No.106206861 >>106206874
>>106206689
WAN and Flux Krea are the most starkly actually-realistic-looking by default, by an enormous margin
that's about it
Anonymous No.106206866
>>106206689
>>106203690
Anonymous No.106206868 >>106207272
>>106204963
>as;inb4
Anonymous No.106206874 >>106206913
>>106206861
Nah, they are biased towards cinematic look. Give me a RAW image.
Anonymous No.106206909 >>106206979
>>106206821
It wasn't done like this in purpose. It's just how these models work. Though I have no idea why it works so well on Chroma but not Flux. I'll try it out on Qwen now.
Anonymous No.106206913 >>106206954 >>106206955
>>106206874
>Give me a RAW image.
here's a Krea gen I posted here a few days ago. It can do any kind of photography quite easily, professional or amateur, just have to prompt for it.
Anonymous No.106206927
>>106206821
well no, I expect to gen at least at SDXL equivalent resolutions by default 100% of the time and upscale PAST that, not generate at SD 1.5 resolutions and upscale to SDXL equivalent ones.
Anonymous No.106206954 >>106207842
>>106206913
There is bokeh in the image though. And if you look very closely at her skin texture, it looks plastic (this is probably not a Krea base issue, but the nature of a distilled model).
Anonymous No.106206955 >>106206965 >>106207034 >>106207851
>>106206851
found some cool artist tags while cooking this up too

>>106206913
that is blatantly AI, still a nice gen though.
Anonymous No.106206959 >>106206993
Anonymous No.106206965 >>106206993
>>106206955
Very nice
Anonymous No.106206979
>>106206909
>I'll try it out on Qwen now.

Also defaulting to slopped output at higher res
Anonymous No.106206993 >>106207029 >>106207034 >>106207241
>>106206965
ty

>>106206959
qwen, right? it really can't do different photo medium styles, huh. it also has a bad tendency to insert parts of the prompt into the picture as text. chroma is way better about only inserting the things in quotes.
Anonymous No.106207020
hidream is such a useless meme model
Anonymous No.106207029 >>106207100 >>106207131
>>106206993
yes
but chroma on the other hand fucks up small details
Anonymous No.106207034 >>106207100
>>106206955
>>106206993
What artists if I may ask?
Anonymous No.106207046
ITS SO FUCKING OVER DEBIAN 13 SLOWS DOWN RTX 3060 ITS FUCKING OVER ITS OVER OVER OVER OVER
I LOST ALL MY FREE LUNCH ITS ALL OVER THE FLOOR
>he pulled
he pulled
>he pulled
he pulled
>>106205401
nice, thanks for sharing, for me time per step is 34s but total time is 150 because of other things like vae etc..
Anonymous No.106207050 >>106207070
Are the Flash versions of Chroma worth using?
Anonymous No.106207056 >>106207100
qwen's optimization is dogshit. my pc freezes even at regular 512x512. i used qwen gguf q3. meanwhile, i can use 1280x1280 on chroma...
Anonymous No.106207070
>>106207050

They make the gens stiffer in my experience, but that might not matter depending on the prompt.
Anonymous No.106207100
>>106207029
I'm considering setting up a chroma or even SDXL i2i workflow just for de-slopping qwen outputs. people have done for flux.

>>106207034
see >>106206851
and
https://danbooru.donmai.us/related_tag?commit=Search&search%5Bcategory%5D=Artist&search%5Border%5D=Frequency&search%5Bquery%5D=impressionism

>>106207056
the GGUFs are fucked or something. fp8 scaled qwen works for me, but the GGUFs nearly crashed my computer
γƒγ‚Ήγƒˆγ‚«γƒΌγƒ‰ !!FH+LSJVkIY9 No.106207104 >>106207114 >>106207150 >>106207221
>>106206599 (OP)
smell ya later <3
never returning to this faggy board ;3
>>106206714
not if you care about loras
Anonymous No.106207106 >>106207156
I'm seeing illustrious XL loras and checkpoints but civtai hasn't made them searchable
is IXL legit or just a meme? [my first attempt to test it looks like shit, pic not related]
Anonymous No.106207114
>>106207104
see you tomorrow
Anonymous No.106207131
>>106207029
There is fucking up small details and there is blurring the background to hide imperfections. I don't think it's fucking up the details per se, just showing you what other models will hide because of aesthetic reasons.
Anonymous No.106207147
Anonymous No.106207150
>>106207104
>never returning to this faggy board ;3
big if true
Anonymous No.106207156
>>106207106
Noob is so good you really don't need LoRAs.
>[my first attempt to test it looks like shit, pic not related]
Requires wrangling.
Anonymous No.106207162 >>106207235
so it's qwen first pass and chroma for 2nd pass refinement now?
Anonymous No.106207209
>Decide to try the "automagic" optimizer in diffusion-pipe.
>It's shit.

Thanks, just wasted 3 hours.
Anonymous No.106207212 >>106207337 >>106207721
Do wan2.2 loras work with gguf variants? I'm trying to use a lora (downloaded both high and low) but produces blurry and poor results
Anonymous No.106207221 >>106207327
>>106207104
damn, another great poster ran off by the resident schizo
Anonymous No.106207225 >>106207328
I like Chroma when it works but I still keep having the issue where it switches between styles no matter the prompt
Anonymous No.106207235
>>106207162
As an i2i refiner it makes no sense. You'd get superior results just taking the Qwen outputs and feeding them to a VLM, then doing raw txt2img on Chroma, if all you care about is the Qwen aesthetic. This will likely teach you how to get the quality of images you like with Chroma so you can eventually drop Qwen altogether.
Anonymous No.106207241
>>106206993
qwen's problem is generalizing the prompt, i.e. if you mention "in a bedroom", your character will end up on a bed, if you say "a bra lies on the floor", the character will also wear one.
Anonymous No.106207245 >>106207381
>>106206777
Skill issue
Anonymous No.106207272 >>106207329
>>106206868
Never mind, I somehow fixed it, seems like increasing the repeats helps it understand styles much better, unlike with characters where it causes it to overfit, in short
>styles: lots of repeats
>chars: 1 is usually more than enough.
Anonymous No.106207327
>>106207221
Anonymous No.106207328 >>106207394
>>106207225

Every few gens just seem to become "deviant art slurry" regardless of the prompt. v50 did nothing to fix this either. When it actually works its wonderful though.
Anonymous No.106207329
>>106207272
Anonymous No.106207337 >>106207347 >>106207508
>>106207212
i'm gguf user. what's the lora? will check it
Anonymous No.106207347
>>106207337
https://civitai.com/models/1307155/wan-22-experimental-wan-general-nsfw-model?modelVersionId=2073605
Anonymous No.106207355 >>106207364
Are there any older epochs of Chroma worth trying?
Anonymous No.106207364 >>106207377
>>106207355
v38 chroma nunchaku cfg 4.5 1024px non-detail calibrated or detail calibrated
hf.co => chroma-nunchaku-test
Anonymous No.106207372
Anonymous No.106207377 >>106207390
>>106207364

ty friendo. The newer epochs just look too much like flux desu.
Anonymous No.106207381
>>106207245
That's not very nice! I'm sure the Chroma spammer is trying his best!
Anonymous No.106207390 >>106207404 >>106207748
>>106207377
its also waaaaay faster with nunchaku especially on a vram limited card like 3060
i think its 2minutes for 45 steps 1 minute for 25 steps but im not sure i fucking forgot ill post some gen speeds later in fact im gonna queue a few gens right NOW
Anonymous No.106207394 >>106207401
>>106207328
I want to try training a realism LoRA so that I don't have to keep fighting the model but I don't know which epoch I would base it off of.
Anonymous No.106207400 >>106207436
do I need to prompt in chinese to unlock qwen's full potential?
Anonymous No.106207401
>>106207394

I'd just stick with v48 for now
Anonymous No.106207404 >>106207455 >>106207479
>>106207390
It's 2 minutes for 45 steps with Nunchaku or without it?
Anonymous No.106207436 >>106207492
>>106207400
No, you need to train a lora. Some redditor did it:

https://www.reddit.com/r/StableDiffusion/comments/1mjys5b/18_qwenimage_realism_lora_samples_first_attempt/
Anonymous No.106207445
Anonymous No.106207446
Anonymous No.106207455
>>106207404
with nunchaku
Anonymous No.106207479 >>106207560
>>106207404
170w
Anonymous No.106207492
>>106207436
Quite interesting. It will be interesting to see which performs better out of Qwen LoRA and Chroma HD, though I suspect Qwen LoRA is not uncensored and also can't do a bunch of different realism styles since it's a LoRA so Chroma will win by a small margin. Is the LoRA out so we can A/B test it?
Anonymous No.106207494
Can you use clip-l with Chroma?
Anonymous No.106207508
>>106207337
did u check it yet?
Anonymous No.106207551 >>106207569
v50 isn't so bad. once I loosened up on the tags and tried some different ideas, it started nailing gen after gen.
I just mean that it's not as much as a steaming turd as it felt like at first, not that I love it or anything, baka.
Anonymous No.106207557
>>106207
works fine for me. check if you have same issues, without this lora
Anonymous No.106207559
Anonymous No.106207560
>>106207479
cool gen
Anonymous No.106207569
>>106207551
>t. no gen
Anonymous No.106207576 >>106207590 >>106207594
Anyone got a WAN2.2 video gen workflow or tutorial that'll run on my TOTALLY ADEQUATE setup?
My specs are:
A 3060 12gb VRAM that is PERFECTLY FINE okay?? It's literally above average (I looked it up) and 32GB RAM that is is PLENTY.

And before you size queens start posting your 4090 screenshots, I'll have you know it's not about the VRAM size! My card gets the job done when it counts!
Anonymous No.106207590 >>106207625 >>106209159
>>106207576
https://files.catbox.moe/kq9ju1.json
it's the one from the rentry guide
Anonymous No.106207594 >>106207625
>>106207576
i have a workflow that works on 3060 12gb + 64gb ram
if mine doesnt work grab ggufs and use comfy native with --reserve-vram 1 and --use-sage-attention and make sure to install linux, and sage attention
150-160s per 5s 640x480 gen btw
https://litter.catbox.moe/kum8tt4727pbcjrb.json
you might have to use a lower gguf quant too rip
Anonymous No.106207624
Anonymous No.106207625
>>106207590
>>106207594
T-thanks anon... I knew my 12GB could handle it! You're real bros for not judging my completely normal and adequate VRAM amount, you don't NEED a massive 24GB card for everything!

My setup is going to run this workflow PERFECTLY. Probably even better than some anons with their oversized, power hungry cards that are just compensating desu.

I'll post results later to prove that 12GB is the IDEAL size for optimal performance.
Anonymous No.106207645 >>106207970
gib me the most photoreal gen
do your best and tell me which model are u using
Anonymous No.106207659 >>106207822
Anonymous No.106207721
>>106207212
BUUUMMPPP
Anonymous No.106207732
Anonymous No.106207748
>>106207390
12GB IT' S AVERAGE, THE MOST COMMON SIZE!
Anonymous No.106207768 >>106208064
what is ModelSamplingAuraFlow and can it be used with chroma?
Anonymous No.106207783 >>106207791 >>106207814 >>106207964
Please God somebody finetune Chroma to fix the hands and style prompting.
Anonymous No.106207791 >>106207801
>>106207783
Tens of thousands of dollars and months of training couldn't fix it. What hope does some rando have?
Anonymous No.106207801
>>106207791
I believe in miracles.
Anonymous No.106207807
wan seriously needs a proper "fire and forget" nsfw finetune or general lora
Anonymous No.106207814 >>106207827
>>106207783
With that money, I buy a better GPU and use a serious model like WAN
Anonymous No.106207822 >>106207841 >>106208555
new lightx2v i2v is even worse with motion than the 2.1 lora
both fail to match no lora >>106207659
Anonymous No.106207827 >>106207855
>>106207814
Until Wan is uncensored I couldn't give a fuck about using it for T2I.
Anonymous No.106207841 >>106207866
>>106207822
how many steps did you use with lightx2v? maybe try to get to a few more refiner steps
my main issue is that it changes faces too much, so I traded the 3 minutes gens to 20 minutes ones
Anonymous No.106207842 >>106208112
>>106206954
nobody mentioned bokeh until you brought it up lmao, thanks for confirming you have a retarded understanding of how photography actually works in real life though. I presume you're the sort of person who thinks it actually makes sense for prompts that state "shot on an iPhone" to be excessively low quality looking and full of JPEG artifacts. Also no, her skin doesn't look plastic by any reasonable metric, fuck off.
Anonymous No.106207851
>>106206955
the point is Chromafaggots will insist it can do better but it fucking cannot unless they intentionally make the output look like unrealistically low resolution blurry dogshit that would never be produced in reality by any remotely modern camera. And I say this as someone who likes Chroma.
Anonymous No.106207855 >>106207918
>>106207827
Kind of sad that the only properly uncensored natural language model is chroma.
Wan is censored unless you use 12 different loras all adding artifacts.
Qwen is censored.
Anonymous No.106207860 >>106207883
incels talking about photorealism is like them talking about sex
Anonymous No.106207866 >>106207876 >>106208250
>>106207841
4 high + 4 low
for comparison this is the 2.1 lora with the same settings, much more motion
Anonymous No.106207876
>>106207866
can you share your workflow with lightx2v 2.2? I'm curious on why it looks so bad
Anonymous No.106207883
>>106207860
literally a significant portion of the entire T2I community seems to believe that "realism" specifically means "the image itself is poor quality and lacking in fidelity". Anything that doesn't look like a fucking 2005 digital camera output re-saved at JPEG quality level 70 at least 25 times isn't "realistic", to here these people tell it.
Anonymous No.106207910 >>106207921 >>106208152 >>106208257
Anonymous No.106207913 >>106207937 >>106208031
Chroma was a cool experiment and hopefully someone will tune it into something good but I don't think I'm going to wind up using it if it keeps playing fucking style roulette with my gens. I could deal with literally everything else--bad hands, bad feet, whatever--but I hate this stupid shit where the model doesn't listen to the most basic prompts I can give it a good half of the time. It's frustrating beyond belief.
Anonymous No.106207918 >>106207929
>>106207855
Chroma is a fucking finetune of architecturally modified Flux Schnell (which was more "muh censored" than Qwen, by default), it's not a unique architecture.
Anonymous No.106207921 >>106207967 >>106208149
>>106207910
GYATT DAYMN WHAT MODELI SHTAT
Anonymous No.106207929 >>106207941
>>106207918
Sure.
Anonymous No.106207930
>>106206754
Anonymous No.106207937 >>106207971
>>106207913
if BigASP guy gets his hands on it it might be good. Asp 2.5 is fucking crazy in terms of what it gets out of the SDXL architecture, it's definitely the deepest model concept wise there is. Somewhat hard to wrangle but it can look REALLY fucking good if you use Clownshark stuff sampling wise, thanks to the Flow technique he trained it with.
Anonymous No.106207941 >>106207955
>>106207929
??? are you saying you don't believe me / think Chroma is something other than that?
Anonymous No.106207955
>>106207941
No, I agree, but that's not exactly refuting what I wrote before.
Anonymous No.106207964 >>106207982 >>106209052
>>106207783
You mean an anime finetune? The cap imo is not Chroma, it is Flux base. Though I hope I'm wrong. We'll see if a team like Pony decides to tune it, though I wouldn't be surprised if they just go with tuning Qwen or Lumina.
Anonymous No.106207967 >>106207975 >>106208149
>>106207921
>https://civitai.com/models/1015498/3d-art-style-noobai-xl-epsilon-v3-update
NOOB VPRED BRO, YOU HEAR OF IT? CUTTING FUCKING EDGE BRO
Anonymous No.106207970 >>106207986 >>106208224
>>106207645
My Chinese Overlords gave me the gift of Mouse Girl screaming in agony at all the super gay ghosts today, and at the Ghost of Anime Girl's Gay Past to leave her body, that's about as real as it gets for me.
Anonymous No.106207971 >>106208030
>>106207937
I think he said on Reddit that he was going to try and uncensor Wan for T2I, and if that didn't work out then v3 would be Chroma. Honestly if he can manage to unfuck Wan's censorship and I never have to touch Chroma again i would be fine with that.
Anonymous No.106207975 >>106208049 >>106208149
>>106207967
>vpred
>lora for eps
erm...?
Thank You anon for posting the lora, I love You<3
Anonymous No.106207982 >>106208129
>>106207964
No, I want something that doesn't give me 2.5D slopped gens 3/5 of the time. 2/5 of the time I get magical gens but the rest of them suck.
Anonymous No.106207986
>>106207970
geg, maybe you should use the anime fix lora
its for 2.1 doebeit
Anonymous No.106207995 >>106208043
Was excited for v50 but it was obviously impossible for the model to make major corrections in just a few weeks. I'll keep v48 around since nothing lets me generate stuff quite as degenerate as it does until a better fine tune is made.
Anonymous No.106208030
>>106207971
what does censorship even mean? Do people believe that these base models were at one time capable of complex multi-person NSFW, that was then DPOed out? If they do they're retarded and delusional, that kind of thing takes specific, intentional, concentrated effort to train for. AT MOST a base model might have at one time been also capable of below-the-waist nudity on top of above-the-waist nudity. But it's understandable even why they'd nuket that, it's hard to get right, you don't want to be the base model that gets known for accidentally putting pussies on dudes or dicks on ladies (which is a VERY common problem in e.g. SDXL finetunes for example).
Anonymous No.106208031
>>106207913
Chroma is bad anon, is only his dev samefaging because weekend and he has nothing better to do
Anonymous No.106208043 >>106208069
>>106207995
wouldn't there be an official announcement / CivitAI release if it was actually "done", though?
Anonymous No.106208046
Do you guys know if there's a workflow/resources for ComfyUI that would allow me to upload a GLB armature and generate 3D animations using my skeleton?
I've seen paid platforms that can do it and I've seen papers so I know that the concept is there, but I can't seem to find it anywhere.
Anonymous No.106208047 >>106208101
DELETE CHROMA FROM RENTRY
Anonymous No.106208049 >>106208113 >>106208152
>>106207975
don't follow the rules bro, carve your own path
Anonymous No.106208064 >>106208100
>>106207768
ok so you can add even more details for chroma with ModelSamplingAuraFlow
Anonymous No.106208069 >>106208244
>>106208043
People on Reddit said that there were more versions coming but I their source was "I heard it on Discord" so who knows. Why lodestone and co. give more info about the model over Discord than on its fucking Huggingface page is beyond me.
Anonymous No.106208094 >>106208101
Delete Chroma from OP in the next bake or I will remove it myself!
Anonymous No.106208096
>>106206599 (OP)
what happened to automatic1111 - is it ded?

dont tell me about other interfaces, i got/use most of them
Anonymous No.106208100 >>106208110 >>106208133
>>106208064
wtf are you talking about
that's a STOCK Comfy node for Auraflow-adjacent DiT models, which Chroma, based on Flux, isn't
anyone claiming this is useful or beneficial in any way is a retard who should rope
you'd either use, you know, ModelSamplingFlux, because Chroma is, you know, Flux
or for something actually third party and a bit better typically, ModelSamplingExponential from the RES4LYF pack, in "Exponential" mode
Anonymous No.106208101
>>106208047
>>106208094
Holy shit shut up you fucking sped. You're more annoying than the faggots who have been sucking Lodestone's cock.
Anonymous No.106208110
>>106208100
*ModelSamplingAdvanced, I meant, not ModelSamplingExponential, that's just the mode name
Anonymous No.106208112 >>106208135
>>106207842
We get it anon. All you can prompt for is slop, and you don't understand what a real photograph looks like.
Anonymous No.106208113 >>106208196
>>106208049
https://civitai.com/models/1232852?modelVersionId=1800674
have yo utried this one? its made specifically for vpred
Anonymous No.106208115 >>106208122
Anonymous No.106208122
>>106208115
posted it again reward
Anonymous No.106208123
I finally finally have an Nvidia card
My suffering is over
Anonymous No.106208129
>>106207982
Have you tried training a LoRA?
Anonymous No.106208130 >>106208190
ITS FUCKING HAPPENING https://huggingface.co/fancyfeast/bigaspv2-5/discussions/1#6883efb9ec640add18b2f156
Anonymous No.106208133 >>106208142
>>106208100
>muh that's not the intended way to use it!
I know
Anonymous No.106208135 >>106208194
>>106208112
exactly the kind of response a retard who inexplicably thinks that early-to-mid-2000s digital camera outputs are the only thing that was ever realistic in human history would say
Anonymous No.106208142 >>106208168
>>106208133
gib actul proof this is useful then or GTFO
Anonymous No.106208149 >>106208196 >>106208257
>>106207921
>>106207967
>>106207975
>https://litter.catbox.moe/jjbtr658jmpcqk4e.safetensors
super drunk and fucked up, that aint' the lora. This is it, I guess it was deleted. actual vpred trained, neat lora overall.
Anonymous No.106208152 >>106208181
>>106208049
>>106207910
if you're using noob vpred you don't even really need a 3d style lora, you can just use style and artist tags
Anonymous No.106208168
>>106208142
I won't.
also, I just learned you can use flux redux with chroma too. this makes the model even more useful
Anonymous No.106208181 >>106208210
>>106208152
let me live my dreams bro, my lora doing work. aside from fugtrup what artists you use? only 3d artist I found that does composition+style, but I don't know the porn 3d artists.
Anonymous No.106208190
>>106208130
if he does WAN the resulting model gets like ~200 downloads on CivitAI ever, I guarantee it, WANNiggers overestimate how popular and runnable in a practical context it actually is for straight T2
Anonymous No.106208193 >>106208201
>being poor
not my problem
Anonymous No.106208194 >>106208207
>>106208135
The slopped output doesn't even look like a real pic with a filter.
Anonymous No.106208195
come on... converge you whore
Anonymous No.106208196 >>106208215
>>106208113
>>106208149
same lora, i'm dumb
Anonymous No.106208201
>>106208193
ASP 2.5 being a Flow-matching SDXL model was already too much for illiterate CivitAI jeets to comprehend, why should AspGuy spend even more money training something even less people will actually ever use
Anonymous No.106208207
>>106208194
ok there bud
Anonymous No.106208210 >>106208218
>>106208181
fugtrup can be a good tag to use, but I just use
>(3d \(medium\):1.5), (realistic:1.5), by [insert artist here], no lineart,
>3d background, blurry background

3d loras and detail booster loras can def be useful too, just saying these models have a lot of power and you can do stuff without a lora
Anonymous No.106208215
>>106208196
3DStyle Locon NoobAI VPRED 1224 V4 RealisticCartoon3D dec0.05 tagdrop01 Fi is what the metadata says, i appreciate you posting it whether or not it's the same
thank you anon
Anonymous No.106208218
>>106208210
Hot
Anonymous No.106208224 >>106209851
>>106207970
Excuse me but who the fuck are you again?
Anonymous No.106208244 >>106208310
>>106208069
there is barely any information on the discord either.
>here is the model, its now done
how?
>cut corner, write later
>k bye off to make a lora for the model
Anonymous No.106208250
>>106207866
but look at adidas logo and text
Anonymous No.106208257 >>106208266
>>106207910
>>106208149
Will you share box this time?
Anonymous No.106208266 >>106208273 >>106208274 >>106208299
>>106208257
i refuse to believe that he is the R poster
Anonymous No.106208273 >>106208290
>>106208266
as a regular, I am hurt. very hurt, by this comment
Anonymous No.106208274 >>106208290
>>106208266
wut
Anonymous No.106208290 >>106208306 >>106208322
>>106208273
why are you hurt, i am saying you are not the r poster
im sorry if i made you hurt, come here for a hug <3
>>106208274
you implied he is the r poster because r poster complained about getting reported because >muh catbox
Anonymous No.106208292
Anonymous No.106208297
the 3d disney style lora is insane though
Anonymous No.106208299 >>106208332
>>106208266
R guy would share, he's not like that.
Anonymous No.106208306
>>106208290
wut x2
take your meds
Anonymous No.106208310 >>106208378
>>106208244
If the model is actually done then fuck that. It still doesn't understand hands and gives you whatever style it feels like regardless of what you prompt, he fucked up somewhere.
Did he say what he was training a lora for?
Anonymous No.106208313 >>106208320
Miss R guy btw.
Anonymous No.106208320 >>106208332
>>106208313
Anonymous No.106208322
>>106208290
oh, I misunderstood. Thanks hommie, good looking out. smooch and cheers
Anonymous No.106208328
Very annoying that SAM2 is so bad at segmenting breasts. The puritanism is so bad that it would even interfere with actual automatic censorship.
Anonymous No.106208332
>>106208320
>>106208299
Anonymous No.106208345 >>106208350 >>106208358 >>106208379 >>106208403
What's the best model I can use to generate photorealistic images of people? I'm trying to make some portraits for a tabletop game. GPT and OpenAI are hamstrung.

I have ComfyUI working.
Anonymous No.106208350
>>106208345
sdxl
Anonymous No.106208358
>>106208345
unironically SDXL
Anonymous No.106208378 >>106208424
>>106208310
reinforcement lora (RL) for the model whatever that means
Anonymous No.106208379 >>106208455
>>106208345
Download a bunch of SDXL checkpoints. Run them with dmd2. Just reroll until you get something good.
Anonymous No.106208403
>>106208345
SDXL and Forge or ReForge, forget Comfy for now.
Anonymous No.106208424
>>106208378

This whole endeavor feels like it's being run by teenagers.
Anonymous No.106208455
>>106208379
actually I think I got that wrong, RL stands for reinforcement learning. is that a lora or a model or what? I have no idea.
Anonymous No.106208504 >>106208663
Anonymous No.106208510 >>106208556
1152

This is code for cope.
Anonymous No.106208555 >>106208608
>>106207822
have the old 2.1 light2xv at strength 3 linked to the 2.2 high at strength 1

then have the old 2.1 light2xv at strength 0.25 linked to 2.2 low at strength 1

makes it better IMO
Anonymous No.106208556 >>106208603
>>106208510
Well prior to this version it was noticeably noisier and less stable for realism gens
Anonymous No.106208603 >>106208695
>>106208556

What epoch?
Anonymous No.106208608
>>106208555
is it really better than just using the 2.1 loras alone?
Anonymous No.106208655
i miss R anon
Anonymous No.106208663 >>106208754
>>106208504
>realism is when the textures and details are completely fucking fried
I can't be the only one who hates this look, right?
Chroma was a mistake.
Anonymous No.106208695 >>106208772 >>106208802 >>106208848
>>106208603
v50. When you prompt at higher res you see less noisy results than say v28, v29 etc... up to v48 on same seeds. Because you are starting at a higher res than 512 the gens just look crisper. Also as I show here >>106198671 it is better at prompt following and multiple subject coherence as well. It also one shots hands and feet more often. Pic rel is just one example, but you can also see it in all of my other 1152 images as well, the images have a lot more detail and less noise than before. Places where it's very obvious now are night time photos, or pool pic which I posted previously had a lot of noise before and now has much less.
Anonymous No.106208754
>>106208663
Not always anon. Most of that looks coherent except for wires, which are understandable weak point. And I'll take that over a blurred background and CGI cinematic closeup any time of the day.
Anonymous No.106208772
>>106208695
>1152
both are 832x1488*, but you get the point
Anonymous No.106208802
>>106208695
just accept it dude, 50 is fucked, stop making excuses
Anonymous No.106208815 >>106208884
Anonymous No.106208848 >>106209052
>>106208695

I'm glad that output worked for you, but I'm just not seeing it on my end. Everything looks more "flux-y" in v50 with rubber skin on everything. Non-realism styles all look blurrier and smoothed over as a side effect. Plus, the model is already slow. I don't want to have to render at a minimum of 1152x1152 to experiment with results worse than v48 at even 512x512.
Anonymous No.106208884
>>106208815
it's time to move on anon...
Anonymous No.106208907 >>106208926 >>106208950
So the general consensus seems to be that v50 is fucked but what exactly did lodestone do in 2 epochs that fried it? Anyone have any guesses?
Anonymous No.106208926 >>106208950
>>106208907

He crammed all resolutions over 1024x1024 into just those two epochs. Not sure what the logic was behind that, but it fucked something up.
Anonymous No.106208950 >>106208977 >>106208993
>>106208907
>>106208926
v50 is just a merge, v49 was the last trained epoch, with resolution bumped to 1024
Anonymous No.106208966
>Gemini 2.5 Pro on Google AI Studio, with a very long and specific preamble that simultaneously jailbreaks
Share pls?
Anonymous No.106208977 >>106209004
>>106208950

I thought annealed was the merge? What even is it then?
Anonymous No.106208993 >>106209014
>>106208950
Did he run out of money or something? First he cut the dataset down from 5m to 1m highres images, now this is the second time I've read that v50 wasn't even trained.
Anonymous No.106209004
>>106208977
It's an optimization of some sort, but even lodestone said it was crap and only released it because someone on the discord said they got better results with it
Anonymous No.106209014
>>106208993
Yes. He said it has cost $150k so far and the money was quickly running out
Anonymous No.106209031 >>106209044 >>106209169
The absolute fucking state of local
Anonymous No.106209044 >>106209048
>>106209031
Go back to /sdg/ and shill SAAS faggot
Anonymous No.106209048 >>106209059
>>106209044
I'm sorry the model didn't work out Lodestone
Anonymous No.106209052 >>106209063 >>106209097 >>106209107
>>106208848
It's not just that one though. I'm not seeing a certain level of graininess anymore.

Stuff like https://desu-usergeneratedcontent.xyz/g/image/1751/31/1751319394305.png
is no longer possible. Now I get this crisp looking at a minimum
>>106204178
(note those are different seeds, but you get the point).

The model is more detailed and looks much better.

As for the pool pic, maybe I should post example of a particular type of noisy image I was getting before so you understand, and it was very strong with this prompt.

https://desuarchive.org/g/thread/106145131/#q106146035

Now, here is not necessarily the same seed, but same prompt in action.

>>106199357

A direct upgrade. Certain types of incoherent or broken images are no longer possible.

I'll go back to a previous image I exclusively one shotted with v50 to show you I'm not just full of crap.

Let's do >>106207964

As you can see it has that weird noise in v48 (pic rel). I've never seen anything like it for v50. I saw that quite often while prompting those older versions.
Anonymous No.106209059 >>106209071
>>106209048
It's a great base model, the overall best for realism and NSFW

Wan is local, Qwen is local, Chroma is local

Two of these are trained with big tech money, one of them is trained with community money, as in a fraction of the cost

Take your SAAS shilling somewhere else
Anonymous No.106209063 >>106209172 >>106209232
>>106209052
What sampler/scheduler/steps/cfg are you using and how are you prompting for realism?
Anonymous No.106209071 >>106209084 >>106209103 >>106209112
>>106209059
>the overall best for realism and NSFW
SDXL does hands, feet, and NSFW better and does it in a fraction of the time that Chroma does.
>Take your SAAS shilling somewhere else
Nobody mentioned SAAS, I mentioned the sorry state of community-led local model training efforts and you started sperging out about it.
Anonymous No.106209084
>>106209071
he forgot to take his meds today forgive him anon
Anonymous No.106209097
>>106209052
As for night time photos, take a look at left and right. It's very noticeable in certain pics too. So to summarize, better multiple subjects, less noise, better prompt following, so it's overall better imo.
Anonymous No.106209103 >>106209602
>>106209071
>SDXL does hands, feet, and NSFW better and does it in a fraction of the time that Chroma does.
No it doesn't.

SDXL finetunes might do that and it's debatable, finetunes that took a long time to come out and lots of money to train for specific purposes.

Chroma is a base model, like plain SDXL, it's way better than plain SDXL.

Stop your insane lying fueled by your irrational hatred of Chroma.
Anonymous No.106209107 >>106209439
>>106209052

I don't think you're full of crap, it's just that the model still has a lot of improvements to make. I also don't ever really prompt realism so for my use case it's a straight downgrade.
Anonymous No.106209112
>>106209071
>Nobody mentioned SAAS
You complained about the state of local, please stop lying

At this point you are so easy to spot, the resident samefagging Chroma hater
Anonymous No.106209117 >>106209716
As an idiot who actually pays OpenAI, I am quite pleased with Local. Sure, it is all held together with duct tape and moxie, but I like moxie.
Anonymous No.106209121 >>106209132
Do not believe your lying eyes, anon. You will spend 4x the time generating dogshit images, and you will be happy and grateful that a rich furry burnt $150k for you to do so. Any criticism to the contrary is all the same person.
Anonymous No.106209132 >>106209176
>>106209121
>and you will be happy and grateful that a rich furry burnt $150k for you to do so
This but unironically. The more the merrier.
Anonymous No.106209137 >>106209150 >>106209162
Most of the complaints I see about Chroma come from non-trainers
I trained a few loras with it and the results were fine
Go train a lora on whatever you like and stop complaining, retards
Anonymous No.106209150 >>106209155 >>106209174
>>106209137
Nobody can even agree which version is the best to train on.
Anonymous No.106209155
>>106209150
probably the last true epoch, v49
Anonymous No.106209159
>>106207590
is there a non-catbox moe one? I just can’t seem to open those links without a timeout occurring, it’s either isp/country blocked.
Anonymous No.106209162
>>106209137
Yes, the lora training is very effective.

Once my current large training is done (in a couple of hours) I will finally be doing some Chroma v49, v50 lora tests, I'll upload the lora(s) here if anyone want to try themselves.
Anonymous No.106209169
>>106209031
they run off all the posters with talent
Anonymous No.106209172
>>106209063
Throughout all my pics, I mostly use res-multistep at 35 steps. I think that anywhere from 30-50 is fine, though I stopped doing 50 steps a while back. Note if you are prompting multiple subjects, I think the more, the better. For realism, I usually start with
"Amateur photograph", ...

Usually I have these negs
"3D, render, drawing,"

If it messes up hands, I take extra step or two before giving up on seed.

Now, in older versions, whenever there was a screw up, I'd switch to dpmpp 2m from res multistep, or heun (30 steps), but haven't really had to do such a thing with v50. So my settings are pretty much almost always stable res multistep at 35. CFG is 4.5, scheduler is beta. I stopped using Euler in the earlier versions because I noticed it wasn't as good at genning hands, feet in one shot. As for the workflow, it's the same you see here >>106198671
which is just the default Chroma workflow.
Anonymous No.106209174 >>106209199
>>106209150
It's almost as if it's subjective and dependent on what you're training and prompting

Eventually there will likely be a consensus as to which is the overall best
Anonymous No.106209176 >>106209204
>>106209132
I'm mostly being hard on Lodestone for the fun of it but yeah, fair enough. I only wish that if something did go wrong with training and he ran out of money or whatever that he'd be open about it so someone else could maybe pick it up. He's been opaque about what exactly has been going on with the training.
Anonymous No.106209199 >>106209212
>>106209174
Come the fuck on now lol. v50 wasn't catastrophically bad like some people are saying but the glazing is getting to be a bit much. Something obviously happened during the highres epochs, otherwise the supposed improvements they yielded wouldn't be a matter of subjective opinion.
Anonymous No.106209204
>>106209176
He said on the discord that he is currently doing a write-up on the Chroma training now that it's done.

So unless there's some furry convention distracting him, maybe it will be out soon.
Anonymous No.106209209 >>106209220 >>106209221 >>106209254 >>106209282 >>106209338 >>106209966
Why does image gen seem orders of magnitude more toxic than text g en
Anonymous No.106209212
>>106209199
Again, v50 is just a merge, it's not a trained epoch.

v49 was the one single 1024 trained epoch.
Anonymous No.106209220
>>106209209
Because we have a crazy Chroma hater doing 100 posts per thread.
Anonymous No.106209221
>>106209209
Text gen mostly attracts developer types since LLMs are much more useful than image gen.
Image gen attracts coomers and basically nobody else.
Anonymous No.106209232 >>106209241
>>106209063
Res_2s with very low steps is also good on chroma. 6-10 steps. Cfg 2.5
Anonymous No.106209241
>>106209232
I'm guessing that's part of the RES4LYFE node pack? I'll give it a try.
Anonymous No.106209250
I just thought of a funny technique to combat quality degradation in stringing together first-frame last-frame videos.

Just do it backwards.

Generate the final clip first (if you're generating 8 clips then make the 8th clip first), then use the first frame of that final clip (which will be highest quality) as the "last-frame" of the 7th clip you generate. And so on.
Anonymous No.106209254 >>106209314
>>106209209
There are far more vramlets (probably jobless people and poorfags) in imagegen communities, probably a lot of horny teenagers, and people are more blatantly coombrain retards.

A lot of people in /lmg/ are richfags who can afford "mining rig" types of machines just to chat with their waifus, and are generally more technical since LLMs are more complex to setup and understand
And there probably isn't really model wars on /aicg/ since they don't have many options in uncensored API llms anyway
Anonymous No.106209266 >>106209272
and lmg types have sex with their beautiful girlfriends all the time and have huge cocks and are so cool and have so many friends and
Anonymous No.106209271
Diffusion models beget and encourage being a VRAMlet. LLMs, being chosen by God, require more VRAM in general and so their users are closer to the divine than we are.
Anonymous No.106209272
>>106209266
Their "girlfriends" have huge cocks too
And some of them are the girlfriend themselves, if you know what I mean
Anonymous No.106209282
>>106209209
this has been an interesting exercise in self fulfilling prophecies
Anonymous No.106209314 >>106209351
>>106209254
I started out in /lmg/ and it's the reason I got a 3090. Over time I realized it's quite an unsustainable hobby and I got left behind anyways. You can CPUmaxx your way into running deepseek though (possibly kimi k2). Though that new ClosedAI model has a 20B and fits into a 3090, so it might not be bad to play with it. But aside from that, with /lmg/ you are sitting around waiting a hoping for a massive datacenter GPU price drop, or for some Chink breakthrough in hardware, but this hopium is not very healthy. With imagegen we are making stride after stride after stride and already outpaced cloudshit in all but our ability to put it all into one model.
Anonymous No.106209338
>>106209209

image gen isn't very useful so most of the community is cumskulled gooners.
Anonymous No.106209351
>>106209314
I got my first 3090 to run SDXL fast enough, as back when it was released, it was painfully slow on my old 3060.
Then I got a second 3090 to run 70b llms and also to run the larger vision llms, and because I had started working with Data Science and I wanted to try things out with more depth. But I stopped there, as like you said, it's unsustainable to "catch up" as the models don't stop getting bigger and bigger and now every lab is competing in the higher end, and I always used proprietary cloud models to code anyway.
Anonymous No.106209412
Anonymous No.106209439
>>106209107
Yes, it'd be nice if the training kept going. Not sure what you're trying to prompt, but hopefully there is an anime finetune from a third party down the line.
Anonymous No.106209523 >>106209530
Now that the dust has settled. There are practically almost no differences between different model versions starting from V30 in Chroma. He could've just done the 1024 training from there and results would be superior.
Anonymous No.106209527 >>106209552 >>106210002 >>106211712
do not update to debian 13, it just killed comfyui while genning (4 out of 4 steps were FUCKING DONE)
i might be going back..
Anonymous No.106209530 >>106209548
>>106209523
>There are practically almost no differences between different model versions
According to what retard, you ?

I've just started training on v49 and v50, and there's a clear difference between the results of those, as well as the results of training against v44 which I did a while back.
Anonymous No.106209548 >>106209590
>>106209530
Okay, but have you tried training v48 and comparing against v30? My point is that the differences are so tiny we have wasted essentially 20 epochs, one of which could've been used to achieve v49 and then more that could've been used to get results better than v50.
Anonymous No.106209552 >>106209569
>>106209527
hop to a real distro sometime like arch
Anonymous No.106209569
>>106209552
to be fair debian is working after the update..
Anonymous No.106209574 >>106209582 >>106209599
New around here. What's the distinction between /ldg/ and /sdg/? Just checking out the OPs I see info about local generation in both?

If this is a massive shitstirring question I apologize
Anonymous No.106209582
>>106209574
/sdg/ is avatar schizos and /ldg/ was made by anti-avatar schizos
Anonymous No.106209590
>>106209548
The drastic difference will be in the beginning of training, then it's very much all about fine tuning, with the emphasis on 'fine'.

If you don't see any particular improvements on prompt A between early / late versions doesn't mean it doesn't have improvements on prompt B.

Also this is not an exact science, you don't want to bump up LR willy-nilly and burn the model, especially at these costs, so you will be careful. If you are a big tech company you can run MANY large scale training tests in parallel, a community funded model doesn't have that luxury.

I don't know what the exact strategy for Chroma was, but I can only assume they played it safe.
Anonymous No.106209599
>>106209574
/ldg/ is for local ai generation only

/sdg/ is for local and SAAS ai
Anonymous No.106209600
>>106206599 (OP)
Whoever made that barbarian woman is based as fuck.
Anonymous No.106209602 >>106209627
>>106209103
>Chroma is a base model, like plain SDXL, it's way better than plain SDXL.

It's a Flux Schnell finetune actually. Trained in special way that unslopped the crap out of it. Even I was surprised that it wasn't a base model. But it works exactly like one.
Anonymous No.106209627
>>106209602
For all intents and purposes (as in the stated purpose from the person who trained it) it is a base model and meant to be used as such.
Anonymous No.106209682
hm, wan doesnt mind hitting people if it's a boxing image
Anonymous No.106209716
>>106209117
We may hate them, but their censorship has caused downpour of millions from corps into local. So in a way they are helping us.
Anonymous No.106209723
Anonymous No.106209804 >>106209847 >>106209869 >>106210013
where can I download the dataset used to train chroma?
Anonymous No.106209838 >>106209858
miku hijack:
Anonymous No.106209847
>>106209804
Anonymous No.106209851
>>106208224
how do you experience this place?
Anonymous No.106209853
Vace for 2.2 fucking when? I need controlnets.
Anonymous No.106209858
>>106209838
This bitch was everywhere, is she god ?
Anonymous No.106209869 >>106209929
>>106209804
The same place you can download the datasets for all the other model finetunes and all the SAAS models
Anonymous No.106209905
comfy should be dragged out on the street and shot
Anonymous No.106209929
>>106209869
s3
Anonymous No.106209966
>>106209209
Visual artists are mentally ill
Anonymous No.106209991
>>106206646
you don't need the padding remover, wtf are you talking about
Anonymous No.106210002 >>106210023
>>106209527
Come back to the warm and comfy embrace of Windows10
Anonymous No.106210013
>>106209804
You think Chroma open srource? no! it crosed srource, no Resteners!
Anonymous No.106210023
>>106210002
mmmmm.. nyo~
Anonymous No.106210040 >>106210045
tfw still no model that beats flux dev
Anonymous No.106210045
>>106210040
>that's an example of a flux dev face
Anonymous No.106210064
an anime style Miku Hatsune walks in from the left and shoots a flamethrower at the right of the screen.
Anonymous No.106210082 >>106210086 >>106210088 >>106210102 >>106210123
Why are there like zero LoRAs for flux kontext? Yeah it's kind of stiff, but you can absolutely train it to do exactly what you want relatively quickly.
Anonymous No.106210086 >>106210088
>>106210082
no need, it can copy stuff and styles by itself mostly
Anonymous No.106210088 >>106210104
>>106210082
They're all on https://replicate.com/search?query=kontex
>>106210086
>styles
Incorrect
Anonymous No.106210102 >>106210108 >>106210114
>>106210082
would love to train one but I want to train utility loras
ie: prompt + input image = output image
I want to feed it prompt + image pairs to have it "intelligently" understand how to transform inputs to outputs
for example converting characters to T-pose, generating sprites from character refs and stuff
to me that has insane potential but I'm not aware of any training pipeline for windows that could let me do this.
Anonymous No.106210104
>>106210088
erm, I can't see a place to download models?!
Anonymous No.106210108 >>106210126
>>106210102
>I'm not aware of any training pipeline for windows
WSL
Anonymous No.106210114 >>106210126
>>106210102
https://github.com/ostris/ai-toolkit
Anonymous No.106210115
kek

an anime style Miku Hatsune walks in from the left and shoots a large fireball at the right of the screen, on the burning buildings.

FIRAGA!
Anonymous No.106210123 >>106210132
>>106210082
There were a bunch of NSFW Kontext loras at first, but they were hunted down and banned from all platforms, and then it seems interest died out, be able to gen quick shitty memes likely gets old fast

NSFW drives interest, just look at the Civitai download stats
Anonymous No.106210126
>>106210108
yeah hoping to avoid this, WSL is just shit linux at that point I'd just go install linux

>>106210114
IIRC this wants models in diffusers format again right, I just redownloaded Qwen Image in bf16 because the fp16 was apparently so incompatible with musubi tuner wtf
Anonymous No.106210132
>>106210123
Pain in the ass desu.
I needed a good style transfer LoRA to test using vace and 3D models to animate clips. Now I gotta train my own I guess.
Anonymous No.106210148
is there a node to clear/unload all vram or whatever, like if you are using models in other apps like reforge, but want to use a full wan 2.2 model that needs like 15gb
Anonymous No.106210151
>>106210147
>>106210147
>>106210147
Anonymous No.106210153 >>106210814
Making a separate general for SDXL-based diffusion: good idea or bad idea?
Anonymous No.106210814
>>106210153
just use /sdg/ you fuckin retard
Anonymous No.106211712
>>106209527
Anon, just delete your venv folder and reinstall your requirements for comfiui and your extensions.
Any time you do a major system upgrade (or even a video drivers at times) you should be doing this.