Discussion of Free and Open Source Text-to-Image/Video Models
GenJam3: https://forms.gle/hWs19H4vTGTdwARq8
Prev:
>>106145131https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows
>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanXhttps://github.com/Wan-Video
2.2 Guide: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y
>Chromahttps://huggingface.co/lodestones/Chroma1-Base/tree/main
Training: https://rentry.org/mvu52t46
>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage
>Neighbourshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>>106146762could be, t2i in wan
Remember to help with the new wan2.2 rentry guide if you can:
https://hackmd.io/RDxlWe8mQCSUi72yUDEzeg?both
Anyone can edit.
>>106146749Slightly better result
>>106146781Arr rook the same
>>106146776>Chinese product not as good as advertisedAre you surprised?
Blessed thread of frenship
What is with the obsession for text? Emad was obsessed with it too
>>106146375Can't post any of my gens here but god dang it's pretty good, nice workflow
>>106146781Alright, so my conclusion is that for photorealism it's at least an open source Mogao. Maybe there's a way to make it look even more photoreal, who knows.
>>106146802Text = model capability!!
It's the easiest thing to fake by overlaying text on training images.
Feet pics also possible but slopped. And of course variance of lighting, pose etc... nowhere near as good as Chroma. Hence the crisp/clean/perfect looking girls. But just like Chroma anyways you can still get bad gens, like with
>>106146690most were bad, though that's not an easy prompt.
>>106146802new shiny "NEVA BEN DUN BEFO'!" type mentality.
Basically people who are only extremely superficially interested in image gen and only wants it for dumb meme shit rather than actually producing anything that normies would see and actually like/not realize was AI in the first place.
>>106146776No idea in what way you're describing it being bad/worse than advertised but everything in the repo text-related looked extremely fried with all text looking extremely digitally super-imposed whenever it showed up.
Anyway as is the base model is useless(just like all base models wow who would have thought). The only important part is how well it takes to being finetuned - which, if the answer is even just "at all" - makes it strictly superior to FLUX shit in every single way.
Either way it's not worth caring about for now and the answer to if it's actually worth caring about or not will be a week to months away, probably.
file
md5: ae41e26a474f431e5fefa6ad13f75293
๐
>>106146839Toes are fucked but I don't mind it otherwise
>>106146853Could her hands actually fit in there?
>instantly crashes when trying to load the model
Welp. Back to wan with me.
>>106146375Feels like the interpolation isn't as smooth on your workflow as the rentry, is there a way to go about that without replacing all the non-wanvideo samplers and shit with wanvideo equivalents like in the rentry? Other than that I really like it
>>106146822Yeah, text has hiccups, but this if fp8 version so can't tell for sure
>>106146892>no chestCould you imagine if femboys actually looked like this
>>106146892This is a man, right?
>>106146896idk why its not giving me big tiddies, maybe i have to prompt for them, prolly cause chinese people dont have tits really
and yeah, thats what they think they look like lol
>post censored video gens on other boards
>keep getting asked for uncensored/catbox
>never post it
Qwen fucking crashes my comfy before it can even try to offload. Aren't there any nodes for this?
>>106146820Loras for sure, but it's a very large model, not sure how many will spend the effort when you might as well train Wan image loras which will have higher reach
Barring some vital information I'm missing. I think I might actually dislike this model.
Is lightx2v 2.2 still shit?
>>106146919multiGPU has an option for virtual VRAM which is better than the native offloading functions in comfy
>>106146865I don't think their workflow has any interpolation. I'm testing it now, and have added a RIFE VFI node as the base gens were 16fps.
>>106146933But that's only for gguf. You can't put safetensors in it.
>>106146903>>106146904doesnt seem to react to 'very large breasts' lol...
qwen image is fun in the same way flux is fun. But I already have flux. And it runs better. So what is the appeal in this model?
>>106146925Yeah, after testing it, nowhere near as good as Chroma. If I want clean images that follow prompt really well I can just use Wan 2.2 t2i, and that would be more realistic too. It's not good at styles nor artists which would've made it good enough to use for those purposes alone. It's hard to tell if this model even is better than Flux. It probably is, but just not enough to justify its size after the release of Chroma and Wan 2.2. This was probably meant to be a commercial product, the only reason it's open sourced is because it sucks and the creators realized it.
Then again, the editing model could change my outlook on the model. Let's hope they're not bluffing those stats too.
>>106146988100% trained using synthetic 4o slop.
Can I fit qwen fp16 into a 48GB GPU with clip offloading?
>this thread
This is why /adt/ will always have it's place.
>>106146983>sdxl is fun but i already have 1.5 and it runs better!you are now realizing local is forced into stagnation because local hardware has not kept up with model developments. these models generate in <10 seconds on saas btw
file
md5: fa28a9ec638d2f647b6224768450c993
๐
I get the same times with or without torchcompile when genning wan2.2 (without lightx), can someone confirm if it's the same with them?
On a 3090 with cuda 12.8.
Wan 2.2 handles emotion prompts well. It seems to obey camera movements a little more too.
Bruhs this model is trash.
>>106147013This model is not an objective upgrade though. It's trading one piece of slop for another. Good would mean SOTA out of the box at something. I don't get that sense from this model.
>>106147047There might be something worth using from the image editing model, but they didn't release that.
>>106147047only Sam GODman knows how to cook image models.
If I dumped booru tagged dataset into a chroma lora trainer, would you still be able to conjure it with the trigger words?
>>106147000I can see the orange hue on some of the outputs, there is no doubt.
>>106147047This. Huge image only model, still not beating Wan at images despite it being a video model.
>>106147066OpenAI's advantage is data, they have an army of Kenyans for annotations.
Data is the only thing that matters. SD v1's UNet is good enough if you have good data.
>Good would mean SOTA out of the box at something.
So we're ignoring the text?
FOUR MORE YEARS OF SDXL!!!!
>>106147094With the image quality it's good only for shitty memes.
>>106147094>So we're ignoring the text?You mean the text that looks like shit? If I wanted overlaid text I'd add it myself in mspaint.
Can these tools be run on Intel ARC GPUs? Is there some kind of translation layer if they can't be run with some native ARC support or something? Help a newfag out.
>>106147107https://docs.pytorch.org/docs/stable/notes/get_start_xpu.html
>>106147101>>106147106>and other things said about flux
>>106147066How does he do it? Seems only he has a dataset that's properly labeled. Everyone gets scraps from him.
>>106147113Nigga it's the same type of slop.
>>106147113Yes because it's just Flux architecture with increased dimensions
Has the wan meta changed over the past 48 hours? Or is the best workflow still kijai's using the 2.1 lightx2v?
I saw some videos using the 2.2 lightx2v t2v and they look like this for i2v gens.
>>106147107Yes, but in Ubuntu. ComfyUI has integrated support, but only in core nodes.
>>106147147Intel GPU work on Windows.
>>106147066kek, knows two artstyles (all piss yellow) and five poses
>b-b-but it does text wellwow
> crying with tears
why tf are they goopy
>>106147161too much the eyes of cum
>>106147156Uncensored Dalle 3 at release goes to show you what happens when they actually try. Unfortunately 4o is censored after the fact, so that's not what it's objectively capable of.
>>106147153Didn't know that. Linux is better anyway.
Qwen feels like absolute benchmaxxed slop. I cannot think of a better way to describe it. There is zero inspiration behind the model.
People need to seriously consider doing away with benchmarks entirely. It's hurting everything.
>>106147115>>106147060These are all different seeds and they legitimately have the same bamboo in each one.
>>106147245lodestone please
"An anime girl with blonde hair and blue eyes wearing a suit and high heels holding a sign that reads Hello 4chan"
>>106147300best gen so far kek
>>106147300I wonder if the word 'slop' shows up in the training data enough to influence the gen
>>106147300Should be benchi suroppu rather than ban suroppu.
The text on top of white rectangle is even worse lol.
Can't even do shit it is supposed to excel at.
>>106147326>white rectangle>white
>>106147330I assume that's the color whatever text encoder this thing uses decided that should be added there before piss filter kicked in during diffusion process.
I'm not crazy right. Some of these images clearly have the 4o piss filter.
>>106147326> Should beonly 1 in 1000000000 will notice
>>106147330>Yes, we trained on GPT 4o images, how could you tell ?
>>106147353You are crazy but you are correct, Qwen is piss slop.
>>106147363https://files.catbox.moe/xo51uc.png
Prompted to be nude.
Those nipples are not looking good...
>Orange lamp? I don't know what you mean.
>>106147423Even the fucking glass is orange
>>106147363>big bro has a ponytail
>>106147447Typical of steroid users before they go bald.
Qwen will have the same uptake as HiDream, as in practically zero
>>106147451Can you prompt it the chinese negative prompt copypasta? This is from wan.
>>106147460The Wan negative prompt?
>>106147468Yeah, just slap into positive
Would you believe me if I told you these were completely different seeds with one word changed?
>>106147451
Before I shit on qwen further. Is it possible there is an issue with the quantization?
>>106146763 (OP)using the ldg fast i2v workflow for wan2.1. any reason i would be getting some really shitty colorful artifacts everywhere when setting cfg in the adaptive guider higher than 3?
also, anyone able to get dicks to not look like an alien scifi prop?
>>106147493I've been thinking it might be a not enough verbose prompt and it needs a wall of text to let go properly, but there's legit tonfo orange slop in it.
>>106147497>2.1Wake up grandpa and get 2.2
>>106147507>Wake up grandpa and get 2.2the fast workflow is built with safetensors, not the ggufs
>>106147507It's just this is supposedly a q4 gguf test I found on hugging face. It makes me wonder if some of the shittyness I'm seeing is a result of the compression being bad.
>>106147161>there's no such thing as too much lub-ACK
>>106147519Where the fuck did you get Q4?
what's the best low-step lora for illustrious type models? need one for fast prototyping
>I'll try prompting in English translated to Chinese, that's a good trick
>>106147530Just a test I saw in a discussion on hugging face.
>>106147537Looks like she has an extra finger. But the text looks pretty consistent.
First run after adding more prompts and more layers
Why is the frog so unhinged?
>>106147534Sorry anon, we're too busy seething over a free model a company likely spent millions of dollars to train.
wan2.2 on 8gb vram doable?
>>106147560Spending millions of dollars on a white elephant doesn't make it useful.
>>106147534>low-step lora for illustriousIllustrious
>>106147567I cant help but feel like I'm missing something important with this model.
>>106147574You're missing the fact that it's Chinese piss slop.
>>106147566Probably, but the quantization needed is likely going to kill quality.
>>106147577No, I definitely didn't miss that.
>>106147537>aces the text>completely fails the handsums of the last year and a half of imagegen pretty nicely
I was trying to play around with Qwen image but my images come out entirely black. The preview looks good but after like 10 steps it just turns entirely black.
Any ideas what the fuck is going on?
>>106147581what if i wanna make animated sticker gifs?
>>106147591Yes. Turn off --fast and --use-sage-attention. It's one of those. I haven't bothered to test which.
>>106147574The editing model MIGHT be good, unless it covers the entire output with piss and plastic skin.
>>106146925wonder if its actually better with chinese prompts, anyone try that yet? A/B comparison with google translate
>>106147593I mean sure, just wait for Wan 2.2 q4 versions to pop up, maybe they already have ?
>>106147591>>106147602I have both on and it works most of the time but I did get a few black images. So they don't totally break it. Maybe a combination of something involving those.
>>106147606>editing modelThe dataset will be from 4o.
Surely openAI notices a single IP in China or a very random country demanding millions of images.
Does not understand reflecting in the sunglasses.
>>106147645Using their outputs for training is against TOS but they don't care, it's free marketing because everyone recognizes the piss.
qwen image is good
chinks continue to carry open source AI on their backs
glory to the CCP
Anons, just put "sepia" in negative and never think about it again. That's literally all it takes with this and every other piss filtered model.
>>106147653Nobody is going to sue over this, because it would open a HUGE can of worms, you saying other people can't use your model images for training while using billions upon billions of images you do not have permission to train on, that's going to be a very unpleasant court case for big tech, since there's no way it can be legally enforceable.
So they will all gladly use generated images from eachother for training, with no fear of legal reprisal.
>>106147675Nobody is going to sue over this because they are in China and they are not subject to the cooky zany rules of the US.
If this model was released even a few months ago you'd all be jizzing right now.
>>106147689I am jizzing over it right now, people here are just addicted to misery. Or coomers who will be eternally disappointed by models not being able to do bobs and vagene out of the box.
>>106147689BEAGHGAHGAHGHAGA
>>106147680All the western models are training off eachothers output as well, nobody will sue because it would A) lose B) put a spotlight on the hypocrisy
>20gb
not local. we need sub 1b models for everything
Can anyone make a comparison of the motion difference between lightx2v and non-lightx2v? Is there really always a slowmotion effect with lightx2v?
>>106147648Workflow/prompt for this?
>>106147729I'm running it locally on a used 3090, a card which came out in 2020, half a decade ago. And I'm on disability and poor, I saved up for months to buy it. If I can have one you've got no excuse.
>heh this model sucks its just another flux!
>running 4step q2lightningfastgguf BTW
>>106147748It's just qwen image default workflow. The prompt is exactly what you see except it should be reflecting in her glasses, not a giant mirror on the front of her bike.
>>106147689the only people seething are the same who seethe at chroma and any other model that simply either cant run on their laptop 3060 at all or needs 5 minutes to do so, every time
>>106147560a free turd is still a turd
these companies want all the glory with none of the safety risk of making it usable out of the box
you should be critical of them so that they see that the bar has risen and they cant just train on complete garbage data and expect everyone to suck them off
>>106147805and btw those same people then say that everything is stagnant (because for their 3060 which cant run anything new, it is)
>>106147805>please ignore the yellow piss filter
>>106147824sepia in negative
It's starting to think I may have been overly critical of the model.
>>106147560The problem is that it's an also-ran at this point, it can compete with Kontext and win, but it's not as if Kontext is good for anything but quick memes.
Add to this that it is huge and slow, and censored, and suffers from the 'plastic uncanny look because we've trained on gazillions of synthetic data', the latter can be mitigated through loras, but why would you bother ? For the slightly improved text rendering ?
It's another DOA model, like HiDream, SD3.
>>106147824Only some images push towards it slightly and rarely but the model isnt anywhere near closedai constant and obvious piss filter
This model is for text and image editing, if you want tranime use noob/illust, if you want realism use chroma, you dont have to actually use only 1 model for everything
vramletretards... your time will come soon with svdq...
there is also
>Thanks to DFloat11 compression, Qwen-Image can now run on a single 32GB GPU, or on a single 16GB GPU with CPU offloading, while maintaining full model quality
https://huggingface.co/DFloat11/Qwen-Image-DF11
>>106146763 (OP)>2.2 Guide: https://rentry.org/wan22ldgguideldg_cc_t2v_14b_480p.json
just leads to a blank file, anyone have that workflow?
https://strawpoll.com/e6Z2Azw1qgN
Thoughts on Qwen image now that we can use it.
>>106147925>now that we can use it90% of the thread doesn't even have 24GB VRAM, lmao.
>>106147930How the fuck is that MY problem?
Made this for a coombait thread on /v/ but forgot to post it before going to bed, u can have it.
>>106147940You used "we" and you gave a poll which is worthless given the fact I outlined.
>>106147942Sorry. Image generation has made me numb to all but the most depraved stuff. Not even a twitch.
>>106147945Oh sorry you're right. Let me clarify it with the fact I don't consider vramlets human.
>good at text
>low seed variability
>slopped
Qwen image is literally HiDream 2.0. It will be forgotten in a week like HiDream did. Screencap this.
>>106147952Given that there is no vramlet test before the poll can be voted on, there is no way to differentiate between them, meaning the poll is worthless.
It's just another bloated image model.
We won't use it.
SDXL currently has the largest community, the most support, and the greatest artistic variety.
Things are moving too fast to choose and train a model from scratch.
Even if all of CivitAI, Tensor, Pix, and Sea decided to train Qwen, the time spent training it would be meaningless because 200 new, better image generator models will be released.
>>106147986>200 new, better image generator models will be released.And you'll hate all of them, and declare each one DOA on the same day it's released.
>>106147986You say that like SDXL users have a choice in what model they use. Do you know how many vramlets there are out there?
>>106147986SDXL was declared dogshit the minute it released, too. And now look where we are.
>>106148048because it was dogshit and it took an eternity from release to make it somewhat ok
>1328x1328
but why this number
>>106148082i don't know who that is, and that's not an appropriate username to use on the internet
>>106148003you and retarded apologists like yourself are the reason why they dont feel the need to improve their models on release, enjoy your synthetic slop every time because thats what they are comfortable with
https://developer.nvidia.com/cuda-downloads
https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html
UPDATE YOUR CUDA 13.0 TECHNOLO/g/Y.
>>106148108Yeah nah I'm not touching that shit until I'm 100% sure whatever improvements it brings outway the issues it will cause 10 fold.
>>106148048I understand, but this is different. It wasn't the era of private models hosted in the cloud, and there weren't large communities formed around SDXL like there are today and as this anon
>>106148035 says. VRAMlets are what really drive the image community. In the grand scheme of things, all these new models don't make meaning and sense because SDXL is what continues to generate revenue and is easy to train and keep up to date and is what keeps the community engaged
We need cross-compatibility if we want to see real change and not a "single-use" model until the next, better, newer one comes along.
>>106148078That's one of the recommended resolutions for the Qwen image model.
"1:1": (1328, 1328)
"16:9": (1664, 928)
"9:16": (928, 1664)
"4:3": (1472, 1140)
"3:4": (1140, 1472)
>>106148108this
>>106148117They always fuck their new release three to four times before they get it right.
Anime style to this model is just 1.5 face.
>>106146778>/ldg/ Comfy T2V 480p workflow (bullerwins): ldg_cc_t2v_14b_480p.json %darkseagreen%(updated 2nd August 2025)%%leads to an empty file
file
md5: 2d07916986caef27136f7deb6b748f13
๐
>>106148082It's on Qwens's example code. Is there some official list of supported resolutions? I've gone all the way up to 2kx2k and it seems to maintain coherency.
>>106148110Damn why is she scratching her leg like that?
Qwen seems alright, probably needs a finetune though.
>>106147566yes, but need 64gb of ram or that merged model and 32gb
>>106147628> maybe they already have https://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF/tree/main
>I'm using Q4-
You didn't run it.
>have to do fucking racial profiling for wan gens because If I don't specify the race it keeps giving me some mixed meat
Thank /pol/
Now that's an interesting qwen output.
>>106148335Seems like this model has no imagination and can just copy.
>>106146936it has that smooth skin but it doesn't give me uncanny vibes, nice
I keep trying to move up to flux but it seems wildly inconsistent and tough to wrangle compared to SDXL. Is this a me problem or is flux generally just harder to work with? Even the loras and controlnets for it seem worse overall. Only thing it's definitely better at is realistic which i'm not interested in.
>>106148347Do you ever get tried of being a cynical cunt?
>>106146983flux is garbage.
It kind of knows one punch man.
>>106148381I'm sort of in the same situation and I attribute it to the prompting process (full sentences are a pain) and the extra time it takes to generate stuff which makes the learning process difficult. To combat this I've built a workflow where you can easily switch between "Nunchaku 8 step TurboAlpha", or "28 step Normal Flux". Definitely install Nunchaku if you haven't, it makes it so much more fun to use Flux now.
>Only thing it's definitely better at is realistic which i'm not interested in.it could be that Illustrious and SDXL are better at "simple anime". Flux seems good at high detail anime, but you need loras:
https://civitai.com/images/73784871
https://civitai.com/images/34333535
I think Flux produces generic anime without loras. So it's purely a lora-usage model, is how you should treat it.
Is it just me or are nunchaku svdquants much worse than q8 and its not even close? Might as well wait a minute for a good gen
Why can't wan2.2 do violence well? Having one character hit another character still produces bad results.
Strange how it can now generate vagina with just the base model but punches still suck
>>106148501>it can now generate vaginaneo vagina maybe.
>>106146928Yes as far as I can tell. 2.1 is better, you can kinda mix 2.1 and 2.2 together but not really sure if that's an improvement over 2.1.
>>106148499They're a little bit worse and that's why you need a workflow that allows you to instantly switch between a fast mode and a slow quality mode.
Don't fall for this "inbetween" mode where you give up some quality for a moderate speed boost. It's a bad approach.
qwen
md5: 0c43e547697dc568f43af418772f8339
๐
Quick test of the usual prompts on different seeds.
fp8 model though until the DF11 works.
I don't know. Seems like a model, alright. The variance is actually kinda wack, though. And what the fuck are those 'green eyes'?
Box: https://files.catbox.moe/o7mzsa.png
>>106148529How are you controlling the inputs? I have a workflow where I have an int constant that I switch between 1 and 2 and it feeds into a bunch of Switch nodes that chooses inputs/values, but it's a huge mess of spaghetti and updating it for new developments can be a pain in the ass. Is there something better?
Wondering if I should just bite the bullet and figure out how to call the comfyUI api so I can finally escape this spaghetti hell (unlikely).
>>106148550I've noticed the similarity between gens too. Sometimes I have to double check the seed actually changed. Also noticed the weird green eyes.
It does not know this character in Chinese or English.
>>106148316kek, Rust is a very subtle man
>>106148562You want to use this "Fast Groups Bypasser", you can configure it to list only the groups that have a certain string in them. And you can configure toggleRestriction to "always one" so you can only have one enabled at a time.
Then you use these "Any switch" nodes, the way they work is it'll choose the input that is not "null", so if one group is disabled then it'll feed a null input and then the Any Switch node will not choose that input, it'll choose the other input.
You have to be careful though, when a node is disabled it can pass in previous nodes that are connected to it, which can cause the input to not be a null input. So make sure the groups (in this case "Load Turbo Alpha" and "Load Normal Flux" are at the start of the flow.
If I didn't just generate this myself I'd say that is 100% a gpt 4o gen.
It seems if the IP was culturally relevant to a high degree it gets into the dataset.
>>106148661You are just imagining things, it's a coincidence
8344
md5: 1717f55aa7d77c911ab75e2c9287b823
๐
Can someone explain to me how to use video loras in kijai's wan2.2 workflow? It looks different from the old rentry workflow and it's not commented.
Do 2.1 loras even work well here?
>>106148590Flying Dutchman?
>>106148688So on the lora node where it says previous LoRA? Just stick a LoRA node on those. 2.1 LoRAs sometimes work. Sometimes they don't.
>>106148550That was with cfg 4.0, by the way. Genning at 1.0 now.
The way it bruteforced all those parts of the prompt into the images (Extremely green eyes, swirls in the smoke literal gimp tool swirls) is awful.
I'll do a big sampler/scheduler plot later once the cfg test is done, but probably only for one of the prompts, most likely titty fairy.
>>106148702More like (((Fleeing Goldman)))
yea qwen image seems overfit, hopefully a finetune can fix it
>>106148705In the rentry workflow there was a comment saying it's strictly required that any additional loras go between lightx2v and the Patch Model Patcher Order node. This is no longer the case?
>>106148732I don't give a shit. I use the native workflow. That just how I strung LoRAs when I use Kij's workflow.
>>106148732don't use the shitty rentry wf, scroll up
>>106148730Yeah I'm just probing some IPs right now so I'll switch up the aspect ratio.
how do i use lightx2v on text 2 vid try using it but video came out all wonky
fuck, I spent like 2 days of compute time training a lora on wan 2.1 for something wan2.2 can already do mostly becuse i only just became aware of it.
>>106148550>>106148708cfg at 1.
Amazing, breathtaking. Stunning, even.
Box: https://files.catbox.moe/21xo8k.png
>>106147602its sage attention, is the only one I had enabled
>>106148745Don't know what you're talking about. The current rentry's workflow uses a slightly modified version of Kijai's workflow.
The old rentry's workflow is pic related.
>>106148812Very much the expected result.
Look guys. It's Miku... but in minecraft.
>>106148835troonku goes to
>>>/g/adt/or
>>>/lgbt/
>wan still thinks panties are just a white tube that wraps around the waist
file
md5: 03f0c27a91bc487af3a503f8ed1112eb
๐
so there a lora training script or whatever for wan2.2
and looks just almost close to what i want it to do with prompting and just needs a little lora push.
vramlet bros, there's light at the end of the tunne!
https://huggingface.co/lym00/qwen-image-gguf-test
>>106148862Can someone translate what this dude saying?
>>106148650This is so good, which UI did you use? How do you learn to make gens like this? Complete noob here.
file
md5: ca3a40d0493c8943de44b7d290c0f616
๐
>what will you be ordering today, anon?
>>106148887Download Automatic 1111 and stable diffusion 1.5 to get started.
>>106148912fuck off >>>/g/adt/
>>106148912Bruh she's like 10?
file
md5: 86a45682871fbe5714ed8a4d6ac50c36
๐
It lowkey knows emma.
>>106148923It knows a lot of things, but it does not know them very well.
file
md5: 9ddd8d7cea42c934887c729e4aaf19bf
๐
>>106148912>deletedBased. Get this loli shit out of /ldg/.
>>106148964ok shoutcollective
file
md5: 2e6d59fb78ca628778655d6087756903
๐
Should I upgrade to 64GB or 128GB RAM for Wan2.2?
>>106148976both 8bit models + T5 + vae + loras take like 50GB so 64GB+ if you dont want to have to wait for slop swap all the time
file
md5: 1674bd0d0860d7b2da3d621c3f9f9ff3
๐
not quite
Doesn't really have a great grasp on streetfighter.
>>106148976i had to settle to upgrading my mobo to 96GB ram instead of 128 since evidently it did not support that much for DDR5
>>106148874how do train lora fir wan 2.2?
>>106148976ComfyUI OOMs with 64GB RAM when changing loras or input image sometimes.
file
md5: d25803e82a4dfe120724fc6419676bff
๐
woah, almost
>>106149016-888 social credits
>>106148501just train a lora
The greatest trick Xi ever pulled was somehow tricking the west into thinking Whinnie the pooh was his achilleas heel.
file
md5: ccd0139a6de8fd4912d6449a878f5c13
๐
>>106149047+888 social credits
>>106149036I have 16gb vram.
sup
md5: 3ecffbc271859b15cc6671ac1f78b83f
๐
qwen is not bad
>>106149064how is it at inpainting?
Oh yeah, definitely looking like their chinese counterparts in this gen.
>>106149079lol thats rough
>>106149060how much ram do you need for training lora?
>>106149064then why post a bland image? or are you just using comfy's shitty prompt?
>>106149094>then why post a bland image? or are you just using comfy's shitty prompt?
file
md5: 80a0b678dac5d4f024a62135e83ba877
๐
I tried Jeffrey Epstein, got some protestant pastor.
not even a genned jak. pathetic
keked
md5: 8f200ab9ab939325c33e1091619772fa
๐
>>106149067still have to test, will do it later and post here
>>106149060rent a H100 gpu and train it, will cost you less than 10 bucks
GGUF Q8 when? I know f8 is ass compared to the might geegoof.
>>106149114my main issue with flux as that it was virutalyl usefless for the inpainting i wanted , freaking SD 2 was better at it.
euler
md5: 8d9e5dfba4081e805078a2cd47249581
๐
>>106148550>>106148812Well. That's kinda disappointing already.
Box: https://files.catbox.moe/tnn1vx.png
file
md5: 8ad0c2581d94046bd2600ff62d89d607
๐
i2832
md5: acd3004860fcd0e15af384b2b1fac126
๐
>>106149132what are you trying to write?
reminder: bloat is bad for your computer
>>106149167>A frog holding a sign that says hello worldUsing T5-small + an adapter though
>>106149159What is the deal with the output variety for this model? It's basically non-existent.
>no submissions for genjam3
It's over.
I was very much hoping that Chroma was going to be good to make the base gens of my oneitis to use in lewd 2.2 I2Vs but it still sucks.
>>106149243it is good, what are you using that is better?
>>106149212even worse gacha that round 2. organizer should just reroll since everyone is just doing the same shit. if the result prompts doesn't leave room for creativity, just reroll
>>106149212>>106149272>still using glowgle forms
>>106149299this too. just have an anchor or reply to the OP like /dmp/ does for their albums
>>106149272>if the result prompts doesn't leave room for creativitysounds like you are not very creative anon.
Truly creative people can extrapolate all kinds of cool ideas from just a few words.
>>106149315>*slops your datasets*why would xi do this? it makes these massive models so shit
file
md5: 8ea8d56d25bb997c5f0bcfb2ce3cfb16
๐
>wan i2v
>after the first frame the face morphs completely
is this because of a lora im using or what? how do i keep it consistent? lower strength on the loras im using?
>>106149335is this 2.1 or 2.2?
>>106149328>gacha character>night sky>jewtube thumbnailnope, nothing creative is coming to mind anon and I'd imagine everyone's result would look way too similar
>>1061493392.1, i should probably get 2.2 but i was uh... waiting a bit longer because i saw some anons complaining about the workflows in the guide
>>106149343yeah that pull is shit
>>106149331good enough, still sota, and still pushing forward everyone else to have to compete hard
if a basement dweller can train a lora in a day or two for a nsfw tune of any model, whats the problem anyway?
>>106149352Imagine refraining from doing something because an anonymous literal who on the internet would have issues with comfyui.
>>106149358sota in... text? what a fucking joke. it's benchmaxxed slop
>>106149335working on a better setup that also uses the phantom fusionx lora and pusa, which both massively help likeness
>>106149362not him but what the fuck do you expect?
if i cant get the light shit to work on text to video and you refuse to be helpful im just going to default with the worflow and model that work.
>>106149362i dont mind waiting, i'd rather let the eggheads sort it out then enjoy the fruits of their labor
>>106149364text is a pretty big use case for a lot of people who dont just want to gen 1girl but the big thing is image editing
>>106149362don't lie, everyone has one issue or another with comfyui. It really went downhill after the org
I've noticed a correlation in Qwen naysayers and Chroma copers.
>>106149362Very few people want to be forced to tinker for hours and reinstall there setup just to get back to where they were. Or just to find out it works never have worked because of X or Y reason.
>>106149358is there a publically avaiable training script for 2.2?
>>106149388the edit model isn't out yet. you are using a bloated and shitily captioned t2i model. the edit model is the only thing I would be interested in so kontext gets dumpstered
>>106149343you could submit a gen of someone throwing all of those themes in the trash can
>>106149397sorry but qwen is overfit and knows way less and it has no style knowledge at all
>>106149410how about no. I just want something fun
I've been getting incompetent Chinese training fatigue for a while now. the xi chad meme just doesn't hold anymore.
>>106149388>text is a pretty big use casewhen gimp, krita and Photoshop exist? you can even arbitrarily add text on paint too. stylized text is the only use for it and I am not in advertising. visual text is the least important thing for a model to know when styles and artists are king
>>106149425>text bleednot a good sign
>>106149434whatever you say 12gb vramlet
>>106149460I'm talking about the boring results. I have 24gb VRAM but I just don't see a point to this model yet. maybe after somebody with talent finetunes it with NSFW and styles I'll care. keep being jaded chinkoid bug
>>106146763 (OP)>https://huggingface.co/lodestones/Chroma1-Base/tree/mainso is that the "finished" chroma now? it doesn't even have a readme
>>106149492nah, two more 1024 epochs
>>106149159Alright, that's the most boring-ass comparison I've ever run. I'll reduce the amount of schedulers to like... 3 and post the big-ass plot once it's done. Or I'll just cancel it and wait for the image edit model, because this is kinda 'meh'.
>>106149328>>106149343I have the same issue I had with the last jam, I have no idea who these characters are. And I don't really want to just 'meme gen'. But maybe I'll think of something.
>>106149255Honestly, nothing else because I gave up lmao. But I keep getting really soft looking images with chroma and then NSFW prompting seems like I'm missing the mark.
>>106149460> 12GB vramI wish...
>>106149498>He literally said "two more epochs"
>>106149502>I have the same issue I had with the last jam, I have no idea who these characters are. And I don't really want to just 'meme gen'. But maybe I'll think of something.That's ok.
If nobody wants to do it, the genjam collage can go ahead with 0 submissions and all the three themes are discarded.
>NOOOO YOU HAVE TO USE THIS CHINA MODEL
>LOOOOOOK I POSTED A XI CHAD MEME THEY ARE SOOOO COOL
>STOP MAKING FUN OF ME! QWEN_IMAGE IS SOTA!!! IT'S THE BEST!!!!
>YOU MUST HAVE 12GB OF VRAM CHUD
China lost the magic. is there a company that is based enough to train what people want anymore?
>>106149558Wan 2.2 came out a few days ago and everyone agrees it's good.
>>106149574what happened to all the text? I thought it was sota at it?
Ok, even better WF is finished. This in 4 steps total, full prompt following / motion quality
https://files.catbox.moe/6lp32g.json
> lodestone just can't stop seething
>>106149584it's an upgrade but the captions they use are slopped. they should hire prompt gods to caption things not let a vlm slop the dataset
>>106149601this. never communicating the problem to the devs will ensure it will never get fixed
>>106149590All I typed was "Screenshot of Windows Desktop."
>>106149591why did you squish the aspect ratio? wan does fine with different resolutions
>>106148381Flux is more for realism than anything. It also has a stronger "slot machine" feel than other models because it can generate wildly varied outputs if you aren't very specific about certain things.
Generating prompts with an LLM is also preferred. You can do it by hand, but you'll never be as autistic as an LLM is when describing an image.
>>106149644for fast testing I kept res low as I tested each tiny change
>>106149591Have it so when he moves out of the way to shoot a Kamehameha he's firing at a planned parenthood clinic.
>>106149651llms go moralfag on me though
>>106149651>Generating prompts with an LLM is also preferred.miss me with that shit
>you'll never be as autistic as an LLM is when describing an imageit's been pretty shit and it will come up with some hallucinations. llms are too autistic about safety as well
>>106149656the aspect ratio is the problem, why would you squish it? just shrink the 1:1 in the same ratio
can I submit multiple gens for genjam or is it supposed to be one per person?
>>106147898they always promise same quality but it's never the case. why are chinks such impulsive liars?
>>106146763 (OP)>Setup comfyUI on my Linux machine>Start downloading loras>All egirl loras have been taken down by (((payment processors))) pressures on CivitAI>Look on alternative sites>None allow download>head to civitaiarchive>Most of the links are dead because the ""archive"" was actually redirecting to civit serversWtf am I supposed to do? train my own? Did they actually succeed in killing the progress towards making new ones? I mean, some have been archived, but nobody is going to spend money on training SDXL new LoRAs if they can't upload it anywhere.
>>106149722Grifting is just a part of their culture
here is this updated to the new WF, much better. Less than 2 mins to gen
>>106149752Have you seen the russian redditor sperging out at you 'stealing' his image and 'ruining' it?
I thought that was pretty funny.
>>106149762lol no. Ah, is it the deleted comment?
Why do you guys defend Kontext?
>>106149791i like it, why what's better for edits?
>>106149772Yeah, that protector111 zigger sperged out about you not genning your own test images because that's the reason people think AI is slop. Apparently.
>>106149722The whole point of dfloat11 is that it is bit identical, it compresses bfloat16 to dfloat11, in other words 16 bits compressed down to 11bits. Of course there's no free lunch, accessed parts need be decompressed as you generate, but it's faster than ram offloading since it does it all in the GPU.
It shaves off ~30% off the model it compresses, so it can make a difference.
>>106149791>Why do you guys defend Kontext?I don't know who you're talking about but people should use any tool that works for whatever use case they have instead of building their personal identity over any one image model. Even if theres one thing a model is good for it's still useful.
>>106149823we already had a quant autist last thread sperging and being proved wrong. I don't want a repeat episode
>>106149730well hidden huggingface repos but yeah, train your own
>>106149818>>106149825I only have 10gb vram tho
>>106149668>>106149672I don't really gen porn, but Gemma 3 27B Q4 with my own Sys. Prompt has never given me any refusals for spicier prompts.
--Qwen-Image criticized as bloated model lacking true multimodal capabilities:
>106143040 >106143057 >106143115 >106143131 >106143158 >106143070 >106143087 >106143097 >106143121 >106143449 >106143453 >106143537 >106143490 >106143540 >106143568 >106144440 >106144456 >106146851 >106143237 >106143313 >106143443 >106143488 >106143527 >106143548 >106143462
seems like /lmg/ doesn't like it. they have China fatigue as well
How much longer until the AI community learns how to upload and download torrents? There should be a torrent repository for adult content.
>>106149903you skipped the part where hallucinations are a thing. it's also fills the prompt with a bunch of superfluous non-token shit. honestly not a fan of full nlp encoders
why is qwen so fucking slow compared to fux, chroma, or even wan?
qwen-image won't write "ไธญๅบใ" (meaning "creampie" in Japanese)
>>106149917Torrenting will never be a solution for this and you need to stop pretending it will be.
Anyone try the qwen image vae with wan 2.2 yet?
>>106149907>half the linked comments arent even against the qwen image model and some even defend itWhat did this historic revisionist gaslighting troony mean by this?
Maybe you should go back to that AGP baker's thread and stay there.
>>106149953>no argumentno problem retard
>>106149917Tried and failed, who will seed the torrents, you? no? didn't think so.
>>106149971You will never reliably be able to seed that many LoRAs for any reasonable amount of time. And volunteering to do so would just make you a server like civit already was.
>>106149991>Tried and failedworks on my machine, sis
I don't think that's a safe way to drive miss...
>>106150003>no argument personal opinionholy shit retard, just stop replying
Are amd users not able to gen at all or they just need to use linux?
>>106149929>superfluous non-token shitUse a sys prompt that cuts down on flowery language. I generate a long, detailed prompt (or image description) first and then condense it down quite a bit to a specific token length (typically 192). It really cuts back on "noise" in the prompt.
>>106149907>China fatiguesam, kill yourself NOW
>>106150005As in it has been tried and it failed because no one will seed the torrents.
>>106150015What's your argument?
>>106149949>chinks didn't train jap kanjito be expected desu
>>106150020You can only do the most basic things and then cope everywhere online that everything actually just works so another sucker buys ayymd before realizing half the projects dont work at all and the rest are so slow you might as well have spent the extra money for a 3090
>>106149991Exactly. Call it something SFW and have another server link to it with obfuscated javascript, it's how stuff like animepahe works.
>>106150003Wrong. How do you think private trackers and usenet indexers work with huge libraries of movies and shows? A lot of these movies are over 50GB in size too.
>>106150020If you card supports rocm then you are fine. I've seen people say they got a 9700xt working in windows but it was really slow (around 3080 speeds).
>>106150021>Use a sys prompt that cuts down on flowery languagedoesn't work. it's always canned safe outputs with stupid filler. just fucking prompt
>>106150030>As in it has been tried and it failed because no one will seed the torrents.Feel free to post a lora that you can't find anywhere to download online
>>106150029jew fatigue is still ongoing as always. asian jew fatigue is just vogue right now
>>106149984thx. looks good
>>106150035Is there any point in upgrading beyond a 3090 at this point? Even the 5090 doesn't seem that much better for the money.
>>106149991>I'm black so everyone else must be too!
>>106150075Unless you have a lot of money or earn from AI, no
>>106150042Usenet backbones are paid services. Private torrent trackers are backed by paid seedboxes.
>>106150059You first, post the torrent.
>>106150033Nah it's censored. It has no problem with "ไปฒ้ๅ
" (among friends)
>>1061500753090 is starting to get the short end of the software stick.
>>106150075just buy a RTX Pro 6000 Blackwell
>>106150086>Usenet backbones are paid services. Private torrent trackers are backed by paid seedboxes.Yes, and?
The price for both of them is absolutely worth having a reliable repository for NSFW loras and models. It's not even that expensive.
>>106150086>"torrents failed!">post a lora you cant find online>"y-you first!"smartest anti-torrenting npc retard
>>106150087Try it without the ใ
>>106150075speed. A 4090 is so much faster than a 3090 it is silly. And if you are thinking about "value" in relation to a 5090 you are being a retard, you don't buy top end gear anywhere because it is priced well lol, you buy it because it is the best and fuck poor people.
>>106150113I knew you were a smelly poorfag.
>>106150097It has no problem with just "ๅบ" (go out)
>>106150127If you want the torrents then you can make the torrents and pay for the seedboxes
>>106150116Main thing is you really do want fp8 support nowadays, it really speeds up using quants.
Best deal is the 4090D 48GB.
>>106150145That wasn't my question you poor smelly idiot.
>>106150116>4090 is so much faster than a 3090 it is silly3+ times higher price for the same vram amount, the same memory bandwidth, and higher speed of only image generation where the compute matters more than in llms that is still not even even double the speed let alone 3 times
>>106150150does fp8 support speed up anything when using Q8?
>>106150048>doesn't workIf you don't know how to use an LLM just say so. No shame in that, they're weird and difficult to control when it comes to output specificity.
>>106150086>>106150094>no reply>seething in the next thread insteadoh no no no
>>106147300It's a combined Mogao (Seedream 3.0) and 4o dataset. The images are practically identical to what those two models give you depending on what style you ask for.
>>106149948It's bigger anon. It ain't that much slower than Chroma anyway. I find it doesn't need as many steps for a "quality" gen. Chroma I usually run at 40 steps, qwen 20 steps is plenty.
>>106147846Kontext is the best way to remove backgrounds from images locally I've found. Way better than rembg or inspyre. You can even tell it what color to change it to, great for chroma key.
>>106149887Could you please share, saar?