Discussion of Free and Open Source Text-to-Image/Video Models
Prev:
>>106185803https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows
>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanXhttps://github.com/Wan-Video
2.2 Guide: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y
>Chromahttps://huggingface.co/lodestones/Chroma1-Base/tree/main
Training: https://rentry.org/mvu52t46
>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage
>Neighbourshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
Chroma thread of disappointment
Chroma thread of happiness
file
md5: 216e5545117b61f46a53b9d99bca4404
๐
Blessed thread of frenship
blessed thread of frenzone ;3
>>106189589<3
looks like he is gonna make a low step version then next is likely wan2.2
>>106190450 (OP)made the catalog again ;3
its like a fun game for me to see if the schizoid detects me or not muehehee
>>106190464>this epoch for sure!>just waitmaxx bro!>this epoch for sure!>just waitmaxx bro!>this epoch for sure!>ok we are out of epochs but finetunes and loras are coming to save us!>just waitmaxx bro!>just wait...>anytime now, it will be amazing!
>>106190464>>106190466So what even is the right Chroma version now?
https://huggingface.co/lodestones/Chroma1-HD/tree/main
https://huggingface.co/silveroxides/Chroma-GGUF/tree/main/chroma-unlocked-v50-annealed
>>106190488no one knows what annealed is but definition wise its prob a version made for finetuning?
>>106190488>chroma-1hdjust a renamed v50
>ggufversion for poorfags who should off themselves
>>106190486Chroma has already been good for awhile
Tried a Chroma v44 trained lora against v49, v50, v50 annealed respectively.
Technically v49 should be best since it's 'closest' to v44, and I think it kind of is, v50 is very close and 'v50 annealed' had the worst likeness.
>>106190450 (OP)>Rinko Kobayakawacute, CUUTE!
>>106190502okay, but no one knows what annealed is?
>>106190509makes sense, its gonna be the most offput from V44
does 2.2 have a fucking memory leak or something? why does it become basically non functional after 2-3 gens
>>10619052164GB+ ram or if will have major issues when swapping
>>106190509okay so 50 annealed looks more realistic, got it
>>106190542? just more smaller details cause that was the entire point, and its HD / V50, annealed I think is one made for finetuning
Hey hey Anon, Anon here.
Quick comparison of Euler on Chroma-HD-Annealed and Chroma-HD.
Since I'm retarded, the annealed version is on the LEFT of the plots.
Got that?
Left: Chroma-HD-Annealed.safetensors
Right: Chroma-HD.safetensors
More stuff to come, surely.
Buh-bye.
Titty fairy: https://files.catbox.moe/0uuist.png
Trash girl: https://files.catbox.moe/f42ft0.png
Rat girl: https://files.catbox.moe/tv35gr.png
Mixed media: https://files.catbox.moe/jl7vfm.png
Oil painting: https://files.catbox.moe/vfjiaj.png
Has any of you had any luck with this version of radial attention? https://github.com/woct0rdho/ComfyUI-RadialAttn if so
- Does gen stay consistent over 10 seconds?
- Is it faster?
I have sage2.2 and its already pretty fast, just curious
>noise? check
>nonsensical details? check
>blurry? check
>vintage analogcore y2k slop? check
it's chromapostin time!
>>106190476nice to see you again
>>106190593said no one ever
This is a popular thread I see.
I thought Qwen-Image could do image editing like Kontext? Why isn't there any workflow for that?
>>106190568>VRit sucks anon
unless your vision is SO shitty you cant see the pixels\lens-smear\bezel
How do I prevent ComfyUI from eating all the RAM when changing wan loras? I've tried different "unload" and "free memory" nodes, toggling smart memory, but it still eats extra 20GBs and I have to restart manually. It's unbearable.
>>106190593>tripcode-userhe never left dumbass
>>106190601the schizo baker just talks to himself to race to the bump limit ;3
>>106190593<3
>>106190562Mixed media errors out.
>>106190600its not that serious mate ;3
>>106190633Whatever helps you sleep at night :]
>>106190615>bro how do I modify the contents of a 20 GB file without using any memory?
>>106190640> When it comes to image editing, Qwen-Image goes far beyond simple adjustments. It enables advanced operations such as style transfer, object insertion or removal, detail enhancement, text editing within images, and even human pose manipulationโall with intuitive input and coherent output. This level of control brings professional-grade editing within reach of everyday users.now what?
Feels like i should pay some visits to n*pt again
Bad doggo!
>>106190509Yeah the third looks the most natural out of the three.
>>106190562vfjiaj is broken too.
>>106190653my favorite is when he debates himself ;3
>>106190615invest in nvme m.2
reboot entire pc in 6 seconds
>>106190483it was the soup spoon ;3
>>106190655you skipped the relevant part
>>106190615>how do I fix a rat's next of poor memory management, dependancy hell and bloat?
>>106190654Are you retarded?
1. this is not how lora works
2. it modifies, not duplicates
3. it eats few GBs after each lora change
>>106190687>>106190638Maybe its my internet I dunno
>>106190562>>106190638FUCK
Mixed: https://file.garden/aIdN6xfH0QVghCy0/annealed_normal_mixed.png
ALERT
chroma-unlocked-v50-flash-heun.safetensors
flash version use heun 10 steps CFG=1
4 minutes ago
>>106190697What do you suppose a LoRA does you retard
Why would you think a LoRA might use memory whenever its changed? Maybe it's because it's doing math on a multi billion parameter matrix?
>>106190692> We are thrilled to release Qwen-Image, an image generation foundation model in the Qwen series that achieves significant advances in complex text rendering and precise image editing. Experiments show strong general capabilities in both image generation and editing, with exceptional performance in text rendering, especially for Chinese.Anything else?
>>106190709>flashis this distilled
file
md5: cbea0609346aba2897a9c345644959eb
๐
>>106190721If qwenimg wasn't slopped it would be seedream/mogao tier
>>106190738>if local wasn't slop it would be good!we knew this since midjourney
>>106190743Well I can't wait for your serious finetune you release for free.
>>106190447>>106190471Thanks. Didn't expect qwen to stubbornly try to make sense of so many purely danbooru tags.
>>106190689What else do you like about the baker?
>>106190743are they still being sued
>testing simple porn i2v prompts
>no matter what, if a woman spreads her legs she begins touching her vagina
>even if i tell it specifically to not have her do that
>thats assuming it listens to the instructions regarding a bare vagina in the first place
am i fucking retarded? is it a lora issue?
>>106190759can she do something else with her hands? the devils playthings n all
>>106190759Do you know the concept of how prompts work? >Don't think about the elephant.
It does not "think". It only knows images that are associated with think and with elephants. Everything should be prompted positively, so you describe her hands doing something other than touch her vagina.
you know the latent preview that shows in a node?
what if it were to the side of it, or above it?
do we have the knowledge to do this?
>>106190713Retard. It does not add new layers and weights.
>>106190714the editing model isn't out yet. learn to fucking read.
>>106190789Retard, when you do math on anything it must put them in memory. Are you really this programming illterate?
x(model) * y(lora)
both must be in memory to do the calculation
>follow 2.2 guide in op
>i2v workflow works fine
>t2v doesnt
do i need extra shit the guide just doesn't mention?
>>106190615if you pulled recently, it has a really comfy memory leak.
>>106190808> what is overwrite
>>106190813I have faith that you can figure out t2i
Just think about it, think hard
>>106190824Think really hard about why this might be a problem if you keep changing LoRAs.
>>106190768>>106190773i must be retarded because my attempts at a nsfw prompt are just creating sfw content you could post on fucking tiktok.
it honestly has really solid motion it's just... missing the point?
>postive prompt [spoiler]The woman spreads her legs, revealing her bare vagina. She has no underwear on. She uses her hands to support herself. Her toes wiggle.[/spoiler]
>negative prompt[spoiler]่ฒ่ฐ่ณไธฝ,่ฟๆ,้ๆ,็ป่ๆจก็ณไธๆธ
,ๅญๅน,้ฃๆ ผ,ไฝๅ,็ปไฝ,็ป้ข,้ๆญข,ๆดไฝๅ็ฐ,ๆๅทฎ่ดจ้,ไฝ่ดจ้,JPEGๅ็ผฉๆฎ็,ไธ้็,ๆฎ็ผบ็,ๅคไฝ็ๆๆ,็ปๅพไธๅฅฝ็ๆ้จ,็ปๅพไธๅฅฝ็่ธ้จ,็ธๅฝข็,ๆฏๅฎน็,ๅฝขๆ็ธๅฝข็่ขไฝ,ๆๆ่ๅ,้ๆญขไธๅจ็็ป้ข,ๆไนฑ็่ๆฏ,ไธๆก่
ฟ,่ๆฏไบบๅพๅค,ๅ็่ตฐ, camera shake, touching vagina, masturbating[/spoiler]
I want my chroma danbooru finetune, NOW
More seriously, we only managed to get an underbaked lumina 2 finetune recently, and an auraflow one later this year, what makes you think someone would finetune chroma with weeb stuff, which would cost much more than both auraflow and lumina put together?
At least for a weeb finetune I'm not holding my breath for a chroma illust, I don't know about realistic stuff, they might have more money
>>106190813>replace I2V models in model loaders with T2V ones>replace I2V loras in lora loaders with T2V ones>replace WanImageToVideo with EmptyLatentVideoI can see what kind of people complain about comfyui being too much
>>106190887Cause finetunes on those needed to make the model learn nsfw from scratch, chroma already knows that and tons of nsfw concepts so they just need like 1000 images of the style
>>106190887It knows everything, a simple LoRA of what's relevant to you will make it do what you want. There's an insane false expectation that any model can do anything for everything for everyone. Even your brain works this way. There is always going to be signal to noise and the only way to get hyper-competent is to FOCUS.
>ram usage goes up 30gb around vae decode with qwen image
cant imagine having less than 128gb ram
>>106190929More than 64gb is for pretentious fags
>>106190882>positive [...] underwear>negative [...] vagina
>>106190851So following your logic weights taking 10GBs in RAM will require another 10GBs to apply a lora in order to prevent the restoring original weights problem?
>>106190908>1000 actually a hundred should be enough I would think
>>10619093764GB is bare minimum for using video models / text models
>>106190937routinely need more than 64gb with video gen
>>106190887auraflow and lumina guys clearly misunderstood how you are supposed to train a model
you just have to train a lora for every single artist style and sex position
>>106190942lets give it a whirl without that
>>106190937>looks at wan2.2, At 8bit that is 15GB for low noise + 15GB for high noise + 12GB for T5 + 1GB for vae + a few GBs for lorasAnd im probably forgetting some stuff
>>106190947Yes, pretty much. If your model takes 10GB in RAM and you want to safely apply a LoRA, you have two real options:
- Keep a copy of the original weights (another ~10GB) so you can restore them without reloading,
or
- Reload the base model every time before applying a new LoRA, which still means allocating that memory again.
If you donโt do either, and just keep stacking LoRAs, the math piles up and you end up with corrupted weights.
The only alternative would be some kind of tensor streaming method to apply LoRAs on-the-fly, but thatโs complex, slower, and easier to break. Most systems donโt bother, it's simpler and safer to reload and no one frankly cares about someone who can't afford 64 GB of RAM.
And then you'll just bitch about how long it takes to change a LoRA.
>>106190986+ a few GB for web browser cause that shit eats ram + a few GB for OS
>>106190986>>106190996oh and if you want to use the qwen prompt extender that is like 5GB more
>>106190969Where's your dataset again with every sex position properly labeled with thousands of examples and every artist style someone could ever want to prompt also accurately labeled? Based on what you said, you must know it exists so could you link it?
>>106191015didn't chroma guy make it public? I heard he used gemini pro + whatever the dan tagger was
>>106191015im sure someone (that is not me) who is going to train all those things into chroma will be right on it
>>106191027Holy shit you actually think that's accurate enough for some anon with a missionary position fetish where the cum is tastefully dripping from her larger than average clit and painted by J.C. Leyendecker from his 1930s grey period?
you can finetune layers of chroma. it is entirely possible to do this on a commercial 24gb gpu
>>106191036I mean yes, I'll make a LoRA of things that are relevant to me and then I won't share it, please understand.
>>106191040what? it does some freaky art styles
>>106191040>missionary position where the cum is tastefully dripping from her larger than average clit and painted by J.C. Leyendecker from his 1930s grey period?I can imagine the kino soul. Best gen ITT.
>>106190882holy fucking sloppa
>>106191040if you've ever used chroma you'd know this is entirely possible and quite easy. just don't be a promptlet and i don't mean that as bait, you just have to know how to prompt by using joycaption or gemini to get a rough idea of how.
>>106191042i will be sure to enjoy your finetune in 2 years
>>106190888this is what i get using the provided workflow in the 2.2 op. ive already tried changing the models anon.
I'm gonna sound like a dummy for saying this, but have you guys ever tried upscaling your videos? They don't just get bigger, they get sharper and better. It's amazing.
>>106190988So ok, I've applied the lora A and now there are W` (W x lora A) and W (original weights) taking 10GBs each in RAM (20GBs in total)
After that I've applied the lora B, so there are now W` (W x lora A), W , W`` (W x lora B) and W again in RAM, taking 10GBs each (40GBs in total)
Correct?
>>106190882>่ฒ่ฐ่ณไธฝ,่ฟๆ,้ๆ,็ป่ๆจก็ณไธๆธ
,ๅญๅน,้ฃๆ ผ,ไฝๅ,็ปไฝ,็ป้ข,้ๆญข,ๆดไฝๅ็ฐ,ๆๅทฎ่ดจ้,ไฝ่ดจ้,JPEGๅ็ผฉๆฎ็,ไธ้็,ๆฎ็ผบ็,ๅคไฝ็ๆๆ,็ปๅพไธๅฅฝ็ๆ้จ,็ปๅพไธๅฅฝ็่ธ้จ,็ธๅฝข็,ๆฏๅฎน็,ๅฝขๆ็ธๅฝข็่ขไฝ,ๆๆ่ๅ,้ๆญขไธๅจ็็ป้ข,ๆไนฑ็่ๆฏ,ไธๆก่
ฟ,่ๆฏไบบๅพๅค,ๅ็่ตฐ>spoilerBEAHAGAHAGAH
>>106191079They always come out with that nasty filtered look for me, what do you use?
>>106190942removed that, the underwear is still present, and she's back to touching her still covered vagina for whatever reason. probably too nsfw to post even though everything is covered for context
>>106191079720x1280 wan is already good enough, no need for more, and if you need more, u use starlight mini with pirated topaz
>>106191100https://files.catbox.moe/kgmpyn.mp4
https://huggingface.co/thaakeno46/ComfyUI-Workflows/blob/main/WAN2.2_GGUF_UPSCALER_14B.json
>>106191124looks like the original video was too shit already, use another workflow
>>106191124https://files.catbox.moe/9flfi8.mp4
>>106191100>VLC play video fullscreen>record screen with OBS done
>>106191089isnt the chinese sloppa pretty standard for negatives to not take up too much space
>>106191142put your trip back on
>>106191124Nice upscale, but how would it look on a human though ?
>>106190701Alright thanks.
>>106191081No, not quite. It's unlikely each LoRA keeps a full copy of the model in memory.
More realistically:
- The base model sits in RAM (10GB)
- Applying LoRA A modifies the weights in-place or overlays them, using some memory but not another full 10GB
- Applying LoRA B likely reuses the same memory used for LoRA Aโs operations, unless itโs larger, in which case it may allocate a bit more
You're not stacking full model copies, but here's the catch: If you donโt reset or reload the base model before applying a new LoRA, youโre doing math on already-modified weights. That leads to cumulative errors or corrupted outputs.
So when switching LoRAs, you have to
- Unload the modified model from memory
- Reload the base model from disk
- Then apply the new LoRA(s) cleanly
And yes, that reload process is rarely 100 percent memory-efficient. Youโll probably get memory creep until the garbage collector kicks in and clears unused objects.
Something Comfy does, and you could call it a feature, is keep models dangling in memory when it can. Thatโs why something like Wan, assuming it fits in RAM, can be switched to instantly, and why you can go from Wan to Chroma nearly instantly as well if you've loaded the models once.
>>106191158https://files.catbox.moe/21sly3.mp4
>>106190498Makes sense to me. I've tested both a bit and I am going to stay with v50 for generation.
file
md5: 88fd3aca46196bcd57c6fc16c2fccd94
๐
chroma bros...
>>106191223just train a lora
>>106190908I don't know, I think the community would rather use an sdxl model with already fucktons of characters in the model itself, and just train loras for specific character/styles, than build the missing 1M of characters and styles from scratch with loras, even if the later is a better model
Dunno about multiple characters bleeding and style mixing with chroma, maybe a weeb finetune would beat NAI
Upgrading my vram has made me realize what insects vramlets are
vramchads of today are tomorrows vramlets
>>106191247>t. went from 1060 3gb to 1060 6gb (laptop)
>>106191257Such is the the circle of life
>>106191247Truer words have never been spoken. Vramlet's are lower than the scum to be scraped off my boot.
>>106191238and? people still use 1.5 cause they are poor with only 4GB of vram
>>106191166> So when switching LoRAs, you have to> - Unload the modified model from memory> - Reload the base model from disk> - Then apply the new LoRA(s) cleanlyBut original weights W in the memory already right next to the modified W`?
>>106190808>>106190851>>106190988So why first you claimed that these extra 20GBs eaten by ComfyUI is how it's supposed to be and now say it's a "feature"?
And how to turn off that "feature"?
>>106191296At this point I don't care. Write a patch since you are so smart. I really don't really care what someone who can't afford RAM for $50 thinks.
using grok to elaborate on simple prompts, seems to work
A sleek, futuristic car interior from the driver's seat perspective, Ryan Gosling gripping the steering wheel with quiet intensity, his face bathed in the vibrant glow of a high-tech dashboard, driving down a rain-slicked street at night, surrounded by towering neon signs in vivid pinks, blues, and greens, casting reflections on the wet pavement, a massive, bold sign with the text "LDG" in glowing white letters dominating the skyline, dystopian urban buildings in the background, cinematic lighting, ultra-detailed, in the gritty, high-tech noir style of Blade Runner 2077, immersive and atmospheric.
>>106191073I'm pretty sure you need fp8 scaled clip if you're using fp8 scaled checkpoints
So is nu chroma good or should I just kill myself
>>106191348Goose is a /ldg/ lurker, I knew it!
>>106191348there is a node called qwen extender that works good as well and runs local. Otherwise gemini is the best cloud model to use if you want to go that route, there is a node for that as well
>>106191358It's good, but you should still kys since you can't run it on your moms laptop
>>106191358yes, the details issue was fixed, it is now by far the best base model, now give it a few weeks for anime finetunes to drop though it already does it well imo
>>106191348another one, "make a stable diffusion prompt for Ryan Gosling making a pizza with a wood stove in a bakery." became this:
A rustic yet modern bakery interior, Ryan Gosling rolling out pizza dough with focused intensity, standing beside a glowing wood-fired stove, its warm orange light casting flickering shadows, flour-dusted countertops cluttered with fresh ingredients like tomatoes, basil, and mozzarella, the air filled with the aroma of baking dough, wooden beams and exposed brick walls in the background, soft natural light streaming through large windows, ultra-detailed, in a cozy, artisanal style, warm and atmospheric.
kek
I have 12gb of VRAM and 32gb of RAM, should I bother with Wan?
>>106191399if you dont mind waiting 30+ mins a gen since you will be constantly loading and unloading your quanted models from the hard drive
>>106191399I did, but now I have 24gb VRAM and 80gb RAM
food for thought
>>106191327Yes, because you don't know how lora works.
> can't afford RAM for $50And how extra RAM will save from the need to restart? These 32GBs will be eaten after changing loras 6 times in the worst case. Not speaking of lack of free slots.
>>106191246thanks. No idea where it says that
>>106191387Hollywood is all fake
Wan really tries to make the prompt make sense, what a trooper
>>106191399At q4 wan 2.1 or merged 2.2 will work almost fine.
>>106191405How do you end up with 80gb system ram ?
>>106191415Guess they have a different readme on huggingface
https://github.com/QwenLM/Qwen-Image
>>10619143832 + 32 + 8 + 8
>>106191352do you know where to find something like that? ive looked at the links int he op and im not seeing anything.
>>106191445Doesn't that fuck up timings ?
>>106191474he will prob see a 10-20% decrease in speed on top of it being the lower out of whatever they are but I guess that is better than not having it, ram is still far far faster than even a SSD
>using woct0rdo radial attention
>have every thing recommended installed (even information that should be on the github read me)
>errors
>"if X error appears just restart comfy"
>restart comfy
>radial attention working
>fuck yeah its at 75% done
>errors with no description
Well, at least his sageattention 2.2 is very good
>>106191534Are you feeling comfortable?
wan is fun but kontext is fun for edits/fast meme images as well, or you can kontext edit an image to use with wan, they are all tools that do something a bit different.
green cartoon frog is sitting in a white beach chair on a sunny beach holding a beer. they are wearing a blue shirt, and red shorts. keep the frog's expression the same. the ocean is nearby, and the sky is cloudy.
source is a regular pepe, kontext q8 model
kontext anon is going to commit seppuku once qwen edit drops
I wonder why they didn't just release an "all in one" version of qwenimg instead of making it two separate releases. Technically you can generate images from the scratch with an editing model, right?
>>106191579because nobody would be talking about it right now
in the industry it's called a teaser, this is a business after all
>>106191553Won't he just switch to qwen ?
>>106191579Releasing model weights is just marketing for SaaS
See
>>106191246 DashScope, WaveSpeed and LibLib are all Chinese companies btw
>>106191590if only things were that simple
>>106191587???
It would generate a lot more buzz than it is currently
If they released a "complete" version right away, people would be spamming social media with "guys look at this amazing open source model that can generate pretty images AND can edit them on top of that!"
Now people will just handle the edit model as an incremental update
>>106191542yea all good bro
>>106191625you seem pretty eager
>>106191468Just google umt5 fp8 scaled? Don't confuse it with t5xxl though, that's for flux
>>106191548fishin
green cartoon frog is sitting on a fishing boat, holding a fishing rod. the frog is wearing a blue shirt, and red shorts. keep the frog's expression the same. the sky is cloudy.
I hope they release a 1.1 version of Qwen-Image with less slopped outputs. It would be easy for them to do it, it's aesthetic tuning/like a lora. They likely won't though.
i'd love to see a "prompt travel" type of wan workflow, if it only does 5 seconds, have say 4 lines/prompts you enter, and it generates each one and stitches them sequentially.
yes, you can do it with ffmpeg or whatever, I mean something you can click "prompt" with and it does all of it.
>>106191579all in one models are shittier
>>106191688ani has to add it. he's the only one to work on it
>>106191688This does exactly that, and it's one of the best extended video nodes/workflows I've seen so far: https://github.com/bbaudio-2025/ComfyUI-SuperUltimateVaceTools/
How do I filter results so I can actually download something on CivArchive? 99% are dead ends with no mirrors.
>>106191828should be taken out of the sticky then
are there any alternatives?
>>106191847I'll give you analternative
green cartoon frog is holding a bag of potato chips that say "SIPS" on the bag in scribbled text with one hand, and his other hand is holding a potato chip. the frog is standing on a sunny beach. the frog is wearing a blue shirt, and red shorts. keep the frog's expression the same. the sky is cloudy.
>>106191909one more
green cartoon frog is sitting in a beach chair on a sunny beach, under a palm tree. the frog is holding a tropical drink in a margarita glass. the frog is wearing a blue shirt, and red shorts. keep the frog's expression the same. the sky is cloudy.
is there a reason t2v takes way longer than i2v?
>>106191876>>106191948is this a ani/comfy erp?
has anyone been running hunyuan world in comfyui yet?
im using the wan 2.2 guide and using self forcing https://rentry.org/wan22ldgguide. i thoguht 2.2 was suppose to be 24 frames per second, but the video i generate seem to be 16 fps like wan 2.1
>>106191982that's not the only 2.2 model they put out
inpaint bros.. whats our model of choice?
invoke vs forge inpainting? anything better
ioinpaint?
krita is kinda limited in some ways but ok
>Chroma still mangles feet and hands
Nice
I know the best inpaint method but I won't share. Enjoy being steps behind me
>>106192132It's OK, this place welcomes browns too
Anon wasn't lying
AniStudio just hard crashes when you try to gen a second picture
Oh well
>>106192120yeah, I noticed that too with the latest release
reducing the parameter count was a mistake
how the hell are you supposed to know what to write to induce animation
>A video of a side-view running animation of a voluptuous elf-like blonde woman.
and she doesn't move.
>>106192139this time with a grok'd prompt:
A hyper-realistic scene of a frustrated man, face contorted in anger, forcefully throwing a pizza against a textured wall, sauce and cheese splattering chaotically, dimly lit rustic kitchen, moody atmosphere, dramatic lighting, cinematic composition, detailed textures, intense emotion, 4k resolution.
>>106192230and a happier variation:
grok 2 is getting released next week, thoughts?
>>106192199how about you post a workfloW?
>>106192254kontext test for fun: A man holding a square cardboard box saying "LDG" in black marker. keep his expression the same.
Recently I saw a couple of gemmy webms where george floyd and that cop were put in famous scenes from movies, the whole thing looked like wonky motion transfer. I think it had a watermark of some cloud service. Any local alternatives?
I don't get it... Chroma still looks like shit!
>>106192440wan2.2
works on 12gb vram 64gb ram
>>106192442pics with WF or it didn't happen
>>106192442chroma is dead, use WAN or Qwen
how many steps for high and low sampler do you usually use for t2v?
>>106192442>>106192494back for more are you?
Asking again. How do I make her punch more like a punch instead of a slow approach? I've tried words like high speed, aggressive, strength, strong, etc. in the prompting but it doesnt do too much
>>106192473I know wan, but it's much more sophisticated than what those seemed to be. Almost puppet warp animation overlaid on original video but more realistic, it could spin around and stuff.
>>106192543what I've resorted to is asking chatgpt to write my prompts for me (you need to tell it that it's a video model for animation though, otherwise it'll give you still-image prompts).
>>106192543edit the video externally or make a lora
>>106192442it looks like shit? just train a lora
it fucks up anatomy? just train a lora
its blurry? just train a lora
it doesnt know any anime artists? just train a lora
just train a lora
punch
md5: 9c51f55b3591c08203d4632076684e7e
๐
>>106192589>>106192580I don't know how to train a LoRA
>>106192543speed it up and cut out frames in the middle where fist is still
i look down at your shitty fetish
>>106192622>i look down at your shitty fetishnigga im new to this and just trying to make reaction images
>>106192673believe me when i say im using it for porn too but im practicing a bit on prompting and learning workflows with shit like this that i found on twitter
>>106192589none of those are issues thankfully
>>106192771exactly, for us chromagods they are features
>>106192440>>106192565ok this shit is called viggle.ai, can I do it on my computer?
Is there any info on what the actual minimum and maximum resolutions for Qwen are supposed to be, in terms of megapixels?
I'd hope the super weird ones on their model card (like who does 1328x1328 as square, why wouldn't it be 1344x1344 lol) aren't the only ones "officially" supported
This is all gonna matter a lot for e.g. Lora training and stuff
A 1024 base Lora wouldn't work that well at 1328
And so on
>>106192543(Fast movement. Quick motions:2)
>>106192846Flux works best at 1448x1448 so what's the difference
is t2i -> i2v generally better or is t2v better?
>>106192984post an example of what u want to make and say the magic word
>>106192950i2v is better. t2v has no use case
>>106193035so i should just generate an image with like flux then use i2v?
>>106193051do you only get one chance at this? why can't you experiment, this is supposed to be fun
>>106192806viggle my nuts lol
>>106193058i get plenty of chances but its obviously time consuming. just wanted to know what was best for making stupid videos or porn to post
anons lora still works just fine
the man is holding a birthday cake with candles. the top of the cake says "LDG" in red icing.
>>106192933No it doesn't lol, you need significant aid from ClownShark samplers to get non-deformed stuff from Flux above like 1344x1344 even
>>106193094/ldg/'s birthday was magical
here's an idea for a jam, MJ refusals
I saw Mayli IRL and laughed
using a source reviewbrah image without holding the pizza box:
>>106193160this workflow works pretty well
it's a json. not mine but I had it bookmarked.
https://gofile.io/d/faahF1
>>106193010Where is my deserved spoonfeeding?
why does ani hate waldorf so much?
>>106191135nfs 2012 remake looking good
>>106193194when did he say that?
>so they took away my favorite fast food item...
>>106192473>64gb ramoh, once again i got filtered
>>106193303post ur setup, u can probably make do with ggufs
i hate comfyui
>i hate comfyui
i hate comfyui
>i hate comfyui
i hate comfyui
>i hate comfyui
https://www.reddit.com/r/StableDiffusion/comments/1mkzdfx/wan_22_14b_image_to_video_seamless_concatenation/
interesting, will try
>>106193395https://github.com/radiatingreverberations/comfyui-workflows/blob/main/wan2.2-i2v-endframe/video_wan2_2_14B_i2v_endframe.md
I LOVE comfyui
Thank you comfyanonymous I love your fennec
I love to hate fennec girls
I LOVE comfyui
I HATE avatarfagging
>>106192950>>106193067i2v is easier to work with and less time consuming due to the source image already establishing the subject, scene, and style. Prompting is easier as well as you can just focus on describing the desired motion and camera movement.
>>106193474who is your favorite /ldg/ avatarfag?
any idea as to why this node wont download / install? ive clicked the button to install, downloaded the pack its from, etc. and it still says its missing
>>106193473the new colossus
>>106193337>>106193359Mirror's Edge SEXOOO!!!
>>106193507some custom nodes don't support windows and only work on linux
>>106193526so this workflow just isnt usable for me then? its one of the few t2i workflows that arent an overcrowded nightmare
>>106193526that is a lie
>>106193535update your custom nodes, press update all, also if its a nightly only thing for some pack you might have to use that instead
>>106193507You need arch linux
>>106193395add the kijai 2.2 i2v lora with 6 steps (3/3) and it works fast too:
>>106193535it is usable for you, just install debian
>>106193545see
>>106193516
>>106193573Sdg tranny, ignore
>>106193584>Sdg tranny, ignoreyou just replied to him
i2839
md5: 7135ab98a3169b6741eff8ddbf10b729
๐
>>106191655isn't that apu tho?
can local into realtime already?
for example like:
https://mirage.decart.ai/
>>106193621that's a world model and they are pretty much walking simulators. also yes local has some
>>106193650no, this does video to video in realtime. no interactive or 3d elements
>>106193669if youre trying to make chroma look good youre failing
You ever look back on old gens and realize how good they were? I'm not talking about anything ancient even, just XL gens from June.
>>106193662oh. idk v2v is lame as fuck but there is probably something out there
>>106193682Newer are always better
>>106193187i said magical word, not to call me a subhuman
>>106193278anime general on /g:
>now he wishes he was waldorf
truly sad
juli
md5: e05ebe1ad5ec0002c120a9bf4ad5c17e
๐
when this niggra not lurkin?
>>106190882How much VRAM do I need to run that kind of model? Does it run inside ComfyUI?
I can run SDXL no problem, but this...
new theory:
if i do TOO well he mass-reports my posts
i will now add generation\workflow data to ALL posts & nothing else.
>report bombing
y i k e s
https://desuarchive.org/g/thread/106190450/
>>106191847Nah, it's actually not 99%, most of the popular celeb LoRAs still work, downloaded many Belle Delphine ones this way.
>>106190450 (OP)Tried v50 (HD) and while the result has better prompt adherence (I asked for 4 women, it gets that right), and though fuzzy the finger count seems about right, the result is very poor in terms of texture compared to regular Chroma. Doesn't happen every image, but I noticed that it's happening to other anon's gens as well. Seems undetrained/blurred in many gens. Maybe the Chroma dev foresaw this downgrade in photorealism gens due to only giving this 2 epochs hence the 2 separate versions.
>>106194334is there V50 non HD? compare that too
>>106194334Okay, so from what I understand the Chroma dev shrank the dataset size for this version. Since now if I remove the neg entirely, I get a very fuzzy looking image for that prompt even when asking for a photograph. So my guess is that I need a much stronger neg (but I tried a bunch of new words and it barely improved), and the v50/v50 annealed are strongly biased towards drawings. Here is one of my usual footfag prompts I test with no changes in neg. Yeah, that is exactly what is happening.