Discussion of Free and Open Source Text-to-Image Models
Prev:
>>105562525https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
>Models, LoRAs, & Upscalershttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
>Cookhttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>ChromaTraining: https://rentry.org/mvu52t46
>WanX (video)https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1
>MiscShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Archive: https://rentry.org/sdg-link
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Bakery: https://rentry.org/ldgcollage | https://www.befunky.com/create/collage/
Local Model Meta: https://rentry.org/localmodelsmeta
>Neighbourshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/celeb+ai>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>still no way to train chroma except with diffusion pipe on linux/wsl
GAY
>>105566558this but unironic.
the easier lora training is the more the masses will come.
On another board anon informed me that this thread is blessed. Can someone tell me if that is true?
Cursed thread of tech stagnation
>>105566569>>105566558no one will train a lora for a model that gets a new checkpoint every 4 days and which just started training with a bigger resolution that will need multiple epochs to have results
>>105566590>no oneI will, fuck nuts
It only takes an hour or so to train flux LoRA's on my GPU
>>105566569>gatekeeping is actually bad hm... i dont fink so anonie
Blessed thread of friendship.
>friendship
Found the tourist.
>>105565292yes.
>>105565315hi comfy. I am NOT a dramaposter, I don't have a grudge against you. I just want to ask, what do you want us to provide so we can merge this in?
https://github.com/comfyanonymous/ComfyUI/pull/7965
IE, do you want an XY chart for more evidence the workflow is better? Update it so both workflows are recommended? great free software maintainers don't just do whatever people ask, but they do set expectations so we know what you require before merging. thanks.
>>105566598sure, you can train a lora for yourself in the meantime but nothing that wont be obsolete when chroma finishes training
Blessed thread of frenship
>>105566619+1 for chroma, sooner or later this will need to be merged anyway
>>105566530 (OP)Thank you for baking this thread, anon.
>>105566634Thank you for blessing this thread, anon.
>>105566658>shitty turd colour>shitty aesthetics>apple logointo the trash it goes, and into the trash you also go. into the trash it all goes.
>>105566619suddenly tsunami waves begin to fill the background, the farmer frank's head explodes into chucks of brain matter and blood.
Oh no anon.
>still no flux dev kontext
>>105566751But you have lots of other toys to play with.
>>105566744i'm getting wood
>>105566751bro even the biggest pro version was mid, let alone the 5x worse one they said they will open source who knows when, what the fuck are you expecting
>>105566777NTA, Your number checks out. I believe you.
There we go, old VACE wan 14b without skip layer guidance looping workflow, frames are blurry (the generated frames are in the middle)
>>105566807vs now with skip layer guidance, same seed
>>105566530 (OP)Is https://github.com/FizzleDorf/AniStudio/ the best UI for someone that wants something that runs bare metal instead of Python or a web UI?
>>105566829Sad attempt at astroturfing bro, just let it go.
>>105566706i was just seeing if chroma could draw an imac g3, jeez
>>105566885i think i'm going to throw up.
>>105566885hans get the sledge hammer.
summary of what does this shit do for wan?
>>105566928Affects motion in a way that Alibaba never really explained except for "leave that shit at the default"
>>105566928Fuzz amount for movement. Higher = fuzzier but moves more
>>105566928controls how fast the gremlins inside your gpu move
>>105566777>trip 7s on a doompostit's over
why skip layer guidance no work with comfy core?
>>105566812holy shit niiiiiiice
>>105566953works with kijais nodes
>>105566812trying to fuck around with the online version ofwan. it never seems to stop generating.
>>105566928Stop picking at it or it might become infected.
>>105566829Wait another couple years for it to be feature complete
>snubbed again
/ldg/ has fallen
Is inpainting an empty space and redrawing/editing existing pixels the same? Or does each need its own type of work?
>>105567092my pic wasn't added in the collage
>>105567078real inpainting fills the masked area with latent noise and attempts to paint in in a way that fits the surrounding area meant to. Problem is most models aren't very good at inpainting and so anons blame comfyui. I real inpaint model just works except that it will need further processing, also real inpaint models require 1.00 denoise strength to actually work.
just set higher batch, walk away and come back and pick the best result and then use controlnet and or just normal low denoise with none inpaint model. to fix the image.
>>105567078there is no real good inpaint though, better off just using a latent noise mask with a decent model imo.
>>105567088Cool, share lora?
a stone statue of Miku Hatsune on a pedestal, in a museum in Tokyo.
chroma 36 detailed, pretty neat, hidream cant get stone right
flux fill is annoying because it tries to fill the whole mask with what you prompt instead of removing crap you were trying to get rid of around the thing
>>105567092I think I posted a good one last thread iirc. Collage nigger snubbed me (unacceptable)
>>105567105aww, don't cwy, widdle smeckles!
dere's awways next twime!
>>105567110I'll prolly just open krita nad redraw the finger...
>>105567139sounds like cfg way too high
>>105567141Which post? I'll tell you if it's good
>>105567078it's the same when you use inpainting model
it's trained so you prompt whatever is in the mask not the whole image
>>105567078>>105567110Look up these nodes for inpainting: Inpaint Crop, and Inpaint Stich. It has some useful settings like a context mask -- this knob expands the context in a separate output so it blends in will with the final output. Its also very fast because it crops a box around the part you want to mask
>>105567144no today was suppose to be my day
For VACE Wan, does sageattention and fp16 accumulation conflict with the 3 nodes below?
>>105567147There are ways and means, some require more involvement than others. You could crop in gimp, paste as new image but increase its size and use inside of comfy, then once done processing, re-scale and paste back into the exact place, finale 2 passes to fix.
but its time and effort.
its far better to get the prompt and model used correct the first time because you're gonna wanna second pass it anyway.
marble instead of stone
hidream also doesn't do this material as well.
133
md5: ff1df42531095b3b17fa7ab45984bcf6
🔍
>>105567222This? I have that, but when I tell it make it into a finger it's either 50% noise vaguely finger shaped or alien 4d matter.
can HunyuanVideo-Avatar be used in comfy yet?
>>105567387i never use that set of nodes, if you mean
>>105567169i've tried that and i really did not like it. Actually inpaint models do better on their own using native comfy nodes but the result will look like it was implanted from another world. Different lighting, skin tone etc. So you have to do more passes and when you do you will lose overall composition.
if i really want to fix something i will import into gimp, then manually using mostly the clone and heal tool paint the damn thing as close as i can. then i will denoise that area same seed maybe. A lot would actually go into fixing a finger because your really are fighting the dataset.
don't fight a models data set, pick the right model, then use another model to convert.
>>105567302Wooden statue? Clay sculpture? Ice sculpture?
>>105567491ice sculpture:
yeah, chroma is pretty awesome, hidream really struggled with diff materials (same type of prompt)
>>105567582prompt was: an ice sculpture of Miku Hatsune with a twintails hairstyle standing on a pedestal, in a museum in Tokyo.
seed: 1061947909606121
what's the best comfy workflow for image to image (denoise) ? with chroma
XCseY0GJ
md5: 76c2eb34c2fd956149deb53ce80685a4
🔍
>>105567613and last, wooden sculpture:
>>105567607Nice work, Fran-anon~!
>>105567653This is just a woman, nigga
>>105567694looks like an old hag
Guys, be generous to a humble guy like me and make mini girls measuring 15-30 cm, please.
>>105567725btw, is it just me or does wan not have a fisheye view?
>>105567880can you do pocket pussy miku?
>>105567880nice refraction
>>105567896I'm seconding this notion.
>>105567880chroma is on another fucking level
and it's not even finished
were there other big lewd models that used natural language captions before?
>>105568149What do you mean? It's version 34 already.
is each version an epoch? how many until it is "done"?
>>105568184He initially said 50, but that was before Ponyfag gave him free access to the clusters he uses, so who knows when it ends
>>105568189Do you think those two furries engage in filthy sex together? I mean he has to pay somehow. Two sodomites are the only hope for community trained models. Grim.
>>105568201honestly furries have spurred some of the most amazing progress in ai so far
their fantasies can never be fulfilled irl so it makes some sense
>>105568189He probably gave him access because the writing is on the wall for pony v7 (it sucks ass) and he needs a better base model to train on. As soon as Chroma fully converges, Pony v8 will start training on it.
best of the last 10 threads imo
>>105511079>>105511695>>105512575>>105513313>>105514073>>105514524>>105514648>>105514721cont'd next post, they auto-detect too many links as spam now which makes this almost impossible to do
>>105568158Omnigen was trained on an unfiltered dataset, including danbooru.
It just never picked up steam cause it's trained using the SDXL VAE instead of the Flux one.
>>105568244Fuck off with the spam, nobody gives a fuck what you think is 'best', retard.
>>105568218Do you think ponyfag will learn anything? His disclaimer about v7 was already alarming enough few months ago.
>>105568244Thank you, I appreciate your efforts.
>>105568149im actually impressed, the ice sculpture one turned out way better than hidream which cant do the materials well. it's a fun model.
>>105568281Tensor based logarithms are still pretty much off.
So is the chroma also some model on civitai that you slap into comfy or do you need an entire new set of tools?
>>105558215>>105564324>>105564509>>105565097>>105565164>>105565504>>105565694>>105565725>>105566004 idk what you prompted but it just looks like turning the cfg up to 11
>>105566224And I can't include two more links but the first Hatsune Miku sculpture and Debo's bloodstained driveway from this thread.
That's the end. Thank you everyone for posting gens
>>105568300i fucking love these, based hudson river anon
>>105568305haha good eye
I've been trying Pieter Claesz too (this is the same seed), he only painted still life and not landscapes but the model seems to do a clever job of transferring his dark classical style even though he never painted this kind of thing
211
md5: 6cf41b81ec1d3f4142ff421cc02a8512
🔍
kek
>>105568316pieter claesz is p. good
can chroma do darker stuff like beksinski? if it can i'm buying a 3090/4090 and it's not even a discussion
>>105568319dunno
they're just comfymaxx
I've hit peak, I can never improve on this, the perfect woman
>>105568332Yeah I think the first image avoids AI-isms like those fractal like repeating little leaves and such which are often so common with painterly gens. It's pretty coherent.
Second image pretty much loses that something.
>>105568339I seen dudes in prison more attractive than this
>>105568355I think I also used a different (less detailed) noise scheduler for the one you dislike so maybe that's what you're noticing, like it's a bit muddier on the leaves because it's less 'complete'
>>105567147Get Krita Diffusion and you never have to leave Krita
686
md5: 6f1d646562a23a429c4102a77a80683d
🔍
>>105568402I guess I have to setup something from within comfy too?
>>105568396second image has some strange leaves around, but man does it do a good job of translating what the artist "would've" done if he were to paint landscapes
i personally prefer russian realists (shishkin in particular which i believe you've mentioned) but this stuff is good as long as the lighting is moody (which is usually the weakest part about HRS paintings, at least in my opinion; the subjects are great)
>>105568396It's just an observation. Can't often do anything but reroll.
I did some paintings with flux (just basic flux style) then upscaling them with sdxl to generate a painterly style but it's a hit and miss.
Too much denoise and it'll fuck up the details, too little and it's not nice enough.
adetailer can really be a piece of shit at times.
>>105568462There are custom nodes involved but it should be able to connect without them I think
https://docs.interstice.cloud/comfyui-setup/
>>105568478try john atkinson grimshaw
dude made some insanely moody works and was obsessed with withered trees
>>105568496Thanks, saving up the name. I'll get back to my setup in a day or two. Can't gen every day or it'll become meaningless and going to burn out.
>>105568496nice, I'll have to try that name later
>A dark, dramatic Frank Frazetta painting. The painting depicts a clearing in a dense, shady forest, surrounded by trees. In the clearing is a rock with a sword hilt embedded in it, the blade of the sword buried deep into the rock. A narrow blue stream trickles past the rock. A bolt of lightning is striking the top of the sword hilt.>neg: cgi, 3d, text, muddy, messy, abstract
Was glancing over Krita Diffusion docs and regarding Cumfy, does it use some predefined node setup whenever user submits something from Krita? I mean what if I want to manually optimize or change its parameters inside Comfy? I'm not going to run some geezer's redditor workflows blindly out of the box. God knows how slow they are.
>>105568561>Frank Frazettayou have kino taste anon, let me repeat
take a favorite irl of mine, by Rafail Levitsky
fits this mood you're going for
First time in a long while I've had to post in here.
>>105568578it's extremely clunky because you have to use the predefined as you said but have to constantly change shit in the krita config menus which glitch out randomly
>>105568583yeah that looks nice, ty
>>105568586Maybe I'll try it out myself. I'm somewhat intrigued. I found this
>https://docs.interstice.cloud/custom-graph/So I guess it's doable.
On other hand, I have already done scribbles and funny images, then used img2img setup in Comfy. Saving an image from Photoshop and then doing a quick img2img or controlnet thing is pretty quick and simple because I have bunch of my own setups in Comfy.
In this sense Krita might be an overkill but let's see.
>"back in the old days, we didn't have none of that controlnet and inpainting crap -- and still we could produce better stuff than any of you gooners ever will!"
any tutorials on how to do multi-i2v with wan vace? all ive seen is V2V, regular i2v, and regular t2v. nothing actually interesting
what cfg do you like with wan i2v bwos
Fuck Im addicted to genning with chroma and wan, Ill leave some gens for the night and instead of going to sleep Ill just wait for the results awake
Ive not been this addicted since the sd1.5 days
https://youtu.be/9FL7IZavt1I?si=DvG103u8Qi5JbCPZ&t=307
comfyui mentioned by fatty again (5:07)
>>105568876to just generate a normal? why? that function is built into all the other software they use including blender
>>105568760She was stunning, jav goat
386
md5: 957171903d931607393a8593b96d3bb8
🔍
>>105568550>>105568493Gay shit doesn't work even though I have it installed
>>105568949Who? I just prompted for a generic "Japanese woman" with a 1980's aesthetic.
how many clips do you need for wan lora
>>105569001For a person, none. You only need images, the more varied, the better. I use 100-200, but 50 or so might work. For a concept, 10-30 low res videos, 50-150 higher res images. Concepts always benefit from mixed datasets, with lower res videos and higher res images.
is this a meme?
>For LTX-Video and English language Wan2.1 users, you need prompt extension to unlock the full model performance. Please follow the instruction of Wan2.1 and set --use_prompt_extend while running inference.
>>105568718You're not going to like it.
https://files.catbox.moe/f2crfm.png
>>105569001me? for the vid loras I've done wanx already had a basic understanding of the concept, it just needed a guiding hand to make it consistent. 15 low res vids up to 80 frames works, with some pics thrown in for detail
>Learn about CauseVid/Phantom for Wan
>Wow, 3 times the speed boost sounds aweso-
>cfg 1
>produces flux slop-tier skin
T-thanks
this one looks good? sampler/scheduler in the filename
>>105568963AND NOW THE KRITA LOCAL SERVER INSTALLS ALL THE PYTHON SHIT FOR ITSELF AGAIN I DONT HAVE ROOM FOR THIS SHIT IAM KMS
"syndrome" in negative helps chroma produce better looking people, reduces weird/ugly face gens
I discovered that trick with sdxl base and some other base-ish models, funny to find it useful again. less filtered base datasets probably include a bunch of medical textbook imagery so it makes sense I guess
>>105569048Looks good to me. How many steps / cfg?
>>105568963>even though I have it installedare you sure you installed all the custom nodes properly? put your reading glasses on and give the docs a nice good read
>https://docs.interstice.cloud/comfyui-setup/>>105569052you can use your existing comfy instance by giving it your server url. it's what i do
>>105569065interesting. gonna try that out. just need to find a way to supress the huegmouf (tm)
>>1055690883.5/25. I went down with the cfg
967
md5: 44e2012f732724c945928b520df60a1f
🔍
>>105569097It's pulling fucking identical shit that I already have and that the log says are missing .
bottom feeding schizo has permeated the thread
>>105569037May I see the comparison?
>>105569200why is it so sdxl
>>105569116did you install those nodes through the manager or manually using git pull? i get issues sometimes when manually installing nodes because of the comfy security stuff
>>105569236because I used sdxl
kek
in the first version of this still life the fat parrot was a watermelon
but then I added an art style lora that seems to have had a lot of birds in the TD images, because it always wants to turn stuff into birds
>>105569267i dunno then bro
does chroma feature the famous buttchin by default
>>105569282replaced by futa cocks
>>105569271what lora? i like birds
>>105569290https://civitai.com/models/726480/fluxcolorscape
what happened to chroma
why is it bad now
>>105569236Because sdxl gets the job done and has been perfected enough. Gone through multiple mixes over the course of many months.
>>105569347it wasn't trained by the chinese
lean
md5: 77455024ac2479158f05db09715ada3f
🔍
>>105569052It's a pain to install first time but totally worth the effort. I haven't touched spaghetti in months.
>>105569236all you have to do is look at the background to see the difference.
sdxl produces such nonsense backgrounds
>Amateur point of view photo
>ask for cosplay of anime character
>get the most SD 1.5 anime slopped image ever
lodestone is failbaking.
>>105569444It's going to suck ass at anime until it's done training and someone finetunes it the way Illustrious did for SDXL. A 5M mixed dataset will never beat a 13M+ pure anime dataset that contains the entire contents of multiple boorus.
>>105569485>It's going to suck ass at animethat's the thing, i'm telling it to gen realistically, but prompting for an anime char realistically instead churns out an anime style, too
>>105569353can you share the workflow of that one anon
>>105569444it's not there yet. too much body horror, too many outtakes. I mean, "girl sitting on a couch", how hard can it be.
Dang this is way better to use than trying to to do anything with images post-gen in comfy.
>>105569636you can't post that and not catbox the image anon
>>105569636IT REALLY WHIPS THE LLAMAS ASS
>>105569609>it's not there yet>14 epochs leftsurely everything is fixed...
for g
md5: aa2dabed746a95d3104cafbdeeee0de2
🔍
>>105569658There's still that inpaint discoloration visible tho eh. Can't post whole since it's /aco/ /d/ stuff.
KwwFgaqU
md5: 16afe83a0de43860249b7ef1fc5f79ea
🔍
>>105569595https://files.catbox.moe/fh5ozh.png
https://files.catbox.moe/9az7hp.png
https://files.catbox.moe/p94rm2.png
>>105569021>(mossacannibalis:0)> 0the fuck are you doing
Capture3
md5: c383b3cc65054363bfcfe6a1e335ef19
🔍
>>105569819I have a preset, default prompts, that I build upon for each new subject. Rather than delete the prompt, I just shut it off.
>>105569839from my testing, (something:0) still influences the prompt. rabbit hole go
>>105569494That should be an issue with Flux, not with Chroma. Try the prompt on regular V36 version, I suspect V36 detailed is more slopped.
>>105569962Also helps to add anime etc... into neg. Never had issue genning photoreal mikus etc..
>>105569543nice. prompts/Loras for this?
>>105569854Gooners in the shell.
>>105569826How many steps is that? It's likely a step count issue, or maybe the prompt itself. Never had such issues.
>>105567147>kritaIts that anon again, i was a bit high and drunk last night and did not see that. it looks interesting, i think i will have to check that out.
>>105569636it looks like something we all need.
uncanny valley mergedtodeath sameface fake smile plastic fuckdoll, take 1065
>>105569971no lora, just the base model.
"an organic portal to the dark world, surrealism, by Zdzislaw Beksinski"
neg "cgi, 3d, text, sepia, fire"
I posted a screencap of the upscaling workflow a while ago.
>>10556999225, it may also be resolution related? I mean I got a few servicable ones. the prompt was really basic but with a photographer wildcard just to see which ones work, might interfere as well.
>>105568250>His disclaimer about v7 was already alarming enough few months agowhat disclaimer?
>>105570146that's just his day job. once he clocks out, it's sex, bugs, and rock & roll.
>>105570146Still prefer him over a jeet
Roaches
md5: 51c9b88d0f0c581a921ae2693cd3c6fc
🔍
>>105570269That's probably a good call.
>>105567618Marina Abramovic as a kid
me
md5: 0e75c61238a87d6befc08701fbcb0378
🔍
>>105566530 (OP)Real talk : Does RTX 4060 or 5060 16GB (NOT 8GB) is enough for Video Generation ?
>>105570467Its night, sirs
Anyway, is it enough ?
>>105570471Please do the needful, Thank You
>>1055704615060 16gb is much better because of memory bandwidth , 16 is ok for video but its not great, especially long term
>>105570461Yes.
Wan Q8 with a bit of offloading.
Wan Q6 with no offloading, I think. Possibly 1-2 for i2v.
>https://huggingface.co/city96/Wan2.1-I2V-14B-480P-gguf/tree/mainPosted this before, but general rule of thumb for loading any AI model is that the filesize is how much VRAM is required to just load the model, though you can offset it a little with virtual VRAM/offloading to RAM/CPU.
That doesn't count loading things like the text encoder or whatever else a model requires on top of that, or inference/generating. The more frames and higher the res, the more VRAM it needs too.
>>105570461> does is> plenkhello sar
>>105567694yeah? did juwanna mann?
>>105570498>Wan Q6 with no offloading, I thinkWith a 16GB 5070ti and 32gb of RAM and using the fp16 t5xxl I still have to offload a little at Q6 (I get AllocationOnDevice crashes after every 4 videos)
You really need 64gb of ram nowadays to have fun with your computer
>reforger
>want to learn comfy instead
>look up tutorials on youtube
>"hey guys, faggotface mcshill here, here's a simplified piece of shit guide that only serves to shill my fucking patreon, subscribe to my patreon for exclusive access to the REAL tutorial
>close video, click another
>"sup guys, amerilard mcskidmark here, here's my poorly explained and meandering six hour long explanation on how to add a single node to your workflow, so you're gonna wanna go ahead and-"
>click another video
>"hello sar, poojat mcstreetshitter here-"
>>105570601>>105567653I need to buy new GPU.
Im sick of being cucked with Online Video generator
>>105570645> meandering six hour long explanationI hate video tutorials for exactly that reason. I wish guides and tutorials were always in writing. Even when a video helps illustrate something, I would have the videos be short, silent, display only and the actual knowledge imparted through words. I can read way faster than it takes to watch a video.
>>105570645>https://rentry.org/localmodelsmeta#comfyuiThere's a good tutorial linked in there with no patreon bullshittery.
>>105570645>>105570670Just use an AI YouTube summarizer
>>105570696Training data for that prompt would be cooking tutorials/shows, stuff like that. Same deal with how it loves to make anime girls talk, because anime girls rarely stand around doing nothing in anime. They're either talking or fighting
>>105570696>why do they always talkWhat's the prompt? I always have my 1girls just standing there and smiling unless I prompt for "talking to the camera"
>>105570696because you didn't give them something else to do with their mouth
887
md5: aeefd46a8e14a7bec9bd90ce9f575259
🔍
Nigga are you for real
>>105570696women tend to do that
>>105570778>adds api node>surprised pikachu face>it costs money?!
>>105570778Comfy must have decided he was through giving free labor to a pack of ingrates
>>105570787now if kijai could only follow suit
>>105570645yeah I wouldn't want to learn how to use comfy now. it's got a steep learning curve and that shitty yt algorhythm doesn't help.
check out:
https://www.youtube.com/@latentvision/videos (ComfyUI: Advanced Understanding) (this guy is behind ipadapter for comfyui)
https://www.youtube.com/@drltdata/videos (maker of the impact pack, lots of shit here, he doesn't talk tho. but good place to learn about how to set up facedetailer)
>>105570778ahahahahaha
>daily Julien seethe sesh
Just stop schizo
https://huggingface.co/gdhe17/Self-Forcing#training
What's stopping someone from renting a cluster of GPUs and training self forcing for the 14B model? Other than waiting two more weeks for someone else to do it and saving your money
For FaceDetailer in Comfy, is there a way to set it to do both face and hand? bbox_detector only accepts one input. Do I really have to have the output image flow into another facedetailer with the hand model instead?
>>105570838afraid so.
>>105570836good idea. I was gonna post more links but seem to draw a blank. control alt ai did a few very detailed in-depth videos for comfyui but they just stopped making content. who else is there anyways (who isn't just about hype) ? sarikas, kamph? lol
>>105570838There's probably custom extensions you could combine to do it but why bother. You want them separate anyway so you can control the denoise individually.
anyone have experience making a chroma lora? how long on average would you say does it take to train?
Guns fucking suck without loras.
>>105571081cool gen lol. is this lora-less? I mean it's a rifle.
>>105571149Yeah but guns are fucking basic shapes slapped together, even thought you get often a pasasble stuff, you also often get SPAS warped with Remington, double mag ARs and bendy AKs. And loras exist only for the most known or meme guns so no comprehensive libraray.
file
md5: 900d4a1ba1d401f88a5524b42645421d
🔍
>>105570838You can concatenate SEGS, then connect both to a "Detailer (SEGS)".
https://pastebin.com/raw/te36DLSR
file
md5: a8827df0b69bb3da312da4f600a1afef
🔍
>>105570838You can concatenate SEGS, then connect both to a "Detailer (SEGS)".
https://pastebin.com/raw/te36DLSR (embed)
>>105571209Based, thanks for sharing. I haven't been using Comfy for too long, I didn't know you could do this
>>105570838The easiest way would be to use swarmui
>>105571171I understand. I did some testing with flux last year and it knows some guns, certainly, but that won't help you here. what is this, illustrious, noob?
>>105571209damn that is.. smart. but what about the prompt? just promptless or dual use
I want some carrot cake god damnit
>>105571209this is the reason I browse these threads. you are the soul of /ldg/
>>105570645yes, ai is grifter paradise, and comfy is the most widely used ui.
What's a good amount of steps for near perfect quality for Wan 14B? I do not care about processing time. I want the most amount of quality until diminishing returns is reached
>>105571276would higher steps start introducing artifacts like multiple limbs,etc? I've had good results with 65
>>105571266>wants to learn about a piece of software>immediately seeks out "tech" influencers on youtube instead of reading the (free) manualI'll never understand this. do zoomers just not know how to parse information larger than a tweet?
>>105571275its not just about the steps, go through all nodes in the ldg's 720p wan workflow and see what can be improved, fp16 weights for everything, 0.13 teacache, output crf 0 and then reencode it with ffmpeg to what you need later
>>105571322your boomer is showing. this is how people get their info now. short form click bait videos
>>105571322>>105571339it's how zoomshits get info, it's not how normal smart people get info
>c-chat are we c-cooked?yes, yes you are :)
>>105568394Yeah, but you're into dudes, so that doesn't really count
Shit's fucking called SEGS lmao
Someone cook before miguman does it pls.
>>105571322Reading a manual is like reading a wiki page on something.
Extremely non-intuitive.
https://files.catbox.moe/juk6sj.json
My modified video 2 looped video "Loop Anything with Wan2.1 VACE" (https://openart.ai/workflows/nomadoor/loop-anything-with-wan21-vace/qz02Zb3yrF11GKYi6vdu)
ComfyUI workflow with Skip Layer Guidance and set up for maximum output quality I could get.
160s/it on a 3090
You can easily speed it up by around 35% with higher TeaCache but in my experience it changes the colors/quality noticably, but might be fine for your use case.
>>105571625Middle frames of
>>105566812were generated by this workflow along with the same prompt used to create the original input video with /ldg/'s i2v Wan workflow.
>>105566812>>105566807Please for the love of god post the workflow or catbox the video for us.
>>105571644I just did right above you
>>105571652Bless you, sorry, i was catching up and getting ready to go out, bless you again Anon!
>>105571015What are you using? Maybe you need to swap for a better model.
>>105571336>0.13 teacacheis this a typo or secret knowledge? the text in the workflow says that 0.14 is the correct option you should go with for highest quality/slowest
>>105571322Manuals are sequential. You generally want to hop from one thing to another out of a long list of possibilities, i.e. you want a walkthrough. Back when it was possible to search for things on the internet you would find a blog breaking down the steps. Now all you can do is search a video title and hope it's not a TTS slideshow or the most Indian man you've ever heard.
>>105570834a few thousand dollars
what the fuck is this?
https://civitai.com/models/1032887/acceleraterillustriouspaseer?modelVersionId=1895163
it claims to just increase speed by 50% with no explanation or anything?
>>105571723>>105571739nvm i saw the screenshot. for i2v the "highest" setting is 0.13, while for t2v its 0.14
otaru sama please choose me.
>there is no XL character lora for Filia from Skullgirls
how in the fuck
How do you get default "masterpiece, etc" shit out of the way on ComfyUI?
On forge I used styles for this.
how do i get the fisheye view in wan? i tried every trigger i knew. it was supposed to be easy
In case people didn't notice a even better wan2.1 speed up merge released
https://huggingface.co/vrgamedevgirl84/Wan14BT2VFusioniX
https://civitai.com/models/1663553?modelVersionId=1883296
>>105571691Still SDXL since most of the coughdiapercough loras is for it. I only make nongoon gens on the side.
otaru when are going to play with me?
Project page: https://seed.bytedance.com/en/seedance
Paper: https://arxiv.org/pdf/2506.09113
>>105571322To be fair, the documentation on some of these repos can be atrocious.
Having a visual aid when devs assume you can read their mind can be useful when they haven't bothered to update the manual.
>>105571924my nigga skullgirls is fairy popular and has decent amount of lora support. check civitai, most of them are trained from pony but seem to work fine for me. Imo its an abomination that's there no proper pine (bombergirl) illustrious lora yet.
https://civitai.com/collections/10048518
>>105572022You could try illustrious, noobai or animagine. They're sdxl-adjacent and illustrious probably caught up on loras the most out of those. Also you might want to mess around you inference settings, looks like your cfg is too high, or the loras you use need lower weights to reduce deepfrying the images.
>>105572234I mean I have Illustrious, which is SDXL.
>>105571733sure, tutorials are great. comfy just felt self-explanatory because of how easy it is to share workflows, add and manage nodes and models, etc compared to some of the other node-based suites. plus there's a ton of info in discussion and issues on github.
it just seems like people might be unnecessarily frustrating themselves in tar pits like youtube to avoid just following the directions, leaving them forever stuck on "how do I get started".
with that in mind, I guess it'd be helpful to explain that part of the point of highly flexibly frameworks is to not get in the way of high-functioning users, and that dumbing it down any more would damage its utility. which is ironic, since one of comfy's problems is that it's too basic.
>>105571939good question. try a prompt styler node. some include presets (in json), others are just CSV loaders.
I personally just create a notes node that I keep a few examples in, but you can also just create a fleet of prompt nodes.
>>105572246Sorry for the missunderstanding. Well, I hope the cfg or lora weight tips help out a bit.
>>105572255I imagine workflows deprecate quickly so nobody can really debug it properly. the comfy environment has been too volatile since the org began which is why there has been more doomers. even ani gave up on comfy's future. it's nothing more than a plugin to more useful apps but even then the experience using comfy in blender and krita is just jank and shitty
>>105572334I do hope ani provides a good solution to this but it looks like a ridiculously hard task especially alone. he got this far already so he really is our only bet on better software in the future
gradio
md5: 72326645a55985126feb00c8a4c27d6c
🔍
for the vramlets lurking: I'm going to host a self forcing gradio so (You) can have fun with it and share the link in the next thread. just finishing up some things
>julien shills his shitty wrapper again in hopes of free devs
>wahhhhh shill comfy more!!!!! don't try to replace it!!!!
>>105572120>shitty closed source model>it's not even as good as Veo 3At this point I do hope Google destroys the competition if they decide to be closed source.
>>105572334>even ani gave up on comfy's future.Pathetic kek
>>105567110>real inpaint models require 1.00 denoise strength to actually workNot true.
>>105567653Me on the right.
what are some good checkpoints or loras that can create indoor/outdoor scenes?
I tried SD/Pony/IL but it cannot make a regular room without many artifacts.
I also tried img2img which for some reason just sucks in comfyui. And I also tried controlnet with canny/depth and some sketches i made but again, it failed.
The models on civitai barely look passable...
Any ideas?
What's the trick to hunyuan i2v taking the image as a first frame reference instead of just doing a sort of img2im and generating something different that just resembles the image?
>>105571234>damn that is.. smart. but what about the prompt? just promptless or dual useWell...I guess promptless could work.
I don't usually use detailer for hands, so I don't know...just came up with this solution since the anon asked for it.