Discussion of Free and Open Source Text-to-Image/Video Models
Prev:
>>105875103https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows/home
>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanX (video)Guide: https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1
>ChromaTraining: https://rentry.org/mvu52t46
>Illustrious1girl and beyond: https://rentry.org/comfyui_guide_1girl
Tag explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage | https://rentry.org/ldgtemplate
>Neighbourshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/celeb+ai>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
Blessed thread of frenship
I'm gonna bless this bread so good if it ends up nice and full of aesthetic gens.
>>105881999>There is a button "unload models" and another called "free model and node cache". These should reduce your vram footprint down to minimum.Where is the button located? I can't see any button on the UI for unloading models.
>>105881996fresh knowledge
>3 pics in collage
I'm just that good
>radial attention waiting room
>>105882001They're fine as long as you keep them low in the 2nd sampler
Secret sauce is... well, doing something https://civitai.com/models/1730935/secret-sauce-wan-21?modelVersionId=1959008
Also, what the fuck is this. All I did compared to the previous vid was change the resolution and alter the prompt a little, but there was nothing mentioning her tits, wtf
>>105882264inflatable booba for her safety out on the water
>>105881840>When my video gen finishes, I notice python is still consuming 10gb+ ram. Is there a way to just quickly stop comfy from occupying so much ram without restarting?Anyone? I'm not seeing the "Unload Model" button.
Anyone know of a good VACE inpainting workflow?
Kontext can be fun but is it stealing souls?
file
md5: fa45c3e4980bb29fb8c4cebf5990a8bb
🔍
>>105882313It's these two buttons. There's also some custom unload model nodes that you can hook up at the end of the workflow, but afaik they only unload vram, not system ram
o begin, the quonflab was meticulously calibrated against the blithering scalar, ensuring precise alignment of its snargle-ports. We then introduced the polymorphic goo, observing its immediate spluttering and subsequent frobniculation within the gribble-chamber. Initial readings from the woozle-meter indicated a curious fizzle-reflux, directly contradicting the pre-established flumph-hypothesis. Further analysis of the glurg-particles suggests an unexpected ditzle-shift in the chroniton field, which may necessitate a re-evaluation of the entire whiffle-paradigm.
>>105882878anime boys mayhaps
>>105882690is this i2v or t2v
is it wan or anisora or watever its called
Do you have to use video loras for wan? Can image loras work?
>>105882908You'd call it a quantum computer and I'm pretty sure that would fool 90% of the population
>>105882690Could you do a Blue Archves image to video?
>>105882969i want to rape her
big
>PromptTea improves caching for video diffusion models by adapting reuse thresholds based on prompt complexity. It introduces PCA-TeaCache (noise-reduced inputs, learned thresholds) and DynCFGCache (adaptive guidance reuse). Achieves up to 2.79× speedup with minimal quality loss.
https://github.com/zishen-ucap/PromptTea
>>105882986obsoleted by self forcing
>>105883000self cucking shits on the quality so much only the 8gb vramlet retards would ever say anything good about it
>>105883011mmmm.. nyo
i remember making videos with teacache in march/april and they were dogshit (at 6/20 steps) and took way more time (+cfg kek) to make
self forcing at 4steps (1 cfg) videos are way better
of course if i had a 3090 or something maybe i wouldnt use it
t. 3060 12gb god
comfy should be dragged out on the streets and shot
>>105883027>nyoah, the retard with the low IQ opinion AND low vram is the AGPtroon, many such cases
Cn one of you make my Saturday a happy one and convert image any blue archive student in a suggestive scene please?
>>105882690please I beg you :(
>105883045
this is either a jeet or a glow bait
>>105883045>>105883050isnt blue archive a 2d game?
>>105883045why doint you do it yourself, flux kontext literally works on 10gb vram with nunchaku
wan video
I can't gen why
it doesn't even matter
how much RAM offload I try
>>105883061>works on 10gb vram with nunchakuworks on 6 gb vram without nunchaku but you'll have to get the Q4 model
>>105883050No anon...I am a coomer, why do you think I use local AI? for cooding?
>>105883061Sorry I misspoke, I meant convert image to video! :(
>torrenting a Lora because it can't be downloaded anywhere now, all mirrors dead
>1 sneed, stuck at 0%
It's over. they won.
e81
md5: 0f64d5c11bf12e8a50ab01b56dd311b9
🔍
>>105883110why dont u share ur setup?
>>105883192What's the lora? Maybe I have it.
What happened to radial attention? Was it another nothingburger?
how do you prevent wan from applying a post-process effect to the final video? Sometimes it just randomly puts a textured glass filter on the video.
How do you solve the issue of something having no seeders? Like let's say that in the future all NSFW models are banned from every site and you have to torrent them. How do you make sure all of them are available to torrent?
>>105883261They've been working on nunchaku (needed to run radial attention) so, kinda yes and no.
https://github.com/mit-han-lab
Man, if we had proper video extension, I'd be happy with never bothering with another video model for a long while. While you can extend skyreels vace, the quality and details vaporize after 10 secs.
i'm free. i don't find image/video ai gens hot anymore. i deleted all my models. goodbye. i'm gonna go outside and touch real tits.
>>105882346can you catbox this please? i want to try making it in impressionist painting style
>>105883454>How do you solve the issue of something having no seedersbe the change you want to see in the world
>>105883454private trackers with seeding ratio
>>105883526The model generated a fake world for you to live and you haven't escaped yet. You need to wake up, anon.
>>105883247negative prompt is important depending on your prompt; if i have cum in the prompt i have to put "splash" and "vomit" in the negatives
i can't do a comparison atm but if you just genned this, can you try another one with adding "minimal motion, perfect lighting, in the morning, criterion collection"? I've been using those to good success
>>105883547Imagine how autistic the interview processes would eventually get.
>>105883454A seedbox. Alternatively you can even forego torrenting and use IRC or even Usenet to share files.
>>105883562>i can't do a comparison atm but if you just genned this, can you try another one with adding "minimal motion, perfect lighting, in the morning, criterion collection"? I've been using those to good successWait, so tags work in wan2.1 prompts? I was under the impression you're supposed to describe what occurs in the video. Where should the tags be placed? A second paragraph after the description?
>>105883614I'm no expert but:
>a tag is just a token expressed without padding language>padding words are also translated into tokens>I have no idea what the captions in the Wan training data look like, but presumably there are captions>prompt bleed still existsThe tags would work for terms which are obviously in the caption set, like "still" or "blocking shot" or "zoom". I also think there are tags like "criterion collection" or "perfect lighting" or "amazing cinematography". I think a sentence or two would still be the best way to describe the actual action, but you could use tags in a similar way to 1.5 where you can fine-tune the output with enough quality tag analogs. That's what I'm doing in my video gens, anyway, and it's working well
what is the best photo realistic model that isnt horny?
>>105883593>A seedbox.Yes, it's not very complicated. Just to make sure each torrent has at least one seed.
Also while the base models are large, those are all available on huggingface, meanwhile loras are ~100-300mb in size, so even if there's many of them, it's not a ton of space compared to other content like games and movies / tv shows etc.
Just a webpage with magnet links and images / text about the loras, that's all you need. Pirate sites have been doing this for 20+ years at least.
>>105883614There is no actual 'tag', it's just captions turned into tokens which are then part of the pattern recognition mechanism of ai training.
From the ai perspective, it doesn't matter if you train with the caption style:
A blonde woman is wearing a polka dot bikini and standing on a beach, it's a sunny day.
or:
blonde, woman, polka dot bikini, standing on beach, sunny day
The training will learn how to associate each of these captioning strategies with the training images as long as you are consistent and will give equally good inference results if you prompt accordingly.
The reason we are seeing a strong shift away from 'tag' style towards 'natural language' is because the latter is the way 99% of people who are asked to write a prompt for what they want, would write it, as the name 'natural language' implies.
>>105883951>The reason we are seeing a strong shift away from 'tag' style towards 'natural language' is because the latter is the way 99% of people who are asked to write a prompt for what they want, would write it, as the name 'natural language' implies.I agree with this post, but another reason for this is that people are using AI LLMs to generate their captions, which are extremely verbose and produce a paragraph-long caption for every picture, and others use the same LLM to write their proompts
>>105883951nothing to do with every new model inserting a LLM to "refine" your prompts btw, purely a coincidence
Does Civitai occasionally just shit itself and refuse to load at all for anyone else?
>>105884164>but another reason for this is that people are using AI LLMs to generate their captionsYes, but this started when the model creators began using ai captioning which generated these very verbose descriptions of the training images.
And since they trained on these extremely verbose captions generated by ai, in order to get the best results when generating, you need to write similarly extremely verbose captions, and so people ask the LLM to write these verbose captions because it's too time consuming.
...
0b
md5: 7bf5c32eba9043acc6b6822011c55b06
🔍
HEY
YOU
FUCKING
FAGGOT
is there a causvid/lightx2v lora version for i2v?
or are these exclusively for t2v?
(that isn't accvid)
>>105884305It's really slow at times, but then again it's probably the least optimized web page on the whole web, coupled with the worst navigation.
occasionally I see someone release two versions of a wan lora, t2v & i2v. if im only doing i2v, does it matter if i use the t2v version or not? I know with wan t2v you can't use i2v loras, but for i2v it shouldnt matter right
>>105884164Careful, anon this is something good!
Do you know how hard it was for me yesterday to create an image of my Waifu building a sandcastle on the beach with a "focused expression"?
Now I am in the same situation.
I want to create an image of my Waifu sewing a scarf, (it is literally impossible!).
There are things that you cannot do with tags, and they end up looking stiff or unnatural.
I hope this change from tags to prose is quick, because prose makes everything less deterministic and captures actions globally. I can say she is sewing a scarf, and the AI can automatically generate the needles, the ball of yarn, and the scarf without me having to hold them.
Or you can say the character is sad, and the AI can not only change her eyebrows to sad ones but also add more body gestures without me having to specify it explicitly.
>>105884279to old
>>105884164to old, 3dpd slop
Please refrain from letting your LLM loose here
>>105884481How many tags are there in Danbooru for "sewing a scarf?" LOL
>>105884512Ok could you please do it with BA students?
THANKS A LOT
>>105884562not him but why dont you do it yourself, flux kontext literally works on 10gb vram with nunchaku
>>105884579Anon, what part do you not understand that I want him to make a video? I already have many images made by myself. Did I miss something? Can Flux Kontext make videos?
>>105884507That's why it's important, so SDXL can think beyond itself and the tags. Surely, he already has 1 million images of people sewing a scarf in his database, but the tag sloppers nerffed him.
When will we be able to generate manga that consistently maintain character designs, and can generate speech bubbles and fx wherever required?
>>105884676Illustrious 3.5 will be able to use a natural language. Sadly, we will probably never get it unless a big donar drops $100k+.
>>105884788At this rate we won't ever even get v3.0. XL is dead.
what are the horse fuckers up to nowadays? baking anything or did they gave up?
>>105884828You mean Ponyv7 ? That ain't ever coming out, he dun goofed picking AuraFlow as base model and now he can't go back
>>105883547this is the only real way and has worked for a decade when it comes to media. private trackers incentivize people to seed. eventually you'll have a dedicated group of power seeders that keep everything alive.
>>105884813Eternal SDXL stagnant thread of death and despair
how would you fix the eyes on an illustrious model? Most of them seem out of shape.
I read to use "ADetailer" but I cannot find out more about it. What is that?
decade
md5: 7fe56237acdef93f7fbd2da2b3e34aa8
🔍
>>105884998inpainting them or hires fix, or wait two more weeks until a Kontext anime Lora, or SDXL 1.5
>>105885010i still havent been able to get in btn or ptp. i gave up. ipt for life
>>105884993>Stable Diffusion XL (SDXL) version 1.0, released in July 2023I started this hobby a month ago.
Are you local anons using the same model for two years??? KEK
file
md5: e6b7a87d8d72af9333f2dc99ba64ade6
🔍
>>105885011>hires fixit's already at 1024, blow it up to 2048?
this thing?
While /aicg/ was using ChatGPT turbo in 2023 and now there are millions of new and better models, we are still stuck in 2023 levels of inteligence....it's cursed hobby
>>105885036Yes, with a base image, it supports even larger resolutions.
>>105882109;3 maybeeee
i liek the friendzone, i like frens!
>>105879469>N0THiNG IS BEY0ND OUR REACH>>105883247avoid liquids, they always look yucky
>>105885079We need more monstrosities and bodyhorror in these threads desu.
file
md5: 7fa170092365996e2ba467b5cb31dc08
🔍
why is it that my middle mouse doesn't move the board anymore but instead moves my text-boxes???
>>105883029H O W
D A R E U
>>105885307can you stick to one name you absolute degenerate.
you don't matter as much as you think you do.
you don't need attention, you need help.
>>105885205works on my machine.
>>105885205You've probably installed the 'middle mouse moves your text boxes' plugin
Rookie mistake
>>105885344Stop stalking this guy, and no more skipping your medicine.
Can image loras be used with wan gens?
>>105885344>noo I can’t look at it! NoooDude seriously?
toobs
md5: a3a339fe458ee2a9d2a45e83177be94f
🔍
>Ask AI for 2B wearing a kimono
>It knows to put the designs from her dress onto the sleeves of her kimono
How does it know to do that?
>>105885551context/data attributes. the model can recognize patterns/attributes with clothing. when transforming her outfit into a kimono, it applies those attributes to the kimono, which is why its black and has the same pattern motif. it tends to do this by default unless you prompt specific features
What is Nunchaku? Why do I see everyone using it for Flux? From my 2 second google search, it seems to be some kind of speed optimization? Does it only work with Flux? What are the draw backs? Any quality loss?
>>105883029Whats the alternative?
A1111 died
Comfyui can do more things better
What is it? Reforge?
>>105885974SwarmUI is okay so far I guess. I can only complain about it having the occasional hiccup.
>>105886036It was horrible bloat last time I tried
>>105886036Nah that's far worse and it's not even close. ComfyUI isn't perfect but nothing else compares currently.
>>105886051Anything you don't need, you can easily hide into the advanced options tab, disable it alltogether, hide just the particular parameter or two you never use, or shuffle around their order.
>>105886068>far worseWdym? It's been doing me a favour some time now, and I've been a long time Forge enjoyer.
8eKB8_Fr
md5: 9a2f9fce90d6e9f729f0353e042781a2
🔍
>>105885974Forge and reForge should both be fine. Personally I just jump between UIs a lot depending on what I need to do
CY5kpUZK
md5: dc11270d7ee8e164ff424cdab0ae4e69
🔍
I just wish Invoke was less retarded, because UI design-wise it looked nice last time I checked. Metastable seemed nice too, but a bit slow to catch up with features already present in other UIs.
MFxUyI_A
md5: 4a5cc81836ed290c15536bccb3b3bc69
🔍
_cDy_dNZ
md5: c67badb5e7ccd1a4108bfd1bf268218c
🔍
Why this community didnt migrate? And we are still sloping with SDXL?
>>105886107I use reforge with gradio 4 main branh and is very good.
>>105886280Migrate to what?
>>105886280Speak for yourself, I went to Flux, then to Chroma and Wan
>>105886293To a newer stabble diffusion
>>105886316That's called Chroma
>>105886280CivitAI is plagued with SDXL Loras and checkpoints because 80% of them are from the third world. And you can SDXL from a phone.
>>105886280I guess you're an anime slopper for saying that. But if you look closely, the aesthetic quality of thousands of users contributing is far superior to the default quality of SD 3.5 Large or Flux, for example. Compare the anime generations in this thread, which were surely made with SDXL, and compare, for example, the image generations of ComfiAnon with the fennex foxgirl.
What would have to happen is a miracle, and someone would have to make SDXL's Loras and Checkpoints compatible with other architectures.
>>105886280It does not feature any copyrighted characters or style, you must create them yourself from scratch, it's tedious.
>>105886391M-maybe ComfiDev could make that?
>>105886107rare high-IQ anon
I only use forge, does reforge has any advantages over forge?
>>105886449couple more samplers and schedulers
>>105886449Yes, it has other built-in extensions, such as skimmed cfg, v pred, and a bunch of more modern samplers and schedulers that are supperior and modern. But it's best to use both.
>>105886280>Sigma near parity with SD3 big if true
Eternal Panchovix love
Eternal Ilyasvel love
Eternal ComfyAnon hate
>>105886467>that are supperiorwe never out-did euler
>>105886481Yes, 2M SDE ancestral (or whatever it's called with ancestral) is better and more creative than Euler.
Right now I'm using SDE 4M something something with a GitHub extension, and the image looks better, but it's riskier.
For general use, I recommend Euler a in the first step for creativity, and then in Hires, I use a modern sampler like the one I mentioned to add detail and consistency.
>>105886454>>105886467cool, thanks for the info, bros
for anyone else reading this, I prefer gradio to comfy as an UI, but the main reason I use forge is because it runs a lot smoother and I can work with other graphic apps while genning, while comfy rapes and collapses the whole system, specially when using flux
>no update on radial attention
i weep
I want to animate my friend's juicy ass getting dick, what should i download?
>>105886534you dont need to post this everyday as if the guy lurks here. its probably not coming until december.
file
md5: 16a37a5480f92a3f628a33225789a45c
🔍
Ai-chan is increasingly clever.
>>105886526Exactly, ComfyUI, the program itself is like running GTA VI without all the image generation processing.
A year ago, Comfy abandoned the idea of creating a local model user interface and focused on creating API nodes.
>>105886526(You) problem. I can use Photoshop + ComfyUI simultaneously without any lag or slow down.
Honestly getting sick of Forge users ramblings. They never post anything good.
>>105886553shut up nerd
*sobs
>>105886582if you're that desperate, DM the guy on git or something and ask for a realistic timeline.
Hell, maybe giving him a few hundred bucks will make him do it faster.
>>105886526Yes! Other than that, there is no tangible evidence that ComfiUI can do anything that Reforge or Forge cannot.
>>105886554Ok now aesthic quality :3
Stabilty MatrixGODS unbothered by this pleb wars
>>105884998face detailer node with bbox detector for eyes
https://civitai.com/models/150925/eyes-detection-adetailer
plenty of workflows and tutorials on using it
>>105886591kek, i'm just hype postan dude, the good things will take time. there's like 5 other optimizations that were announced months ago that probably wont see the light of day. in the mean time
*whimpers
>>105886572>>105886598the main issue with comfy is that it's a tool designed by a computer nerd (coomer) for other computer nerds (coomers), so it's specifically crafted for people that want to make animu waifus and have absolutely no clue about how to use any other graphic application
it's an assembly-line for people who want to generate 1000 images to get a few decent ones because that's their intuitive workflow, and that's where it all ends, so connecting 54 nodes to automatize this process makes total sense to them
meanwhile graphic artists have a complete different workflow/process, and the gradio interface is a lot more straightforward and convenient because we onlu use AI for the things AI is really good for, and we use other more appropriate tools for the rest
that's regarding the UI itself
and then there's the performance part, which obviously depends on everyone's own personal set-up
>>105886280I'm genning with neta-lumina however? Just two more weeks until they train more artist styles.
>>105886280>flux modelsToo large and slow. And for dev, the license is written broadly enough that it arguably bans NSFW.
>SD3The models are better, as base models, than people give them credit for. In particular SD3.5M is far more powerful than SDXL. But the license explicitly bans sexual content and doesn't allow for commercial use. Complete non-starter
>AuraflowTotal ass, I have no idea how it is this high on this mememark.
>PixartBretty good, but undertrained. Will probably always be limited by how small the model is.
>CascadeNot enough better than SDXL
The models that actually have the potential to replace SDXL are Lumina, Wan 1.3b, and Cosmos 2b. But nobody is doing large finetunes of them. That is slowly starting to change with very first anime finetunes of Lumina currently in progress. We'll see how that goes.
@mossacannibalis
Pixtral I'm using for captioning is more retarded than its gemma.
>>105886765>But the license explicitly bans sexual content and doesn't allow for commercial use. Ouch, and they can't generate an ecosystem with cryptos and XMR?
>>105886773Okay, now use tags and grim yourself with how well they understand the tags, and the prose is pure bloat that makes the model stupider.
>>105886765AAAAIIIEEE why are we downgrading in parameters?
>>105886761Ok and the aesthic? I can't goon with that.
>>105886856I know. But it's there for me to wait for. And RouWei dev said he'll frankenstein both a 16ch vae and a normal modern llm instead of clip. Let's say, this year. Don't know if he has a solid lead for the former, sounded like that. But for the latter, see https://github.com/NeuroSenko/ComfyUI_LLM_SDXL_Adapter it already works and he dropped some examples in the cord after asking for prompts.
>>105886888This shit is going to be the new Illustrious, in a good sense. When they're finished, however that happens, we have to wait 2 more weeks for NAI and WAI et cetera.
License? Yeah ... Artistic License.
>>105886902What depresses me is that we continue to build on SDXL, which is a 2023 model, and that's really poor. Its like keep building in llama1 for local textgen, it doesnt make sence there are structural flaws in SDXL.
I used RowWei with natural language, but unfortunately it doesn't work as shown in the CivitAI photos utilizing very easy to udberstand prompts, so I ended up using the old and reliable tags.
Don't like SDXL? Fine tune your own model.
>>105886902Great! Can I use that with any model? I want my waifu to knit a wool sweater with a stressed face, but right now it's impossible to indicate that to her using tags.
>>105886967Nice, looks like a 70s sci-fi illustration
>>105886995sorry, yeah . kinda. end of a too long img2img rescue mission.
>>105886996sort of. thanks. it is kind of slopy though. just trying to kick a slow thread.
>>105886967nice. what model is this?
>>105886988sdxl 1girl, solo, sitting, indoors, white t-shirt, distress, scared,
knitting, knitting red sweater, holding sweater, holding knitting needles,
I think it's more that it has no ability to depict sensible knitting, not that it can't separate or merge the concepts.
>>105883543https://files.catbox.moe/pjcw42.png
>>105887069>what modelflux1-schnell
sort of old school at this point
for our resident miku poster ;3
>>105887102Nice, what happened to mainframe girl, was she assimilated by the evil computer ?
>>105887127>mainframe girlthts lain
>>105886765>>105886339I will be fair with you, I'm not a thirdie and i have the resources to run up to 48 gbvram but I'm still using 1.5, pony, xl and barely started using illustrious.
Why?
Thousand of ready made loras and sliders for body shapes, hairs, etc.
It is also easier to debug what goes wrong when you have thousand of resources on top of it.
1.5 was good to me and i know how to use it. It also works well with controlnet.
Xl and pony were so-so. The quality got better but i didn't need 4k images. Thwy were very bad with controlnet.
Illustrious seems promising so far.
The thing i am seeing the most are enthusiats who barely even count as hobbyists that just go for the latest tech and really just dick around in it.
Right now, ai should focus on better efficiency rather than another model and to refine the existing models.
If a new model comes out, i cannot do much with it other than dick around with the basics.
>>105887194>'m not a thirdiestop acting like one then. If you are using 1.5 in current year you really need to expand your horizons and get out of your safe space.
>>105887194Ok you are candidate to use 3.5 large if size its not your concern. But yes CivitAI its plagued with thirdies that run A1111 in google colab.
>>105887214What do you mean?
>>105887221Good lighting, why gif though ?
>>105887290you are clinging to outdated tech because it is familiar. Flux and Noob vpred are so far superior to anything 1.5 and pony can do it is absurd to even use those models. Both models require you to rethink how you prompt, and you'll get a ton of gens that are trash, so you act like a bitch and stay with the familiar and safe old shit.
>>105887373Explain to me.
Noob, illustrous, RouWEI and Pony, there aren't SDXL finetunes?
>>105887380>i follow the chain of repliesI didn't look before, I just responded to your comment about being a bitch on 1.5 and pony. Now I see you are just a retard and I am angry at myself for wasting my time.
>>105887385Im different anon, an slopy one
>>105886902What discord are you talking about, anon?
I also want to be at the forefront. I'm bored of being here watching photo montage videos of Donald Trump or US celebrities.
Sex havers hobby, thats why there is no one on Saturdays,
Is there a place online where I can learn about new updates, like new samplers, noise schedulers, or integrated tools like CFG skimmer or v pred? How can I stay informed? I rarely see any news about new samplers here.
>>105887527its tough because so much shit is out there. I use grok to get news, just ask it what people are talking about, when something pops up you just ask more questions on what it is, how it works etc. I think oai, mr claude, and grok can all do it for you.
>>105886280Flux has been my primary for a year now
>>105887553But they have actuak news? For example if tomorrow RouWeI publish a decoder for blah blah, Claude wills earch the AI gens news?
>>1058875613dpd slopper, share your taylor swift gens
https://drive.google.com/file/d/1gwnEBM09h2jI2kgM-plsJ8Mm1JplZyrw/view
new flux kontext 2 image workflow with nunchaku, can swap it to multigpu gguf too I think
it's just a .json file.
>>105887633ehh
disregard, https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev#flux-1-kontext-dev-grouped-workflow still works best for 1 or 2 images.
>>105887385yeah i'm not this guy
>>105887400I get your point and I'm working on it to get out of the comfort zone.
>>105887592it can, yes. Grok is actually good at it because so many people post shit to twitter. Just learn how to prompt it, like any llm.
>>105887409omg. How does it just KNOW asuka?
>>105887672i did that to her
>>105887661Bad limb proportions
>>105886280I went to Flux. It's not big/too slow if you have a 4090 (which is basically the minimum now) and it's so easy to train compared to everything else, you can just set it and forget it. It's very malleable if you're willing and able to train it.
I do wonder about Chroma though and where it will ultimately sit on that graph. I have my doubts it will train as easily as Flux (Schnell doesn't) and it's output looks hit and miss... but that's just Flux in general if I'm being honest.
WELL YOU GOT YOURSELF CAUGHT
>hey hey ho ho
>schizoid baker has got to go
>beyblade ho ho
>>105887919you're a big warehouse...
>>105887912I run Flux q6 on my 2060 12GB and it's fine
a big fulfillment center for you.
Asking again about Kontext training. Does it just werk with a regular flux dataset and training parameters?
>>105888062No you need to use flux gym's training repo. It's different.
>>105888062kontext training is all about a/b. Primary image and then target. So, woman and then same pic but nude woman. Image and then same image but now pixel art. You train on a set of that kind of thing.
>>105888048that a BIG baker
for u
>noob knows pepe without needing a lora
>>105887767>he doesnt understand the thick squat warrior buildngmi
>that is one stinky shitty/farty BAKE!
..for you
>>105888148thats not pepe! I DONT KNOW YOU!!!
>clean it up jannies
crazy how many meme characters noob can do, yet this capability is rarely used. people would rather do kontext failgens I guess
>>105888182it's a single anon who also replies to himself. he uses the thread as a dumping ground for his garbage and says they are "examples"
chroma v44 is being uploaded.
>>105888212god I wish I were an ugly bastard in anime too
>>105888212I love this so much
>>105888205chroma is shit after v40. prompts are not respected. hope v44 is better
Tick tock chromasomes. Only a few days left until training is complete and it's looking a little... undercooked.
>>105888212how did you capture this video of my soul?
>new pony soon
wanna bet it's worse than noobai/illu (which is great)
Is it normal for a 5-second 480p video to take 8-10 minutes to generate on an RTX 5070 Ti? My card is also undervolted in MSI afterburner.
>>105888328It's almost guaranteed to be shit.
>>105888334How could anyone give you a meaningful answers without seeing what settings and workflow you're using? You have more money than sense. Give your GPU to me so it can be appreciated.
>>105888335doesnt matter to me regardless, I have wan for video, noob for anime (wainsfw does everything even before loras), and kontext for edits.
>>105888334use the rentry workflow and use the new lightx2 lora, it's super fast for generation. 4 steps, far lower gen time.
>>105888359https://rentry.org/wan21kjguide#lightx2v-nag-huge-speed-increase
how do you prompt wan? Literally I just want some light shafts and water ripples.
>>105888334a wan gen can take anywhere from a couple minutes to an hour or more based on what snake oil you're using.
>>105888371That's exactly what you got.
f
md5: f738bd106171178fd9b046738fddfe49
🔍
>>10588823haven't really noticed any particular loss in prompt accuracy around v40
what are some must-have extensions? I have adetailer for faces/hands, tagcomplete, ultimate upscale and ar-plusplus for aspect ratio.
>>105888371do a gen with the light shafts with noobai (ie wainsfw v14, it knows asuka and everyone) then prompt "light shines on the anime girl and the water ripples" or something.
>>105888413civitai helper in forge/reforge for quick lora management. also reactor for face swap stuff.
https://github.com/Gourieff/sd-webui-reactor-sfw
can make it nsfw capable by editing a single line of code, reason is github wont allow it if it is nsfw.
f
md5: f738bd106171178fd9b046738fddfe49
🔍
>>105888232haven't really noticed any particular loss in prompt accuracy around v40
>>105888383well yes I suppose it is. It is also shit. Maybe I'm asking too much but I want some more subdued animation. I keep prompting in that direction and shit just gets worse, I feel like I am missing something.
>>105888428like so:
https://www.reddit.com/r/comfyui/comments/1i43l3v/guidehow_to_disable_comfyuireactors_nsfw_filter/
>>105888430can you post the full pic please?
>>105888359I am using the rentry workflow, but not the "lightx2v + NAG" workflow because I'm generating at 480p and not using slow motion.
chroma
md5: e81e0a126b95b5f3af4146127b896dfb
🔍
>>105888430be specific: stuff like "water rippling gently/slowly", wan can do anything and now that there is a speed lora, iterating isn't that bad when a gen takes a minute or two instead of 15.
pepe
md5: 49c076eb8cb2ea8fcb2afda58623a17e
🔍
Anyone got the breast/nipple fixer lora for kontext?
>>105888445use it, it works great for 480p or 720p. the lora is like a better version of causvid with better results. you can get fast motion just fine.
this has 2 stitched clips.
Is it a bad idea to generate more than 5 seconds? Even just one extra second to go to 6? I tried generating a 6 second video; it took about 200 seconds longer and one of the characters had an extra leg; is that what to expect when generating too many frames?
>>105888497the video/vhs node can pick the last frame from a wan video you made, use that as the first frame for your next prompt then stitch them together if you want 10/15/20s clips.
file
md5: f93fd5a0eb5c9236ea9b64a2a8a66a55
🔍
>>105888497it should be about linear in terms of how much time it takes to gen except if you exceed your VRAM and swap more
but depending on the model you get nonsense. that said some WAN models ARE designed to try and extend the output
IMG9266
md5: d943aeac88e42bfdaf6af4cd22f598b0
🔍
Any good?
>>105888497no i don't think the extra leg is because of that. i go above 81 all the time and it works, especially if repetitive motions are intended. ignore the retards. pic related is 161 length
why do you guys hate pony so much?
>>105888559once you see the bug eyes you'll never unsee it and all pony gens will forever be shit. and it will happen to youuuuuuuuuuu
>>105888559It's not hate. It's more of a "glad to not need it anymore".
Now WAI... I hate WAI.
the woman on the left is wearing a bikini top with an image of the pink hair anime girl. keep the woman's expression and pose the same.
2 img workflow from comfy org
>>105888599now if I use the clothes remover lora, I get a proper bikini, since it isn't afraid to do lewds.
didn't say to keep the anime girl expression the same but you see the difference with the lora.
>>105888608actually stylish, if no bocchi
Is there a list of Booru artists who draw art in exactly the same style as the official art or the anime?
I was looking at stuff from keihh and that's right on the mark.
>hey hey ho ho
>schizoid micromanager wannabe janny has got to go
>hey hey ho ho
best way to control exactly where a subject is in the frame, where they are generated, pose, et c? for images.
preferably something without much impact on generation speed.
used outlines for i2i, and controlnets with lineart and depth maps way back.
chroma
md5: 403fe38dfb50f42428672d9f700f121d
🔍
>>105888559I don't hate pony, but it's technologically less good than hidream, flux/kontext, wan, [...] so I'm glad people are training these and more.
>>105885344 deleted
>>>105885307>can you stick to one name you absolute degenerate.>you don't matter as much as you think you do.>you don't need attention, you need help.
>>105886733Exactly, with ComfyUI you can make an assembly line, feed some input and get output. Just like what artists need. Krita plugin, blender plugin, etc: show me such plugins for gradio.
Gradio is bloated shit with which you have to click multiple times to get through dozens of tabs, menus, pages scrolled just to change one thing. With ComfyUI all necessary is here easily accessible.
>>105885952>not sphincter cellngmi