17gb of VRAM for a 4b Diffusion Model Edition
Discussion of Free and Open Source Text-to-Image/Video Models
Prev:
>>105673353https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Models, LoRAs, & Upscalershttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
>Cookhttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanX (video)Guide: https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1
>ChromaTraining: https://rentry.org/mvu52t46
>Illustrious1girl and beyond: https://rentry.org/comfyui_guide_1girl
Tag explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage | https://rentry.org/ldgtemplate
>Neighborshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/celeb+ai>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
cursed thread of hostility
file
md5: 50778acba2df287dff9e19b753a4db48
🔍
not bad, 3 minutes per image on 3090 as the other anon said
i just reused the venv that i use for comfyui instead of creating a new one and it just worked out of the box
diffusers 0.33.1
onnxruntime 1.22.0
onnxruntime-gpu 1.22.0
torch 2.7.0+cu126
torchaudio 2.7.0+cu126
torchvision 0.22.0+cu126
>>105678577>not badyep, and the most interesting part is that it's apache 2.0 licence, I feel the BFL guys will delay their release even more, Kontext dev needs to be much better to that one for people to care about a distilled model with a bad licence, it is how it is
>>105678577>3 minutes per image on 3090it'll be much faster with sage, sage2++ claims it's 3.9x faster than flash attention
Cant wait to get a 4090D. Knowing my luck right after I buy it the 5090D 64GB would available, kek
>>105678597What's this? Looks way better than inpaiting on comfy using gimp...
Blessed thread of frenship
>>105678729>What's this?https://github.com/VectorSpaceLab/OmniGen2
>>105678727oof, it didn't understand my request and going for 2 images is twice as long, that one took 06:24 mn
How to fix wan2.1 vagina gens, always adds a penis or hella loose lips
>>105678802>06:24 mnwhy is it so slow? it's just a 4b model
>>105674616What values should I use then for the convolution?
Maybe this is a retarded question but usually the formula goes
>((Number of images in dataset X number of repeats)/batch size)X epochs = total number of steps
Is this right?
Also, with less images you have to train more, but is not a direct relation right?
If you have 100 images and you train for 2000 steps with good quality, you wouldn't get the same quality with 50 images and 4000 steps, you would have to do more than 4000 steps right?
>>105677998nta but thank you so much for posting those workflows
great gens, i'll stop there.
>>105678746You VILL be a fren
>https://huggingface.co/rocca/chroma-nunchaku-test/tree/main
v38 is out, vramletchads its our time to shine
>>105678999oooooh yeahhhh.. it's sloppin' time!
>>105678727>>105678802only works with chinese prompts mister, please understand
>>105678999>>105679022>The quality of outputs produced by the models in this repo are not as good as they could be, probably due to bugs in my code. You may need to wait for official Nunchaku support if you want good quality outputs.
>>105678999kek, I'm running chroma on bf16 and I regulary got some anatomy abominations, can't wait to see some images made with some Q4 tier quants kek
>>105679040i've been posting images made with svdquant chroma v29 since a month ago ITT
2 of my images from the last got into the current collage
try to pick them out, you probably can
>>105679050>chroma v29that was before the distillation lobotomy though, it went downhill after that
>>105678999I have the Q8 version. This one is worse?
>>105679056it's waaaay faster than even q4 gguf, and it is very very good quality
Can you anons give me feedback for my training settings please? I'm trying to train illustrious.
pastebin.com/pdnQG4fj
file
md5: c56bb709f534d34add6e6833034acdcf
🔍
>>105679056heres a quick gen
https://files.catbox.moe/h9d8op.png - v37 from some anon
keep in mind im using fp8 text encoder too..
>>105679056>This one is worse?everything is worse than Q8, but we don't know exactly how's the quality of svdquant is, it's probably between Q4 and Q5
Is omni 2 censored? Does it work if you give him an nsfw pic and tell him to use the second image character for it?
>>105678802Does kontext understand that though? Struggled on Gemini with that. Anyways, it's quite obvious this model isn't as capable as Kontext.
>>105679075oh also i picked v38-detail-calibrated-32steps-cfg4.5-1024px out of instinct, maybe the other versions roccieGOD uploaded are better
>>105679075Can you post the prompt for this? I'll compare the two models.
file
md5: 22b6d423feb92bf76abb203422b15a05
🔍
>>105679095the whole workflow of the anon's gen is in the catbox
seed: 391147357278020
but here's the prompt:
amateur photograph,
A close-up photograph of a Coca-Cola can, featuring a pink and white cherry blossom design on the label. The can is positioned in the center of the image, with a white background and a blurred pink and white evenly cherry blossom-print-patterned fabric in the background. The lighting is soft and even, highlighting the details of the label and the can. The image has a shallow depth of field, with the can in focus and the background out of focus. The overall composition is simple and elegant, with a focus on the cherry blossom design and various cherry blossom petals scattered around the can itself. The can has condensation droplets dripping on it.
>chroma-unlocked-v39-detail-calibrated.safetensors
It's out.
>>105679084>Does kontext understand that though?we can't test kontext dev (the model we'll get) yet and kontext doesn't seem to be taking multiple image inputs?
4o nailed that shit btw
>>105679102https://www.youtube.com/watch?v=JsmAcFY5AHQ
>>105679114>kumiko stabbed herself
>>105679114Nah, 4o changes the entire image
>>105679121She killed herself after realizing that Reina doesn't want to be her girlfriend kek
>>105679127yeah, but it understood the task at hand at least, that's way better than this shit
>>105678802
>>105678999Where do I put the unquantized layers?
I just always wanted a field medic character that fights with syringes and scalpels.
>>1056791324o will always understand prompts better though. It's autoregressive, so does to images what LLMs do to text.
file
md5: cb4f760781be1a4c7d4c4fe30ee2e5db
🔍
>>105679137you put them all in the same directory, for example like picrel
https://litter.catbox.moe/mkhhhxkwddipqzbj.png
here's an example workflow, albeit with a shitty prompt
>>105679161you were warned
file
md5: e724f5165e18855f16582ea1fdda8b8a
🔍
i warned you :^)
Any SDXL speed ups for upscaling? that's the longest part of my gen
>>105679206there's linux, asides from that I'm afraid I can't help You, Anon.
Do you combine with text gen? I imagine I'd have one of the models be something tiny if I run the likes of Comfy for image and text gen on webui?
Ideally just accompanying image for each "chapter" of what I'm genning elsewhere.
>>105679244using an entirely different OS is not a solution. sageattention2++ needs to hurry up.
When I put the chroma into unet like the q8 gguf version, the unet loader doesn't see the new one.
When I put it into diffusion models, comfy says it can't tell what model type it is.
>>105679259did u install nunchaku?
>>105679099Why use the SD clip and not chroma?
>>105679296no idea, never thought about it, tried chroma with my nunchaku wf, errors out
eventually ill use 4bit awq t5 once i upgrade nunchaku to 0.3.2
>>105679315I am sorry but as a vramlet i cannot help you
file
md5: 5b86a9268aa2bcc582bcd78d2ef7317a
🔍
So uh, anons.. How are those chroma v39 gens going?
>try chroma
>gui reconnecting
>the the nuncharku chroma
>gui reconnecting
>try sdxl
>crash outright
Was there some new update that I missed in comfy?
>>105679368have you tried turning it off and on
>>105678914I don't know, I usually train with batch 4, and make as much as epoch as it takes until I reach 2000 steps, even if I have 20 or 100 images
>>105679368this is useless information. why don't you look at the console to see what's actually going on
>>105679368>comfywhat were you expecting
file
md5: 24b8cb95dff89a5c341b1f43be9457ad
🔍
soulppa
>>105679406Prompt executed in 0.07 seconds
got prompt
model_type FLUX
and then the gui crashed
Ok, the big chroma decided to go through now
Is this ever gonna just be available to easily set up locally or what
https://magenta.withgoogle.com/magenta-realtime
And the nunchaku now hangs on the padding removal node if I use chroma clip and crashes when I used SD clip
>Half of /ldg/ are 8GB vramlets
Say it ain't so, bros...
>>105679546i'd never bother with ai if i had such a low amount of vram.
>>105679546I'm a 24GB vram chad tho
>>105673777can i get the metadata for this pic if anon is still around?
>>105678558 (OP)Do you ever edit your generated images in a graphics editor? You know, removing extra fingers, fixing scuffed lines, blurring/sharpening and so on?
>>105679789I have krita set up with comfy to inpaint:
https://github.com/Acly/krita-ai-diffusion
>i wake
>no sage 2 ++ update
>i weep
>>105679741> 24GBThat's vramlet-tier these days, come on.
I just started using chroma with the default workflow and getting around 24 seconds per gen ona 5090 is that normal?
The "detail calibrated" chroma looks a bit off compared to the normal version. Often gives more artifacts and the backgrounds are broken. I don't understand the name/why it's a thing?
>>105679949It's an experiment where instead of training on 512x512px images he trains on 1024x1024px images. The intent is to improve how details look but it's not a huge success.
>>105679898its not tho.
24GB is still a vram chad.
>she show bob and vagene
>my beautiful girlfriend
>I am doing the fucking
>>105679945Yeah it's pretty slow. The main culprit is that the negative prompt does something since it's undistilled, so it's almost doubling the flux gen time. Usually getting around 6s/it on my 3060.
>>105679991noted, thank you anon. also is there a hiresfix node I can yoink from anywhere for chroma?
>>105680028you just chain "Upscale Image By" -> "VAE Encode" -> "KSampler" (with <1 denoise). You don't need a node.
>>105678956mages casting spells, the images you posted dont look dynamic enough. Also i'd prefer something stylized
>>105680087replace the girl with a sci fi bbw milf with bedroom eyes, smirk and huge saggy breasts
to test omnigen2, of course
>>105680126I'm not inputting any of that.
>2 character lora exists
>hmm, which one to use
>try both
>1st one barely works. art style is generic and hair isnt even the same
>increase the model strength. still doesnt work the way it should
>try the other one
>works perfectly at default weight
>maintains every subtle detail about the character
>works excellent on every checkpoint I've tested it against
I love people that know how to make loras.
00
md5: 859ccd578ca2603abe3d3e551bf368c2
🔍
Now it makes sense.
what is the reason we can't use multiple cards to gen pics/vids? is it a hardware limitation?
>>105680688Yeah well trained lora makes whole checkpoint feel more stable. Drop a like for the guy.
>>105680716Which one was better?
>>105679789Yes, sometimes it's easier than trying to regen or inpaint, sometimes it's the only option cuz the AI can't get it.
file
md5: 1b7247bbe928505ff575e654047cf6cb
🔍
>>105680716Which one works better, lol?
10k steps is crazy.
>>105680732>Which one was better?The bottom by a long shot.
>>105680716Some people put their blood & soul into making the perfect lora. Others just dump the minimum amount of images in a dataset with ai generated captions and call it a day.
Anyone using Omnigen 2
Im trying to see if it can match the same quality outputs as Flux.1 Kontext.
Example 1: Desire is to isolate items/characters (subject matter) using prompts and using ai to fill in any missing details of subject. Fully restoring subject and isolating in a white background. Keeping all details intact
Example 2: Removing text, watermark, and speech bubbles in comic. Keeping all details intact.
Flux kontext is pretty cucked when it comes to any subject that is lewd, holding weapon and doing anything slightly nsfw. Its able to do it just that the company filters all request and refuses to do it/ provide a black image. Im wondering if there is a way in Omnigen 2 to achieve the same thing that is fully uncensored. It could be a neat way making the process of forming a dataset faster for lora training and so forth.
>>105680738>>105680743Yeah 10k seems pretty nuts. Could it be that Illustrious Prodigy needs marathon for convergence
>>105680761Some people make loras because they want to use them; others make loras because their source of happiness in life is CivitAI upvotes.
>>105680716I don't understand why even use "repeats" if every concept uses the same amount, 18. Isn't the amount of training same even it's epoch 50 instead 2 due to inflated training data
>>105680777getting one (You) from an anon, now that is the true source of happiness
>try NAG
>error error tensor shape
>error error
nice
>>105680833Have one, very good vid.
>>105679991>negative prompt is undistilledWhat does this mean and how do I fix it?
file
md5: 24b127155dc99cd621290992b5182c82
🔍
>>105680897Some people can actually do that
>>105680928Post one real example.
what exactly does ExpressiveH do?
>>105678999Yeah I dunno about that one
v29-32steps-randCfg vs v38-32steps-cfg4.5-1024px vs v38-detail-calibrated-32steps-cfg4.5-1024px
>>105680897LET JESUS FUCK YOU!!!
>>105680738>>105666637>how do i stop the video from dimming on wan?I'm not sure but maybe it happens when the image is too saturated or too dark
>>105681028Do you use any loras? Some, both concept and "optimizations" ones, can cause glitches
>>105666637I have this exact same problem, I would have thought more ppl were having this too
anything cool release for wan since lightx2v + NAG? anyone test MAGREF?
any more obfuscated chroma tags found?
file
md5: ac25776beeaec23c34e9447defcd6b62
🔍
>>105680957https://www.youtube.com/watch?v=0UheoQaVmQE
>>105681055that one was lightx2v, bounce and ultrawan 1k
For Wan now we need the ability to properly generate and train 10+ second long-context clips. lightx2v is amazing though, it fixes the main problem of it being too slow
>>105681284Try to disable the torch compile node, since that was increasing my gen times like crazy.
>>105681295My gen times are good, I just want more than 81 frames and really what I want is 20+ seconds of coherent video
>>105681208>ultrawanIsn't it for the 1.3b model?
>>105681344fug you're right
Couldn't you let Wan make a 5 sec video, then let it continue its train of thought with the prompt, make another 5 sec video, and then stitch the clips together?
>>105681688CCXL my beloved
>>105681688That's a tiny river.
>>105681713just testing regional stuff with the cond pair set props nodes and it works. 2 different loras at work yay
>>105681441workflows exist that stitch 5 second clips together but forget doing any complex scene with continuity. it just doesnt work
>>105681284>it fixes the main problem of it being too slowAnd in return gives you neutered slow motion, making it useless for anyone that wants actual quality. I can point out every single wan video using lightx2v. that's how badly it fucks the motion.
>>105681757works fine for me with loras
>>105681766the problem isnt relating to it working
>>105681812loras don't have slow motion for me :) no need to reply, I can tell you're a faggot
>>105681833We already discussed this and it was proven.
>>105681179Wait, its real? He really hid artist tags and shit??
>>105680719imggen fags aren't smart enough to figure out sharding like based llm chads
>>105681868no, anon just didn't know that gens could produce coherent results without dictionary prompts.
eg- picrel is an example of what comes out when chroma is fed random vectors.
>>105681008>v29-32steps-randCfg vs v38-32steps-cfg4.5-1024px vs v38-detail-calibrated-32steps-cfg4.5-1024pxuse this script so you won't have to write what is what on 4chan but it'll be written directly on the image, that's more convenient
https://github.com/BigStationW/Compare-pictures-and-videos
>>105681008The lightning is much more natural on v29, he really slopped that model after going for that "low step" nonsense. Sad
>>105681868>Wait, its real? He really hid artist tags and shit??no I don't think so, he noticed he got anime pictures by writing random shit, but that's more because the model is just biased to output anime images no matter what
>>105682135Not consistently true.
file
md5: e20fd09593710952fd0d5274fa89d0a1
🔍
>>105680766It just doesn't do the former. And for the latter, it did this.
>>105682082Why is her ass acting like it's something from a rhythm videogame?
>>105680766>Im trying to see if it can match the same quality outputs as Flux.1 Kontext.bad idea, we won't get kontext pro in the first place so it doesn't matter, what will be important will be the comparison between omnigen 2 and kontext dev
I am starting to think Kontext Dev won't even happen considering how useful and unmatched the model is, so they will jew out with the API model for longer
>>105682198That in context generation of Musk on their github is terrible though. Kontext pro is better at that. Doubt we'll get anything that bad from Flux Kontext dev.
Stupid Wan, just make her uncross her legs
>>105682135Just fyi I was prompting for the vintage 80s magazine look so the lighting didn't necessarily have to look natural at all. Still, v29 handled it a lot better, the other two are oversaturated to hell
Imagine the potential if Chroma were fine-tuned on instructions like Kontext...
1 - "Make her nude / remove her clothes"
2 - "Make her boobs bigger"
3 - "Make this woman futanari and make sure her cock is long and erect"
4 - "Fill this person with cum"
5 - "Make a blowjob scene featuring this woman"
(...)
>>105682173Consistently true.
https://www.reddit.com/r/StableDiffusion/comments/1lin7p6/comparison_chroma_prev295_vs_chroma_v3638/
>>105681298Nice, bit of Robert Mcginnis feel to it, weird vertical lines though
>>105681008Nice third leg on the v38 detail calibrated kek
>>105682308Without knowing what the exact prompt is, this comparison is kind of useless. Maybe it may be more biased by default, but when testing my usual prompt where the first thing I ask for the for photo be amateur, I have found the opposite to be true. The model (barring the calibrated version due to being behind in merging as discussed in the Discord screenshot) is actually better at consistently producing photoreal skin than before.
>>105682398>Without knowing what the exact prompt is, this comparison is kind of useless.you didn't share the prompt on your image aswell lol
>>105682392>Nice third leg on the v38 detail calibrated kek>>105682398And nice third arm on chroma v36 detail calibrated, I think the calibrated one is broken no?
>>105682298Is all you can think of coom?
>>105679546constraint breeds creativity as evidenced by this thread
>>105682298Omnigen is a 4b model and it's eating me 20gb of vram, there's no way we'll get something like that out of a 8.9b model, we don't have powerful gpus enough for that size
>>105682406I did share catbox including prompt and tested it more than one seed. Also same holds true for v38
https://desuarchive.org/g/thread/105586710/#q105589249
>doesn't know what the communist logo is
DOA :(
>>105682464>communist logolel its called the hammer and sickle anonie
>>105682308I think it's possible that optimal settings also change while the model is being trained. Might be more informative to find good settins for 39 and then test it with 29 etc
>>105682472based commie knowledger
>>105682512锤子与镰刀, 共产主义的象征
>>105682512try to reference more commie shit like soviet union. Prompt it as a coat badge or something
https://xcancel.com/ostrisai/status/1937211561682354206
>Focusing solely on the no CFG version moving forward as it appears to be converging faster. Adding some training tricks to target the high frequency detail more. The high detail artifacts are hard to get rid of as the model just never learned that data.
>>105682614>no CFGso no negative prompts? nah bruh this ain't it...
chroma 39 is actually a lot better from my brief testing, over prior versions
might be placebo tho
>>105682677Negative prompts are a cope like inpainting and loras
https://imgsli.com/MzkxNzQz v29 vs v39 comparison
Seed: "696969696969"
Positive: "Attractive Chinese woman standing in front of a house"
Negative: "low quality, bad anatomy, mutated hands and fingers"
>>105682363It's Michael Whelan actually, but yeah those lines are annoying.
>>105682841If your flux/chroma prompt is not at least two paragraphs long, you're doing it wrong.
>>105682917this anon is right, it must be 69 paragraphs long to see the magic
>>105681008The hands look terrible, I don't get how this model got any hype in the first place.
>>105682917this. you need to get the ai into a good mode with an interesting story before genning
>>105682886>Michael WhelanBased...
>>105682967>tfw you have to read a little story to the AI before you convince it to generate quality images
>>105683023Damn, I hadn’t realized how different Chinese and Japanese people can look.
>>105683062all the east asians have subtle differences
chinese have bushy brow
koreans have giga chins
Hell yeah Chris-Chan! What a hunk of a man he is.
>((Number of images in dataset X number of repeats)/batch size)X epochs = total number of steps
Is this right?
Also, with less images you have to train more, but it is not a direct relation right?
If you have 100 images and you train for 2000 steps with good quality, you wouldn't get the same quality with 50 images and 4000 steps, you would have to do more than 4000 steps right?
Also, can you anons give me feedback for my training settings please? I'm trying to train illustrious. I used both standard and lycoris and convolution network at 32 with lycoris
pastebin.com/pdnQG4fj
>>105683145>artificial boobsthat's funny because he recently became a troon so that fits
>>105682955go gen a picture where someone has seven fingers on one hand and three on the other. bonus points if either of them are shooting the bird and you don't use either a "shooting the bird" or "seven / three finger hand" lora.
i'll keep the meter running, but not even the big models can do this, anon.
kek
md5: abd8c2960e23e7dd4a9f87c561dde089
🔍
>>105683154>seven fingers on one hand and three on the otherUse case for such prompt?
>>105683184oh shit it's been 11 years? I really thought that was more recent than that
>>105683151>Also, with less images you have to train more, but it is not a direct relation right?Smaller dataset = less to learn. No training setting can turn bad dataset into a good one. Garbage in, garbage out.
Does Chrome know more celeb names than Flux or is it still cucked in that regard?
>>105683262It knows a bit more celebrities (Will Smith, Di Caprio...) but I don't think it knows that much more, like characters and artist tags, Chroma doesn't seem to have much more knowledge than Flux (it's a shame really)
how long did it take until you got bored of this. Will we just go outside eventually? Though maybe it's like calling dial up internet a fad.
>>105683364https://www.youtube.com/watch?v=J3lUWiZHqDY
>>105683364For videos about 5 gens.
For images I've yet to grow bored. Chroma coming out with a new checkpoint every 4 days helps as well.
i mean base omnigen 2 is obviously lacking but since it's apache-2.0 surely we can finetune this to do some good shit right?
>>105682182really wondering if it's possible to prompt it somehow to get the proper gen output of retaining the quality/style with no degradation and removing speech bubbles, text and watermarks.
>>105682198when i meant in quality, i just meant following/adhering to the prompt in editing. it's able to remove backgrounds and text from images and isolate objects. It's just has to prompt some way to achieve it and i just want to find out if anyone has a clue how. Kontext dev (downgrade version) might never come out or be delayed for an year who knows.
>>105683364if you've grown 'bored' you're either creatively bankrupt or never were that interested in the first place. Imagine having a tool at your disposal that lets you generate basically anything you want and calling it a fucking fad.
How in the
you know what, here's your (You)
>>105683392im really hoping so, this way it would be super easy to fill in missing parts of an image like characters missing hands because of speech bubble/text, remove unnecessary text watermarks and so much more. I really want chroma dev or any other dev like pony to just drop their time in wasting it on making those models and work on this. It would be much better to have this editing/photoshop/inpainting tool model that is like kontext locally than another txt 2 img model that will fade in a month.
im out of the loop. what's the difference between detail-calibrated and regular chroma? why are there two versions now
>>105683392obviously, it's not a bad model at all, and who knows how bad kontext dev will be, so I believe this'll be the SOTA tool to use until a new challenger comes along (desu I wouldn't mind for them to keep their effort and go for an Omnigen 3)
>>105683468the former has more diapers in the dataset desu
>>105683170imagine you knew someone who could draw exquisite, detailed hands, so you asked them to draw a hand with more than the typical number of fingers, and they not only continued to draw only anatomically perfect five fingered hands, they couldn't even imagine any other type of hand, even when given explicit direction.
you might wonder whether they were just an idiot savant and not really a great artist, right?
that's everything, right now. some are just more passable fakes than others.
>>105683262basically none, yet. that'll converge late, if ever.
>>105683446sorry I should have specified this genre. I use these tools for art everyday. Have fun with your degenerate furry imagination if that's what's creative to you.
>>105683515frig I meant civitai type 5s goonslop not my down syndrome sdxl gens
>>105683532I could teach her...
>>105683544>>105683515what graphics card do you use?
>>105683392I tried it to remove the censor bars of an hentai image and it didn't work :(
hmm not sure if this is the right answer
>>105683636she might be using american common core math
>>105678558 (OP)offtopic but does any ai voice to voice actually work for audio porn? I tried using a comfyui workflow for rvc voice to voice but it sounded very bad
How do you guys get pony-based checkpoints to produce hands that don't look like ground meat?
>>105683968my eternal soul in a faustian deal
are there working omnigen2 nodes for comfyui yet? the Yuan-ManX repo is totally nonfunctional even with the relative path PR fix. there's like a dozen other showstopping bugs after that one
>>105683694The technology is not there yet, none can handle moans and screams that well. Besides, most newer local models aren't even for voice to voice but for text to voice. Some can do voice cloning but it mostly sucks. RVC is only ever good if you've used some other text to speech model and it sounds ALMOST like your desired voice (with similar intonations and whatnot), but not quite.
We doing redheads again.
I wish Chroma would take more of my Flux LORAs.
Oh well.
>>105684110Scaled looks better
>>105684134scaled has much more background bokeh though, ugh
>>105684110>t5 fp8why? if you can run Chroma, you can definitely run t5 fp16
>>105683968by switching to illustrious
>>105684145Can't you just unblur it with negs? I thought chroma isn't as cucked in that regard as flux
>>105684197>Can't you just unblur it with negs?nah it doesn't work that well, maybe with NAG but I didn't try that
>I thought chroma isn't as cucked in that regard as fluxit's better than Flux, but it's starting to look more and more like Flux through epochs
>>105682308
>>105684162Yup, plus unlike quantizing Chroma itself, quantizing the text encoder makes a massive difference in output quality
>>105684204it's gonna be funny as hell if he spent $50k to gradually transform Schnell back into Schnell again at 50 epochs
>>105684145it's supposed to have slight background blur because I prompted for it
>Camera details include a focus on subject in a medium close-up shot with a shallow depth of field blurring the background slightly and highlighting her figure and smiling expression.>>105684162it's the q8 gguf ran with 16gb. Just testing out of curiosity
>>105684134No it doesn't? Just slops the output
>>105684213he's so fucking retarded, Schnell was so slopped because it was running on a few steps, and what did he do? he transformed Chroma so that it also works on a few steps, what's the fucking point to undistill a model if at the end you're distilling it again?
>>105684246>what's the fucking point to undistill a model if at the end you're distilling it again?yeah that's absolutely retarded but I partially blame people whining about the speed
low information morons want a non-distilled model which means cfg and a good amount of steps
but they don't want to wait the extra time that entails
>>105684264>I partially blame people whining about the speedBut that's their fault for being vramlets, why did the modelmaker who likely has his personal 5090 concern himself with those users?
do you use the detail tweaker loras or are they snake oil?
>>105684278Plus at the end of the day you've got SVDQuant and other future things that would improve it. It was shocking that he doubled down on "improving" gen times. Though I don't think the model is worse or distilled like that other anon says.
>>105684246He's added a shitton of furfaggotry to the model. I think that was his main goal
>>105683220But what I mean is if the training settings isn't directly proportional to the number of images
>>105684246what's the max amount of steps chroma uses now at v39?
>>105684264>yeah that's absolutely retarded but I partially blame people whining about the speedI don't, he should've trained the model correctly until the end and then distill the v50 if he wants, what he did is completly braindead
>>105684278I dunno, popularity and visibility if your model can run on potatoes? But it doesn't matter. Most people can't put in the hard work to get FLUX running in any variant to this day but it didn't stop the model from gaining steam. Obviously, you don't want it so you need a server or supercomputer with Blackwell to run but even 5080-5090 levels of hardware required means you'll get a community that will do that downscaling work for you. It doesn't matter how slopped or loss of quality that will get your output, the morons will eat it up regardless. Catering to them early does you nothing.
>>105678914With Flux.1 Dev, my best results are always between 2700-2900 steps regardless of how many images I use (25 to 150+), so I never train more than 3k. Style LoRAs (no captions) usually top out at 500-700 so I never go past 1k.
No idea what you're actually trying to do with what model though.
>>105684315>Most people can't put in the hard work to get FLUX running in any variant to this day but it didn't stop the model from gaining steam.this, Chroma got popular before he did this distillation stuff, I don't know why he wanted to cater to those impatient vramlets, fuck them, quality first
>>105684334how can you fucking simp over an e-celeb. go out, plenty of cute women everywhere
ARE WE FEELIN WEREWOLF DICK'D OR SNAKE OILED TODAY?
>>105684308>what's the max amount of steps chroma uses now at v39?The max I've needed to use is 40, minimum 16. Cfg from 3.5 to 6.
>>105684412there's no way that's a chroma image, it's so slopped lol
>>105682420>I think the calibrated one is broken no?Not necessarily broken, just doesn't like my prompt and seed. Look at the whole seed, it's actually closer to the slopped v28. Now for v38 the detailed calibrated is a bit closer to the non detailed version as intended, but I have found it's still behind. Haven't tested v39, probably improved a bit more on that aspect as well.
file
md5: c762cd4083f9166480afac812c3df950
🔍
>>105684412I see what you did there
>>105684450it's exactly what I prompted for
>>105684473gj anon. Interrogated with JoyCaption beta one
>>105684368Gen your face when I drop an $8k GPU into my $7k PC to batch images of my babygirl.
>>105684473haven't seen this image since i was in middle school, which was back in like 2002? you must be a fossil
Any good SD models or anything else that I can use for img2img2 upscaling or reimagining of equirectangular planet textures like the ones NASA hosts? I wanna try it on some game assets, tried ChatGPT so far but it's junk. I'll probably end up manually painting it anyway later but I'm curious.
>>105684657try kontext or omnigen 2
Looking to buy a better pc than what I have now (mainly for stable diffusion generation, I don't care about gaming), which of these is best value for money?
https://allegro.pl/oferta/bsg-raptor-komputer-gamingowy-intel-core-i7-32gb-ram-1tb-ssd-rtx-3060-win10-15845303509
730$ intel core i7-4770, gpu GeForce RTX 3060 12GB, 32GB RAM, 1tb ssd
https://allegro.pl/oferta/intel-core-i3-14100f-rtx5060ti-16gb-nm620-512gb-32gb-17527702172
1000$ INTEL CORE I3-14100F, gpu RTX5060TI_16GB, 32gb RAM, 512gb SSD
>>105684681>which of these is bestthe one with a 24gb vram card
555
md5: e3fd6431bbbe6b3b98abf798652cccda
🔍
>>105684681>buying prebuilts>here>on allegro of all placesanon waht are u doing
stahp
one of those literally has an old ass haswell cpu in it
fuck chroma, back to svdq flux. 8 seconds for this shit @ 30 steps on my rusted 3090 from 1982. (distant shouting can be heard: "lora hellhole!" ..wha?) gonna slop the living daylights out of this. cats, elves and tigers. chroma can suck my binary balls
>>105684727>fuck chroma, back to svdq flux. 8 seconds for this shit @ 30 steps on my rusted 3090 from 1982. but you can run chroma with svdq though?
>>105684681>intel core i7-4770, gpu GeForce RTX 3060
Total noob here. If i want to use real pictures of girls to blend it together and want it to look as realistic as possible, which free service should I use? LMArena is decent but it changed the faces too much when I ask it not to.
516
md5: eca5448b7e58bf8df4ff9aeaf749dadd
🔍
>>105684681Ain't no way nigga lmao. This is actual e-waste, not even memeing. Idk how is the second-hand market in poland but here in Czech Rep. I had no issues hunting used parts one by one or in small bundles and building a pc that way. Shit like what you posted is a joke and you are overpaying a fuckload.
file
md5: e4d0dfc36c5740c41e665b73bd734813
🔍
>>105684513nipple warning
>>105684850I know asians love to do plastic surgery but even with their standard, that skin looks way too plastic lol
file
md5: e94e34c272a80434cdcf2a60be76522d
🔍
>>105684908the world is plastic
>>105684977>the world is plastic*microplastic :^()
>>105684748you can. rocca made new chroma svd quants but it's very much experimental.
>>105684972all equally replusive lol. have you tried that pixelwave schnell model ?
>>105684804>laughs in i7 8700 and 4070tis
>>105683151Please any help? For what I see more than 4000 overcooks the model.
Images look to perfect also. Like the illustrious model still has much influence vs the database. something missing, is there anything I could change in my training data, in my database or something I could tweak?
>>105685008>have you tried that pixelwave schnell modelNo, haven't. Is it good?
I'll try that tomorrow then.
>>105685092I did OOM several times, that much I can say. it was the full model tho and an upscaling workflow. need to test it again - 8 steps is a solid argument.
>>105685124i posted the pancake one in the /b/degen thread and got like 20 (You)'s, kek
file
md5: a8f4966555bd41494a0eab010f41b475
🔍
>>105685124these are so good
file
md5: dcaf5193ca70900b4e7f7b89e4aa9ad0
🔍
too realistic tone it down chroma
>>105685124would
>>105685160Thief!!!!
Mind sharing a link to the post?
>>105685228it was a few days ago, sorry its long gone and /b/ threads aren't archived
>>105685228NTA but can confirm many keks were had in degen
>>105684825>which free service should I use?The one connected to the graphics card in your computer.
>>105685228It was also posted in /v/, but that one got deleted in 12 minutes by the jannies.
https://arch.b4k.dev/v/thread/713346891/#713369668
>>105685292i was wondering if the jannies would delete a pulpy strawberry
>>105684513What was the prompt for this?
file
md5: 3d9ed8c78e5ea9b23dbf58646f1067f2
🔍
I prefer the other pancake video
https://files.catbox.moe/853lyo.webm