Interesting Exercise in Self Fulfilling Prophecies Edition
Discussion of Free and Open Source Text-to-Image/Video Models
Prev:
>>106206599https://rentry.org/ldg-lazy-getting-started-guide
>UIComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.com
https://civitaiarchive.com/
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows
>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanXhttps://github.com/Wan-Video
2.1: https://rentry.org/wan21kjguide
2.2: https://rentry.org/wan22ldgguide
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y
>Chromahttps://huggingface.co/lodestones/Chroma1-HD/tree/main
Training: https://rentry.org/mvu52t46
>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage
>Neighbourshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>qwen image vs chroma
verdict?
>>106210153Split the general as many times as it takes for the jannies to relent and make an AI board desu.
>>106210158Qwen: Slopped
Chroma: Noisy and busted
Nobody wins here.
Blessed thread of frenship
>>106210164>Chroma: Noisy and bustedWhere?
>inb4 no reply againExactly
>>106210148"clean vram used" node from comfyui-easy-use pack
find the node that freezes and lags, then put the "clean vram used" node in between whatever's feeding into that node. (for example maybe an "Image" input)
Miku Hatsune singing at a concert, she waves hello.
wai v14 for gen, wan 2.2 to make it move
>mfw I finally figure out how to prevent women from getting schlongs in my videos
>>106210153>Making a separate general for SDXL-based diffusion: good idea or bad idea?It's not SDXL that should have a separate thread, it's all the models that came after that should be contained
grainy + low-quality + shit lighting + muted color tone + random incorrect high frequency details = best photoreal
does the new qwen lightning lora not work with qwen gguf's? doesn't want to load the lora using q8 qwen.
also is there a fast way to unload models from ram/vram in forge/reforge, wan wants to use a shitload of both during video gens.
>>106210218Why do you think there are a gazilion Flux loras for analog photo look ? Because it's the way to make things look more photoreal, particularly with something like Flux which is super plastic due to being trained on synthetic data.
With Chroma you get it out of the box.
>>106210223nm, this works in other -> actions
>>106210180>Matted straw like hair with weird clumps under the ear>nonsensical bedding>Either earring or drop of cum, who can say>weird bra tanktop hybrid merge on right shoulder>Strange white disruption on fabric on left should>stain or half baked attempt at logo on left breast>Armpits look a little eczema-like to meYeah it's noisy.
>>106210246>>106210248Remember, this is the gen he uses to showcase the quality.
there, now I can do noob gens and gen in comfy without memory issues.
>>106210242Miku Hatsune firing an ak-47 to the right.
not the best trigger on first try but still neat, even got smoke plumes.
>>106210223since when the fuck does forge run Wan?
>>106210265it doesn't. I am making images with forge (noobAI/wai v14 model) to animate in my comfy tab. but, if the model is loaded I dont have enough vram for wan 2.2. (have 16gb)
>>106210248You know you could do this with literally any model, yes? I'll hate on anything but you're being a turbotroll without purpose.
>>106210264there we go, and her fingers are back to normal.
>>106210274Nah chroma has particular issues with this. That and even without the obvious errors it IS grainy and noisy.
>>106210248>>106210246Outside of the hair clump under ear which is a photoshop heal tool 1 click tier fix, the rest of the cope is hilarious, especially given this is a result of an image that's not even upscaled.
A crease in a pillow is a "problem", lmao. Antichroma vramletkids really are absolute retards.
>>106210282>upscaled*Highres fixed
>>106210282No. The people ignoring the obvious issues with chroma that will never be solved are the retards.
>>106210299Yes lil bro, a pillow crease is an issue and 2 1click healtool fixes make the model not realistic anymore. Retard.
>>106210282your usage of "cope" is ironic
Miku Hatsune eats a piece of cake with a spoon.
it was a whisk in the photo but doesn't matter.
>>106210311>no uGood one sis
>>106210314the eye contact is actually really good, wan 2.2 must have been trained with more anime in it.
>>106210241why are you angry?
Total Chinese Dominance
Resistance is futile
>>106210314why is she in a permanent shocked/surprised state
>>106210353>i2v>permanent
>>106210353because she doesnt know if the cake is good yet
>>106210344>When you have no counter argument, make desperate irrelevant claims
>>106210368why do I need counter argument?
>>106210346it was game over from the start.
Miku Hatsune is driving a car that drifts around a corner, the camera tracks the car as it turns.
not the biggest drift but she's trying!
So, finally got the chance to do some Chroma v49, v50 testing, and from said quick tests I found that I preferred v49, v50 looked good as well but the images often came out overexposed compared to v49 where they looked normal unless you prompted for overexposure, otherwise they were very much the same quality-wise.
Will try v48 (the last epoch only trained on 512) later to see how it does vs v49.
Here's the Bryce Dallas lora I trained for these tests:
https://files.catbox.moe/pdoru8.7z
Here's a simple workflow:
https://files.catbox.moe/yrcjjg.png
0.8 - 1.0 seems to be a good lora strength from quick tests, I didn't do much epoch testing but this one looked fine.
Trained at 512 resolution so don't expect miracles, particularly for full body gens.
>>106210248>>106210246>NOOOOOOO LE UNFIXABLE MISTAKES IT HAS TO BE PERFECT YOU CANT INPAINT TO FIX IT AL-AAAAACKUh huh
>>106210405Using photoshop and pasting in real objects cheating, chud.
>>106210390>I found that I preferred v49Same here. Better details with face, teeth, ears, jewelry. Small difference, but it's definitely there.
>Here's the Bryce Dallas lora I trained for these testsGreat man thanks! Is there any chance for Christina Hendricks one? Rewatching Firefly
>>106210405I mean, you can't just claim something isn't noisy and then post a very noisy example.
>>106210405Notice how when
>>106210164Mentioned qwen had flaws. Nobody leaped to its defense?
But the moment he mentioned Chroma had flaws too you literally started sperging out?
What exactly is your investment in this model that makes it above valid critique?
>>106210274Any model makes mistakes, yes. But this is a very simple image, and most modern models starting with sdxl can be coaxed to produce a very cherry-picked image that features a clean bra to cleavage transition when there're no lace, knits or background in sight.
>>106210435Notice how that's not a non sequitor? Notice how those two models are not the same and nobody has to defend both?
Notice how you don't know basic logical fallacies and strawman with
>makes it above valid critiquewhich nobody argued for? NPC retard.
>>106210431>Our Mrs. ReynoldsYeah, I could probably do that, like any other straight man in existence I do have a penchant for busty redheads.
When I see Chroma outputs I'm reminded of this very old AI image where you can kind of see things but if you look to closely at any one thing you become unable to identify.
>>106210448No, what made you think so?
Miku Hatsune is driving a car that launches off a ramp, high into the sky. She smiles.
no ramp, but the jump is there.
>>106210470okay, now I got a ramp.
>>106210405i'm not one of those pserging anons but what in the hell is that black line attached to her right shoulder strap that just goes off into the sky?
>>106210463spoken like someone who doesn't care about quality. the perfet gen doesn't exist. some amount of inpainting can only increase coherence. i get you're doing the whole troll thing but goddamn. this general really fucking sucks every time a new model comes out.
americanized us vs. them mentality will kill us all.
>>106210460i unironically love these images, do you have more from wayback?
Why can't the chroma guy update his model card. What is supposed to be the difference between 50 and 50 annealed. Does it mean the learning rate was slowly tapered out to nothing. Something else? How is it supposed to affect the result?
>>106210539Access to the model card was seized by the bank as collateral in order to pay off his training debts.
>>106210539Annealed is the last 10 epochs merged together or some shit.
>>106210528>2000 years later
>>106210572using skyrim/oblivion footage with vace? not sure what's going on but i'm enjoying it.
>>106210521That was the only one of that kind of image I am aware of.
yep, now that I can unload ram/vram in forge easy it's a lot simpler to gen and then use it in wan 2.2 (comfy)
>>106210576Low noise model at a low denoise.
>>106210518which model? awoo lora?
Any anon have version of metadata viewer userscript that works on archived.moe, desuarchive or even archiveofsins for threads that was archived and deleted from main 4chan? Main userscript only pop-up in console, said that thread was archived and thats all :/
>>106210589A locally trained LoRA based on ohayou faces reaction pics.
For sloppa image and video generation I have a choice between 9070XT and 5060Ti. Which one would be better?
>>106210613AMD for video gen always.
>>106210609neat, reminds me of these:
Some of these are truly ass.
>>106210622Well, that's an ohayou face too.
make noob/illu gen (waiv14), make it move with wan 2.2
red hair anime girl is singing and waves to the camera, as Miku Hatsune watches.
didn't specify miku's eye color as the initial image had them closed, but you get the idea.
>>106210684there, being specific fixed it.
red hair anime girl is singing and waves to the camera, as Miku Hatsune with cyan color eyes watches.
>>106210691You might see some retarded shit in there.
https://files.catbox.moe/r30852.png
do I need a 50series card to enjoy this hobby?
>>1062107234070s is purelly enough
anime girl wearing a pink bodysuit points at the camera and smiles.
same idea, make a gen and then prompt with wan 2.2,
>>106210723no, 16gb vram will let you do mostly everything though including the best video models.
>>106210752also note the shadows when the hand moves, wan is a neat model
>>106210508wan2.2 camera control???
>>106210717had no idea you can just v2v that easily, really cool wf. thanks anon.
>>106210717>retarded shit> wan nag> no neg suppliedhearty kek
anime girl wearing a pink bodysuit turns around and does a 360 spin.
neat, it worked.
>>106210192care to share with the class?
>>106210817https://www.youtube.com/watch?v=kHeCt6tHJQg
same vibe
So Iโve run into a bit of an issue, Iโm currently following the wan22idgguide with my 3090. I thought everything was fine till I tried running it, then they tell me that fp8 isnโt supported and I need something like f16
Am I screwed from using the 480p json there? Or can I download a different model to salvage the time I took?
>>106210581Redpill me in ram/vram un loading
>>106210693>>106210752I
>>106210850Shouldn't be any issues with a 3090. You're not screwed. Idk what your specific issue is though because I dont use kij nodes. Maybe it's one of the weight options on the model loader?
>>106210853do this in settings -> other -> actions in reforge
>>106210617But why then everyone recommend Nvidia cards for image generation?
Are you trolling me there, buddy?
>>106210850i have a 3090 and i'm using kijai's workflow no problem
>mention a reflecton
>wan will force the other man's face to look the same
It's ogre...
>>106210880Oops, uploaded the .png instead
oh right, nikke is also a shooting game.
>>106210872Just to ask, is your pytorch automatically 3.8 with a matching triton?
My main error is fp8e4nv not supported with this architecture, the supported fp8 types are fp8e4b15 or fp8e5
>>106210871See. You already knew the true answer.
>>106210908no it says pytorch 2.8.0+cu126 in comfy. i don't know what the triton version is
>>106210928go to your python_embeded folder, type
python -s -m pip list
and see what version is listed there
>>106210928My bad 2.8 I meant, but Iโm still confused of that error, Iโm pretty sure I followed the guide exactly.
Upon loading the json, we just give an image and prompt (Iโm using i2v) and it should work right?
For now Iโll try using the wan_autoinstall again
let's test wan physics shall we? (it has no physics model, but the data learns how it works)
anime girl pours maple syrup from a jar on her breasts.
Okay last Migu crush for tonight but holy shit lol at this one.
>>106210956ask https://x.com/JustinLin610 and he will probably answer
>>106210975honestly i would drink all that honey
>>106210723yes, you need a 5090 minimum for wan video and it already takes several minutes for one video
and no, lightx2v is a cope that sucks at large motion
>>106210975look at the refraction on the glass jar with the hair too!
>>106210928You need triton 3.4.
>>106210723everything except 3090 is cope unless you want to burn money
do you think the next series of Nvidia GPUs will give us 24GB as a minimum or will they fuck us yet again
>>106211033just buy the 5070 ti super
>>106211033Not a chance. They will push 8gb until the sun burns out.
>>106211033nigga the 5080 didn't even have 24gb, the fuck are you smoking
>>106211026>buying a 4x or even more expensive gpu to get shittier results than Q8https://www.reddit.com/r/StableDiffusion/comments/1gc0wj8/sd35_large_fp8_scaled_vs_sd_35_large_q8_0_running/
seems training qwen lora at 1024 takes around 39GB vram, that's with fp8 quant and cached text latents, getting about 6s/it on an A40
>>106211044you must be doing something wrong for the motion to be so grainy/blurred.
>>106211033there will never be another high gb card for the consumer. not with local ai taking off. it would cuck their own ai card solutions.
at least imo.
the 5090 has 32gb but the 6090 will at best have 32 or less just to fuck the consumer.
now the real question is why the fuck they are still making 8gb cards in the high end range.
1
md5: 0bc78d8900115e0f18c1b1e4f1f9cc2f
๐
Redid everything, still the same error. Anyone encountered this before?
On an RTX 3090, I freshly open the workflow json given without changing anything and use the example image and a small prompt
>>106211064gguftards btfo lmao. also they gen 2x slower because of the extra math
file
md5: c0210719fe35b5f8d029180a92fae976
๐
>>106210717fyi, you can use this to run any video without going instantly oom when using 4k/8k videos
>>106211064Theres barely any motion and every quant looks like shit there including fp16 where the bird just floats before landing, worthless comparison
>>106211064I thought scaled was supposed to be better, no? Also, what happened, Kijai?
>>106211075get the fp8 scaled e5m2 one from kijai or turn off torch compile (slower speed doe)
>>106211024I would suggest upgrading your python version and all packages so you can run 3.4 and be up to date with everything.
Maybe another anon can provide a better solution.
>>106211084I'm pretty sure torch compile fucks the output anyway
>>106211092thats an extreme cope, it gives a huge speedup and doesnt fuck up output
>>106211076>gguftards btfo lmaoLow IQ NPC
>>106211080>2x slower I'ts not 2x slower, and even if it were, the point is quality, pajeet, not that someone who thinks a bird levitating in the air at fp16 is a good gen would know anyway, lmao.
>>106211092>>106211098right... i need to do more comparisons but that would cut into my kino porn genning time
>>106211084Iโm gonna try that now, a part of me just didnโt want to redownload 30 gigs of stuff but letโs see if that fixes it
>>106210158qwen is the future, chroma isn't.
the qwen ecosystem will be more mature. do not underestimate the power of china incels
>>106211033Unless actual competition forces their hand, they will keep offering as little vram as possible.
>>106211076gguf is the closest to fp16. fp8 either have significantly different bird (comfy) or missing the vertical bamboo at bottom right (KJ). Only gguf Q8 has both.
>>106211117it will fix it, thats because we are 3000 g0ds
>>106211120gguf is too fucking slow though
>>106211064Sasuga, great benchmark you got there, retard-kun.
>yes yes the wan video benchmark is totally irrelevant, yet the sd3.5 large plebbit benchmark is SOOOOOOOOOOOOOO important and true yesssss
>>106211130>gguf is too fucking slow thoughNot a problem unless you're both poor AND brown.
this time no syrup, just a sunny beach!
>>106210987not clicking on that obscure porn website
>>106211142The quant tech itself didn't change, it doesn't matter what model it's quanting, brainlet.
And the benchmark was DOA day 0 given
>>106211080
>>106211130It's only like 5-10% slower for me. Not significant enough to switch to fp8.
is there any nsfw or photoreal lora for qwen image?
anime girl picks up a tropical drink in a margerita glass and drinks it.
the glass isn't quite that, but still pretty cool
could we do requests here?
>>106211220I can do you Miku getting crushed by a hydraulic press or nothing.
>>106211209>downloading Q4 right nowhow fucked am I?
>>106211220only if you request anime agp gooner approved images.
>>106211228could it be a brazillian miku?
>>106211215one more: not a full spin but she passes the test.
now this is pretty good
anime girl sits on a beach chair, puts on black sunglasses, and waves hello.
with wai v14/noob for source gens and wan 2.2, you can do anything.
So qwens a little slopped. chromas a little noisy, but hey. We got Wan 2.2 and that's more than I was expecting.
https://github.com/tianweiy/CausVid is this any good?
>>106211156>The quant tech itself didn't changeActually it has, for example quantization uses scaling these days, to better distribute the available bits when lowering precision
>>106211296causvid was okay before the lightx2 t2v and i2v loras which work much better (and faster).
time came to upgrade my gpu, and main reason to do so is to be able to run llm locally to help me with some coding.
i'm choosing between 5070 ti and 5080, which one should i choose?
i2v of a woman with her legs spread, her vag in full view:
lightx2v 2.1 at 3.00 on high, 2.1 at 1.00 on low: her vag stays the same
lightx2v 2.2 at 1.00 and 2.1 at 3.00 on high, 2.2 at 1.00 and 2.1 at 0.25 on low: in the first few frames her pussy turns into a mangled mess
lightx2v 2.1 at 3.00 on high, 2.2 at 1.00 and 2.1 at 0.25 on low: mangled vag
lightx2v 2.2 at 1.00 and 2.1 at 3.00 on high, 2.1 at 1.00 on low: no mangled vag
>>106211178just saw this
civitai.com/models/1851673/mcnl-multi-concept-nsfw-lora-qwen-image
>>106211296I saw someone saying the causvid LoRAs did a better job at preserving characters when using vace if that's your specific niche. Otherwise light kicks its ass.
>>106211316have u tried lightx2v at 1 low 4 high?
>>106211313Both have 16gb, so just look at the % difference in CUDA cores and measure that with the difference in $
>>106211302There are recipes that optimize things a bit more but fp8 scaled didn't reach near Q8 for anything other than very basic images. Unless you want to gen without caring for quality too much it doesn't make sense to use. But if you already spent at least 1500$+ for 4090+ what's the point of settling for shittier quality images with it.
>>106211319great
chinamen to the rescue
Update, so when I swapped to the e5m2 one I passed that error, now itโs time to see if I can make something in a decent time.
I wanna see how long the default of size and frames are before I modify em
>>106211362it's just wan 2.2 + 2.2 i2v loras
https://github.com/kijai/ComfyUI-WanVideoWrapper/blob/main/example_workflows/wanvideo2_2_I2V_A14B_example_WIP.json
https://huggingface.co/Kijai/WanVideo_comfy_fp8_scaled/tree/main/T2V
get high/low i2v loras from there, use them in the workflow for the high/low loras respectively
image source was just reforge, using wai v14 (noobai/illustrious model)
>>106211362stop talking to yourself. it's mentally unhealthy
>>106211231It depends...
You won't know until you find it out
fp32 is no guarantee that each and every gen is a masterpiece
>>106211209retard
change the seed an it will gen good, but other bad
>>106211411you are right, it's not as bad as I feared.
I wish I had the money to set up some enterprise tier AI gen server.
ann saying hello to (you):
>>106211325which version...?
anyway I don't see the point of using the 2.2 loras at all. with the 2.1 loras I have insane prompt adherence and coherence. what the fuck happened?
last beach ann:
anime girl holds a camera up and takes a photo.
pretty good, nice lens.
>>1062115082.2 loras literally just work on my machine..
2
md5: 135eb7234989c27b4d330211b6ff6550
๐
>>106211075>>106211361Man, it always crashes here, I swapped the model in the WanVideo Model Loader, but it still has the same error as last time. Is there another node I have to switch somewhere? My Triton is 3.4.0
>>106211508use these, high and low
https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Wan22-Lightning
why? because they are better and for 2.1 you needed a workaround like using 3 strength for high. these use 1.0 for both and have better results.
>>106210869but when it is nessesary? i have never done it, teach me!
>>106211531post ur workflow
>>106211531e4nv?
try with these
https://huggingface.co/Kijai/WanVideo_comfy_fp8_scaled/blob/main/I2V/Wan2_2-I2V-A14B-HIGH_fp8_e4m3fn_scaled_KJ.safetensors
https://huggingface.co/Kijai/WanVideo_comfy_fp8_scaled/blob/main/I2V/Wan2_2-I2V-A14B-LOW_fp8_e4m3fn_scaled_KJ.safetensors
>>106211531i just realized yo're on windows.
wan doesnt work on windows you need to install lonix
>>106211544it's only necessary if you have both open in 2 tabs, and you need to free up memory for wan 2.2 gens. otherwise, in reforge it's holding it in vram or memory. I made the ann photos for example then cleared memory, to gen with wan. (which uses most of my vram)
>>106211526I'm sure they work for "1girl wiggles her ass and waves at the camera" just fine
>>106211534no, they suck my fuckin dick
>they are betterproof?
>>106211559proof is they were trained for this 2.2 model and not 2.1, and 2.2 has many improvements over 2.1.
>>106211563*and use them at 1.0 strength for both, 3.0 on the high lora was a workaround for 2.1 lora on wan 2.2.
3
md5: a44dec0409449933c30922dde0947257
๐
>>106211563oh, you're just retarded. k
4
md5: dba37f200e2ce85665ea4abb04dee38d
๐
>>106211579If it wasn't already obvious, but this is my first time interacting with diffusion like this
>>106211579>>106211586Also wait, im realizing that theres another term in quantization as im posting these, is that really it?
>>106211585why would you use loras trained for an older model instead of 2.2 loras for the 2.2 model
okay, enjoy the old ones, they work but theres literally no reason not to upgrade.
file
md5: 705fe17446d5741ce2d01784b33d7ebd
๐
>>106211579i want to ravage your bussy, anon
>>106211579Change the quantizatuon setting to match the model in wanvideo model loader. Wtf are you doing?
>>106211608Yea I realized that as they were getting posted, im blind as fuck
qwen img + wan 2.2 for enhancing has interesting results, both use same vae / compatible latents
>>106211553the fuck are you talking about?
that's not even remotely believable bait.
please come up with something better next time.
>>106211601yeah, I get that you're autistic, you're the kontext turbo-autist, aren't you? anyway, the results speak for themselves
>try genning SDXL on my /ldg/ rentry comfyui installation
>instantly OOM
>gen SDXL using my old comfui install
>barely uses any vram
what the fuck?
>>106211508>prompt adherencenone of the lightx2v loras are good for this
>>106211690because they all suck. this is like when the first light2x lora came out for wan2.1 and it destroyed all movement. we need a v2 of these loras, probably going to be another month or so.
Anyone tried the qwen 8 step lora, what are the best settings for gguf? https://huggingface.co/lightx2v/Qwen-Image-Lightning/blob/main/Qwen-Image-Lightning-8steps-V1.0.safetensors
So far, I'm using 8 steps, 5 CFG, dpmpp_sde, simple. Parts of the image is great but produces super blurry edges
Chroma isn't made for this type of genning but that's what I already have downloaded and the point is to showcase quant closeness to fp16.
Same prompt as that old quant comparison:
"A drawing of Hatsune Miku with dreadlocks and light black skin skateboarding in New York at night. She is holding a smartphone on her left hand and a multicolored ball on her right hand, she has a red t-shirt with text on it that says: "MIGU". A pikachu can be seen on the top of her head. Her speech bubble says "Hard to keep me in style huh?", neons, 50's comic book style."
Negative: "low quality, ugly, unfinished, out of focus, deformed, disfigure, blurry, smudged, restricted palette, flat colors"
Scaled version downloaded from: https://huggingface.co/MaterialTraces/Chroma-V50-fp8/blob/main/chroma-unlocked-v50_float8_e4m3fn_scaled_learned_svd.safetensors
>>106211645autistic for using a 2.2 lora with a 2.2 wan model, not quite
but you seem to love the 2.1 so keep using it!
miku hatsune sings at a concert while a red hair anime girl cheers her on. (gen in wai14, wan to animate)
>>106211733Why did it make the drawing look like shit
>>106211690>doesn't post the promptwith the 2.1 loras I can do pretty much anything, get the characters to do any action, make the camera move however I want, and if it doesn't work then I just simply need to adjust the prompt a little
>>106211690>>106211706the result is not better on t2v. we have identical and cloned faces. light on 2.2 is a failure
>>106211733fp8sissies... our cope? spending 3 to 10 times more money for fp8 support so we can gen shittier images than a 500 to 700$ 3090 but at least its double the speed, right?
>>106211804>doesn't post genapparently you're a prompt god so try to replicate the no lora result with lightx2v 2.1 lora
i will wait
>>106211876you'll be waiting a while because I'm away from my PC. what is the prompt, she floats into the sky?
neat, it actually worked despite the smol teto.
the red hair anime girl hugs Miku Hatsune who is singing.
wan 2.2 is magic.
123
md5: 2598684d52ef93ad13de2d86bb7f85ec
๐
>>106211733v50, 28 steps, euler/beta. also using fast and sageattention.
left to right: bf16, e4m3fn, e5m2
>>106211861> our> we> sissiesyou must stop and go away.
>>106211976too close to home huh?
>>106211930>the girl flies upwards into the sky while the camera is zooming out and the girl becomes shocked with mouth wide open and the sky turns from dark blue to dark orangewan doesn't have perfect coherence here but lightx2v fails worse:
>girl flies way less>girl isn't becoming smaller due to camera zoom out>sky is mixed dark blue and dark orange, instead of pure dark orange
>>106211649ldg rentry has a bunch of stupid third party nodes doesn't it
also use tiled vae decode if you get memory issues
>[Low GPU VRAM Warning] Your current GPU free memory is 1308.66 MB for this diffusion iteration.
>[Low GPU VRAM Warning] This number is lower than the safe value of 1536.00 MB.
>[Low GPU VRAM Warning] If you continue, you may cause NVIDIA GPU performance degradation for this diffusion process, and the speed may be extremely slow (about 10x slower).
AAAAAIIIIIIEEE Im using Forge, text2img 1 checkpoint, 2 loras, and 2 controlnets. Is there a way to avoid that? I have a 3060
>>106212195ah, and 2 Adetailers
the red hair anime girl standing on a stage starts playing a guitar to the music.
starting to get the hang of qwen
>>106211690try to set the last frame
Whatโs the best way to interpolate frames of a wan video after genning, and is there any best way to slightly crush the bitrate of a video to make it look more like a snapchat video? Iโve just been clamping the file size to get it just a tiny grainy but I was wondering if anyone was using any tricks to make outputs more โphone likeโ at least to hide some of the AI look of I2Vs
With wan using the kijai i2v workflow, can someone explain what causes the color settings to randomly change?
I know it's related to input images. With most of them, there is no problem. But with some input images, there will be a transition in color-grading information. It could be exposure, contrast, white balance, even hue sometimes. I am genning dozens of times with the same input and 90% of the time the contrast increases over .5 seconds, usually at the beginning.
Is there a specific type of input image I should avoid to prevent this from happening? I'm thinking it might have something to do with images with a lot of white in the background.
>>106212350Are you using the lightx loras by any chance? That contributes a lot to it, that and sometimes frame count being over like 89 or something.
>>106212302how do you get the last frame when you don't have one in the first place?
>>106212366Yes, Kijai's workflow uses it.
Most of the time it's fine though. Again, only certain input images.
I am currently genning at 97 frames, but it was happening in all the 81 frame versions too. I'm just trying to figure out exactly what type of input images cause this so I can work around it. Again my best guess is that it's something to do with bright whites.
1
md5: ebc699657318689762e49646fde4fadf
๐
>>106212195Sure, that's a very simple and common problem. Come here! I have the solution for you! :3
>>106212010NTA but I think you need to prompt better. This is with lightx2v 2.1.
Hello everyone.
Does anyone know of a good detailer node (or denoiser) that I can use in Wan image generation flow? Those wisps/scratches that it generate in some images screams "AI Slop" and I wanna remove that. Using ComfyUI in a RTX A4500.
Thanks in advance
https://files.catbox.moe/n8s8l2.png
>>106211598>>106211605>>106211608I have the same issue on a 3060 using the correct models and quantizations. Never had issues with fp8. The actual problem is Triton. Disconnecting the Torch Compile node fixes it
>>106212452https://files.catbox.moe/wmv0fv.png
The miku spammer is starting to be more annoying than the Rocketfag, and that sure is something.
This shit run out of its novelty, your gens are not that interesting
2
md5: c2e7256c23670c9209a219e451dcde53
๐
If you have 3060 or less, please come here!
We have the soultion for you! ^^
>>106212490I have 3060
and that too the laptop GPU (6 GB)
>>106212475Post something yourself you whiny faggot, or better yet, just gtfo
>>106212446>no shocked mouth>movement suddenly becomes smooth midway instead keeping anime style>weird blur in the end
>>106212452>>106212464Another example is the whispy/scratchy text on this image. I'm aware that it'll generate nonsense text, but I'd like for it to be sharper
https://files.catbox.moe/0je08f.png
>>106212346I am using the old school method, vapoursynth combined with MVTools
3
md5: 98a9fd2fe71eb5ef4d589c1f1227ecdf
๐
^^!
https://files.catbox.moe/k5rfvs.png
https://files.catbox.moe/nosrop.png
>>106212587wan2.1-t2v-14b-Q4_K_M.gguf
https://files.catbox.moe/hue0bi.png
>>106212616meant to reply to you, apologies.
here's an image gen for you because your epic
https://files.catbox.moe/4ukt7s.png
I love WAN for image gen, it's so artisticaly superior to any model.
I have 32gb of ram and 12gb, to make it 1frame image do I need the same requeriments to load the model as for a 5 sec video?
>>106212711That's the gist of it, yeah.
My gens have a wan image gen flow for ComfyUI, just drag and drop
faces on this one are very busted, i'll do a face-fix pass but i'vent the time yet
https://files.catbox.moe/9z4cop.png
>>106212527didn't prompt the wide mouth, and the 3 samples you shared with the loras were way worse. The blur at the end was because of the prompt: "her outline becoming progressively obscured until she completely vanishes from view."
>>106212735oh thanks for sharing I will try it. I love you! <3
>>106212771>and the 3 samples you shared with the loras were way worse. it's still worse than undistilled which is my whole point
>The blur at the end was because of the prompt: "her outline becoming progressively obscured until she completely vanishes from view."yes, the blur prompt bleeds into the background and adds a layer of grey mist onto it, and i also showed prompt bleed for sky color in my original
thanks for proving my point
>>106212490>you see goldberg>when unkosher vram hording evil nazi's weren't looking>i used godly promised glitch in latent space>to draw myself out of the barbed wire>and survived death camp
>>106212735Thank you for this simple workflow
>>106212785You're welcome, fren
(this gen was made in FLUX fyi)
https://files.catbox.moe/ihdh1l.png
>>106212856You're welcome too, fren
https://files.catbox.moe/71t1xw.png
Isn't there a place where I can use WAN for free, or the service give x frames per day available to try out this txt2img thingy?
>>106212814and my point was you just needed a better prompt. You're getting triggered for nothing anon.
>>106210723Ive done it with an 8GB 3060ti
Low framerate and low resolution or you will be waiting a long time for a render.
You can also rent gpus online
Are AMD cards useless for this? On one hand I want to grab an AMD to prepare for a move to Linux and on the other I want Nvidia for the extra power.
>>106213189What's wrong with Nvidia + Linux?
>>106213215my wallet is wrong
>>106213189>I want to grab an AMD to prepare for a move to LinuxDon't make this mistake. Nvidia drivers are just fine on Linux.
>>106213215>>106213231I always see people saying NVidia drivers are shit on Linux. Have they gotten better and people just parrot old info or what?
>>106213231For AI they're more than fine. Practically all AI research and production is done on Nvidia gpus.
All AI research and production is done on Linux.
Making Linux drivers optimal for AI is a number one priority for NVidia.
>>106212490Does the 3060 12GB also have to go into the camp ;-;
>>106213266people still parrot 3090 is the best gpu so idk
>>106213266people here still parrot 3090 is the best gpu so idk
>>106213373>>106213311it is the most optimal gpu before you are hitting insane diminishing returns for $
>>106212962no, you will never get the lobotomized prompt coherence back and you're coping with tardwrangling. it's similar cope for low bit llm quants.
here is another example:
>the girl hears wolf howl and is startled and looks to her left, then the camera pans right, it shows only empty forest without any animalslightx2v again shows wrong look direction and prompt bleeding, while undistilled understands that wolf should not be shown at all
Is there a guide to training Chroma loras?
>>106213511No not yet. One anon has figured it out pretty perfectly using diffusion-pipe and using huge 128 rank. You'll have to browse prevous threads.
>>106210431I started with annealed, tried 50 and went back to 48 very disappointed. Now I am testing 49 and I find it acceptable
>>106210147 (OP)What tools should I use for CPU only image gen setup? I tried fastsd, but I am having problem e.g. where to set the Euler sampler, the result is awful, it's always downloading things, and so on.
I am new to this, but I think I can understand the basics. Debian btw.
>>106213511Yes, for Diffusion-Pipe: https://rentry.org/mvu52t46
>>106213561>using huge 128 rank.Why, unless you are training a LOT of images at ~1024, why would you go that high ?
>>106213189just get a cpu with integrated graphics. reserve your dgpu for compute.
i used to buy up discount radeon pros when i needed to add cheap graphics to a server, but today it doesn't matter. i915, amdgpu, nouveau/nvidia drivers are all fine.
>>106213266nvidia were naughty, not bad (especially considering how rough ATI was in the catalyst/fglrx days), and complaining about proprietary nvidia was a way to stick it to them while helping AMD, the open underdog. the idea transformed into big reddit cope through the 2010s when linux interest was growing for why it was wise to stay on windows 7 forever (intel+nvidia, the combo gamers trust). i feel like most FUD now is spread by linux people who don't want to help screeching panicked zoomers troubleshoot their first installs.
you shouldn't use linux, by the way. it's way too hard, you'll spend all of your time fixing your systems instead of using it, and i hear nvidia gpus don't even work on it.
>>106213578It's pretty strange how well some loras trained on Flux work with Chroma and they also work faster than loras trained on Chroma. I don't get it.
>>106213594>Why, unless you are training a LOT of images at ~1024, why would you go that high ?No clue.. but it works really well. Only anon who trained it can answer.
>>106213584What are some good gpu cards that are actually available for poor anons? All I see is VRAMlet 5060/ti where I am. There is also 3060 12GB, is that OK?
>>106213614>and they also work faster than loras trained on Chroma???
>>106213644Yeah doesnt slow down generation as much
>>106213654I've never noticed that, if true, weird indeed.
Does AI-Toolkit just suck? I don't know why every lora I've tried with it so far seems like ass.
>>106213786Some people have been complaining about the AI Toolkit results on the Chroma Discord as well, but then it's typically about horizontal line artifacts.
I've only used Diffusion-Pipe, I have no complaints.
>>106213819I'm having the hardest time getting diffusion-pipe to run with the modified .toml to run chroma.
>>106213862Really ? What errors do you get ?
>>106213633The one with the most vram, also consider getting more system ram as well (the more the better).
does riflexrope work with wan 2.2?
I tried it and basically got a low contrast painting result thing
can you go above 720p when genning t2i using wan2.2?
>>106213633I suspect the 3090s will go down in price a lot by the end of the year with the 24GB SUPER 5080 release, so get one used once it's like 400-500$.