Slop Must Flow Edition
Discussion of Free and Open Source Text-to-Image Models
Prev:
>>105626899https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
>Models, LoRAs, & Upscalershttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
>Cookhttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>ChromaTraining: https://rentry.org/mvu52t46
>WanX (video)https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1
>MiscShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Archive: https://rentry.org/sdg-link
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Bakery: https://rentry.org/ldgcollage | https://rentry.org/ldgtemplate
Local Model Meta: https://rentry.org/localmodelsmeta
>Neighborshttps://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/celeb+ai>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
What's the best upscale method for Chroma? Feels like it's really hard to balance between unnecessary details and higher resolution for this one.
Trigger words can be pain in the ass you know.
>>105629743USDU 0.3 denoise > facedetailer 0.33 denoise. Two upscaling passes at 1.4x each adds more detail, but I find you only really need one at 1.4x.
Need controlnet tile to do anything more right now.
>>105629762To add to that, 4xNomos2_hq_drct-l is a really good intermediary upscaler for USDU and Chroma, but it's resource intensive if you're already running bf16 Chroma
>>105629757i store all my positive prompts in a big text file, and comfyui just loads a line from it. so I never have to remember anything.
they also include the <lora:blah:1.0> text, so I don't have to mess around with manual lora loaders.
>>105629757everything was built around fetching info from civitai lol
>>105629762>USDU 0.3 denoisewhat's that?
>>105629796>USDUUltimate SD Upscaler
Blessed thread of frenship
>>105629711 (OP)>WanX>Self Forcingthey know their audience
> be euler
> win by default
You guys have any prompting tips for Wan2.1?
Both positive and negative prompts
>>105629916default wan negatives. don't really need to go too ham on it unless the model is sperging and refuses to play right.
for prompting it depends what you're doing. you have to talk to the model like it's a retarded 5 year old: "the woman picks up the tray from the table. then she turns around and walks away. then she glances back at the viewer and winks" very beat for beat type of prompting if you don't want the cohesion to be wrong.
can anyone comment on if this is better for quality than just a normal wf? i'm using it and honestly i don't see it. https://github.com/ClownsharkBatwing/RES4LYF/
>>105629963for wan/chroma specifically i mean.
>>105629938Gonna try it with the defaults. I've been changing them. I fed the wan2.1 prompting guide to gpt and it's actually doing mostly what I want it to now
>>105630012yeah. i found that keeping the description of the action free of GPT-like fluff words works better. i posted on one of the previous threads (be4 it went to shit again) on how i usually prompt, using that girl aiming a gun at the viewer as an example.
>>105630062jesus she's having a heart attack on the bottom one
>>105630062That one on the bottom is really looking to sue the owner.
>>105630062I like how the girl in blue hair is noticeably concerned
>>105630062what do you use to stitch two video together like that
>>105630141probably a python script, that's easy enough to create with chatgpt
>>105630062Middle girl's uppper torso got snapped
>>105630150>>105630141I don't know have you ever heard of video editing software?
>>105630169what's the point? with a python script you get what you want in seconds
reaper
md5: 8d651514663d5dfa64617615f2e29a7a
🔍
>>105630141video editing with reaper is very autistic
>>105629743second this
>>105629762. decent tile overlap (120-ish) and a tile size to account for that. no seam fix bs, low steps (like 10-14). can use detail daemon too, uItimate sd upscale has a custom sampler version. went to up 0.4 denoise for full body shots and chroma stays on target suprisingly well even w/o cnet.
>>105629781 eh siax does just fine unless certain fine patterns are on dispay. DAT's just too heavy. if I want a bit more color (and a nice brown tint) I go 4xNMKD superscale
>>105630169that seems silly. maybe i worded the question wrong. i was asking if perhaps he was using some node in comfy that lets you do this.
>>105630189this might be the most cursed thing posted in /ldg/
NAG released the SDXL code
https://github.com/ChenDarYen/Normalized-Attention-Guidance/blob/main/src/pipeline_sdxl_nag.py
>>105630141Can be easily done with ffmpeg (ChatGPT will readily tell you the command)
I have over 32gb of vram should I download a bigger non quant version of the models in the guide?
>>105630276why does it has to be a personalized code for each model? CFG's single code works for everything
>>105629938NTA, but do you have any advice for camera controls? I want to do a [pedestal down] motion, yet wan just creatively finds ways to do anything but.
Prompt: The camera drops to the floor. The girl is looking at the camera.
>The camera backs awayPrompt: Dynamic anime scene. The girl is looking at the camera. The view drops to the ground, flying toward her bare feet, camera pedestaling down.
Negs: static shot, presenting feet, lifting legs
>she stands up, camera tilts down to her feetetc.
>>105630329stick to the guide, offloading sux. ram = turtle speed
>>105630358honestly for camera controls you might be better off using VACE and a control video as i find camera movements still too RNG. you can get lucky sure but it's still annoying if you have a super specific camera movement in mind.
though having said that your prompt isn't great. i'd def try to simplify it with more generic terms like: "The camera view moves/pans towards the floor while the girl looks at the camera/viewer."
something closer to that, but it will still be lucky seed/rng based probably.
Some anon in one of the last threads mentioned FusionX being better for lightx2v than regular Wan. So I tried it.
It's around the same gen time for both. I did notice some improved physics/motion compared to normal WAN on some gens, but one of the LoRA's they baked into it tends to completely change faces. Why the hell they'd bake in a LoRA that does this is a mystery, because the model has some merits to it. But there it is.
>>105629963Looking at the RES_3M vs. Uni-PC example they provided looks good. Didnt even know this existed, I'll try this out tonight, thanks
>>105630444damn, the changing face is brutal, this model is unusable
>>105630062>Soulless>Sovl
>>105630181You can do a lot more with a video editor if you know how to use it. Sometimes I get tired of explaining stuff to that retard (clod).
>>105630478>tries to justify using a video editor to stick 2 videos verticallylmao what a fucking braindead anon, please tell me that's debo, that would be so funny
>>105630492I'm not that guy but it takes 10 seconds. I have premiere pro for my job. Literally drag, drop, click, click. Claude would fuck up the overlays or something. Not worth the time for a one off shit post.
>>105630519>I'm not that guy but it takes 10 seconds.my fucking ass, you have to open up the software (that takes time), then you drop the videos, then you have to write some shit and put them in certain positions... who the fuck you think you're fooling?
>>105630519I got Gemini Pro 2.5 (which is free) to make this
>>105630444 in about five back and forths to fix issues and add features. I asked it to specifically add gaps, told it where the text inputs had to be, etc.
It autoscales the end canvas to fit 4chan's max limit (2048) in either aspect, it outputs a lossless version and a version just under 4MB and it lets you use a custom font file.
LLM's have come a long way when it comes to coding.
The lightx2v+nag workflow gives some weird results. It really wants to make my cartoon girls real women. I guess I need to tweak the lora strengths.
>>105630141https://github.com/BigStationW/Compare-pictures-and-videos
So, is there are any way for Advanced Live Portrait work with Anime Face?
If its only have 1 line as mouth, its fucked
Does anyone make longer videos using the last frame and stitching the result?
>>105630639No, because it doesn't understand the context of preceding frames so the motion is always janky between video transitions. You also get feedback loop style image degradation every time you feed the video into itself, so you can't do it too much.
VACE can do it better, I've seen longer videos that looked almost seamless, but I never checked it out since I feel like 5 or 8 seconds with Rifle is enough for me
>https://x.com/viccpoes/status/1934983545233277428
I don't see why you'd post this unless it was already decided that you were gonna open source it. It says they trained it alongside BFL though, that's a real red flag. Distilled open source, "Pro" version for API? A new model is nice regardless, but local is really lacking a good base model to finetune off, barring Chroma when it's done of course
>>105630678>barring Chroma when it's done of courseI gave it up on Chroma, the anatomy is still as bad as v20 even though we're reaching the end, his model hasn't improved at all
>>105630565thank you! that's the one i was looking for that another anon posted a few threads ago
>>105630639I've done it but it kinda sucks, especially with anime/cartoon gens. The transitional frames are very noticable.
>>105630692NOOO, SWEET JESUS, DON'T MENTION CHROMA ANATOMY BEING LACKLUSTER
YOU'LL SUMMON *HIM*
Can someone tell me why in some guide, FLUX Checkpoint placed in checkpoint folder, while the other placed in diffusion folder?
>>105630678>I don't see why you'd post this unless it was already decided that you were gonna open source it.it's simple, they won't release it, but they want people to know about their model release anyway, the open source community is the best when it comes to give free advertisment to a product kek
unet
md5: c14c3fc5fca0d65570a11b748028e8a9
🔍
>>105630713there are different versions of the flux models. the original version, the quantizations. mine are all in models/unet and that works
>>105630639Can but it really isnt worth it. Best options are VACE (if you have good reference videos) and RifleX. Since we have amazing speed optimizations, hopefully someone can create a way for pure 1 min gens and prompt traveling so we can escape 5 second hell. Theres 1 project here >https://github.com/DualParal-Project/DualParal but its been pretty quiet
>>105630719>it's simple, they won't release itnot only that, but since they worked with Black Forest Lab to make this model, I suspect this is a finetune of Flux pro or something, why the fuck would they release something this powerful? they would kill BFL's business, that would be a treason lol
>>105630789ai companies have gone the route of any pr is good pr. them alluding to open source is just bait.
screenshot this if you want, but they will not opensource it WITHOUT cucking it into the ground.
>>105630819>but they will not opensource it WITHOUT cucking it into the ground.they don't even need to do this effort, look at tencent, they did much worse than that, they literally said they were going to open source their image model, and they didn't, and nothing happened to them, they can disrespect their users as much as they want without consequences, so yeah, there's no bad PR, there's just PR
>>105630789need to get my hands on it to see how strong it really is. payed or not, its gonna be competition either way. probably some sneaky anti poach contracts involved.
anyone tested crystal clear flux? it's out there.
>>105630789>hey worked with Black Forest Lab to make this modelthat alone is the definite proof it'll never be open sourced, imagine the Krea team going after BFL and saying this
>hey can you help us train a model, we want it to be really good, and then we'll open source it and make flux and your whole company irrelevant, how does that sound?
>>10563018999% of my video work over the last 15 years has been with reaper
>>105630917so do you just hate yourself?
Question: does anyone else's head genuinely hurt after looking at just a few of these at a time? it's like a strong tension in the back of my eyes that takes a few minutes to go away once it happens. it's like my brain can't fully accept what it's seeing and has some kind of averse reaction to it. it usually happens with those image-to-image gens but also with a lot of videos. anyone knows what i'm talking about?
>>105630954I believe that's called a brain tumor
are people uploading wan loras to sites other than civitai? theres so little activity.
>>105630954A lot of AI videos are uncanny. And some are almost indistinguishable from real people or things, so that they're just over the uncanny valley, and that can be sometimes worse, because the brain still picks up on the fact that something is off. Probably feel the same if you saw some one to one doppleganger of yourself in real life knowing it's not supposed to be possible. That what you mean, a feeling of unreality?
>>105630954bad ai has a lot of optical illusions, like a character in the foreground's hand touching something in the background
I choose not to doom over Krea.
>>105631024people like you are the reason why they're doing this bait and switch in the first place, because they know the vast majority of people are ok being betrayed and then move on, weak ass cuck you should be ashamed of yourself
>>105631005>That what you mean, a feeling of unreality?No, I'm talking about literal physical discomfort in my head, somewhere behind the back of my eyes, that is usually brought on by looking at certain AI gen images/videos. You know when you orgasm hard and your eyes roll back? Kinda like that, except without the orgasm or the eyes actually rolling.
>>105631045Oh, uhh... yeah, nah. Can't say that's happened to me with AI videos
>>105629711 (OP)there were some nice pictures in the last thread that would deserve to be in the collage more than some of these videos. don't make such errors again.
>caring about the faggolage
>>105631052>NOOOO YOU HAVE TO ACKNOWLEDGE MY ARTFAGGOTRY, LOOK AT THE ANIME GIRL LOOKING PENSIVELY INTO THE DISTANCE, IT LOOKS JUST LIKE A REAL AUTHENTIC HAND DRAWN BY A JAPPY ANIME MANGARINO!
>>105631065i wasn't talking about anything i posted. but the video meme is getting out of hand
>>105630954It's called slop for a reason
>>105631077Do you bitch when there's good video gens in a thread but the next collage is all images? No? Then shut the fuck up.
>>105630678They didn't train it "alongside" them, they used BFL's hardware to train it. They don't owe BFL anything
>>105627389How's the progress? Does injecting latent noise improve the quality?
>>105631077>crying this hard for a faggotlagelmao
>>105631116Nope. I tried a few different methods (latent movement, warping, zooming, etc), but it was no dice. The distilled scheduler using the proper code looks about the same as the FlowMatchSigmas scheduler I posted the other day too, so I didn't end up posting it. I'll probably take a look at the code again tomorrow and make sure I implemented it correctly though.
>>105631115of course they owe BFL a favor, they lent their GPUs to make it happen after all
>he didnt make it into the collage with his lazy still image slop
LMAO
"someone" is insecure today
>>105630553wdym? are you using other loras?
>the fairy flies towards the camera, getting big in the frame, her wings flapping rapidly and hair flowing. her breasts are bouncing shaking or wobbeling up and down. she leans into the camera and puckers her lips and closes her eyes to give a kiss.rip prompt adherance
>seething because hes unable to gen an anime girl pensively looking into the distance that looks just like a real authentic drawing by a jappy anime mangarino
>>105630954now that you mention it, last night my eyes randomly started to hurt intensely. then the video gen I was looking at looked all fucked up and choppy when before it seemed fine. very strange
where are all the undress models now at?
there was a cool lora that was like a clothes disintegration or something but i don't remember right.
>>105631295wan, obviously.
>>105629711 (OP)Any upcoming Wan projects you goobers looking forward to?
>>105631295>https://civitai.com/models/1423479/rip-her-clothes-wan21-i2v?modelVersionId=1608935GONE
cuddle
md5: 8ebbe4eb0f9f9bdc09c72134e549fa30
🔍
Makoto loves cuddling with her sister. God I wish that were me....
>>105631332well that's some shit. i hope some autismo who had more foresight than me backed it up on hugg
>>105631360well that was fast to find. https://huggingface.co/dnad244/wan_random_loras/tree/main
get fucked civitai. why don't you bend over further.
>>105631182Yeah I tried some with the same stack I used for vanilla 720 and the people have a greater tendency to morph and do unprompted actions. Maybe I just need to lower all the weights.
>>105629711 (OP)I playing with WAN light2x lora in rentry
>When using i2v on Wan, start by selecting the i2v 720P model in the model loader. Next, adjust the width and height settings of your input image to 1280x720 or 720x1280.>Set Teacache coefficients to i2v_720 Where is teache node in the sample workflow, I don't see it? Can you run teache with light2x lora thou?
>>105631295I have two:
"clothes_rip-shura-000006.safetensors"
https://litter.catbox.moe/uzp930oghq2c5srr.safetensors
"woman_undressing_v1.0_wan14b_epoch048.safetensors"
https://litter.catbox.moe/2k65cy14opwvxwv4.safetensors
I don't know the trigger words/phrases, sorry
>>105631377Why are there so many loras for the same thing
>>105631408differently trained, idk. undress is obv different to rip clothes
>>105631398https://huggingface.co/dnad244/wan_random_loras/raw/main/WAN_LORA_IMAGE_SAGE_TEA_v1.0.json
that's someone's workflow that has all the triggers in notes.
>>105631391teacache is not in the light2x workflow for whatever reason
>>105631437some anon mentioned it's because you only have 4 steps so there's nothing to cache.
you really should be using Mag cache instead as it's superior in every way. desu teacache should be removed. just a shame about it being a hard requirement for skip layer
>>105631454>shame about it being a hard requirement for skip layerDo the initial SLG releases that worked without TeaCache still work (if you had them), or has there been too many changes under the hood to go back?
>>105631477i have no idea. don't even know where to find that version of slg.
i'm a complete tard when it comes to making nodes and integrating papers but i assume slg can't be that hard unless it has some fucky dependencies.
>>105631505Disturbing, yet enticing
>>105631505why my light2x lora has snail pace movement?
any trick to make video natural?
>>105631391That doesn't apply to lightx2v, since you don't need TeaCache with a 4 step distill. I'll make a note on there tomorrow about that.
Does anyone else have the problem where the model's denoising becomes weird after a couple of gens? Like, it stops following the reference image after a while
I suspect it has something to do with caching, but I am not using teacache or anything like it
>>105631665Is you video FPS properly set to 16?
>>105631690I am obviously talking about Wan btw
>>105631665increase frame rate
>>105631690yes, an anon previously said it happens after ~30 gens. i've experienced it too
>>105631712>yes, an anon previously said it happens after ~30 gens. i've experienced it tooHas anyone investigated the culprit yet? This shit is forcing me to completely restart Comfy every time it happens
this is after many gens, right before it completely breaks
>>105631505kek
requesting this one melting onto the pancakes or something
>>105631726FWIW, and this was years back, there was another model early on in local genning that would have this issue on A1111. There was some way to clear your GPU cache or something to get it working again until the model/A1111 was fixed.
Not super helpful, but I do feel I remember this issue being present before and there being somewhat of a workaround if you can find it
>>105631752and then it gets more schizo
>>105631775>here was some way to clear your GPU cache or somethingOh you mean reseting gpu driver? on windows: Windows button+Ctrl+Shift+B
>>105631802Maybe? Clearly A1111 shows the age of the information, but this is worth trying
I suspect it may have something to do with either Kijai's implementation of TorchCompile or something broken with Wan and Comfy Core. I never had this happening without the lightx2v Lora, though. So another hypothesis is that there is a regression somewhere (I pulled git changes last week). Is any anon genning without the distill lora?
Wait, if I have aa 4070S and it displays pytorch 2.8.0 that's bad?
>>105631726i removed torch compile/teacache/mag cache and it fixed it.
or at least it hasn't happened since. some anon a while back said it's because of random pytorch/cuda bulllshit
how do you use wan's loras for light2x?
I put lora model only between light2x lora and model patcher order but result has 0 lora movement.
>>105631813yeah exactly. but it also happens with the other compile models.
>>105631833lower the ligh2x lora strength and make sure there's nothing in the prompt that the model could hyperfocus on instead.
i had a gen with the cowgirl lora have the woman completely spazz out because i had something like "her tits jiggle" or something like that.
light2x lora is cool but it needs some tard wrangling. hopefully the official self-force will fix this.
>>105631827>>105631849Weird. What does the lightx2v lora does that torchcompile "dislikes"? Again, this never happened to me running Wan without the Lora before
>>105631833try without torch compile
>>105631833Isn't the position in the rentry example workflow the correct one?
file
md5: 7b9bc013cb321581fd554b2a830e7d86
🔍
fuck krea dogshit model
the mercedes posting will continue until i get accurate leggings
i should also try making lance vance and the lawyer
>>105630444ok but seriously, if you degraded the middle video and put it onto film and dated the film how could people tell its a fake?
i dont mean for this hitler i2v but in general, or if you made a continuation from footage of hitler that was cut short and presented it as the
"lost extra footage"
>>105631759>syrup drips and flows. the yellow butter on the top of her head slips and falls and she makes a surprised face.
>>105631690There is some weird shit that can happen with loras + patch order + torch compile, where it gradually degrades things after a certain point. It is conditional though, with some configurations of settings it never happens. Try without torch compile and the patch order and see if it still happens.
>>105631996goddamn that's clean for a 8bit videogame style.
did you use a lora? can i ask for a catbox?
>>105631997teh most uncsored evaar!
lmao very funny anon's actually believed that.
>>105631997Why do these companies still try to pretend that the models won't be used for nsfw?
file
md5: 0492e5390f8c2e54ac6222178fc6acd8
🔍
>>105632044some pixel art lora to generate it with flux, i can dig it out if you need, and for video i just did this to make it look more 8-bit
>>105632051they have to because of payment processors and advertisers
the united states government unironically published a house committee report on their influence and censorship
1856
md5: fdc3f2e62d912d2092b4bb2940834678
🔍
>>105632055aw come on man you know what i mean.
>>105632058ooh nice, i'll have to try that. all good, i can make pixelart pictures, was just surprised how clear and actually pixelarty that video came out
>>105632075remove torch compile. imo it breaks more shit than it actually ends up saving time.
>>105632075Result. Looks nothing like the one on civitai.
Erotic outdoor scene: in a city, beautiful woman and huge breasts is dancing towards the viewer, her breasts movement is bouncing, her breasts rhythmically bouncing up and down. she embodies erotic behavior. Her breasts bounce as she dances.
>>105632012I think one of the reasons Wan is not up there with the sota saas video models is because its prompt alignment is pretty bad.
I'd be very happy to get a new Wan model of even just a fine-tune trained on a better captioned and more nuanced dataset
>>105632051NOOOOO. AI IS FOR *checks sora* MAKING GRUMPY CAT
>>105632084anon here
it also doesn't seem to play well with the light2x lora though i don't have hard facts to prove that aside from the quality degeneration and caching issues it causes
>>105632098i think the light2x lora fucks the prompt adherence
Correct pattern, incorrect color. this reminds me of color bleeding in early SD where if you mentioned a color (e.g green eyes) it shows up everywhere like in their hair or clothing too
Its coming from "black leather jacket" and now as i write this i just realized i dont need the black part at all because women dont really wear any other color leather jacket
>>105632106>doesn't seem to play well with the light2x lora though i don't have hard facts to prove that aside from the quality degeneration and caching issues it causesmay we see your soft facts? the "watermarking in the last few frames" schizo from last thread chickened out
>>105632141Maybe. But base Wan is not too amazing in this regard either, you need loras even for the simplest things like camera rotation and dance moves
>>105632084I still got same result
>>105632088but slower
3090 anons, what values are you using for virtual_vram_gb when genning at 720x1280? The ldg reentry doesn't have a recommended value as a baseline.
>>105632165watermarking? wat.
all that would happens is after a few generations the quality would eat massive shit, i've deleted them all but basically every consecutive generation would have more and more noise. after ~4 generations it was just pure noise being output until you restart comfyui.
granted, i believe this is just a torch compile issue and not actually related to the lora.
but i removed compile from all wf's and shit has been fine since.
file
md5: a26527b74afa81e67abcbe2bcbae77f0
🔍
>>105632182Im using 8-9
go to nvidia cp > manage 3d settings > program settings >add > (find your comfy's python.exe)
Set CUDA system fallback policy to prefer no fallback
This way, python with gives OOM error when you have it, instead of silently switching to system ram. Easier for you to know when you OOM.
>>105632249>arr rook the samelittle bit racist model
>>105632257Can you elaborate? I was under the impression that setting that value avoids execution being interrupted by GPU OOM. Wouldn't preferring no fallback defeat the purpose of setting virtual vram gb?
i fucking knew it
now the issue is WAN is too horny so it makes them see through instead of opaque
>>105632231>watermarking? wat.schizobabble, dont worry about it
>>105632278when you OOM nvidia will switch to system ram by default. You may not notice it because nothing's happening except speed is slower.
Change it to no fallback, it will just give you error and stop running.
Then you can just adjust virtual_ram settings until there's no OOM error.
>>105632381comfy is shit at making models. see sd3. let us have it to do something
>>105632304im not racist (in the sexual way at least), but i am face-ist
>>105632381comfy, having access to the early kontext weights, will surely provide us with a comparison of the early version and final release, right?
as a local-first developer, he will represent us and make a point to call out any censorship or nerfing of the model BFL attempts to pull before release, right?
he isn't just a SaaS API sellout whose UI should be removed from the OP for blatant API advertising, right?
>>105630931Yes
I'm only doing basic nondestructive editing (very rarely do i export) and I use the actions feature to bind a whole slew of commands to a few hotkeys. It can do everything I need and still works on my aging rig.
>>105632454this is why ani needs more influence. he does not pull punches. fuck comfy for licking up slop from his masters
What's the current to go realistic model(s)? Most shit on civitai is stylized nowadays.
>A watercolor painting of a ceramic mug with "Mystery Drink" printed on the side in a floral text pattern.
Flux, Hidream, Lumina 2, GPT-4o, Seedream, SD3-m, Chroma v33, Cosmos 2b.
Which is which?
>>105630678What if they do release Pro? They become the top dog and force BFL to release a better model. Competition is healthy. Besides, distilled dev version of their model wouldn't be as good as Pro at photorealism. It'd very likely pick up the fake skin from Flux, MJ, etc... still nice aesthetics for artistic images, but other than that the model could still be forgotten in the sea of other plastic skin models we already have, unless someone comes around and does what lodestone did with Flux, but destroying distillation is expensive to get model back on track as lodestone shows.
>>105632494unironically wan with 1 frame or some of the older SDXL meges maybe
how do i get the most out of NAG? i dont like the lack of movement, so i tried setting lora to 0.5, but then its all blurry. i bumped steps to 12 and this is the result
>>105632102cat go bleargh
>>105632494YOU'RE NEXT (to get generated)
>>105632494no idea what you are into but SDXL wise lustify, biglust and cyberrealisticXL5.7 are as good as it gets. zavychroma for fashion (last ver. was trained on that shit specifically). quite a few decent models floating around. moving up the ladder, flux (can do ok & can be fast af with svdquant (nunchaku) and chroma (very promising but dataset already fucked with huegmouf infencer lips and furryshit and landwhale crap)
>>105632499pls no guessing game. 1 chroma?
>>105632567bro how do you not immediately recognize 4o
>>105632494Depends what you want. Like the other dude said, I use BigLust a good amount. Chroma is coming along though and will likely be the answer when it hits v50
>>105632012Is that Aunt Jemima? LMAO
>>1056324994o, flux, seedream, 3m, HiDream, Chroma, lumina 2, cosmos 2b
>>105632581oops. to my defense, just never looked at anything 4o
i think video models are going to struggle with realistic water movement for a long, long time
if they're not trained for 24 or 30fps its already a non-starter
>>105632655what's "realistic" to you? 1/30 shutter speed? 1/1000? there's far more to it than frame rate.
>>105632681my point is more that if you have to interpolate, water will never look correct and always be too "fast"
but i dont think we will ever have to deal with interpolation again. why would you ever train at something other than 24fps for realism or 12fps(?) for anime?
did the wan devs ever justify their 16fps choice? even mochi could do 24fps and that was like 9 months ago now
that being said, oceans seem to look better than pool water when interpolated, so maybe im actually complaining about the interpolation model's intelligence
>>105628298very well written, thx for posting
>>105632608close, it's:
4o, Flux, Seedream, Chroma, Hidream, Lumina 2, Cosmos, SD3-m
or maybe im just looking for something to complain about. there's bigger issues in terms of realism than water
>>105632843interpolation fucks up a lot of my gens. i wonder if its better to prompt for slow motion then manually speed up the frame rate in a video editor, but then you are left with short clips
Is there a custom node in comfy that'd create an image for masking in a resolution I put in? I don't want to constantly open paint to create new blanks anytime I change gen resolution.
>>105629750i honestly wouldn't mind fucking this..
>>105632902One way is to make a gen, then right click the output in the "Save Image" node, select "Copy Image", then click your "Load Image" node meant for masking, press "CTRL-V".
>>105632890its better to wait 6 months for the next gen of video models so we never have to deal with interpolation again
16fps must have been a budget limitation on the WAN teams part
file
md5: 25d57b1bcca78fd4cd117ec99eece8fc
🔍
>>105632902You mean something like this?
>>105632902>>105632927why not just use the load image-> open in mask editor? so much easier.
>>105632951eh too much work
i want to run automatic
>>105631780>>105631752ive genned over 50 gens in one session on linux, with torch compile, fp8 wan and lightxv and havent had this issue
kijai btw
>>105632951Because I want to mask first, since you can't rely on the model to generate two characters .
>>105632528maybe share the workflow, and what you're trying to accomplish? nag is just cfg without cfg
poll
https://poal.me/jxzjeh
https://poal.me/jxzjeh
https://poal.me/jxzjeh
https://poal.me/jxzjeh
>>105630954It's the interpolation. It's 100% a meme. Normally the original non interpolated vid would be better.
>>105631665I'm still experimenting with this shit but:
1. Four steps is just way too little, even with self-forcing lora at full strength. I'm currently jumping back and forth between 6-8 steps.
2. Splitting your gen between two ksamplers, with the first one doing a couple of steps at high CFG helps out a lot.
3. lcm sucks, unipc is still king. ddim with ddim uniform is also pretty interesting.
4. Selfforcing lora by itself kills the motion, I'm trying to experiment by mixing it up with accvid and causvid at various weights. Hell, I also throw in the HPS/MPS rewarding loras at low weights for the first ksampler, which seems to help out a little.
5. Not only do you have to "boomer prompt" - you have to explain shit like you're talking to a toddler with a down syndrome. Every action, every movement has to be described as simply and at the same time as thoroughly as possible. Even then, I just can't make this one work
>>105630954it will get worse before it gets better
>>105633110yeah some things it will just never understand without a lora.
i consider this the sd1.5 of videos. it will take a while until we get a video model that understands so many more random concepts.
for the cat throwing up a hairball i posted, the prompt i started with was:
"a cat throws up a hairball into a bowl".
this didn't even remotely work as a hairball would materialise from out of frame and float gently into the bowl.
only the prompt "a cat vomits a hairball into a bowl" worked.
krea1_2
md5: c6ea5eb8d770b0b3db154c45d7c63185
🔍
krea looks dope but fuck this shit
should I even try to do video gens on a 4080 or should I wait until I get a non vramlet card?
>>105633183I do 480p on a 4070S
>>105633210how long does it take?
>>105633183a 4080 is pretty decent. you can definitely give it a whirl
>>105633215With NAG and 20 gigs offload, about 5 minutes.
>>105633175Can you try a photoreal prompt? Idk, like
"Amateur photograph, a cute Japanese woman in her room. She is sitting in her gaming chair with her leg up on the chair. The room is decorated with anime posters."
Just trying to see what triggers the photoreal look.
>>105633051im following the rentry. im trying to make high quality porn videos and then ill make a patreon. so far this NAG solution seems to take away too much movement even though its much faster. getting a good output is what matters to me, i just use sage and 720p takes 40 minutes, but its worth it having a nice clip forever. i always regret cutting corners when i view clips in my gallery
>>105633163isnt that more a text encoder thing? as in the text encoder isnt moving "throws up" in the "vomit" direction as well as "vomits" is due to the extra semantic meaning in "throws up"?
because its not the video model understanding the text, its the text encoder encoding the text in a way the video model can understand, right?
>>105633183no torch compile for you (i think) but 16gb of vram is fine (king of the vramlets, if you will.) even a 3060 12gb is reasonable enough for 480p sloppa now
>>105633274ran out of juice. you get 11 gens for free daily and it does them in batches of 4. it already draws the line at micro bikini, pfft. that floating porsche on their main page looks so dope, ugh.
>>105633332 thumbus longus but really cool, the face&makeup look really good.
>>105633331>even a 3060 12gb is reasonable enough for 480p sloppa nowtruth. genning on a 3060 power limited to 100w i can gen one 480p vid in 200seconds
when not power limited (170w) its 170seconds
>>105633393>thumbus longus(small delicate hand) etc seems to help little bit
>>1056334195 second video, 81 frames total, 16fps
>>105633110my problem is it doesn't seems to work with any wan lora.
There is like 20% movement at 1.0 strength
>>105633403remember the doom attitude towards video and its compute requirements only a month ago? I think this is finally the SD moment for video. Maybe prepare for a wave of legislation and outcry, but maybe people have better things to worry about in the real world right now
okay
https://files.catbox.moe/c5dgm0.mp4
this futa took 2 minutes. idk. for some scene the lack of movement is okay. i have had problems with making girls dance though, without lightxv and nag it will look awesome where you can see the facial expression change and hey body jiggle. but for very static images nag is okay
>Anon you're so sweet, tee hee!
>>105633464yeah i remember in march a single 480p 5 second video would take 20 minutes on a 3060, with teacache and torch compile, no matter if gguf or fp8 was used
it looked way shittier than the gens i can make in around 3 minutes now, until recently i was happy with accvid+causvid which took 6-7 minutes
the chinese are wizards
>>105632902what, you mean like creating an empty latent image in the chosen resolution and then vae encoding it to an image?
>Han Solo will never escape
>>105633467>futano vagoo = man
>>105633510Anime Screenshot Merge - VPred
https://civitai.com/models/1298771?modelVersionId=1767684
https://files.catbox.moe/mrtea8.jpg
https://files.catbox.moe/rwo51t.jpg
Why are all the older models unable to generate oneshota content? They always depict girls in submissive role and any mention of a boy results in a disgusting old black man. The fuck was wrong with our ancestors if the early anime datasets only contained that?
Literally all of SD1+2 and even SDXL models trained to replicate Anything do that. Only noobai and pony are cultured enough for femdom.
>>105633537nta but if you were trying to post the workflow, jpeg's dont save the workflow
t. downloaded your jpegs to make sure they dont work
>>105633475how
t. new to vidgen in general
>>1056335671) install linux
2) https://rentry.org/wan21kjguide
>>105633110based, thank you anon
>>105633589>based wan removes the hag lines on her forehead
>>105633548what the fuck is wrong with you. what a disgusting fetish, just like those bestiality enjoyers.
I want to add local image generation to my telegram bot, so I need a model that can be prompted in natural language (preferably uncensored) that can oneshot pretty good results (loras, upscaling etc are fine as I'll just preset them in the workflow, shouldn't take literal minutes for an image on a 4080 super however).
Any recommendations?
I'd appreciate it, /g/entoomen.
>>105633583>comfynevermind, not touching that meme
they're like the apple of gens, became the default standard for no reason
how do you apply RiflexRope node in Kjai wrapper workflow?
>>105633567wan lightx2v workflow, no torch compile, spinning everything lora
>>105633617what's wrong with apples?
>>105633617https://github.com/deepbeepmeep/Wan2GP
don't use pinokio
>>105633555am using reforge and prompts and settings are in jpg file when you drag it to png info.
hey it's anons tinder match
>>105633589I dare you
>>105633681we've reached the age of near-instant video gen replies. that's pretty big actually.
>>105633630I have the same question. I used that shit the other day and it basically made 8 sec videos be a slow motion version of 5 sec videos.
>>105633682why isn't this in the OP?
>>105633682Thanks, this looks like something I might actually use.
>>105633682why not pinokio? it works pretty alright for me but slow because i have 4060ti 16gb. Is there a performance difference between the package installer via pinokio and regular custom install.
>>105633741That shit occupied 25Gb of my space before even getting to downloading any models.
>>105633741pinokio is just more bloated shit you don't really need. the repo alone installs like auto pretty much, you don't need anything other than that
How do I stop Wan from zooming in or out in the gens, without changing angles? How do I force it to keep the same static take reliably? "Long take" does not seem enough
>>105633797i use these for negatives
>vibrating, slow motion, camera zoom in, camera zoom out, and in the positive you can say
>camera is steady
>uncut penis lora for wan just dropped on civitai
>on civitai
stopped reading
>>105633850she is ready top dive in
I need a comprehensive guide on facial expressions
being an autist I've always struggled with this, I need a look that says "I don't want to be here, but I have to be here", what's that called
>>105633298good shit
[X] gooned
>>105633885RIP emojis for expression in natlang models
>>105633885Try a combination of those: unamused, annoyed, deadpan, bored, expressionless,
>>105633885it doesnt really matter because models dont understand them either. you can prompt "angry expression" and one model may make them shocked, they other may make them crying. prompting is more about trial and error and getting intimate with your model. for yours you may want to try some combination of "annoyed, frustrated, skeptical"
>>105633298is this with lightx2v?
>>105633758>>105633761not sure what the 25gb from the pinokio app is for but if space is an issue in the future than i would consider nuking the install. I'm not ultra tech savy but the app functions appropriately to my needs. While not overly optimized, its noob friendly and still getting updates.
what a horrible day to have eyes
maybe page 1 bakes aren't that bad of an idea
>>105633938I think we found the big stinky
>>105633816Sadly not enough =(
Damn, I really hope they train another wan and add proper tags to the data
>>105631505I have a new fetish
>>105633968Why is Tommy 60 years old but Lance gets to be young
>>105633885I use solemn a bit, models seem to recognise it as just a blank stare, and it can be intepreted lots of ways
kek I forgot to resize the image
>man walks towards a bonfire
>>105629711 (OP)>>105627141Please post the original image for this one (the first frame)
>bread resurrected after anon got the hammer
>>105633885https://danbooru.donmai.us/wiki_pages/tag_group%3Aface_tags
>>105632001>presented it as the >"lost extra footage"There's AI models that can spot other model outputs via identifying patterns and artifacts unique to machine generated content. For now
Eventually though, a time will come when the models are so advanced that they don't have those patterns. Then all bets are off.
>hasn't hit bump limit
>multiple new threads
:|
file
md5: 089b35c5f42dce7eae2ab5a69b88afc5
🔍
>>105636605I do apologize, sir
file
md5: a70d1dff0ac3aca1e0919c42f74e15ae
🔍
file
md5: 4e6d42e09f47f6136fda7ebcbd76940d
🔍
file
md5: 9335ade3bb472bf420d02001cb19e1a3
🔍
file
md5: 9c6ae392a8ff5806d3adebcdcf8330f5
🔍