God Has Abandoned Us Edition
Discussion of Free and Open Source Text-to-Image/Video Models
Prev:
>>105774047https://rentry.org/ldg-lazy-getting-started-guide
>UISwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, & Upscalershttps://civitai.com
https://civitaiarchive.com
https://tensor.art
https://openmodeldb.info
>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe
>WanX (video)Guide: https://rentry.org/wan21kjguide
https://github.com/Wan-Video/Wan2.1
>ChromaTraining: https://rentry.org/mvu52t46
>Illustrious1girl and beyond: https://rentry.org/comfyui_guide_1girl
Tag explorer: https://tagexplorer.github.io/
>MiscLocal Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Samplers: https://stable-diffusion-art.com/samplers/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage | https://rentry.org/ldgtemplate
>Neighbours https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg>>>/b/degen>>>/b/celeb+ai>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
teach me how to train a lora on a 6 gb card
thx
>>105776992Die, in the next life do better life choices so you can buy a better card.
anon I found that you may apply vanilla flux dev loras to flux kontext dev if you set the steps below 10. I tried some nsfw clothing and it worked. can anon confirm?
what was the password on that rar from last thread
>>105775547removed an entire burqa lads
nice lora
>>105777018ldg
>>105777040For example, I could tell you to kill all the Japanese...
>>105776992Get a job and work for 100 hours
>>105777014Some redditor found a much better method
>https://old.reddit.com/r/StableDiffusion/comments/1loyav1/important_psa_you_are_all_using_fluxdev_loras/n0r9sex/
i'm starting to realize the per-parameter value of SDXL is much higher than flux/chroma...
I provide a workflow for those who want to switch between the Image Stitching and the Latent Stitching method with a simple press of a button, it also allows you to switch between 2 image inputs and 1 image input if you disable both switches.
https://files.catbox.moe/q3540p.json
>>105777144Stitching seems like an emergent behavior that needs to be properly finetuned.
reposting for new thread
>>105775770hello rtx 6000 pro enjoyer. please share your workflow
>>105776994Yes, but I did not expect 3060 to be better than a100.
>>105777170dude I already asked him, don't make it weird.
>>105777085>>105777010okay I'll just use civitai to train a lora, faggots
at least tell me how many epoch and other shit I'll have to run it for it to be good
>>105777200it's still a gaming card, graphics accelerators have a tonne of ram but are slow as shit
>>105777223Everything depends on the signal to noise ratio of your dataset and how much the base model already knows. Some things train fast some train slow. If your dataset is ass and it's a difficult concept it may never train or it will learn the wrong things. Also as a poorfag you must have lots of time on your hands, maybe spend some of that time learning by doing. :)
>>105777087I got oom on a 16gb card
>>105777279>16gbSTOP BEING SO FUCKING *POOR*
MY GOD, ANON
>>105777283Wah why does this 4K 3D render crash on my shitty GPU!
so far I tried
>euler
>use prompting weight:2.0 on motion parts
>NAG with slowmo, slowmotion ,slow
>0.8 lora weight
Doesn't completely eliminate self forcing's slowmo but I get good seed more often than not.
genning with 24frames/sec also help but then you get shorter video.
I tried 2 samplers workflow but didn't like it verymuch, because gen time was too long, which kills the purporse of self_forcing
Anything else I can try? Preferably ones that have good gen speed
>>105777223Go for 100 epochs, you haven't even told us what model you want to train since you are retarded, so I will assume flux, which means don't train the text encoder at all and use a unet training LR of ~0.0003 .
Start with a resolution of 512 to see the results before you start cranking up resolution if needed, also if you train at higher resolutions, always train on lower as well, like set the trainer to do at least 512, 1024.
>>105777332A good motion Lora counteracts the slowmo. More or less when I was doing my own Lora as far as I could tell it would match the pacing of the training clips.
Guys I'm going to paint, how long will it take?
- average spoon feed beggar
>>105777351>tfw the best Nazi representative of the 21st century is a black mankek
>>105777355>I'm going to paintjust do it anon. It isn't hard
And remember, there's no such thing as mistakes, they are just happy little accidents.
That's what you mom said when she gave birth to you :^)
Do you guys know of any benchmarks comparing the image generation performance between the 5070 Ti and the 5060 Ti 16GB? Thanks.
>>105777380islamic dillema dodging poohamad sissies... our response?
>>105777346Start with the defaults you get from Civitai, go for 100 epochs, 0.0003 LR, there is no magic bullet, there are only 'good enough defaults'.
There are endless variables depending of what type of data you are training and the model in question, does the model already know this type of data well, if not you probably need more epochs and/or higher LR, and if it knows it well you may need to lower LR and perhaps introduce regularization to slow training down.
For the best results, it's not a 'press button', if it was, there wouldn't be a gazillion knobs to tweak.
wtf
md5: f84166ab9f5d101662589c7e68e70eeb
🔍
Why the FUCK is MaskEditor so SHIT?
I click a spot to draw the mask, and where it draws is offset from the fucking cursor guide.
>>105777418niggers just go look up the memory size and bandwitdh and you have your "benchmark" for the speed you will get
https://www.techpowerup.com/gpu-specs/geforce-rtx-5070-ti.c4243
https://www.techpowerup.com/gpu-specs/geforce-rtx-5060-ti-16-gb.c4292
>>105777424I would never tune LR, it's one of those settings you really shouldn't change, same with an optimizer. Really the main variable to me is how many epochs and the best way to do it is to let it run and test epochs over time and find the best that does your objective while maintaining generalization. From experience the thing that matters most is your dataset, the quality of your images and the quality of the captions you used.
>>105777427reforge/forge still has a better inpainting ui than comfy
grim
https://civitai.com/videos
how do I browse only nsfw gens?
I don't see any in this
>>105777463It's for a project, not gooning
I'm asexual
>>105777448Comfy has been phoning it in for the past year
>>105775547Thank you for this lora. Spent some time playing around with it, and so far it's much better than the other one from Civitai. I've been noticing some boob inflation in some images despite prompting like in the examples, but I imagine that'll be fixed with the newer dataset.
>>105777472Start gooning.
>RifleXRoPE extends WAN video output by an additional 3 seconds, increasing the frame count from 81 to 129. However, it comes with some limitations: Higher VRAM usage, Longer generation time, A tendency to revert/loop the scene to its original state by the end of the video
How is this any different than just increasing the length?
>>105777481if you just increase the length, it'll look like shit, if you increase the length while activating RlfeX, it'll look less like shit
I just want 30 second video clips that can actually tell a story and not just be variations of the prompt.
>>105777124https://huggingface.co/Freepik/flux.1-lite-8B-alpha if you scroll down here, you'll find some testing regarding how much influence individual layers have on the output, and that was the basis for those models with pruned layers like the one linked or Flex
Then Chroma prunes the modulation instead
So there's a lot of fat on Flux, it's not the most efficient model
>>105777451found it
https://civitai.com/videos?tags=5146&view=feed
>>105777493Basically this. RifleXRope is only good for repetitive actions, like making some fat assed slut twerk. It's good for some sex LoRA's too, since again, that's just repetitive movement and it doesn't matter if it tries to loop or revert to the starting position. But if you try to do anything complicated, it falls apart. That's what I basically put in the note.
Once RadialAttention is released and if it's not some shitty snake oil, that'll be the go-to for increasing frames.
>>105777512We can tell Flux is overly bloated because we see even video models fit in 1B. Wan 1.2B is actually decent for its size which begs the question of why we don't see properly trained 4B models.
>>105777514now, how do I download all of them
>>105777478what is the trigger word?
just curious. why are female genital wan loras so bad? is it the same flux situation?
Guys my hands don't work so you're going to need to hold the spoon for me.
>>105777548As far as I can tell Wan can do perfect nsfw with a good dataset. The reality is the majority of people who publish Loras are really bad at it.
>>105777478No problem, glad other anons are getting some use out of it.
>I've been noticing some boob inflation in some imagesNot enough images, more images of women with big tits than small in the dataset. I've mostly been using it on images with big titties, so I haven't really noticed desu.
The 800 image set is a little more balanced, and it'll train longer, but I still need to go through the captions and make sure the LLM I used didn't fuck any of them. And training will probably take awhile, fucking ai-toolkit uses a little over 24GB for a paired dataset on Kontext @ 768x res, so I have to offload. It takes so god damn long to train.
>>105777547It's in the .rar file, in prompt.txt.
>>105777548>read description of any wan lora>"yeah i trained this on 5 clips"gee i wonder
>>105777472>I'm asexualhow do you know you just haven't found what makes you sexual yet
does your penis not get hard if stroked/stimulated
>>105777677it does and I can cum too but it's purely physical
I feel nothing
Another day of jannies not doing their job and fixing perma bans. Just can't wait for email verifications to be required.
thats some suggestive leg movement
>>105777405can you share workflow,
i tried but why it keep changing image composition
>>105777514>even gallery-dl won't let me download everything on the pagewhat should I use reeee
>>105777589I've also been testing it on some anime images as well, so the lora just might not be translating too well onto those. It also recognizes "topless" in the prompt like 20% of the time, but I'm pretty sure you haven't captioned for that.
Given that it has changed the boob size of some of the images I tested, it makes me think it would be nice to see some sort of breast slider lora specifically for Kontext, but that would be a whole different undertaking.
>>105777481Never managed to make Xrifle work with the rentry's workflow, I always keep getting people appearing out of nowhere once the video reach 5s
We're in an era where you can ask AI anything and it will produce a useful response in seconds and retards still beg for spoonfeeding on random forums hoping someone happens to be browsing who is willing to waste time on their retardation.
>>105777735it costs AI companies money to answer retarded questions
>>105777735Ai knows nothing about any info after it was trained
so chroma and kontext is dead now right.
>>105777749>write a python script that can browse a web page with videos and download them
>>105777732Yeah, there's no "topless" captioning, the paired dataset is just "fully clothed > naked", so it doesn't understand the concept of topless/bottomless. Surprised it even worked for you 20% of the time, actually. It's all real women too, btw, so it might struggle with anime.
Someone could make an anime version the same way, take naked anime pics and use Kontext in a batch folder run to clothe them all for your paired dataset.
>>105777754why are they in the matrix (green filter)
>>105777729>too oldthis general specializes in women who look 16,26, and 36 at the same time
>>105777749>Ai knows nothing about any info after it was trainedchatgpt, claude, gemini... can search on the internet though
>>105777735this wont change until every phone mic is gonna be hot 24/7/365 not just for tripple letters but for everyone too so normies can just say "make this work" to literal AGI who will set up everything needed on any of their devices to just make it work for them
so many still dont even know what an ad blocker is and so many can find this site and post a comment somehow but dont know how to search the internet with basic keywords
>>105777749thats why basically all ai can search the internet now or why you give them the page you want info from, npc
>>105777775better luck next time
>>105777771I thought they were supposed to deprecate search engines and now they hinge on them?
>>105777811I can't believe how stupid this conversation is. As if AI needs to know exactly how Civitai lays out their HTML because it has it scraped in the dataset, why the fucks does AI need to use a search engine to scrape video elements from a web page. Are you fucking stupid? I mean this seriously. How fucking useless, skill-less and incompetent are you? Scraping web pages hasn't changed for a decade.
>>105777589requesting more small booba as well
also more pale skin, too much tan
why some civitai videos dont play and are served only as images?
https://civitai.com/images/84458308
https://civitai.com/images/83638933
can anyone catbox these if they work for you?
No wonder he's a fucking doomer, he's such a manlet kek
>>105777735>>105777849smartass, script will only download public videos using the api, I want the videos hidden behind the filters
>>105777849why couldn't AI just make us a desktop app so we don't have to use comfy anymore?
>>105777508We'll have radial attention soon and hopefully, it wont increase the ram usage too much. 30 sec is the sweet spot for vid gen. I can see a future where Wan 3.0 has 30 sec as default and Pro versions have something wild like 2 minutes and prompt traveling like animatediff
file
md5: 1a227ad198b42d6868ae8deba5894e16
🔍
>>105777914Why don't you talk with ChatGPT, they have scrapers that actually web browsers. Anything you can see it can see and interact with. That means browse the webpage, even scroll down forever, and find all the data sources for each video which you download. You should seriously consider killing yourself, it doesn't get easier.
>>105777945chatgpt is saying I'll have to login through selenium and I don't want to give up my account
>>105777754kontext has a lot of potential but nobody is going to train it. it's way too difficult and the only decent lora i've seen is the pixel one anon posted here. everything else so far is just obvious GPT-slop ported over to kontext. the entire kontext ecosystem will just be saaaars asking GPT to ghiblify their SDXL outputs and then feeding the pairs into ai-toolkit.
>>105777965... you do know it uses your Chrome.exe on your computer right? Never mind, no more replies, you may talk to ChatGPT now, ask it to be rude to you and treat you like a retard.
>>105777936You can see right now how much VRAM it'll require by just bumping Wan's frames up and hitting gen. It'll try and cache for that many frames (even if the result would be dogshit), but you'll still see the VRAM requirement.
10 seconds = 162f
20 seconds = 324f
30 seconds = 486 f
>>105777968>it's way too difficult and the only decent lora i've seen is the pixel one anon posted here.I can't imagine the amount of work the BFL fags had to put to get good quality training pairs, jesus...
file
md5: 95e2bed0a028734abe4375765b9881b6
🔍
I truly worry about how incapable zoomers are at even tying their shoes without outsourcing to 4chan.
import os
import time
import requests
from tqdm import tqdm
from selenium import webdriver
from selenium.webdriver.firefox.service import Service
from selenium.webdriver.firefox.options import Options
from selenium.webdriver.common.by import By
from webdriver_manager.firefox import GeckoDriverManager
# === CONFIG ===
URL = "https://civitai.com/videos?tags=5146&view=feed"
SAVE_DIR = "civitai_scraped_videos"
SCROLL_PAUSE = 3
MAX_EMPTY_SCROLLS = 3 # Stop when no new videos after these many scrolls
# === SETUP DOWNLOAD FOLDER ===
os.makedirs(SAVE_DIR, exist_ok=True)
# === SETUP FIREFOX BROWSER ===
def setup_browser():
firefox_options = Options()
firefox_options.add_argument("--width=1200")
firefox_options.add_argument("--height=800")
# firefox_options.add_argument("--headless") # Uncomment for headless mode
return webdriver.Firefox(service=Service(GeckoDriverManager().install()), options=firefox_options)
# === MANUAL LOGIN ===
def manual_login(driver):
print(" Opening Firefox browser for login...")
driver.get("https://civitai.com/login")
input(" After logging in, press Enter here to continue...")
# === SCROLL PAGE AND COLLECT VIDEO URLs ===
def scroll_and_collect_video_urls(driver):
print(" Navigating to video feed...")
driver.get(URL)
time.sleep(5)
video_urls = set()
last_height = driver.execute_script("return document.body.scrollHeight")
empty_scrolls = 0
while empty_scrolls < MAX_EMPTY_SCROLLS:
print(" Scrolling...")
driver.execute_script("window.scrollTo(0, document.body.scrollHeight);")
time.sleep(SCROLL_PAUSE)
new_videos = driver.find_elements(By.TAG_NAME, "video")
new_found = 0
for video in new_videos:
src = video.get_attribute("src")
if src and (".mp4" in src or ".webm" in src or ".gif" in src):
if src not in video_urls:
video_urls.add(src)
new_found += 1
>>105777994 if new_found == 0:
empty_scrolls += 1
print(f" No new videos found ({empty_scrolls}/{MAX_EMPTY_SCROLLS})")
else:
empty_scrolls = 0
print(f" Found {new_found} new videos")
new_height = driver.execute_script("return document.body.scrollHeight")
if new_height == last_height:
break
last_height = new_height
return list(video_urls)
# === DOWNLOAD VIDEOS ===
def download_videos(urls):
print(f" Starting download of {len(urls)} videos...")
for url in tqdm(urls, desc="Downloading"):
try:
filename = os.path.basename(url.split("?")[0])
path = os.path.join(SAVE_DIR, filename)
if os.path.exists(path):
continue # Skip already downloaded files
r = requests.get(url, stream=True)
r.raise_for_status()
with open(path, "wb") as f:
for chunk in r.iter_content(1024 * 1024):
f.write(chunk)
except Exception as e:
print(f" Error downloading {url}: {e}")
# === MAIN ===
if __name__ == "__main__":
driver = setup_browser()
try:
manual_login(driver)
urls = scroll_and_collect_video_urls(driver)
print(f" Total videos found: {len(urls)}")
download_videos(urls)
finally:
driver.quit()
>>105777903doesn't work for me either (trannyfox)
>>105777980i mean it could just all be color-corrected gpt outputs too who knows. they did it already with midjourney outputs to train the original flux
>>105777980I think these capabilities can be trained emergently, you first start with very simple and programmable transforms and as the model gets smarter you pile on more complicated transforms. They already have a very good inpainting model so you can use that for the in-place fixing. It also helps when you run an API service because you can steal people's decent outputs and inputs and then it's just using an AI captioner to caption the difference.
>>105778029nah, 4o changes the image a lot when you ask for edits, Kontext is much better than that
https://www.youtube.com/watch?v=Ot_aYxptzJ4
>>105777979>tfw OOM anything over 17 seconds>tfw reached 20 seconds once and never againit hurts...
>>105778097Does it even remain cohesive at those lenghts ?
>>105778127It's just a test for when RadialAttention is released and allows longer wan videos which do remain coherent.
>>105778127nope quality degradation is significant.
>>105778139How does that work? Isn't wan trained in 5 second clips? How does using some optimization make it perform better outside of the parameters in which it was trained? Explain as if you're speaking to a golden retriever.
use "character" if you want kontext to keep the looks, man/woman/they/they/LGTV will turn it into generic cosplayers
Where can I get sageattention 2.1.1? I want to compare it to 2.2.0.
>>105778214Who wants to bypass 5 sec?!
Whooooo wants to bypass 5 sec?!
It bypass 5 sec, yes it does! ~
>>105778214Idk, those chinks are fucking math magiciens lol
>>105777823damnit she dodged it
>>105778258you build it yourself or
https://github.com/woct0rdho/SageAttention/releases
^windows only
>>105778214BARK BARK BARK
BARK BARK
>>105778276>you build it yourselfI want to do that but the branch for it isn't there?
>>105778214>How does that workmagic
I still don't understand self forcing and why it works or why it speeds up generation time so much
I dont think the researchers do either, because self forcing SOUNDS like something that shouldve been discovered for the first video model, but took until now to find
Its like when you build a website on a super high quality computer, and then when you test it on your moms laptop nothing works because you didnt need to optimize on your hardware so you didnt even think about it
>>105778258Both Windows and Linux wheels:
https://huggingface.co/Kijai/PrecompiledWheels
>>105778290the branch is the main branch, you can build it from there
https://github.com/thu-ml/SageAttention
>>105778214Just because the training clips are 5 seconds doesn't mean it can't infer motion from prior frames (ultimately that's the attention). And say they did 5 second clips, they're often from a larger video which the entire model was trained on. e.g. a 60 second video split into 12 5-second clips. If img2vid works then anything is on the table. If it can handle longer contexts it might be able to infer motion from all that predicted frame data.
Something someone should consider in the future is supporting multi-frame context inputs for video. (Say you input a 1 second clip at 4fps which is 6 frames) and then infer the next series of frames at the regular 24fps. This could give you forever stitching.
>>105778293>magic >I still don't understand self forcing and why it works or why it speeds up generation time so muchto me, NAG is just as impressive, I won't be surprised people will train models with NAG instead of CFG now
>>105778324the sageattention paper is literally magic too if you look at the images and comparisons
>NAGall of these little iterative improvements just go to show how unoptimized the whole thing is. there is so much low hanging fruit and we don't know what we don't know
after seeing WAN go from 15 minutes on a 4090 for 480p down to 2 minutes I have stopped pretending like I can predict trends anymore
>>105778300I tried those and they don't work for me.
>>105778302The main branch gives me 2.2.0 unless I'm doing something wrong.
>>105778366Try to switch to an older commit that has the 2.1.1 version
like commit 1718ddc06dbc694bcf3c6b49ac28c1921aa2d8bd for example
https://github.com/thu-ml/SageAttention/commits/main/
>>105778362>there is so much low hanging fruit and we don't know what we don't knowWe should compile a list of quality and speed optimizations that everyone just ignores or have bad support.
>NegPIP>Golden Noise>a bajillion different attention optimizations
>>105778489ive never even heard of any of those kek
who knows, mochi might secretly be the #1 video model right now with all the optimizations that technically work for it but no one will ever try it again so we'll never know
>>105778489I like Sliding Window Guidance
>>105777087>much better method>the method is just getting lucky but doubling gen time.anon,
>>105778701YOU'LL DRINK THAT FUCKING SNAKE OIL AND YOU'LL *LOVE* IT
>>105778489honestly, it wouldn't be forgotten if it actually was fucking implemented in main applications instead of letting retards maintain extensions
>>105777976it just doesn't work
I tried selenium, it just won't let me log in because of clodfare
I tried cookies but cookies aren't working either
>>105777087So basically you could modify those loras and save them so that you can load them normally after?
>>105778489There are so many optimizations out there and there will always be more because no matter how great your hardware is in your big tech company, there's always the need to lower the time to generate / train.
Which means there's always funding for ai optimization projects in academia, and really smart people working on them. Which of course is great for us with consumer level hardware as well.
Deepseek for example wasn't even academia, just a bunch of really smart of chinese stock traders dabbling with LLM optimization on the side.
>>105778723I already gave you another hint retard.
>>105778770just make me a ripper, baby girl, I'll love you long time and it'll come handy to other users as well
>>105778801You mistake my position, I think you're a waste of space and only worth mocking for my own amusement, I gave you what you need to do it yourself, I have no interest in doing anything for you. I think Civitai is a cesspool and the videos you want to download as Pajeet trash, you can think for a second what I think of you for wanting to download them, let alone begging me to make something for you to download them.
fuck u
md5: b904adf7fc856faea39fb1231a3f4509
🔍
why are there so many hyperspecific nsfw wan loras ffs
constant body horror, too
image
md5: cbcba6bde7fcbb9eaa8bd70ec3a329a9
🔍
What would you prompt to achieve this aesthetic?
>>105778958https://civitai.com/models/263107/sdxl-flux1-d-matte-vantablack-experiment
the power of wan. meanwhile veo 3 slop floods normie internet. local doomed?
i'm going to be controversial.
nag is not that great.
>>105779380mag is amazing fuck you
>>105779390i took the snake pill and died
>>105779370sure but who has time and tism' enough to make actual art with this besides bouncing boobas of 2000 year old witches.
>>105779352it's much easier to make porn. wan is a smut generator, not a normie gacha
>>105779406a question for the ages
>every local diffusion general on the red boards is just gooner central where they ERP and the same few full time gooners keep reposting their image sets
>no one will share prompt tips or models/loras
>just ERPing
>>105779411true, and for that, I am grateful
>>105779380nah you're tripping, this shit is literally a replacement of CFG for models that can't handle CFG, it's kind of amazing really
>>105779448Imagine if every time someone posted art you required a full video tutorial. We're not in SD 1.5 days any more, anything can be created with minimum effort and if you're gatekept by this you'll never be able to do it no matter how many video tutorials you're given. And most people aren't interested in you literally duplicating them.
cute
md5: 791eefbf322693a2b11faa797acda38d
🔍
Recommended checkpoint if I want to generate oil paintings that look like something Rembrandt would paint?
>>105779554a1111 days were peak because everyone understood how worthless this shit is and would just share all the time. cumfart age is disgusting shill grifter behavior
Will wan 2.2 raise the 5 second limit? God damn I hope so...
>>105779468is this seriously what hag enjoyers find attractive? none of these women are above a 5/10
the girl is holding a silver revolver. keep her expression the same.
>>105779682can you post on /v/ instead? I just really hate your images
>>105779663what are you even talking about, all the "impressive" workflows I've seen here are bloated messes made by retards, some of which could be solved with a custom node making it redundant or a external workflow script that properly managed inputs and outputs
it's actually impressive how some of you are proud of the shit sculptures you smushed together
>>105779693thank you for encouraging me to post more!
>>105779682it would be more impressive if you actually maintained image quality instead of making it look like a bad photoshop collage
>>105779693better gun output, feel free to use one for yourself!
>>105779693>can you post on /v/ instead?can you eat some dicks instead?
>>105779703what are you talking about about? If anything I am agreeing with you but making more fucking shitty nodes to pollute the already garbage infested community workflows on civit. comfy isn't retard proof like auto was except that both are retarded python applications that require tinker tranny autism to keep running
ayo
md5: fe8ae8600f76537f3beb9e3ce8d4b877
🔍
>>105779593Base SDXL and for example: "portrait of a noble woman in the art style of William Dyce, Baroque, Impressionism, portraits, realism, Renaissance, romanticism"
>>105779693>I just really hate your images
>>105779755A1111 is awful and couldn't be easily extended
Comfy actually works even as an API
>i-it's not my problem
>keeps replying to the post that threw him into a melty fit
your images are dog shit lol
here, so the baby crying over pauline can stop sobbing:
running through some stuff I've saved from past threads
>>105779352While i can apricate the existence of wan, ltx, skyreels and framepack being local open source. Frankly those models an absolute joke when compared to closed source api models.
>>105779810that shall remain nameless
>>105779794>has a melty everytime he sees an image he doesn't like>says it's the others who have a melty>too much of a pussy to respond directly to otherslike poetry
>>105779792extensions aren't that hard compared to nodes. the API is shit yeah. comfy installs viruses to your computer and has invasive telemetry. again, extensions were a place for people to contribute but nodes are just mass manufactures snake oil slop 80% of the time
>install nunchaku
>automatically downgrade some libraries and break teacache
>install missing dependencies to try to correct the issue
>for whatever reason comfyui uninstall torch
>install torch back on the most recent update, dev129
>break wan
>downgrade to a recent dev128
>wan work again but break sageattention
I love comfyui so fucking much...
>>105779830I don't like your post. That is all, good day.
>>105779847>not creating an entirely new ComfyUI installation before trying anything at allskill issue
>>105779847why not make your life easier.
https://github.com/deepbeepmeep/Wan2GP
>>105779671No clue, there seems to be zero real information on it other than that one tweet
>>105779671Even if they don't do that, I will still be happy if they improve the prompt alignment which currently sucks for the most part
>>105779979Radial Attention will allegedly raise the 5s limit "up to 4x" with no downsides
>>105779847find . -type f -name 'requirements.txt' | parallel -N 99 'uv pip install --torch-backend=auto --upgrade -r $(echo "{}" | sed "s/ / -r /g")'
Is it wrong to do this? I handle all dependencies at the same time, instead of doing it sequentially...
How much can I choke my card on powerlimit before it starts affecting gen speed negatively? Currently at -20% Also does core clock affect cuda cores or it doesn't matter?
>>105778489There is a bunch here:
https://github.com/pamparamm/sd-perturbed-attention
I only use smooth energy guidance though
>>105779675as a hagGOD, no, those arent even hags and yes, they are below mid
>>105779468Interesting how the (I assume kissing lora) interprets the different images and turn it into a wall, you can basically throw anything at Wan and it does a good job.
Just how far ahead are the chinks if this is what they give away ?
how many years after AGI will python, torch and cuda management hell be solved? is it physically possible given the current known laws of physics?
soonTM
md5: 7717b33fe3095bba37fc97ef178e53c3
🔍
>>105780534one day we'll cet Comfy.cpp, trust the plan
>>105780551and then we'll have to wait a few years until koboldfycpp so it all 'just works" and then a few more years until ik_comfy.cpp so we can run the biggest models at better speed
>>105780534>python, torch and cuda management hellPeople will accuse me of larping, but I've been doing open source ML stuff since like 2014, and I have never once had major problems with Python environment stuff. Sure, I've run into some problems at times, but they have always been relatively easy to fix. I see people claiming they've spent hours trying to resolve errors and dependency hell and I just don't understand how. Skill issue tbhdesu.
>>105780584I think you are just naive to think doing an isolated python project is the same as trying to adapt hundreds of shitty python written repos all with different dependency versions
>>105780584It's not about dependencies but performance.
>>105780605anistudio is our only hope unironically
In case anyone was wondering, sageattention2 isn't worth it for chroma. It makes the quality of gens noticeably worse and needs pytorch 2.7 which makes chroma run slower when you turn sageattention off compared to pytorch 2.5.1.
>>105780584because you're probably not trying to make some bleeding edge project from a team that only used h100 with old as fuck versions of everything work on a local gpu or you dont do it much for most new releases every week
>>105780605>muh python performanceProjects like Exllama and vLLM are as performant or moreso than llama.cpp. The underlying kernels are very efficient, it's just a matter of tying them together efficiently which can easily be done.
>>105780656I don't think you know what you're talking about if you think Python is faster than compiled languages.
I noticed that neither sageattention nor --fast has any performance impact on nunchaku
is it by design?
>>105779352bytedance was supposed to release something great for local. but these monsters released online seedance
>>105780678wait until he finds out why torch compile is fast kek
>>105780681snake oil is by design, yes
>>105780678>if you think Python is faster than compiled languages.I'm not claiming that in general, I'm just stating that for running LLMs, Exllama or vLLM get as much or more tokens / sec than llama.cpp does. All of the compute is in the underlying CUDA kernels, and if the Python part is correctly written, the Python overhead rounds down to 0.
C / C++ is not good for ML projects. It is a fact that the llama.cpp codebase is a massive incomprehensible clusterfuck, the project has major issues with onboarding new devs and adding support for new model architectures. It continues to exist purely off of momentum. There is a reason all this ML stuff is written in Python.
>>105780734Yeah sure. Afaik ik_lama fork is supposedly the hot shit now too.
>>105780534Python is shit, but it is so entrenched in academia that it will never be dethroned there.
CUDA itself is not tied to Python in any way, you can build a Torch ecosystem in any other language, but that will a lot of take time and/or a lot of money. So in order to do so there must be some real motivation, and so far there's not.
anon i found the foot closeup vlog prompt
i need 20 second videos with no degradation
i need 20 second videos with no degradation
i found the foot closeup vlog prompt anons and the AI is a genius and put her on a swing so her foot rocks back and forth
i need 20 second videos with no degradation
i need 20 second videos with no degradation
where the FUCK is radial attention
this will save the children i NEED 20 second videos NOW
>>105780678You literally think they're writing the billion parameter loops in Python? Or maybe, just maybe, Python is a wrapper for bare-metal GPU code...
>>105780814sir this is a christian minecraft server
>>105780734>Python overhead rounds down to 0.>still have to download GB worth of garbage to a fence>still have god awful memory bloat>garbage collector overhead>garbage UI runtimesyou're a fucking script kiddie kek
Oh look jannies failed yet again to perma ban ensuring we'll get email verifications because he can just as easily post CP and change IPs.
>anons are so cute when they are arguing
>>105780831You actually think Pytorch is Python allocations to VRAM?
>A fake child on a swing is CP
>>105780815you can't access the bare metal GPU in python to tinker with optimizations more effectively. litteral applekike behavior
>>105780856And yet your ban will say "Pedo shit".
>>105780862Yeah no shit you fucking retarded it's compiled C code, exactly what you wanted. Go make changes in the pytorch library and recompile.
>>105780850libtorch runs the necessary functions but the pytorch wrapper returns the memory for python to cache. so far all you've been doing is making a better case for going low level, not stinky jeet their high level
>>105780706A veo-tier model would probably require more than 48gb vram to run, even quantized.
>>105780876and what? make a wheel then uninstall the version I already have installed, install the new one and test to see if it even works? why add all these extra steps? no wonder research is in a slump right now
never make the mistake writing "holding a camera" versus "holding THE camera" if you're trying to get selfie style outputs
i now have 30 "holding a camera" gens queued up and comfyUI sucks ass and there's no way to manage queues
I have access to email verified contexts so email verification doesn't really matter
but this is the last webm
>>105780934great now how do i
- see the prompt that is queued
- rearrange my queue
When they require verified emails it will be Apple, Microsoft and Google accounts. Let's find out how many you'll burn through.
>>105780784>ik_lama fork is supposedly the hot shit now toolol just searched this up. props to ikawkrakow for saying "fuck you gerganov" and making his own llama.cpp with blackjack and hookers. i dont remember the drama well but ika was always trying to add SOTA stuff and gerganov was pushing back and then there was some licensing faggotry
>>105781003i will pay a service up to 100 dollars a month to take care of that for me. itll probably cost 10 a month max though
>>105781014They'll ban you after the first CSAM report hits them.
>>105781003>>105781014how about to post on 4chan you have to deposit 100k USD and if you get banned the jannys take that money.
>>105781036learn what a "verified context" is and what "orbs" are because you're starting to embarrass yourself. there are email-verification only generals on this site already. there already are services to dodge these things.
>>105781039you're gonna get hired by DOGE with ideas like these
Managed to recover everything but nunchaku, though I will need it since it's where the radial attention guys want to implement it
I hope it doesn't break something again...
>>105780052Never tried this
>>105779963I will give it a look, thanks
>>105778873so funny to see APIkeks suffering
file
md5: 1a97b779b2c90e03f5ce7b25d26de45b
🔍
>>105781014>ikawkrakowNah, it was just the license stuff. But he also has a 15 years vendetta against ggerganov...
>>105781014>this fucking guyThe man is basically a genius, he invented k-quants and bunch of other stuff I think, but holy shit he is insufferable. I remember some spat about license or code attribution or some bullshit. Like, no license was actually being violated, but he felt his contributions weren't being acknowledged prominently enough or something. So he completely sperged out, attacked other devs, and demanded all his llama.cpp contributions be removed.
Fuck people like this. Doesn't matter how smart they are if they perpetually cause drama and are impossible to work with. To this day I believe he still throws fits when anybody tries to take stuff from ik_llama.cpp and upstream it, or even just consider doing that. Even though both projects are MIT licensed, and ik_llama.cpp is constantly copypasting code from upstream. Even comfy is nowhere near the level of egotistical douche as this guy.
Is a gaming laptop with an RTX 5070 TI sufficient for running locally?
>>105781129>talking to AI>HEYYYlol?
>>105781255ego is required to take on ridiculous projects
>>105781325can't wait to see how ani will btfo comfy
>>105781343ego also creates great ash piles
>>105781302are shoes number 44 (eu) gonna fit my feet? do you know? ?
should i be prompting differently with NAG compared to CFG?
>>105781469nope, it's the exact same thing
>>105781313its a fake screenshot, everything is a fake screenshot now :(
Damn, that nsfw wan finetune an anon posted last time got my prompt right on the first try even though I had to pray to the gacha gen gods when using loras before to get something somewhat ok with NSFW
>dataset of 30,000 explicit videos and 20,000 high-quality images simultaneously
Well, not surprising it gets most stuff right, though I wish they did an i2v finetune too, but doesn't seem to be in their plan sadly
Dunno if they have weebshit in their dataset, doesn't seem like it
>>105776972 (OP)>Guide: https://rentry.org/wan21kjguide>Decreased visual quality, though its less of an issue at 720p>Reduced motion fluidity and prompt adherence>Favors slow motionI'm getting back to this stuff now that I have a dedicated 3090 pc.
Last time I tried using wan with 4-5 steps optimizations, the output was coherent but it looked horrible, so I guess there is nothing yet that doesn't have a big impact on quality yet helps speed like teacache for example?
It's either 5 steps looking like shit or 25+ slow but ok looking?
Just made a lora of my oneitis, now I just need good chroma settings for realism
>>105781516>but doesn't seem to be in their plandamn it
>>105781353so who is more egotistical? cumfart or TRANI?
>>105781534yeah especially knowing that current loras shared on civitai are too specific and usually trained on not nearly enough material
>>105781516>that nsfw wan finetune source?
>>105781530>>105781575https://huggingface.co/NSFW-API/NSFW_Wan_14b
Seems like there is at the very least the common anime style in their dataset
Mind you I didn't test a lot since it's t2v, so I could be wrong about it's capacity
this shit is expensive huh?
>>105781516Have you just tried the full finetune or the lora ?
>>105781665wait there's a lora? can it be used with i2v?
>>105781516the lora should hopefully be able to be used on image to video
>>105781656>100 grandif that's true holy shit...
>>105781671>nsfw_lora_wan_14b_e15.safetensors 1.25gbNo idea if it can be used with i2v
>>105781665>>105781671>>105781680Of course I tried the lora with i2v, it turns the whole image into blurred shit, even though you can still somewhat see the movement seemingly working fine behind all the blur
it ain't working yet but it would be great for i2v if it did
>>105781656>I had to make sacrificeIdk man, I would've been ok with him undertraining the model but in a non distilled mode, going to the finish line no matter what isn't always a good thing
>>105781701>turns the whole image into blurred shitshit
>>105781701you tried both high and low weight? Maybe they messed up extracting the lora, normally those work fine.
>>105781708>non distilled modewhat do you mean? Its not distilled
>>105781683Well, at least a new release every 4 days is likely going to get a delay
>>105781718he has been training a distilled version since 29.5
file
md5: 7f5a21bbea6b0c732db5cf5ef202b715
🔍
>>105781718>what do you mean? Its not distilled
>>105781718>what do you mean? Its not distilledThat guy is retarded, keeps posting a blurry flowchart image and claim the model is distilled
x
md5: 0d132af3dcc0f122046c6ddf53192905
🔍
>>105781684yea, he did it somewhat efficiently for a small independent training, it's just an arsepain to get/rent the hardware.
>>105781727from what ive seen he has several tests runs testing different things to try later on the model when its more mature, the main model is not distilled
1984
md5: 2ddfeb4f578f1533cc7a45162a48ff21
🔍
>>105781727Show me where lodestone says the model is now distilled
Come on...
>>105781748look at the chart again, everything is mixed with "fast" now, ther's no pure undistilled main models anymore
file
md5: f227125e4a798b426e0c23d49bc3ffd7
🔍
Do I have to add things in Chinese here?
>>105781760thats wrong though https://huggingface.co/lodestones/chroma-debug-development-only
file
md5: 9f7a80b4b1c907bca0b15207e55de8e6
🔍
>>105781690Is he retarded? This is from hyperbolic's website for their on demand instance for H100s. Even if he rents H200s, he isn't even remotely close to that $700+ figure a day. Is he being retarded and renting Blackwell B200s?
>>105781656$100,000 dollaridoos.
I assume that if he has this much disposable it's not a problem.
>>105781775i think he's just ESL and means malaysian dollars or some SEA currency
>>105781772no it's just almost ritual posting from what alibaba showed they were using when the model was out on their site (before hiding it), and since it works, well people continue using it
>>105781774So he wrote a wrong chart is that the hill you want to die on?
>>105781727
>>105781782In any case he should treat the training as an investment. Shit in = shit out.
Don't train if you can't train as simple as. Makes no sense to waste resources if you don't have them in the first place.
>>105781787did he say that was what he was already doing?
>>105781760Mixed with fast is not distillation you stupid fuck
Here is what lodestone wrote a couple of days ago on his Discord:
>i will let it complete first then i can train my own bias on top of it later on>i want a good uncensored and unbiased base model>no DPO bs, no locked in distillationThere, dumbass
I thought the password for the .rar file was "ldg". Part1 opened fine, but the part2 file shows an error, saying it is the wrong password. Can anyone confirm?
take it for what it's worth
>>105781523is there a list somewhere of all wan optimizations and their impact on speed/quality?
>>105781718>you tried both high and low weight? Maybe they messed up extracting the lora, normally those work fine.Tried both yeah, not working
We were so fucking close...
>>105781806>Mixed with fast is not distillation you stupid fuckall right, let's start again for the slow in the back of the class, what is fast?
>>105781826literally just told you
>>105781809
>>105781806>staging fast = low stepsthat's step distillation, that's like flux schnell, only retards would believe otherwise, like this low IQ for example ->
>>105781806
>>105781838he explained before it wasn't, you will have to comb through the discord yourself, also I don't think its in the base model
>>105781848>also I don't think its in the base modelit's everywhere, you're completly retarded anon, holy fuck
>>105781809
>>105781838What the fuck are you babbling about, it's a model merge of several branches which have different training parameters such as resolution and batch sizes, it has nothing to do with distillation.
You are just so full of shit.
>>105781867again, you said "I don't think it's in the base model" yet "fast" is absolutely everywhere, you don't know what you're talking about, you're wasting my time you retarded monkey
staging fast is a high LR, not distillation
>>105781870>>105781888>it's not distillation it's just a low quality training that destroys the overall quality of the model, ahah I winhow do you think this is a win anon?
i think staging fast is for initial distribution preservation not low steps
file
md5: bf42030ee27d962990d4ea7a8dae7e5e
🔍
>>105781903>i think staging fast is for initial distribution preservation not low stepsyou thought wrong
>>105781918>without distillation
>>105781923>"it's not made for low steps">*shows evidence it was made for low steps*>yeah... well... *time to move the goalpost*I knew you would do something like that, here's the answer for that ->
>>105781899
>>105777087Anyone proposing general solution for AI anything is talking out of his ass.
>>105781859Let's try to make you understand 'distillation', using Flux dev as an example as it is distilled.
It was trained on output from Flux Pro, that makes it 'distilled', BFL also took it a step further and generated the Flux Pro images at a small range cfg settings, and then faking actual cfg in Flux dev, which speeds things up.
Lodestone has talked about testing locking cfg to ~4 ONCE THE MODEL IS DONE since it would speed up inference a lot. It is NOT done now since the model is training.
They have gone slightly crazy and changed the VAE recently, that could really be something that forces a rollback.
>>105781918that does not mean its wrong, as he says in the image, its not distillation
>>105781936>It is NOT done now since the model is training.it has been done since v29.5, with the fast branch that is everywhere and makes the model work on less steps, he decided to kill the quality of the model to please the vramlets who couldn't be bothered to wait for 50 steps
>>105781918>>105781727
file
md5: b08f6c95be966b36ea45f54c3c364caa
🔍
>>105781936>They have gone slightly crazy and changed the VAE recently, that could really be something that forces a rollback.Interesting. They changed it for what VAE?
>>105781918Posts image claiming it's distillation despite being proven wrong time and time again.
The actual image says 'Without distillation'.
How can you even be this dumb and still remember to breathe ?
>>105781953Except he never said he did that and the latest models sure dont act like it
>>105781964>Except he never said he didexcuse me? do you have eyes? can you even read?
>>105781918>babe wake up we managed to create low step modelhello? is there something in that empty head of yours?? hello????
>>105781953he never said that was the case, he stated before he was testing things alongside it with different instances for when the model was more mature
>>105781979Can you?
1. He says that it is not distillation
2. He says he is making a based model without any distillation / bias
3. He said he was testing multiple things alongside the main training run separately
>>105781956I don't know, I just skimmed the lodestone channel and they talk about the 'new AE' they are switching to, I think it has better compression capacity, hopefully without losing quality.
>The new AE is already underway and the recalibration has been running for some days now
>>105781959>well yeah he's making the training quality worse by increasing the LR, but that's called a different name compared to what you said, therefore everything is fineif you want to argue about semantics please do it, I'll focus on the quality of the model instead
>>105781993>basedbase model, but that as well I guess
>>105781997for anything complicated / artsy it beats everything else atm and its not finetuned for aesthetics / details or anything yet
retard
md5: b75140d2cf93b33ab696f9b582c46141
🔍
>>105781993>He said he was testing multiple thingshe said he had to made "sacrifices", and the "fast" model is one of them
>>105781656even him admitted that he had to make the training in a worse quality to make shit faster, if you don't want to admit that, why should I give a fuck, I'll take the word from the guy who actually trains the model insteas, who the fuck are you anyway?
>>105782013when did he state the "fast" model was some compromise? stop making shit up already. He said he couldn't make the dataset as all encompassing as he wanted for artists / poses
>>105782025>when did he state the "fast" model was some compromise?so you're telling me that increasing the LR to make the training faster (and shittier) is not a compromise, all right you're wasting my time, go act like a retard somewhere else
>>105781997What the fuck would increasing LR have to do with distillation ? You are beyond retarded.
Distillation (in AI terms) is when you use a large models output to train a smaller model, typically to make it more efficient on the specific output you choose from the larger model.
It has nothing to do with learning rate.
>as he gets angrier, the ESL starts coming out in his words
>>105782031>stop saying that he's dead, he's in a COMA!! THATS NOT THE SAME THING, therefore, the guy is completly finenah, the guy is still in a bad shape anyway, if you want to think this is any better, you're more retarded than you thought
>>105782030>so you're telling me that increasing the LR to make the training faster (and shittier) is not a compromiseIncreasing the LR could indeed lead to worse results, it has NOTHING to do with 'distillation' which you keep yapping about like an absolute moron.
>>105782043>Increasing the LR could indeed lead to worse results>could take a load of this faggot
>>105782030I did not say that
>>105782043not really, depends on batch size and other settings, increased LR does not make it automatically worse somehow
>>105782052Yes, could, sometimes it leads to better results, it depends on how hard the data is to train, but also on your batch size, if you increase batch size you need to scale your LR as well.
You really don't know jack shit about AI training do you ?
>>105782065>increased LR does not make it automatically worse somehowhttps://www.youtube.com/watch?v=hpbGz9JPadM
>>105782076doubt all you want, its true
>>105782065>>105782074if he noticed that increasing the LR on fast lead to something similar (or something even better) he would've trained the model only on fast mode, yet that's not what he's doing, he keeps doing fast and base, because he knows base is still the better quality model
can't believe I have to explain something this basic but hey, even retards can be educated on some point, I believe that
>>105777332i find the image quality itself is the problem. self forcing gives it the ai sloppa look with plastic skin, it really ruins realistic gens
>>105782013He's training on horrible captions I'm surprised it learns anything. You have to remember an artist tag is like 0.1% of the blob the model is learning from. It could take 500 epochs for it to figure out which token is the artist and how it relates to the image. If he really wanted the model to learn artists he needs to do a finetune task with a high signal to noise on the artist and a caption related to the style separate from the blob.
>>105782109from what I know its a combo of gemini captions when it was less censored and tags, that is like the best possible captions you are going to be able to make atm
>>105782109yeah, I think he went too optimistic with his 50 epochs prediction, the model will probably take at least 150 to truly learn anything
>>105782120Ive been maining it lately, what do you think is better than it? Besides illustrious for artist tag 1girl
>Lodest One
>Lode Stone
Is all this some convoluted way of telling him that his website hosing AIM metrics is broken at the moment?
>>105782120I mean it's doing well at general concepts but it's optimistic to assume out of a 300 token blob it's going to see "by McCumface3999" which has 25 images in the dataset. SD 1.5 trained well because many of the images were ONLY the artist name.
>>105782091>increasing the LR on fastWhere has he ever mentioned he is using high LR on fast ? He could just be using a very high batch.
>>105782141that is what caption dropout is for
>>105782148>bro just randomly drop tokens you'll definitely get McCumface3999 for sure
>>105782147>Where has he ever mentioned he is using high LR on fast ?good question, that anon said the LR got increased, maybe he has more details on that?
>>105781888>staging fast is a high LR, not distillation
>didnt gen much today because it was so fucking hot outside and the heat is coming in
yikes, when is it winter again?
>>105782155with how many times it has seen the images by now? yea.
>>105782162this summer is fucking atrocious, I start to believe in climate change again kek
>>105779832> comfy installs viruses to your computer and has invasive telemetryProofs?
>>105782166It will see an image once per epoch, unless the image is duplicated that means it has seen each image 41 times at this point (v41).
>>105782190exactly, have you trained for flux before?
>>105782187https://youtu.be/Ly3j2UF8LrA?si=VqGuYo_ospueY8AL
>>105777903video encoding issue. Fixed them for you
https://files.catbox.moe/rv0xtw.mp4
https://files.catbox.moe/mvwst8.mp4
>>105782508thanks but it was civitai servers, after trying after a few hours it worked normally