← Home ← Back to /g/

Thread 106593104

325 posts 110 images /g/
Anonymous No.106593104 >>106593142 >>106593152 >>106593180 >>106593180 >>106593869 >>106593914 >>106594382 >>106594687 >>106595849
/lmg/ - Local Models General
/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: >>106582475 & >>106575202

โ–บNews
>(09/14) model : add grok-2 support #15539 merged: https://github.com/ggml-org/llama.cpp/pull/15539
>(09/11) Qwen3-Next-80B-A3B released: https://hf.co/collections/Qwen/qwen3-next-68c25fd6838e585db8eeea9d
>(09/11) ERNIE-4.5-21B-A3B-Thinking released: https://hf.co/baidu/ERNIE-4.5-21B-A3B-Thinking
>(09/09) Ling & Ring mini 2.0 16B-A1.4B released: https://hf.co/inclusionAI/Ring-mini-2.0
>(09/09) K2 Think (no relation) 32B released: https://hf.co/LLM360/K2-Think

โ–บNews Archive: https://rentry.org/lmg-news-archive
โ–บGlossary: https://rentry.org/lmg-glossary
โ–บLinks: https://rentry.org/LocalModelsLinks
โ–บOfficial /lmg/ card: https://files.catbox.moe/cbclyf.png

โ–บGetting Started
https://rentry.org/lmg-lazy-getting-started-guide
https://rentry.org/lmg-build-guides
https://rentry.org/IsolatedLinuxWebService
https://rentry.org/recommended-models
https://rentry.org/samplers

โ–บFurther Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

โ–บBenchmarks
LiveBench: https://livebench.ai
Programming: https://livecodebench.github.io/leaderboard.html
Code Editing: https://aider.chat/docs/leaderboards
Context Length: https://github.com/adobe-research/NoLiMa
GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

โ–บTools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler Visualizer: https://artefact2.github.io/llm-sampling

โ–บText Gen. UI, Inference Engines
https://github.com/lmg-anon/mikupad
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/ggerganov/llama.cpp
https://github.com/theroyallab/tabbyAPI
https://github.com/vllm-project/vllm
Anonymous No.106593110
โ–บRecent Highlights from the Previous Thread: >>106582475

--Paper: Steering MoE LLMs via expert activation/deactivation for behavior control:
>106586569 >106586649 >106586696
--Papers:
>106589525
--Node-based agent circuit for multi-model daydreaming experiments:
>106591301 >106591335 >106591411 >106591447 >106591518 >106591560 >106591683
--DDR5 RAM purchase recommendation for glm air over waiting for Arc B60:
>106585865 >106585907 >106586028 >106586157 >106586691 >106587973 >106588740 >106588044
--MoE architecture enables larger models to be faster through selective parameter activation:
>106587275 >106587302 >106587405 >106587419
--glm 4.5 air setup issues in Silly Tavern template configuration:
>106586816 >106586886 >106587013 >106587027
--Qwen model dataset imbalances and performance tradeoffs:
>106582623 >106582643 >106583124 >106583138 >106583143 >106583155 >106586595 >106583147 >106592024 >106592033 >106592110 >106592242
--VibeVoice model availability, quality tradeoffs, and reverse-engineering challenges:
>106585909 >106585930 >106585940 >106588461 >106586039 >106586587 >106586610 >106586647 >106587720 >106586704 >106587007 >106587090 >106588243
--CPU offloading performance trade-offs for mid-sized MOE models:
>106583262 >106583338
--IndexTTS 2 speed and interface improvements for text-to-speech:
>106585295 >106585756
--Grok-2 support merged into llama.cpp:
>106587526 >106589842 >106589942 >106589949 >106590115
--Critique of flawed AI-generated writing despite model advancements:
>106592247
--ROCm 7.0 RC1 boosts AMD's AI performance, challenging NVIDIA dominance:
>106589235 >106589359 >106589362
--Parameter tuning suggestions for K2 model version differences:
>106584425 >106584478 >106585603
--Miku (free space):
>106584024 >106584226 >106584417 >106587589 >106587800 >106589360 >106589741 >106589764 >106592033 >106589913

โ–บRecent Highlight Posts from the Previous Thread: >>106582480

Why?: >>102478518
Enable Links: https://rentry.org/lmg-recap-script
Anonymous No.106593132 >>106593164 >>106593180 >>106593187 >>106593223
qwenext goofs???????????????????????????????????????
Anonymous No.106593142 >>106593187 >>106593196 >>106593208 >>106593223
>>106593104 (OP)
I want to do backpropagation with her, if you know what I mean.
Anonymous No.106593148 >>106593187 >>106593196 >>106593208 >>106593223
>OP image is a random non-slop Miku I posted a few threads ago
Anonymous No.106593152 >>106593187 >>106593196 >>106593208 >>106593223
>>106593104 (OP)
PANTYHOSE FEET
Anonymous No.106593164 >>106593173 >>106593187 >>106593196 >>106593208 >>106593223
>>106593132
Boil some rice, put it on a plate and let it dry and then eat it for a similar experience
Anonymous No.106593173 >>106593196 >>106593208
>>106593164
yes but
WHERE ARE THE GOOFS?
Anonymous No.106593180 >>106593186 >>106593188
>>106593104 (OP) (OP) #
I want to do backpropagation with her, if you know what I mean.

...
Anonymous
09/15/25(Mon)14:36:05 No.106593148
1.66 MB
1.66 MB JPG

>OP image is a random non-slop Miku I posted a few threads ago

...
Anonymous
09/15/25(Mon)14:36:32 No.106593152

>>106593104 (OP) (OP) #
PANTYHOSE FEET

...
Anonymous
09/15/25(Mon)14:37:42 No.106593164

>>106593132 #
Boil some rice, put it on a plate and let it dry and then eat it for a similar experience
Anonymous No.106593186
>>106593180
goofs for this feel?
Anonymous No.106593187
>>106593132
>>106593142
>>106593148
>>106593152
>>106593164
>
Anonymous No.106593188
>>106593180
I look like this, say this, and also fail to quote posts
Anonymous No.106593196 >>106593203
>>106593173
>>106593164
>>106593152
>>106593148
>>106593142
Anonymous No.106593202
Jaks are a sign of a diseased mind.
Anonymous No.106593203
>>106593196
Anonymous No.106593205
what prompts this schizophrenia? I just want my hecking wholesomechungus 'THE CAKE IS A LE LIE' qwen 80b goofs
Anonymous No.106593208 >>106593228
>>106593173 #
>>106593164 #
>>106593152 #
>>106593148 #
>>106593142 #
Anonymous No.106593223
>>106593132 #
>>106593142 #
>>106593148 #
>>106593152 #
>>106593164 #
Anonymous No.106593228
>>106593208
Me when using gpt oss
Anonymous No.106593234
this thread ggoofy af
Anonymous No.106593238
The melting man is back
he's much softer than before
did you borrow a personality
or did you steal it all on your own?
Anonymous No.106593247
/QA/ WON
JANNIES LOST
Anonymous No.106593253
.
Anonymous
09/15/25(Mon)14:44:53 No.106593234

this thread ggoofy af

...
Anonymous
09/15/25(Mon)14:45:17 No.106593238

The melting man is back
he's much softer than before
did you borrow a personality
or did you steal it all on your own?

...
Anonymous
09/15/25(Mon)14:45:42 No.106593247

/QA/ WON
JANNIES LOST
Anonymous No.106593259 >>106593266
Anonymous No.106593266
>>106593259
>
Anonymous No.106593301 >>106593353 >>106593368 >>106593575 >>106593590 >>106593593 >>106593602 >>106594419
>decide to take a break from /lmg/ and doomscroll on twitter for a bit.
>it's not X, it's Y
>the smell of stale cigarette smoke and regrets
>fake greentext pasta spaced into paragraphs
>you hit on the core of the issue
>shivers, ozone, Elara, emojis
how do I unsee
Anonymous No.106593302 >>106593314 >>106593324 >>106593325 >>106593329 >>106593353 >>106593358 >>106593368 >>106593377
https://files.catbox.moe/eegitb.jpg
Anonymous No.106593314 >>106593353 >>106593358 >>106593368 >>106593377
>>106593302
I fell for it last time, ain't happening again.
Anonymous No.106593324 >>106593353 >>106593358 >>106593368 >>106593377
>>106593302
thicku miku
Anonymous No.106593325 >>106593353 >>106593358 >>106593368 >>106593377
>>106593302
nigga that's nuts
Anonymous No.106593329 >>106593358 >>106593377
>>106593302
Meh.
Anonymous No.106593353
>>106593301
>>106593302
>>106593314
>>106593324
>>106593325
>
Anonymous No.106593358
>>106593329
>>106593325
>>106593324
>>106593314
>>106593302
>
Anonymous No.106593368
>>106593301 #
>>106593302 #
>>106593314 #
>>106593324 #
>>106593325 #
>
Anonymous No.106593377
>>106593329 #
>>106593325 #
>>106593324 #
>>106593314 #
>>106593302 #
>
Anonymous No.106593386 >>106593412 >>106593420
https://old.reddit.com/r/LocalLLaMA/comments/1nhgd9k/the_glm_team_dropped_me_a_mail/
lol glm has employees doing social media engagement
wonder if one of them is among the people shitting this thread right now
Anonymous No.106593393 >>106593404 >>106593413 >>106593558
OP just delete thread if you can
Anonymous No.106593397
>lol glm has employees doing social media engagement
>wonder if one of them is among the people shitting this thread right now
Anonymous No.106593404
>>106593393
nah, fuck qatroons
Anonymous No.106593410
>OP just delete thread if you can
Anonymous No.106593412
>>106593386
You are even more gullible than reddit.
Or something worse.
Anonymous No.106593413 >>106593420
>>106593393
Let the retard seethe. It's not like he can do anything.
Anonymous No.106593420 >>106593424 >>106593462 >>106593539
>>106593386
why would GLM shit up the thread where their models are praised?
>>106593413
what's the shitter even angry about? Is it the thread mascot debate again?
Anonymous No.106593421 >>106593427
>>106591301
I was thinking of fucking around with those sorts of workflows to see if I can make a smaller model perform better by making it go through steps before providing a final response. Almost like a thinking workflow that tries to extract as much information from the big picture to then focus on the relevant details and the like.
I got caught up with other projects and ended up forgetting about that.
Anonymous No.106593424 >>106593427
>>106593420
fuck your thread culture bullshit
Anonymous No.106593427 >>106593444 >>106595008
>>106593421
What's the UI in the quoted reply? Seems cool.
>>106593424
fuck you I didn't even advocate for "thread culture" I was just asking a question you dork
Anonymous No.106593429 >>106593441 >>106593443 >>106593448 >>106593453 >>106593454
I ask again , just in case. Can "Mistral-Nemo-Instruct-2407-GGUF" handle beyond 16K context?
Anonymous No.106593441
>>106593429
Try it. Only you can know if it can handle it to your satisfaction.
Anonymous No.106593443
>>106593429
Technically yes but realistically no. Just try it out for yourself, the model could fit on a 6G card ffs
Anonymous No.106593444 >>106593487 >>106595008 >>106595722
>>106593427
>What's the UI in the quoted reply?
Not sure, but I know of two UIs that can do that kind of thing, NoAssTavern (simpler and recommended), and astrsk (don't even download it, has telemetry and shit).
Anonymous No.106593448
>>106593429
it creates mustard gas
Anonymous No.106593453
>>106593429
No.
Anonymous No.106593454
>>106593429
Yes, of course.
It will perform worse than it does at, say, 4k context, however.
Anonymous No.106593462 >>106593482 >>106593487
>>106593420
>why would GLM shit up the thread where their models are praised?
you assumed I was talking about the meme spammer. I don't even pay attention to his image spam, it doesn't register in my eyes, image posters are to be ignored.
I was talking about people who praise this garbage model like you, you are the reason this is a garbage thread
spammer is just a minor annoyance that will go away after a b&, the retards never go away though
Anonymous No.106593482 >>106593572
>>106593462
>image posters are to be ignored.
sir this is image baords
Anonymous No.106593487 >>106593494 >>106593504 >>106593572
>>106593444
Huh, I stumbled upon another interesting UI called "talemate" mentioned in one of the NoAssTavern's issues.
https://github.com/vegu-ai/talemate
>>106593462
Every model smaller than Deepseek is garbo, get a grip. Smaller models like Air are the only thing most people can run. Fucking hell, you see how often Rocinante gets mentioned here? What is there to discuss with "non-shit" models if nobody can run them you dickweed?
Anonymous No.106593494 >>106593572
>>106593487
>talemate
Alright, that looks promising.
Anonymous No.106593504 >>106593511 >>106593520 >>106593572
>>106593487
>if nobody can run them
then let's close this so called local model general if no one is even doing local?
Anonymous No.106593511 >>106593524 >>106593525 >>106593572
>>106593504
>if no one is even doing local
Nobody is using anything smaller than deepseek? news to me...
Anonymous No.106593520
>>106593504
I am running the local sir
GLM chan very large
Anonymous No.106593524
>>106593511
deepseek 8b
Anonymous No.106593525 >>106593553 >>106593561 >>106593582
>>106593511
>Every model smaller than Deepseek is garbo
you said it yourself it's time to stop
Anonymous No.106593526 >>106593533
After I stopped shitposting in this thread the quality of it became even worse. I can't believe it.
Anonymous No.106593533
>>106593526
You're absolutely right! This really delves into the tapestry of how shit lmg is!
Anonymous No.106593539 >>106593550 >>106593566 >>106593582 >>106593929 >>106596067
>>106593420
like kids need a reason to be angry
Anonymous No.106593550
>>106593539
>itt raises the kid experince
Anonymous No.106593553 >>106593559 >>106593582
>>106593525
It's garbo compared to large, cloud-hosted models but it's still fun. If the only car you have is a shitbox, do you throw it away? Come on, man.
Anonymous No.106593558
>>106593393
Delete your posts
Anonymous No.106593559 >>106593588 >>106593602
>>106593553
>If the only car you have is a shitbox, do you throw it away?
yes, take the bus and train (API) like a normal person
Anonymous No.106593561
>>106593525
maybe I love garbo
Anonymous No.106593566 >>106593574 >>106593578 >>106593593 >>106593593
>>106593539
While it doesn't change my position on it at all, I suddenly understand where the proponents of age verification are coming from.
Anonymous No.106593572
>>106593511
>>106593504
>>106593494
>>106593487
>>106593482
>
Anonymous No.106593574 >>106593591 >>106593602
>>106593566
That wouldn't help tho as clearly an adult is helping and encouraging the corruption
Anonymous No.106593575 >>106593587 >>106593602
>>106593301
You cannot close your eyes once they've been opened
Anonymous No.106593578
>>106593566
lmao you actually think age checks are to protect kids?
Anonymous No.106593582
...
Anonymous
09/15/25(Mon)15:24:12 No.106593559

>>106593553 #
>If the only car you have is a shitbox, do you throw it away?
yes, take the bus and train (API) like a normal person

...
Anonymous
09/15/25(Mon)15:24:18 No.106593561

>>106593525 #
maybe I love garbo

...
Anonymous
09/15/25(Mon)15:24:37 No.106593566

>>106593539 #
While it doesn't change my position on it at all, I suddenly understand where the proponents of age verification are coming from.
Anonymous No.106593587 >>106593612 >>106593617 >>106593640
>>106593575
anon is you okay, you can close the eyes
Anonymous No.106593588 >>106593612
>>106593559
Nah I think I'll stick to my shitbox. I can drive it when and where ever I want, and it won't suddenly change routes and timetables. But I support your ability to choose, just don't pretend like the only options are public transport or a lambo...
Anonymous No.106593590 >>106593612
>>106593301
If you get into imagegen, you'll see it everywhere.
Anonymous No.106593591 >>106593612
>>106593574
It wouldn't, but I get the emotional reaction.
Anonymous No.106593593
.
Anonymous
09/15/25(Mon)15:26:00 No.106593574

>>106593566 #
That wouldn't help tho as clearly an adult is helping and encouraging the corruption

...
Anonymous
09/15/25(Mon)15:26:02 No.106593575

>>106593301 #
You cannot close your eyes once they've been opened

...
Anonymous
09/15/25(Mon)15:26:19 No.106593578

>>106593566 #
lmao you actually think age checks are to protect kids?
Anonymous No.106593602
...
Anonymous
09/15/25(Mon)15:27:05 No.106593587

>>106593575 #
anon is you okay, you can close the eyes

...
Anonymous
09/15/25(Mon)15:27:06 No.106593588

>>106593559 #
Nah I think I'll stick to my shitbox. I can drive it when and where ever I want, and it won't suddenly change routes and timetables. But I support your ability to choose, just don't pretend like the only options are public transport or a lambo...

...
Anonymous
09/15/25(Mon)15:27:16 No.106593590

>>106593301 #
If you get into imagegen, you'll see it everywhere.

...
Anonymous
09/15/25(Mon)15:27:23 No.106593591

>>106593574 #
It wouldn't, but I get the emotional reaction.
Anonymous No.106593609 >>106593629
Thanks this is very helpfuls.
Anonymous No.106593612 >>106593616
>>106593587
>>106593588
>>106593590
>>106593591
>
Anonymous No.106593616 >>106593629
>>106593612
I do not like this miku
Anonymous No.106593617 >>106593629
>>106593587
Im fine. Thanks for asking
Anonymous No.106593629 >>106593640 >>106593640 >>106593693
>>106593609
>>106593616
>>106593617
>
Anonymous No.106593640
do not like this miku

>>106593629 #
...
Anonymous
09/15/25(Mon)15:31:05 No.106593617

>>106593587 #
Im fine. Thanks for asking

>>106593629 #
Anonymous No.106593664
/QA/ WON
/G/ LOST
Anonymous No.106593690 >>106593694 >>106593698 >>106593704
can i get a short stack miku pls
Anonymous No.106593693
>>106593629
Anonymous No.106593694 >>106593711
>>106593690
cope
Anonymous No.106593698 >>106593711
>>106593690
best xhe can steal is shart miku
Anonymous No.106593704 >>106593711
>>106593690
No. You get a baby Miku instead.
Anonymous No.106593707
>best xhe can steal is shart miku
Anonymous No.106593709 >>106593743
Is NoobAI still the meta or have things moved on
Anonymous No.106593711
>>106593694
>>106593698
>>106593704
my day is ruined
Anonymous No.106593743 >>106593756 >>106593764 >>106593774
>>106593709
ponyv7 releases this month
Anonymous No.106593756 >>106593813
>>106593743
oh? can it be downloaded or is it online only?
Anonymous No.106593764
>>106593743
back to your board barney
Anonymous No.106593774 >>106593777 >>106593804 >>106593813
>>106593743
more sdxl slop?
Anonymous No.106593777 >>106593787 >>106593790 >>106593793
>>106593774
as opposed to what then?
Anonymous No.106593787 >>106593794 >>106593796
>>106593777
you haven't heard about the current best local model called chroma?
Anonymous No.106593790
>>106593777
idk, I haven't kept up with image gen, I wish we had something integrated with LLMs instead of CLIP
Anonymous No.106593793
>>106593777
Chroma SOTA 4futures!
Anonymous No.106593794
>>106593787
Can it match noobAI/pony for character stuff?
Anonymous No.106593796
>>106593787
That's just a rip off of ligma
Anonymous No.106593804
>>106593774
Wasn't it gonna be based on some random shit nobody has ever used
>AuraFlow
Yep.
Anonymous No.106593813 >>106593820
>>106593756
weights
>>106593774
it's based on auraflow
Anonymous No.106593820 >>106593829
>>106593813
>weights
ok, can it be downloaded or is it online only?
Anonymous No.106593829 >>106593832
>>106593820
Yes you will be able to download it
Anonymous No.106593832 >>106593857 >>106593862
>>106593829
Thank you.
Anonymous No.106593857 >>106594031
>>106593832
You're not welcome
Anonymous No.106593862 >>106594031
>>106593832
You're free to leave
Anonymous No.106593869 >>106593881 >>106593924 >>106593942 >>106594021 >>106594037 >>106594305 >>106594439
>>106593104 (OP)
Good morning /lmg/ frens. I've got a question:

So it it pretty much confirmed and fact that you HAVE to use at least a 12B model I order for it to be "smart"? (Not forgetting important details mentioned earlier in the content)? Based on my own testing 7B - 8B models struggle immensely with this. What has your experience been like with the different sized parameter models?
Anonymous No.106593881 >>106594021
>>106593869
If you don't train on The Entire Internet a simple 4B is more than enough for the narrow use case of ERP.
Anonymous No.106593914 >>106593919
>>106593104 (OP)
mikubutt
Anonymous No.106593919 >>106593927
>>106593914
should've been a miku short stack
Anonymous No.106593924 >>106594021
>>106593869
I wouldn't say smart, but 12b models are about the starting point where you don't need to hold their hand for every reply to get a usable output.
Anonymous No.106593927
>>106593919
*miku shart stacked
Anonymous No.106593929
>>106593539
He's just like me except I'm using a pc
Anonymous No.106593930 >>106593935
VRAMlets:
>image generation
pretty good
>voice cloning/TTS
okay
>text generation (simple)
decent
>text generation (advanced)
really bad
Anonymous No.106593935 >>106593941 >>106593949
>>106593930
What is this (advanced) thing about?
Anonymous No.106593941
>>106593935
DeepSeek K2 4.5
Anonymous No.106593942 >>106593953 >>106593958 >>106594021 >>106594034 >>106595077 >>106595608
>>106593869
I don't think 12B is enough, Nemo is pretty dumb too. GLM-air often mistakes who did what and struggles with theory of mind (secret keeping test and such). I'm not cool enoguh to run larger models though.
>Not forgetting important details mentioned earlier in the content
This one in particular is about specific context training and architecture, not really about parameter size.
Anonymous No.106593949
>>106593935
not brain dead
Anonymous No.106593953 >>106594021
>>106593942
>GLM-air often mistakes who did what and struggles with theory of mind (secret keeping test and such)
Mistral Small 24b and Gemma 27b are guilty of both these things as well.
Anonymous No.106593958 >>106594021 >>106594032
>>106593942
>GLM-air often mistakes who did what
sounds like prompt format issue that nemo used to have early on, probably broken implementation as usual
Anonymous No.106593960
Holy schizo
Anonymous No.106593964
Cursed schizo
Anonymous No.106594021 >>106594037 >>106594044 >>106594049
>>106593953
>>106593958
>>106593942
>>106593869
>>106593881
>>106593924

So I guess we have to accept that ALL local LLMs will make fuck ups in some way shape or form? What contribute more to how BADLY It fucks up: perimeter size, architecture, and/or training methods?
Anonymous No.106594031
>>106593857
>>106593862
Bawww.
Anonymous No.106594032
>>106593958
I mostly run it in text completion mode
can't have prompt format issues if you don't format your prompts.
Anonymous No.106594034
>>106593942
>GLM-air often mistakes who did what and struggles with theory of mind (secret keeping test and such).
Funny. I find that it does pretty well in keeping secrets.
Granted, I do prefill the thinking block with instructions to consider exactly those things, which might have some adverse effects in other areas I guess, but still.
To me, the one strong point about GLM is that it actually follows its thinking, instead of something like Qwen that might draft a whole plan in the thinking block then reply with something completely different, even with guidance.
Anonymous No.106594037
>>106594021
And for clarification I'm mostly referring to forgetting details right after you mentioned something, temporal coherence (if a system prompt or previous prompt mentions there in a park, they should stay in the park until stated otherwise or the LLM makes a transition that makes sense), not randomly switching the genders of main characters (this one really likes doing that: >>106593869 , ) etc
Anonymous No.106594044 >>106594687
>>106594021
>What contribute more to how BADLY It fucks up: perimeter size, architecture, and/or training methods?
yes
Anonymous No.106594049
>>106594021
>What contribute more to how BADLY It fucks up: perimeter size, architecture, and/or training methods?
Training on The Entire Internet will do that to you.
Anonymous No.106594111 >>106594126 >>106594146 >>106594262
has someone scrapped AO3 to create a dataset?
Anonymous No.106594126 >>106594319
>>106594111
it's already on most models and yes they did to creators dismay and threats
Anonymous No.106594146 >>106594230
>>106594111
IDK if they specifically from AO3 or from other sites to but here's The closest thing I could find to something like that that hasn't been nuked

https://huggingface.co/datasets/mrcuddle/NSFW-Stories-JsonL

It's not formatted to actually be useful for training but it does have a bunch of raw stories.
Anonymous No.106594230
>>106594146
https://archive.org/details/AO3_final_location
Anonymous No.106594262
>>106594111
its better to just do it yourself so you can filter it how ever you like. its like 40% gay porn by tag. and 50% Harry Potter by universe. it needs balancing if you want it to be useful.
Anonymous No.106594302 >>106594355
I thought I could get away with running unquanted <4B model CPU-only on an old machine.
Nope, absolutely unusable.
Edge AI Status: Meme.
Anonymous No.106594305 >>106594324 >>106594421
>>106593869
Again, your prompting format is all wrong, if that's Llama 3.
Anonymous No.106594319 >>106594331
>>106594126
Gemma 2/3 and Mistral Small, that I've tested didn't appear to be trained on the ones explicitly tagged as "Explicit" or "Underage".
Anonymous No.106594324 >>106594353 >>106594356
>>106594305
It isn't. Elaborate further if you're certain it is. If you're going to tell someone something is fucked up with the hopes they will unfuck it, at least explain WHY....
Anonymous No.106594331
>>106594319
i mean obviously, why train on low quality illegal shit, the classifier correctly said hell no to that sick shit
Anonymous No.106594353 >>106594408 >>106594421 >>106594461 >>106594666
>>106594324
https://www.llama.com/docs/model-cards-and-prompt-formats/meta-llama-3/
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are a helpful AI assistant for travel tips and recommendations<|eot_id|><|start_header_id|>user<|end_header_id|>

What is France's capital?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Bonjour! The capital of France is Paris!<|eot_id|><|start_header_id|>user<|end_header_id|>

What can I do there?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Paris, the City of Light, offers a romantic getaway with must-see attractions like the Eiffel Tower and Louvre Museum, romantic experiences like river cruises and charming neighborhoods, and delicious food and drink options, with helpful tips for making the most of your trip.<|eot_id|><|start_header_id|>user<|end_header_id|>

Give me a detailed list of the attractions I should visit, and time it takes in each one, to plan my trip accordingly.<|eot_id|><|start_header_id|>assistant<|end_header_id|>
Anonymous No.106594355 >>106594401
>>106594302
>CPU-only
Yeah, that's going to be pain. Not so much the token generation, but prompt processing is so slow.
There's a reason we use MoE models the way we do, generation on CPU, PP on the GPU.
That said, does whatever device not have a GPU you could use for PP with vulkan?
Anonymous No.106594356 >>106594408 >>106594421
>>106594324
just look right at the middle of the screenshot, man.
Anonymous No.106594382 >>106594387 >>106594394
>>106593104 (OP)
Can someone recommend best Mistral model? Preferably abliterated
Anonymous No.106594387 >>106594585
>>106594382
The biggest you can run. Any.
Anonymous No.106594394 >>106594585
>>106594382
Medium 3 or Large if you know where to look.
Anonymous No.106594401
>>106594355
Old machine was promoted into a home server after I got new one. I like my home servers to be quiet and low-power, so I don't feel like sticking a GPU in it.
Anonymous No.106594408 >>106594421
>>106594353
>>106594356
That's a fuck up with how axolotl inference outputs. It likes to duplicate portions of text. Here's the correctly formatted text file i inference off of

https://files.catbox.moe/fozpkz.txt

Nothing I thought it is fucked up as far as I can see....
Anonymous No.106594419
>>106593301
Enjoy the wonderland and see how deep the rabbit hole goes
Anonymous No.106594421 >>106594435 >>106594439 >>106594461
>>106594408
>>106594356
>>106594353
>>106594305
It either way it completed in The exact fashion it was supposed to complete in so I don't see what the hyper fixation on that is.
Anonymous No.106594435 >>106594446 >>106594461
>>106594421
A single extra space can make your model drop 90IQ
Anonymous No.106594439 >>106594446 >>106594461
>>106594421
>I don't see what the hyper fixation
>>106593869
>Not forgetting important details
>Based on my own testing
Anonymous No.106594446 >>106594461 >>106594470
>>106594435
>>106594439
Nta. So what was stopping you from pointing that out the first time?
Anonymous No.106594461 >>106594470 >>106594475
>>106594421
>>106594446
Nta it's technically formatted correctly but also not really. It has duplications of the assistant token towards the middle and the end. remove those and then try again. Not quite sure why ultra autists >>106594353
>>106594435
>>106594439
Were so unwilling to point that out
Anonymous No.106594470 >>106594499 >>106594738
>>106594446
The assumption that anon can google "llama3 chat format".
In that much, I admit I was wrong.
I don't care either way. Anon wanted info on how his chat format is wrong. I provided it.
>>106594461
>it's technically formatted correctly but also not really
It is or it isn't. It is not.
Anonymous No.106594475
>>106594461
>That's a fuck up with how axolotl inference outputs
Anonymous No.106594495 >>106595369
>GLM-4.5-IQ2_M
is it even worth using or would i be wasting my bandwidth?
Anonymous No.106594499 >>106594514 >>106594565
>>106594470
They understood how the formatting works it just had duplicates for some reason. He probably ran The prompt to AI or something and it injected the duplications and they didn't realize.

A simple "hey you have duplicate assistant tokens you might want to remove that" what have sufficed instead of being condescending. You know it's exhausting going out of your way to be that way right?


Not that it would have made much of a difference anyway since anything below 12b is retarded regardless.
Anonymous No.106594514 >>106594522
>>106594499
>anything below 12b is retarded regardless.
completely wrong though that is the fault of training on too much data
Anonymous No.106594522 >>106594527
>>106594514
Who are referring to?
Anonymous No.106594527 >>106594559
>>106594522
every lab right now cramming too much into small models instead of making narrow use case ones
Anonymous No.106594559 >>106594575 >>106594576 >>106594625
>>106594527
You mean something like
>https://huggingface.co/allenai/Flex-creative-2x7B-1T
Anonymous No.106594565 >>106594583
>>106594499
Anon is assessing the quality of models and can't use google, read or follow instructions.
>they, he, they
Be consistent.
I posted the example in llama's site. With his carefully constructed tests, eagle eye and attention for detail, I would have expected him to notice all the empty space between the chat format tokens and the content, which his catbox post clearly doesn't have. The other anon pointed out the template dups.
Anonymous No.106594575 >>106594598
>>106594559
>data owners can contribute to the development of open language models without giving up control of their data. There is no need to share raw data directly, and data contributors can decide when their data is active in the model, deactivate it at any time, and receive attributions whenever it's used for inference.

What?
Anonymous No.106594576 >>106594598
>>106594559
no what the hell is this abomination fuck allencucks
Anonymous No.106594583 >>106594616
>>106594565
I used the format though, it just had duplications. The only error where the duplications....
Anonymous No.106594585 >>106594609
>>106594387
>>106594394
Ty, I just saw a lot of focused tarins... focused on some specific stuff like RP or philosophy, but I was looking for good one for general purpose research and deep thinking. So wander maybe someone know a good one that is stands out
Anonymous No.106594598 >>106594625
>>106594575
>>106594576
There's also a literal reddit version.
>https://huggingface.co/allenai/Flex-reddit-2x7B-1T
Anonymous No.106594609
>>106594585
What da fak I just spit out lol, I mean *trainings
Anonymous No.106594616 >>106594630
>>106594583
>The only error where the duplications
You're missing the empty lines.
Anonymous No.106594619 >>106594626 >>106594636
does Linux have an alternative to sillytavern yet
Anonymous No.106594625 >>106594648
>>106594598
>>106594559
It claims they can contribute to training without providing the user data.... How the fuck does that even work? Am I misunderstanding what they're saying?
Anonymous No.106594626 >>106594650
>>106594619
does window?
Anonymous No.106594630 >>106594666
>>106594616
Which followed after the duplications right? Removing those should have fixed the incorrect formatting
Anonymous No.106594636
>>106594619
llama.cpp HTTP server + curl
Anonymous No.106594648 >>106594780
>>106594625
You basically train a smaller domain specific model (expert modules) that can later be part of the larger final product.
>https://www.datocms-assets.com/64837/1752084947-flexolmo-5.pdf
Anonymous No.106594650
>>106594626
I don't use windows
Anonymous No.106594652 >>106594670 >>106594745
beg me to shitpost again so this thread stops being dead.
Anonymous No.106594655
stfu im zorking it
Anonymous No.106594656
Just give me the goof
Anonymous No.106594666 >>106594687
>>106594630
Look at this >>106594353 or llama's site.
After
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

there's an empty line. Every other line is an empty line. Those are not in your catbox file.
Anonymous No.106594670
>>106594652
i beg of sama-sama please just let us rest in piss
Anonymous No.106594687 >>106594699 >>106594729 >>106594738 >>106594780
>>106594044
>>106593104 (OP)
I'm asking this to everyone: what's the bare minimum parameter size someone should use if they want to have decent RP where the "assistant" isn't retarded?
>>106594666
I don't think those are strictly necessary given that it autocompletes correctly without them. How do you know that's not done just for ease of readability?
Anonymous No.106594699 >>106594831
>>106594687
4B with proper training.
Anonymous No.106594729 >>106594743
>>106594687
you'll have to accept retardation and learn to live with it
Anonymous No.106594738 >>106594748
>>106594687
>How do you know that's not done just for ease of readability?
>>106594470
>I don't care either way. Anon wanted info on how his chat format is wrong. I provided it.
Anonymous No.106594743 >>106594756 >>106594786 >>106594795
>>106594729
I wonder if the deepseek api users over at /aicg/ have to suffer with it anywhere near as much as we do.
Anonymous No.106594745 >>106594998
>>106594652
i dare you to do it again
Anonymous No.106594748
>>106594738
>Doesn't answer the question
Anonymous No.106594756 >>106594774
>>106594743
Yes, I don't recommend reading their thread for your sanity but even they complain about all their models even Opus and such.
Anonymous No.106594774 >>106594794
>>106594756
Damn... So the retardation is inescapable no matter how big or "smart" the model is?
Anonymous No.106594780 >>106594817
>>106594687
The thing is, retarded is a spectrum.
Some people will have more tolerance for certain errors and certain magnitudes of errors than others, so the lower boundary us fuzzy as hell and a model can be perfectly serviceable in one scenario while fucking up another.
Some people will tell you 12B is enough, others will say 70B dense, other's will tell you to not bother unless you can go for the biggest best-est thing because retardation exists even in the best models, just to a much lesser extent.
Etc etc.
tl;dr : There's no consensus and I'm not sure there can be, at least for now.

>>106594648
Reminds me of CUDADEV's idea of training a bunch of different models on a subset of the full training set, running them in parallel, then averaging the logits, although in that case it was more about getting the results equivalent to a model trained on
>[number of models] x [training tokens each model sees]
tokens than specializing models.
Anonymous No.106594786 >>106594817
>>106594743
Deepseek as to deal with theirs. A much worse fate.
Anonymous No.106594794 >>106594817
>>106594774
Correct, this is the LLM blackpill there are zero non retarded one currently.
Anonymous No.106594795 >>106595943
>>106594743
i am a 4 bit cpumaxxing coper
llama_model_loader: loaded meta data with 52 key-value pairs and 1096 tensors from models/Kimi-K2-Instruct-0905-GGUF-smol-IQ4_KSS/Kimi-K2-Instruct-0905-smol-IQ4_KSS-00001-of-00011.gguf
llm_load_print_meta: model ftype = IQ4_KSS - 4.0 bpw
llm_load_print_meta: model params = 1.026 T
llm_load_print_meta: model size = 485.008 GiB (4.059 BPW)
llm_load_print_meta: repeating layers = 483.197 GiB (4.053 BPW, 1024.059 B parameters)
llm_load_tensors: offloaded 62/62 layers to GPU
llm_load_tensors: CPU buffer size = 420246.00 MiB
llm_load_tensors: CUDA_Host buffer size = 927.50 MiB
llm_load_tensors: CUDA0 buffer size = 13632.97 MiB
llm_load_tensors: CUDA1 buffer size = 18510.81 MiB
llm_load_tensors: CUDA2 buffer size = 18668.47 MiB
llm_load_tensors: CUDA3 buffer size = 19280.69 MiB
llm_load_tensors: CUDA4 buffer size = 5382.00 MiB
Anonymous No.106594817 >>106594822 >>106594839 >>106594853
>>106594794
>>106594786
>>106594780
Are we at least in agreeance that The higher the perimeter size, be lower the retardation generally is? Or is that not a reliable way to gauge?
Anonymous No.106594822 >>106594839
>>106594817
Generally somewhat, but then there's stuff like Llama4.
Anonymous No.106594831 >>106594856
>>106594699
do you have empirical evidence of this claim? what 4b model is best for rp? how come 4 and not 3 or 5?
Anonymous No.106594839
>>106594817
>>106594822
dataset quality matters a bunch. garbage in garbage out..
Anonymous No.106594853
>>106594817
Generally, yes. Although training data and procedure plays a large role in it too, and there's also dense vs sparse to consider, etc.
Basically, there are not enough scientific comparative experiments for us to tell how much each component matters (general architecture, depth, width, training data, training procedure,e tc) and there's a good chance that the ffinal result also varies with usecase.
Meaning, it's a clusterfuck.
Anonymous No.106594856 >>106594867
>>106594831
That's the best I can run. So it HAS to be the best size and everything anyone could ever need.
Anonymous No.106594867 >>106594882 >>106594907
>>106594856
What do you use your 4B models for?
Anonymous No.106594882
>>106594867
I was joking. I'm not that anon. But I think the sentiment is still the same.
Anonymous No.106594907
>>106594867
I can run and currently cope with 12-24B but models are so bloated it's implausible we can't do better with less trash and more use case data.
Anonymous No.106594924 >>106594955 >>106594974 >>106594977
So what I'm getting here is that LLMs RP. What else can they be useful for? I feel like the main reason they don't hit the mainstream is because you need beefy graphics cards to even consider trying them. And tonight if you consider attacking the train them yourself.
Anonymous No.106594955
>>106594924
code and math is the only other use case
Anonymous No.106594974 >>106594996
>>106594924
>I feel like the main reason they don't hit the mainstream
Claude, chatgpt and gemini are mainstream.
>What else can they be useful for?
>And tonight if you consider attacking the train them yourself.
They could be used to correct text before being sent. Other than that, simple translation, google replacement for simple verifiable things, spamming image boards, replying to corporate. You know... the usual...
Anonymous No.106594977
>>106594924
Also non-generative use cases like classifying data.
Anonymous No.106594996 >>106595041
>>106594974
>Claude, chatgpt and gemini are mainstream.
Was referring to local LLMs. Also forgive that last part of the last post. I'm writing this on voice to text.
Anonymous No.106594998 >>106595017
>>106594745
i said beg you maggot
Anonymous No.106595008 >>106595114
>>106593427
>>106593444
The UI is in the Regions repo, and makes flows for it. Deleting and renaming nodes is jank, but it works otherwise.

https://github.com/dibrale/Regions
Anonymous No.106595017 >>106595115
>>106594998
ya that's what i thought pussy
Anonymous No.106595041 >>106595477
>>106594996
>Was referring to local LLMs
Then yes. Lack of GPU, not knowing how to compile stuff, terminals are scary and all that. A tech-literacy gap, if you will. Not that anons here are much more tech-savvy.
>git pull. thing broke
>he pulled
Anonymous No.106595077
>>106593942
The workflow from the last thread is supposed to help with that, but I'm not sure what the best way of testing it is. Might be cool to turn it into a server script if it helps.

>>106591301
Anonymous No.106595081
llama.cpp changed the metal backend and made it eat way more memory, I'm OOMing with the same params that left me with 10GB of headroom on the last commit... curse you gerganov
Anonymous No.106595114 >>106595515
>>106595008
That's pretty sick.
I might scrap the shit I was working on and use that as a reference to start over.
Or maybe just use that as a middleware between the LLM backend and my app. Either or.
Anonymous No.106595115
>>106595017
fine. enjoy your dead thread.
Anonymous No.106595242 >>106595261
shitposters won
Anonymous No.106595261 >>106595370 >>106597194
>>106595242
One kike throwing an endless temper tantrum over this thread hardly counts as winning.
Imagine a parent, their child is having a full, flailing on the ground, pant shitting tantrum. Are they proud? That's you. Your "pride" is but a cope.
Anonymous No.106595334
reddit won
Anonymous No.106595369 >>106595787
>>106594495
I was running iq2_kl since it fits on my 5090 + 128 ram setup and yea it's not completly retarded sure beats air... if you can fit that then you can alternatively get away with qwen 235b at iq4
Anonymous No.106595370 >>106595418
>>106595261
funnily enough I don't think I've ever had a pants shitting tantrum
I imagine it's rare?
Anonymous No.106595418
>>106595370
I remember pissing myself a few times but it wasn't because of a tantrum.
Anonymous No.106595477
>>106595041
I just want an EXE, not any of that hacker shit
Anonymous No.106595515
>>106595114
What were you working on? Also, deletion and renaming in the Regions GUI is allegedly fixed as of the last commit?
Anonymous No.106595608
>>106593942
I feel like most of the schizo retard moments from glm air come from using cope quants. I switched to using q8 from q3 after upgrading my ram and the difference was immediately noticeable in the way that it remembered and incorporated details from context. Still not perfect and still somewhat slopped, but definitely better.
Anonymous No.106595722 >>106595786
>>106593444
>astrsk (don't even download it, has telemetry and shit).
The only non-local host domain it connects is Google Fonts. As far as I understand, you can enable analytics by setting an API key during the build. But it doesn't seem to have one by default. This was a normal site that became open source later.
Anonymous No.106595758
https://outsidetext.substack.com/p/how-does-a-blind-model-see-the-earth
moesissies don't look
Anonymous No.106595786 >>106595824 >>106595847
>>106595722
a single glance at the readme is enough to close the tab instantly
Anonymous No.106595787
>>106595369
thanks downloading them now
Anonymous No.106595824
>>106595786
It has the correct license.
Anonymous No.106595847 >>106596881
>>106595786
Someone posted this one in another thread.
https://github.com/onestardao/WFGY
Anonymous No.106595849 >>106595882 >>106595960 >>106596292 >>106596362
>>106593104 (OP)
Many normies are claiming that AI is "eating itself to death". What do they mean by this?

https://www.tiktok.com/t/ZT6ofKC5U/
Anonymous No.106595865 >>106595917 >>106596059
>Someone in r*ddit built a DDR4 server with 8 MI50 (256gb vram) for the price of a single 5090
>400w idle
oof
Don't build it if you don't have solar panels.
Anonymous No.106595882
>>106595849
Sounds like this shitjeet has no idea what the fuck he is talking about and has no fucking idea how pretraining works. And by "this shitjeet" I mean you.
Fuck off back to whatever normie shithole you crawled out of.
Anonymous No.106595917
>>106595865
You forgot about heat and noise too
Anonymous No.106595943 >>106595953
>>106594795
>models/Kimi-K2-Instruct-0905-GGUF-smol-IQ4_KSS/Kimi-K2-Instruct-0905-smol-IQ4_KSS-00001-of-00011.gguf
When you load first part, does it mean you just using first part or it's automatically know where to look to next one on the load process?
Anonymous No.106595953 >>106595970
>>106595943
>or it's automatically know where to look to next one on the load process
That.
Anonymous No.106595960 >>106596180
>>106595849
Retards who believe AI is a living being that constantly feeds of the internet instead of simply being a file that can be backed up
Anonymous No.106595966 >>106596053
Grok-2 impressions: (running IQ4_XS)
>*Yawn*
Not sure if it's just impatience from only getting half a token per second in generation, but really not worth the fuss. Would run Llama-3-70B over it any day of the week-
Anonymous No.106595970
>>106595953
Ty
Anonymous No.106595976
im backed up rn
Anonymous No.106595985 >>106596053 >>106596168
Whats a good uncensored LLM? No politically correct bullshit and refusing to give answers. I have low VRAM, I don't mind if its a bit laggy and I don't care about it being 'smart' on programming tasks etc. Most important is just that it chats well and is uncensored in its responses.
Anonymous No.106596053 >>106596139 >>106596174
>>106595966
I actually like grok 2(Q8) and think that it's a hidden gem. Their official prompt on lmarena sucked and made me undervalue it.

>>106595985
I'd suggest grok2, but you are a ramlet...
Anonymous No.106596059 >>106596110
>>106595865
Just turn it off when you're not using it.
Server motherboards come with baseboard management controllers so you can even turn them on and off remotely.
Anonymous No.106596067
>>106593539
Why are parents like this?
Anonymous No.106596106 >>106596134
>check thedrummer's page on hf
>still finetrooning command A
>only uploaded Q5_K_M goofs
why is this the state of finetuning in 2025?
Anonymous No.106596110 >>106596163 >>106596192
>>106596059
My amd workstation takes forever to boot if I don't turn off ram training.
Anonymous No.106596134 >>106596191
>>106596106
Be the change you want to see
Anonymous No.106596139
>>106596053
It's decent at Nala
It's less slopped than most open models, but it comes up pretty dry in soft mommy RP, sadly.
Anonymous No.106596163
>>106596110
5 minutes is not a long time, just make some coffee in the meantime. make a script that makes a coffee at the exact time it takes for you to walk to your kitchen plus five minutes and while you are at it have it write an email that tells kumar that he's an asshole.
Anonymous No.106596168
>>106595985
>uncensored
>low VRAM
Mistral Nemo, always and forever.
Anonymous No.106596174 >>106596220
>>106596053
isn't grok2 8 experts 2 active? you can't run it decently with dual channel
Anonymous No.106596180
>>106595960
You should think of AI is an industry that needs to churn out new models in return for investor money.
Anonymous No.106596191 >>106596222
>>106596134
Some people's time is too valuable to be a glorified data entry and sanitation monkey.
Anonymous No.106596192
>>106596110
>turn off ram training.
turn off what
Anonymous No.106596220
>>106596174
Sadly not, but I have 12+12 channels
Anonymous No.106596222
>>106596191
Opinion discarded then
Anonymous No.106596292
>>106595849
Not entirely wrong tho I didn't look at the asstok link, new models are more and more poisoned by the gpt slop being poured all over and the labs themselves doing synthetic data and amplifying bias for more slop
Anonymous No.106596305 >>106596334
which one does the best lolis
Anonymous No.106596334
>>106596305
gemma3 closely followed by gpt-oss they're the only ones with the proper knowledge
Anonymous No.106596362
>>106595849
It is inbreeding, not eating itself to death.
Anonymous No.106596402 >>106596412 >>106596453 >>106596514 >>106596522 >>106596568 >>106597029 >>106597518 >>106597628
Why are vibe coders like this?
Anonymous No.106596412 >>106596514 >>106597029 >>106597518 >>106597628
>>106596402
Ugh...
Anonymous No.106596420 >>106596426
grandpa crying about zoomies again
Anonymous No.106596426 >>106596515 >>106597029 >>106597518 >>106597628
>>106596420
https://github.com/ggml-org/llama.cpp/pull/16016
Aaaaaaaa
Anonymous No.106596453 >>106596600
>>106596402
It's funnier this way. As long as you don't have to deal with them yourself, anyway.
Anonymous No.106596514 >>106596600 >>106597029
>>106596402
>>106596412
>https://github.com/creatorrr
Anonymous No.106596515 >>106596600
>>106596426
https://www.startupgrind.com/events/details/startup-grind-hyderabad-presents-diwank-singh-tomer-thiel-fellowship/
explains a lot actually
Anonymous No.106596522 >>106596600
>>106596402
Literally all they have to do is change the remark and nobody will ever be the wiser.
Anonymous No.106596542 >>106596713 >>106596739 >>106597303
What will happen to Mistral AI now that ASML bought it for $1.3B?
https://www.asml.com/en/news/press-releases/2025/asml-mistral-ai-enter-strategic-partnership
Anonymous No.106596568 >>106596622 >>106596674 >>106596690 >>106596882 >>106596991 >>106597029
>>106596402
He's probably trying to build his CV to find a job in America or Europe.
Anonymous No.106596600 >>106597029
>>106596453
Someone will have to.
>>106596514
>>106596515
Oh. I had forgotten what puke tasted like. I didn't want to know that much. Thanks.
>>106596522
Yeah. It wasn't obvious. Like that other one....
Anonymous No.106596622
>>106596568
honestly don't think he needs to, sounds like he's already making decent money living in the US
Anonymous No.106596674
>>106596568
>Diwank
Dam Son...
Anonymous No.106596690
>>106596568
sounds like a nguyen
Anonymous No.106596713
>>106596542
>https://www.asml.com
Oh...
Anonymous No.106596739 >>106596784 >>106596793
>>106596542
Holy shit.
I suppose that does make sense, but still.
Holy shit.
I wonder if the idea is to diversify in case their monopoly on high end lithography machines ever comes to an end or if the intent is to somehow improve their existing business.
Anonymous No.106596784
>>106596739
Lower your temp
Anonymous No.106596793 >>106596816
>>106596739
>if the intent is to somehow improve their existing business.
No way...
Anonymous No.106596816
>>106596793
Companies do invest in things other than their core businesses, to the point where sometimes they shift completely away from it.
I doubt ASML will stop selling EUV machines to become an AI lab, but the point stands.
Anonymous No.106596881
>>106595847
That's so fucking funny.
>Tutorial: How to Awaken the Soul of Your AI in under 60 seconds โ€” by the WFGY Engine
Is this what all those "awakened AI" tick toks I've been hearing of are about?
Anonymous No.106596882
>>106596568
>em dash in his two sentence description
bros....
Anonymous No.106596991
>>106596568
Hello sarrs I have build very AI system for you
Anonymous No.106597029 >>106597043 >>106597053
>>106596426
>>106596412
>>106596402
>>106596514
>>106596600
>>106596568
What am I looking at? I see a bunch of shit that looks like it was written by AI. Not even code related to the software. What the hell are these merge requests? I've never merged anything on an existing project in my life so maybe there's something I'm missing here
Anonymous No.106597043 >>106597057 >>106597072
>>106597029
Thanks for reusing this dumb image, MD5 filter works well
Anonymous No.106597053 >>106597071
>>106597029
Guy used AI agents and pushed the files the agent was using to keep track of the work into the repository.
Or something like that.
Anonymous No.106597057
>>106597043
Does it now?
Anonymous No.106597071 >>106597126
>>106597053
And he couldn't do that shit on his own fork of the git repo instead of the official one? He doesn't deserve any attention or employment or consideration for anything if he is this self-centered.
Anonymous No.106597072
>>106597043
https://github.com/woltapp/blurhash
Anonymous No.106597126
>>106597071
Looking at the image again, it's worse, the commits were made on his own fork, and he created a merge request.
Hell, in all likelihood, it wasn't even him, he just gave the AI agent access to git commands too.
Anonymous No.106597194
>>106595261
>shitposting is throwing a tantrum
>4chan is serious business
I would have said that with that the transformation into reddit is complete but this place has been a reddit since forever. Enjoy your dead thread you dumb faggot.
Anonymous No.106597252 >>106597260 >>106597284 >>106597285
Do I need to change something else aside from the GPU / power supply?
CPU : 5500 w/ stock fan
RAM : 32G 3200 CL16
MB : B550-PLUS
GPU : GTX 1050
PSU : 400W 80PLUS Gold
Case : Antex P101
512G M2, 3*4T WD Red Plus
Anonymous No.106597260 >>106597271 >>106597281
>>106597252
wrong thread?
Anonymous No.106597271
>>106597260
No?
Anonymous No.106597281 >>106597312
>>106597260
No, I just want to know what component I should change if I need to run a language model locally.
Anonymous No.106597284 >>106597334
>>106597252
What do you want to do exactly?
I'd tell you to get at least 64gb of ddr5, but ideally, you'd go for a server platform with a ton of memory bandwidth.
Anonymous No.106597285 >>106597334
>>106597252
You can manage with a new gpu and larger PSU. I'd get 64GB ram too or more. Plus fast nvme drive.
Anonymous No.106597303
>>106596542
Same thing as always pinky. They will release another incremental update to 24B small that would have been impressive if everyone wasn't running 2bpw+ fuckhuge moe's.
Anonymous No.106597312 >>106597334
>>106597281
>what component I should change
Don't need to change anything. You can run one right now if you want to.
Anonymous No.106597334 >>106597347 >>106597354
>>106597284
>64gb of ddr5
Ryzen 5 5500 is AM4 kind sir.
>you'd go for a server platform with a ton of memory bandwidth.
That would be a lot of money.
>>106597285
>new gpu and larger PSU
>I'd get 64GB ram too or more. Plus fast nvme drive
That's reasonable enough.
>>106597312
Won't it run like shit?
Anonymous No.106597347 >>106597382
>>106597334
gpt-oss 20b would run very blazings
Anonymous No.106597354 >>106597359 >>106597382
>>106597334
>Won't it run like shit?
A definite maybe. Post a Miku
Anonymous No.106597359 >>106597364
>>106597354
>Post a Miku
kill yourself
Anonymous No.106597364
>>106597359
no u
Anonymous No.106597371 >>106597392 >>106597400 >>106597462
Do people actually use GPT-oss?
Anonymous No.106597382 >>106597452 >>106597471
>>106597347
As long as I can talk in loop at it about how miserable my life is.
>>106597354
>A definite maybe
Still better than a sure no.
Anonymous No.106597392 >>106597420
>>106597371
why not?
Anonymous No.106597400
>>106597371
I tried using the 20B in place of Qwen 30B. It wasn't very good at all.
It spit refusals for no reason at all and it was dumb as shit otherwise.
And yes, I was using the correct chat template since I let llama.cpp deal with that.
Anonymous No.106597420
>>106597392
The refusal reasoning was funny, but I got bored with it.
Anonymous No.106597426
Good morning recently I try out new AI Chatgpt-OSS for very impressed so far!!!
Anonymous No.106597452
>>106597382
It'll run like shit yes. Get yourself a used 3090 and you're set
Anonymous No.106597462
>>106597371
Yeah, it's the best one around ~100B.
Anonymous No.106597471 >>106597499
>>106597382
Run Q8 or Q6K of this with koboldcpp: https://huggingface.co/TheDrummer/Rocinante-12B-v1.1-GGUF/tree/main Should be fine on your current machine for most chats, with partial offloading to CPU, to see if you like local models at all.
If later you want more speed or quality, get minimum of one 3090 and 128GB of DDR5 for GLM 4.5/lite
Anonymous No.106597499
>>106597471
go black drummer
Anonymous No.106597516
pm me when the local jannies kill themselves. then i will revive this thread.
Anonymous No.106597518
>>106596402
>>106596412
>>106596426
Saaar can you redeam report please?
Anonymous No.106597628
>>106596402
>>106596412
>>106596426
See? This is what "AI is eating itself" looks like.