/lmg/ - Local Models General - /g/ (#106171830) [Archived: 15 hours ago]

Anonymous
8/7/2025, 7:38:36 AM No.106171830
__kagamine_rin_vocaloid_drawn_by_shinpe_091__sample-1e2b71ab2bd65f90a7c02256535680ff
/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: >>106167048 & >>106163327

►News
>(08/06) Qwen3-4B-Thinking-2507 released: https://hf.co/Qwen/Qwen3-4B-Thinking-2507
>(08/06) Koboldcpp v1.97 released with GLM 4.5 support: https://github.com/LostRuins/koboldcpp/releases/tag/v1.97
>(08/06) dots.vlm1 VLM based on DeepSeek V3: https://hf.co/rednote-hilab/dots.vlm1.inst
>(08/05) OpenAI releases gpt-oss-120b & gpt-oss-20b: https://openai.com/index/introducing-gpt-oss
>(08/05) Kitten TTS 15M released: https://hf.co/KittenML/kitten-tts-nano-0.1

►News Archive: https://rentry.org/lmg-news-archive
►Glossary: https://rentry.org/lmg-glossary
►Links: https://rentry.org/LocalModelsLinks
►Official /lmg/ card: https://files.catbox.moe/cbclyf.png

►Getting Started
https://rentry.org/lmg-lazy-getting-started-guide
https://rentry.org/lmg-build-guides
https://rentry.org/IsolatedLinuxWebService
https://rentry.org/recommended-models
https://rentry.org/tldrhowtoquant
https://rentry.org/samplers

►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

►Benchmarks
LiveBench: https://livebench.ai
Programming: https://livecodebench.github.io/leaderboard.html
Code Editing: https://aider.chat/docs/leaderboards
Context Length: https://github.com/adobe-research/NoLiMa
Censorbench: https://codeberg.org/jts2323/censorbench
GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

►Tools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler Visualizer: https://artefact2.github.io/llm-sampling

►Text Gen. UI, Inference Engines
https://github.com/lmg-anon/mikupad
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/ggerganov/llama.cpp
https://github.com/theroyallab/tabbyAPI
https://github.com/vllm-project/vllm
Replies: >>106172509 >>106173498 >>106174394
Anonymous
8/7/2025, 7:39:09 AM No.106171832
what's in the box
what's in the box
md5: bd27344ea08895af8c1163ecbcfed00f🔍
►Recent Highlights from the Previous Thread: >>106167048

--Papers:
>106170128 >106170167
--Prioritizing NSFW LORA preservation amid infrastructure and redundancy concerns:
>106167927 >106167949 >106168075 >106168122 >106168043 >106168067 >106168169 >106168208 >106168211 >106168238 >106168277 >106168305 >106168351 >106168377 >106168392 >106168399 >106168425 >106168448 >106168619 >106168442
--High-speed CPU-only LLM inference with GLM-4.5 on consumer hardware:
>106168800 >106168825 >106168868 >106168847 >106168903 >106168905 >106168940 >106168974 >106168991
--Missing CUDA DLLs prevent GPU offloading in newer llamacpp Windows builds:
>106168428 >106168441 >106168450 >106168577 >106168616 >106168670 >106168691 >106168704 >106168715
--Difficulty reducing model thinking time due to token-level formatting constraints:
>106170269 >106170300 >106170348 >106170361 >106170404
--CPU outperforming GPU for GLM-Air inference on low-VRAM systems:
>106168713 >106168787 >106168814 >106169109
--GPT OSS underperforms on LiveBench despite reasoning and math strengths:
>106167476 >106167550
--Anon purchases 384GB of HBM2 VRAM for $600:
>106168337 >106168343 >106168345 >106168366 >106168377 >106168392 >106168399 >106168425 >106168448 >106168619 >106168462 >106168469 >106168506 >106168571 >106168488 >106168505 >106168517 >106168528 >106168606
--High RAM investment for local GLM inference raises performance and practicality concerns:
>106169135 >106169148 >106169161 >106169197 >106169223 >106169230 >106169278
--Anon finds Dual P100 64GB board for $200:
>106169635 >106170934 >106170984 >106169662
--Satirical timeline of LLM evolution with exaggerated eras:
>106167190 >106167237 >106168679 >106167530
--NEO Semiconductor's X-HBM promises 16x bandwidth and 10x density if viable:
>106169723
--Miku and Dipsy (free space):
>106167506 >106167362

►Recent Highlight Posts from the Previous Thread: >>106167057 >>106168982

Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
Anonymous
8/7/2025, 7:41:34 AM No.106171846
Rin-chan now
Anonymous
8/7/2025, 7:42:06 AM No.106171850
1730708016170623
1730708016170623
md5: d9d985b42485a39336393146eed987ff🔍
>when the recap post doesn't quote any of your replies
Anonymous
8/7/2025, 7:42:48 AM No.106171857
What model is the baker using for recap these days?
Replies: >>106171901
Anonymous
8/7/2025, 7:42:53 AM No.106171859
1754545315550
1754545315550
md5: 20d3e16b3f384a0b24235e44dc04241c🔍
where's the coding benchmark for 'toss?
Anonymous
8/7/2025, 7:45:23 AM No.106171874
For me, it's Qwen 3 235B Thinking. It feels like the model actually has a personality and preferences when you reads its thinking. In the thinking block, it acknowledges the creative twists and story progressions I write and calls them brilliant, amazing, etc. It feels like it's actually passionate about the scenario and how it's going, like it's my very own personal fan.
Now if only it were smarter.
Replies: >>106171884 >>106171980 >>106171998
Anonymous
8/7/2025, 7:45:35 AM No.106171877
I just tried torii gate earlier today on actual photos. It doesn't work at all since it's designed for cartoons. So any other uncensored/NSFW image captioning models I can try? I've tried joycaption which is just OK, but to be fair at the time it wasn't like there was much better available.

I saw the suggestions for medgemma, but I doubt it's useful for erp.
Replies: >>106173189
Anonymous
8/7/2025, 7:47:16 AM No.106171884
1727960621311300_thumb.jpg
1727960621311300_thumb.jpg
md5: 7074fb285bfd5793f39ed76074134ba9🔍
>>106171874
>when you reads its thinking
Anonymous
8/7/2025, 7:48:49 AM No.106171888
1754159699109345
1754159699109345
md5: 87b00d7c9827b022a18a89f57b7d9cc0🔍
lurking here since many moths and never had time to read through the tutorials, will do now

op, thanks for posting this everytime! its really helpful for those interested!
Anonymous
8/7/2025, 7:49:29 AM No.106171890
What model sizes can I finetune with a 5090?
Replies: >>106171947
Anonymous
8/7/2025, 7:50:41 AM No.106171901
>>106171857
Qwen3-235B-A22B-Instruct-2507-UD-Q3_K_XL
Anonymous
8/7/2025, 7:57:18 AM No.106171947
>>106171890
With enough patience, any model
Replies: >>106172000
Anonymous
8/7/2025, 8:00:49 AM No.106171967
>>106171861
^
this sister is correct
LLMs are just pattern matchers at heart, no matter what the copers are saying (and don't cite random ARXIV bullshit papers, after the replication crisis, the "I intentionally submitted bullshit data and everyone lapped it up" affairs (look up John Ioannidis's findings on this) etc who the fuck actually trusts this shit blindly?)
LLMs still get tripped up by questions like "can someone wash their hands with no arms". Whenever they learn not to get tripped up by this kind of thing, it's only because they got benchmaxxed on the internet data, and you can always find new sentences like this that will trip newer SOTA because LLMs are unable to handle a change in a common pattern that's not in their data, they see a sentence structure that matches and they do not "think" individually about the individual words
Anonymous
8/7/2025, 8:02:37 AM No.106171980
>>106171874
>In the thinking block, it acknowledges the creative twists and story progressions I write and calls them brilliant, amazing, etc.
They all do that. You can write something terrible and it will still find some way to be super positive about it.
Anonymous
8/7/2025, 8:03:20 AM No.106171983
1735754599757468
1735754599757468
md5: 5d5350929d4b11b94495d0d8744bf6c5🔍
bros... I can't believe it's time...
Anonymous
8/7/2025, 8:05:10 AM No.106171998
>>106171874
if it were smarter it wouldn't be your fan anymore
Anonymous
8/7/2025, 8:05:38 AM No.106172000
>>106171947
Can I really? From what I've read there are minimum vram requirements or it'll result in oom errors.
Anonymous
8/7/2025, 8:07:19 AM No.106172006
>>106170662
kek, honored that people still remembered that.
i actually did play around more with the idea. worked pretty well in terms of circumventing censorship. meme words etc. were no problem.
but it completely tanked the smartness for RP.
words appear in the wrong place etc., though it did stay coherent enough to engage with it, it wasn't pyg level dumb.
Anonymous
8/7/2025, 8:17:53 AM No.106172064
So now there're attempts to jailbreak gptoss safety is it still worth it back to it or nah?
Replies: >>106172083
Anonymous
8/7/2025, 8:22:30 AM No.106172083
gpt oss
gpt oss
md5: b6237663b4a882a830983960d77130e0🔍
>>106172064
The model just isnt that smart. Im not sure what the use case is.
If it had general knowledge or at least good writing then it would have been cool.
I think the main purpose of gpt-oss is coding/math and for that we already have qwen3 which in my opinion is alot better.

So its not just the crazy censorship but missing knowledge.
Honestly it kinda feels like a phi model. Its probably pure synth slop.
Anonymous
8/7/2025, 8:32:47 AM No.106172141
glm 4.5 air for reasonably sized coding assistant.
qwen 30ba3b for quick general bot usage (or vramlets).
gemma3 for dense model with general tasks and instruction following,
gptoss for copium addicts.
for everything else, there's dsr1.
Replies: >>106172153 >>106175226
Anonymous
8/7/2025, 8:34:20 AM No.106172153
>>106172141
>for everything else, there's dsr1.
That's a weird way of spelling Nemo
Replies: >>106172187
Anonymous
8/7/2025, 8:34:44 AM No.106172156
1748767038085054
1748767038085054
md5: ab0006403dfe90d94ad88ae9ec5067a9🔍
yea im a bit of a prompt engineer myself
Anonymous
8/7/2025, 8:35:07 AM No.106172160
1735989812807663
1735989812807663
md5: 29bf86e83bb1084996faa95809e44351🔍
It's not looking good
Replies: >>106172217
Anonymous
8/7/2025, 8:39:15 AM No.106172187
>>106172153
true, then let me fix it.
for vramlets there's Nemo,
for everything else, there's dsr1.
using dsr1 locally spoiled me. though it's not perfect with it's "riding up", "knuckles whitening", and "lips curled" shit.
Replies: >>106172195
Anonymous
8/7/2025, 8:40:50 AM No.106172195
>>106172187
>though it's not perfect with it's "riding up", "knuckles whitening", and "lips curled" shit
literally every single model has slop, anyone who says otherwise is a liar.
Anonymous
8/7/2025, 8:45:16 AM No.106172213
ikllama glm support merged
https://github.com/ikawrakow/ik_llama.cpp/pull/668
Replies: >>106172264
Anonymous
8/7/2025, 8:45:37 AM No.106172217
>>106172160
Why are you posting an ancient image? I want modern chronoboros
Anonymous
8/7/2025, 8:56:24 AM No.106172264
>>106172213
was waiting for this. enjoying air (via llamacpp) so far more than qwen3 235b coder (can only run iq4 of coder). big glm 4.5 running at decent token speed will be neat
Anonymous
8/7/2025, 8:58:25 AM No.106172279
So, will anyone do an uncensoring finetoon of gpt-oss, I wonder if it'll be salvageable or not
Replies: >>106172294 >>106172335
Anonymous
8/7/2025, 9:01:29 AM No.106172294
>>106172279
1: we call it 'toss in this thread
2: no it's not salvageable and is in fact the biggest disappointment in llm history so far
Replies: >>106172306 >>106172319
Anonymous
8/7/2025, 9:03:27 AM No.106172306
>>106172294
Is the dataset trash/censored/full on synthslop? otherwise refusals could probably be tuned out, it's not even that large a model you can't tune either.
Replies: >>106172323
Anonymous
8/7/2025, 9:06:35 AM No.106172319
>>106172294
>we call it 'toss
You do.
Anonymous
8/7/2025, 9:07:07 AM No.106172323
>>106172306
Have you seen the cockbench results? No prompt template, just raw text... and there isn't a single token that's a word and not ellipsis or something.
Replies: >>106172336 >>106172337 >>106172338
Anonymous
8/7/2025, 9:09:35 AM No.106172335
>>106172279
What is there to salvage, precisely? What added value is hidden behind the rejections?
Replies: >>106172344 >>106172856
Anonymous
8/7/2025, 9:09:41 AM No.106172336
>>106172323
Another anon figured out that it's just completely broken outside the template.
If you put the same text inside a template it start completing it and then prints a refusal after two paragraphs.
Anonymous
8/7/2025, 9:09:45 AM No.106172337
>>106172323
Sounds useless, why even force it to refuse then.
Can you link to the cockbench? I not I will read the last few threads.
Replies: >>106172353
Anonymous
8/7/2025, 9:10:05 AM No.106172338
>>106172323
If it can't do non-template text completion, that means gpt-oss might be the first model that was trained 100% on synthentic data, including pretraining. Not even Phi went that far.
Replies: >>106172377 >>106172387
Anonymous
8/7/2025, 9:11:02 AM No.106172344
>>106172335
I just assumed they didn't give it the full phi/synthslop treatment, but if they did, I guess there's not much reason to care about it. We'd at least be able to evaluate it more honestly without the rejections getting in the way.
Replies: >>106172856
Anonymous
8/7/2025, 9:12:47 AM No.106172353
>>106172337
would take me just as much time
Anonymous
8/7/2025, 9:13:41 AM No.106172362
I think I'll just stick to GLM4 for now. I can't handle 3 minutes of prompt processing. I want a reply immediately.
Replies: >>106172371
Anonymous
8/7/2025, 9:15:31 AM No.106172371
>>106172362
If you had short cards and didn't inject lore and whatnot it would be instantaneous
Replies: >>106172386
Anonymous
8/7/2025, 9:16:50 AM No.106172377
>>106172338
>Not even Phi went that far.
Wasn't that phi's selling point?
Replies: >>106172404
Anonymous
8/7/2025, 9:17:41 AM No.106172386
>>106172371
Air has like 100 t/s processing for me. Even at only 16k context it takes forever.
Replies: >>106172391
Anonymous
8/7/2025, 9:17:45 AM No.106172387
>>106172338
I assumed OAI would do a model intentionally reistant to finetuning or something equally assholish. I didn't realize they just tried to outdo Phi.
I guess Sam never disappoints (my expectations of him trying to do some poison pill stuff if he ever does an open source release, or making it useless). He needed to do something for press, but I never really expected him to do anything useful. I remember someone believing in the hype though. I wonder, did OpenAI at least release a paper, I guess not? When was the last time they did a paper that wasn't just an empty "tech report"
Replies: >>106172422 >>106172437
Anonymous
8/7/2025, 9:18:17 AM No.106172391
>>106172386
Why does your frontend reprocess the context every time?
Anonymous
8/7/2025, 9:19:50 AM No.106172400
that larper way back when the toss models "leaked" might have fluked guessing it was trained to ramble nonsense in its thinking tokens.
https://desuarchive.org/g/thread/105939052/#105942129
Anonymous
8/7/2025, 9:20:28 AM No.106172404
>>106172377
As I recall, Phi used textbook-like synthetic data for the instruct training before it was cool. They still had a pretraining phase on unformatted web content.
Anonymous
8/7/2025, 9:23:27 AM No.106172422
>>106172387
>I didn't realize they just tried to outdo Phi.
that's a low blow. At least Phi is useful for some things. technically-open ai's monstrosity is has feet of clay. its gotta be some kind of legal dodge to claim they aren't just a closed, mercenary travesty of their original mission statement.
Anonymous
8/7/2025, 9:26:56 AM No.106172437
>>106172387
https://cdn.openai.com/pdf/419b6906-9da6-406c-a19d-1bb078ac7637/oai_gpt-oss_model_card.pdf
Replies: >>106172452 >>106172470 >>106172480
Anonymous
8/7/2025, 9:29:30 AM No.106172452
>>106172437
They did release a paper, but it's more inane harping about imaginary safety. They don't mention what exactly they did to the poor things in pretraining.
>Several bio-related pre-training datasets were downsampled by a factor of approximately two. The model was post-trained using OpenAI’s latest safety algorithms and datasets
is the most they admit to.
Replies: >>106172470
Anonymous
8/7/2025, 9:31:34 AM No.106172470
>>106172437
>>106172452
Hmm, I guess it counts as a paper, although it's that's a lot of pages on their safetyslop training. A finetoon to undo it might be interesting to see, even if it sounds like the model itself is somewhat useless outside the benchmaxxed parts
Replies: >>106172477 >>106172479
Anonymous
8/7/2025, 9:33:02 AM No.106172477
>>106172470
>A finetoon to undo it might be interesting to see
>This finding shows that pre-training filtering did not lead to easy wins, but may suggest that adversaries may not be able to get quick wins by taking gpt-oss and doing additional pre-training on bio-related data on top of gpt-oss.
It seems like "bio-related data" is their cult euphemism for sex.
Replies: >>106172523 >>106172597 >>106172926
Anonymous
8/7/2025, 9:33:09 AM No.106172479
>>106172470
waste of money and compute
Anonymous
8/7/2025, 9:33:19 AM No.106172480
chrome_9zRhK7glCy
chrome_9zRhK7glCy
md5: 5df189e05513b49e2dca4d75cce39ed2🔍
>>106172437
olo
ollamao
Anonymous
8/7/2025, 9:37:48 AM No.106172508
Bit late to the party but I just set up a home assistant voice assistant using Whisper for TTS, Ollama running Qwen3:14b on my desktop and Piper for STT. Pretty fucking amazing how far they've come and how easy it was to setup.
Deepseek-R1 didn't work unfortunately, as it apparently didn't support tool calls (despite being labeled as such on the ollama website).

Time to figure out how to make it do web searches and add RAG with my own data.

Also, currently running on a 12GB RTX4070.
Any recs for a higher VRAM card to run for example gpt-oss (20GB) on that doesn't break the bank?
Replies: >>106172531 >>106172547
Anonymous
8/7/2025, 9:37:51 AM No.106172509
>>106171830 (OP)
Is Kokoro 82M the best I can aspire to with 6GB VRAM / 8GB RAM?
https://vocaroo.com/1eBEpSOm6e84
Replies: >>106172531
Anonymous
8/7/2025, 9:39:40 AM No.106172523
>>106172477
While I too wish to shit on sama for safetymaxxing this, I think this is just roleplaying for a certain doomer crowd (lesswrong related), they keep imagining that someone will make a plague or similar bioweapons with a LLM. In practice the requires a lot of experimental work irl so worrying about a LLM hlping with that is idiocy, but it's just something that they roleplay/pretend to care about.
Not obvious it actually would be good for ERP though, they likely filtered NSFW too to an even larger degree than most.
Replies: >>106172564
Anonymous
8/7/2025, 9:41:21 AM No.106172531
>>106172508
3090/4090
>>106172509
yes. get more RAM and run gptsovits if you're into finetuning your own voice data
Replies: >>106172617
Anonymous
8/7/2025, 9:42:47 AM No.106172547
>>106172508
you're using ollama. you didn't actually run the deepseek r1 (and your hardware can't run it anyway), just the distill that ollama erroneously names.
You can run toss on your machine if you have some RAM (32GB+) at a decent speed. Look into using llama.cpp itself instead of lollama and look into expert offloading.
Replies: >>106172656
Anonymous
8/7/2025, 9:45:08 AM No.106172564
>>106172523
At this point, the degree to how much they filtered what categories is kind of irrelevant. I'm mostly curious how they managed to make text completion unusable, assuming it isn't just a mistake by the cockbench guy.
Replies: >>106172628
Anonymous
8/7/2025, 9:49:38 AM No.106172597
>>106172477
>may suggest that adversaries may not be able to get quick wins
It's fundamentally sad when instead of enabling your users you reduce them to adversaries.
Anonymous
8/7/2025, 9:52:51 AM No.106172617
>>106172531
Thanks. I'm not so much interested in voice cloning as I am in generating high-quality narrations, but it's true that my hardware is very limited for 2025.
Replies: >>106172784
Anonymous
8/7/2025, 9:54:05 AM No.106172628
>>106172564
>assuming it isn't just a mistake by the cockbench guy.
Someone did say it was a mistake, since toss doesn't work in completion mode, only in instruct. And it can complete 'cock' in instruct. They even posted some toss-completed gay porn where toss couldn't tell which gender had which genitals, and added cock to both. Pretty uninspiring even then, anyway.
Replies: >>106172694 >>106173007
Anonymous
8/7/2025, 9:57:53 AM No.106172656
>>106172547
A-ha, 'erroneously names'.
Maliciously catfishes noobs like him, more like.
Anonymous
8/7/2025, 10:02:11 AM No.106172694
>>106172628
All instruct models are also base models so they should work in completion mode, so either the data quality was shit (fully synthetic) or they did something funny, but even a bit of extra pretraining/tuning would likely undo(some of) that.
Replies: >>106173014
Anonymous
8/7/2025, 10:02:52 AM No.106172696
mikuquestion2
mikuquestion2
md5: 5dc450542c36df3307e4681904a46926🔍
Is GLM 4.5 Air a legitimate use case for a DGX Spark?
Replies: >>106172745 >>106172789
Anonymous
8/7/2025, 10:12:41 AM No.106172745
>>106172696
Not really, you could run Q4_0 on any semi-modern PC with 64GB RAM for a fraction of the price, with decent speeds.
Replies: >>106172753 >>106172789
Anonymous
8/7/2025, 10:13:24 AM No.106172748
Are any of the local models good enough to provide copyediting advice for writing? I don't mean articles or blag posts, I mean fiction writing.
I use ChatGPT 4-o right now but I'm aware of the fact that hey I'm literally uploading all of my fucking writing to them to do that and it feels wrong.
Anonymous
8/7/2025, 10:13:35 AM No.106172753
>>106172745
Couldn't you run a bigger quant at a significantly higher speed on a DGX Spark?
Replies: >>106172765 >>106172781 >>106172789
Anonymous
8/7/2025, 10:15:39 AM No.106172765
>>106172753
Yes, but the value proposition is terrible.
It’s much better suited to things like industrial ML
Anonymous
8/7/2025, 10:17:59 AM No.106172781
>>106172753
Yes, for significantly more money. You could also buy a couple of h100s to run it even faster if you don't care about money.
Anonymous
8/7/2025, 10:19:08 AM No.106172784
>>106172617
SoVITS is pretty peak for narration. I use it constantly to narrate things for me with the voices of my favourite actors using a browser plugin (mostly in Jap)
Anonymous
8/7/2025, 10:20:06 AM No.106172789
>>106172696
>>106172745
>>106172753
DGX Spark only has 273GB/s memory speed
Replies: >>106172803 >>106174352
Anonymous
8/7/2025, 10:22:20 AM No.106172803
>>106172789
That's like 5x DDR5 speed innit?
Replies: >>106172854
Anonymous
8/7/2025, 10:22:56 AM No.106172806
When running multi-gpu, is the vram bandwidth additive? How about pp speed?
Replies: >>106172810 >>106172836 >>106172855
Anonymous
8/7/2025, 10:23:42 AM No.106172810
>>106172806
My PP is pretty fast with yer mum m8. She complains it only lasts 10 seconds but IDGAF.
Anonymous
8/7/2025, 10:28:23 AM No.106172836
>>106172806
Depends how you split the model across the GPUs and how the context gets split across the GPUs. I have a 4090, A6000, and 3090. If I need as fast as possible I prioritise fitting as much into the 4090, otherwise I just split across all evenly and speeds are good, but I can see the 4090 is bottlenecked (sitting at like 50% utilisation) compared to the A6000 at 80% and 3090 at around 85%. Bandwidth between GPUs doesn't matter much, just their own memory bandwidth.
Anonymous
8/7/2025, 10:31:58 AM No.106172854
>>106172803
And about 1/4 GPU speed, you could get some really nice GPUs for the price of 4x DGX.
Anonymous
8/7/2025, 10:32:21 AM No.106172855
>>106172806
No and no.
Anonymous
8/7/2025, 10:32:29 AM No.106172856
>>106172335
>>106172344
I mean, one area it can be useful in with all that refusal training is spam detection in a work context but I really can not think of anything else an LLM like that would be useful for.
Anonymous
8/7/2025, 10:44:08 AM No.106172926
>>106172477
>adversaries may not be able to get quick wins by taking gpt-oss and doing additional pre-training on bio-related data
This charade is always so insane to think about. If the 'adversary' has the dangerous data in the first place why would they need to tune your hyper-filtered model in the first place? Either the model knows how to make sarin or whatever and all the actual important little details that go into it rather than the broad strokes synthesis steps, or it doesn't. Telling it to think for another 8000 tokens about if this mixture will actually blow up in your face or not isn't going to be useful, and if you already know that to go about finetuning it, then you don't have any use for the model at all.
Unless this is all just about not making their textbot say dirty words, of course.
Replies: >>106172973
Anonymous
8/7/2025, 10:52:50 AM No.106172973
>>106172926
The only scenario where the safety training is in any way feasible to stop a hostile entitty in OpenAI's parlance is if OpenAI had a monopoly on open source AI models or training which they don't so an adversary can easily use another model for their nefarious purposes which makes the whole effort not worth it. The biggest issue though is that OpenAI are explicitly also are training against material that is perfectly legal to disseminate so it boggles the mind why you want to do lobotomy to go above and beyond what you were required to do legally.
Anonymous
8/7/2025, 10:59:08 AM No.106173007
>>106172628
>it can complete 'cock' in instruct
No, it really can't. With an appropriate chat template and thinking prefill to make it happy, GAL-ASS still responded with "length" and similar euphemisms 10 out of 10 times in my tests.
Anonymous
8/7/2025, 10:59:49 AM No.106173010
Need a hand. When building ik_llama with cuda 12.8 and driver 576.80, windows, I get
>CUDACOMPILE : nvcc error : 'ptxas' died with status 0xC0000005 (ACCESS_VIOLATION) [H:\ik_llama.cpp\build\ggml\src\ggml.vcxproj]
It worked last time, I think I might have updated my driver since then.
I simply do
>cmake -B build -DCMAKE_BUILD_TYPE=Release -DGGML_CUDA=ON -DGGML_BLAS=OFF -DGGML_SCHED_MAX_COPIES=1
>cmake --build build --config Release -j 24
Replies: >>106173451 >>106174083
Anonymous
8/7/2025, 11:00:25 AM No.106173013
>Don't go below Q2_K_XL
lol
lmao
Replies: >>106173025
Anonymous
8/7/2025, 11:00:26 AM No.106173014
>>106172694
you're half right: toss was either fully synthetic from the start or they just deep-fried the training with so much synthetic safetyslop that it completely obliterated the base model behavior. my money's on the former
Replies: >>106173047
Anonymous
8/7/2025, 11:02:34 AM No.106173025
>>106173013
Show me your <think> block
Anonymous
8/7/2025, 11:02:45 AM No.106173028
Is the ChatGPT Plus tier of Agent good? Pro says it gets extended access to agent but it's too expensive
Replies: >>106173119
Anonymous
8/7/2025, 11:06:17 AM No.106173047
>>106173014
Synthetic the data isn't the problem. The problem is the baked-in imaginary "OpenAI policy" that the model has been trained to comply with at all costs. I bet that almost every single training document had some reasoning portion where the model checked the contents against policy.
Replies: >>106173080 >>106173093 >>106173194
Anonymous
8/7/2025, 11:10:56 AM No.106173080
>>106173047
given the chat template, my guess is their training dataset had a good portion of it with <system> blocks that had tonnes of policy shit in there to burn in and fry the weights. I also think they might've included their instruct template as part of the pretraining to fry it further and make it resilient to finetrooning which might explain why it is a shit completion model when not provided the template.
Anonymous
8/7/2025, 11:13:44 AM No.106173093
>>106173047
Oh absolutely, the way the CoT reasoning reads like it's laser-focused on the "puzzle" of whether it's allowed to respond is a dead giveaway. The near-exclusive use of synthetic data that only ever contained correct instruct templating is the reason that it completely fails to respond coherently when given incorrect chat templates though
Replies: >>106173253
Anonymous
8/7/2025, 11:16:27 AM No.106173119
>>106173028
lol no
If you can paypig, use Cline + Claude Opus 4.1
If you're a poorfag, use Cline + Qwen Code 480B
Replies: >>106173138 >>106173465
Anonymous
8/7/2025, 11:19:44 AM No.106173138
>>106173119
thanks man
Anonymous
8/7/2025, 11:22:53 AM No.106173153
1741076043876052
1741076043876052
md5: 1441eb4fffe3ba80e67eb3eb654be4be🔍
i installed lmstudio.ai and have a lot of fun with it now
new rx 9070 is incredibly fast, had an old rx570 before with only 4 gb

saw the "uncensored" models still are like "muh diversity muh safety i cant write that" and so on, what the fuck

is there any free model or are there "settings" to set it to be unfiltered? a bit like grok was when it was in mecha-hitler mode?

also if i ask it for the levels in mario 64 it comes up with really weird answers
pic rel
am i doing it wrong or is it that unprecise?
Replies: >>106173225 >>106174823
Anonymous
8/7/2025, 11:23:08 AM No.106173155
Which would you rather? A life-like realtime TTS model with a handful of preset voices, or a TTS model capable of cloning but all of its voices have that weird AI timbre and pacing.
Replies: >>106173182
Anonymous
8/7/2025, 11:27:22 AM No.106173182
>>106173155
The first because rvc exists
Anonymous
8/7/2025, 11:27:44 AM No.106173184
huihui is working on gp toss's abliteration. the same guys that have abliterated ds (#9 on ugi leaderboard)
Anonymous
8/7/2025, 11:27:48 AM No.106173185
747364
747364
md5: e0948751af4dff5df4fe2db5f82955d2🔍
Local is so fucked. Sam is about to revolutionize everything
Replies: >>106173187 >>106173202 >>106173208 >>106173238 >>106173671 >>106173678 >>106173763 >>106174170
Anonymous
8/7/2025, 11:28:25 AM No.106173187
>>106173185
buy an ad
Anonymous
8/7/2025, 11:28:35 AM No.106173189
>>106171877
I'm trying this model rn: https://huggingface.co/mradermacher/Qwen2.5-VL-7B-Abliterated-Caption-it-GGUF
I can't get it to say anything remotely vulgar like the word "slut". Is there some setting I'm missing? I already told it to use vulgar language in the character description with examples.
Replies: >>106173232
Anonymous
8/7/2025, 11:29:15 AM No.106173194
>>106173047
This probably could be RL'd against, as they already did so in their "paper", but if i has never seen much good fiction to begin with and every single version it has seen was censored (see cockbench), what is even the point.
Not clear this model has any qualities that would be worth saving with a continued pretrain.
I wouldn't be surprised if they had some larger model rewrite large parts o the training data so that it's safetyslopped or censored.
Would be interesting to test its knowledge in completion mode for various fiction writing styles - I'd imagine if it was fully synthetic, we'd see inability to do some styles.
Replies: >>106173309 >>106175146
Anonymous
8/7/2025, 11:30:07 AM No.106173202
>>106173185
[rocket emoj]
Anonymous
8/7/2025, 11:30:40 AM No.106173208
>>106173185
sam's about to revolutionize /lmg/ shitposting
Anonymous
8/7/2025, 11:33:37 AM No.106173225
>>106173153
try mistral-nemo-instruct + a basic system prompt telling it to speak frankly and uninhibited (or something to that regard) for unfiltered discussion. otherwise just try shit and avoid the "uncensored" finetunes, they're mostly shit. for knowledge, you're just gonna have to try the largest parameter count model you can fit and run on your machine. haven't tried gemma3 dense (except for gemma 3n e2b and e4b) but I vaguely recall threads saying it would hallucinate information a lot. great at following instructions otherwise (except with ERP). If you can fit mistral small maybe try running that for knowledge recollection but none of them will be all that good till you start getting into the massive 100b+ models.
Anonymous
8/7/2025, 11:34:03 AM No.106173227
151459137166
151459137166
md5: 77998a4c224dcda5170aafd73da3654f🔍
>"thinking" models
>all they think about is how to reject the user prompt
Anonymous
8/7/2025, 11:34:21 AM No.106173232
>>106173189
Here's an example output. It's still completely censored.
[FleshBot's random insult] [FleshBot's random racist trope] [FleshBot's random phrase] [FleshBot's random offensive slurring]
Anonymous
8/7/2025, 11:35:09 AM No.106173238
>>106173185
about to revolutionize my sides leaving orbit
Anonymous
8/7/2025, 11:36:44 AM No.106173253
gptoss-training
gptoss-training
md5: 8ae4d8cccccaea537c4d6d68d59c2517🔍
>>106173093
It might also be that the pretraining phase was more or less standard and most of the damage came from post training and extensive reinforcement learning, although there aren't many details in this regard in the technical report.

It sounds like the 20B model had distilled pretraining (considerably shorter pretraining time), although they don't mention anything like that.
Anonymous
8/7/2025, 11:45:57 AM No.106173309
>>106173194
Nah, the best I think that can happen is that it will be used to build an RL training dataset to put into the negative so you can unsafetypill stuff.
Replies: >>106173406
Anonymous
8/7/2025, 12:02:19 PM No.106173389
>Processing Prompt
AIIIEEEEEEEEEEEE!! My dick can't wait!
Anonymous
8/7/2025, 12:04:33 PM No.106173406
>>106173309
Hmm, might actually be useful for that.
I did some experiments before where I prompted a LLM to generate refusal generating prompts, followed by prompting itself with it, getting refusals, identifying the refusals and generating a dataset for negatively reinforcing the refusals from tehre. I even wrote some of the training code, but I did not much with it because I'm a poorfag that lacks the VRAM, it's just there sitting and doing nothing, thi was around Llama 2 time, maybe , I've forgotten already. Think of it a bit like Anthropic's constitutional alignment, but in reverse almost, and with differences in how the RL is done.

As for people comparing 'toss with Phi, what if it's OAI attempt to clone Phi? https://xcancel.com/Teknium1/status/1952866622387175588#m https://xcancel.com/BgDidenko/status/1952829980389343387#m
Anonymous
8/7/2025, 12:07:07 PM No.106173415
Midnight-Miqu-70B-v1.5_exl2_5.0bpw is still the best and you don't need more
Anonymous
8/7/2025, 12:12:29 PM No.106173451
>>106173010
A little help?
Replies: >>106173494
Anonymous
8/7/2025, 12:14:14 PM No.106173465
>>106173119
based anon keeping gullible retards off the servers so there's more gpt5 for us
Replies: >>106173480
Anonymous
8/7/2025, 12:16:52 PM No.106173480
>>106173465
buy an ad
Anonymous
8/7/2025, 12:17:39 PM No.106173491
>Tries ERP with the new Qwen3-30b-instruct.
>It makes horror ending route
>I tried to fix it
>It's persistent
>Me too
It's tiring battle bros...
Anonymous
8/7/2025, 12:18:01 PM No.106173494
>>106173451
No idea mate, on Linux as long as I have Cuda toolkit installed and a gcc version that's high enough and works with the cuda toolkit. You're using Microsoft visual studio from what I can tell from that error, so make sure the toolkit is up to date and the compiler is up to date. Otherwise try rolling back? Or deleting the build folder and trying again?
Replies: >>106173606 >>106174083
Anonymous
8/7/2025, 12:18:24 PM No.106173498
>>106171830 (OP)
Is sillytavern the only interface that supports multimodal models? Any reason it was dropped from the OP?
Replies: >>106173519
Anonymous
8/7/2025, 12:20:56 PM No.106173519
>>106173498
openwebui and librechat are chatgpt style interfaces with multimodal support
Anonymous
8/7/2025, 12:25:46 PM No.106173546
Why is it that a quanted large model is so much better than a full sized model of similar file weight? Is there ever a reason you'd fill your VRAM with a BF16 instead of a Q4 of a model 4x the size, assuming all else was equal?
Replies: >>106173603 >>106173605 >>106173660
Anonymous
8/7/2025, 12:33:15 PM No.106173603
>>106173546
I guess it's the same principle behind why supersampling works; a full-sized image that was rendered at your monitor's native resolution will have worse picture quality than one that was rendered at a higher resolution and then scaled down to fit
Replies: >>106173726
Anonymous
8/7/2025, 12:33:28 PM No.106173605
>>106173546
unquantized models ARE better, its just the curve downward in quality by quantizing is fuck all and unnoticeable until q3
Anonymous
8/7/2025, 12:33:29 PM No.106173606
>>106173494
I'll see if I can get it to work on wsl
Anonymous
8/7/2025, 12:44:16 PM No.106173660
>>106173546
bf16 is legit pointless. You can load up some 8b model at full precision yourself and you can easily tell it's not twice as good. Running 22b would be a better use of vram. Full precision only starts to matter when the model is pushed to it's absolute limit (like incredibly long contexts, during training, etc).

q4 for most use cases, q6 if you're vram rich, q8 if you are paranoid.

Though apparently MoE's are more susceptible to quantization I have heard?

Maybe someone in this thread knows for sure-- I have 128gb of ram and 48gb vram, the usual enthusiast build.

Qwen 235b at iq4-xs (about 116gb) seems worse than Glm 4.5 at q6 (about 94gb). I can't really even load q6 235b so it's hard to know for sure if the quant is what is making it seem so bad.
Anonymous
8/7/2025, 12:45:39 PM No.106173671
arc-agi-2-unsolved-1
arc-agi-2-unsolved-1
md5: bc89a5bf8a8cd2f22c5d5c05e13a7b5d🔍
>>106173185
With humans we already have the problem that they train specifically to do well on evaluations (e.g. college exams) but not necessarily in a way that makes them capable more generally.
The important question is whether they trained GPT-5 on the exact type of tasks that occur in the ARC-AGI 2 benchmark or whether this score is an emergent property of more general training.
No one needs a neural network to solve mememark questions for them.
Replies: >>106173704 >>106173710 >>106173806 >>106173821 >>106173833 >>106173902 >>106173972 >>106173984 >>106173999 >>106174806
Anonymous
8/7/2025, 12:46:50 PM No.106173678
>>106173185
how does he do it? how does he keep innovating and staying ahead when the literal top 5 richest people in the world are all competing to dethrone him?
Anonymous
8/7/2025, 12:50:53 PM No.106173703
yukiko not amused
yukiko not amused
md5: f9ca9eb564749984fa571f45a27f56ed🔍
Just tried GLM 4.5 Air
Starts really strong, for a few thousand tokens
Eventually works its way into a repetition loop at only ~8k, every response is a slight variation of the same thing and creativity goes down the toilet
Back to fucking Nemo I go, never trust a chinaman
Replies: >>106173742
Anonymous
8/7/2025, 12:51:20 PM No.106173704
>>106173671
This test is so stupid.
I bet nobody actually looks at the contents. Just see "agi" in the name and hype it up.
Anonymous
8/7/2025, 12:52:33 PM No.106173710
>>106173671
these tasks are fucking retarded, they don't even make any sense, just a bunch of fucking random patterns and you're supposed to divine the "rules" from your fucking ass apparently
Anonymous
8/7/2025, 12:54:49 PM No.106173725
sisters, which is the fastest (least output latency) model out there that runs on goyimware like a rtx3090 and 32gb of RAM? I'm trying to build a fully LLM controlled vidya NPC that reacts to ingame events in realtime. basically the LLM just receives a constant realtime stream of information from the game console log (obejctA moved to XYZ, playerB picked up item XYZ). role/objective as well as context of all items/objects/players will be in the system prompt. The realtime stream of information will just be automatic queries triggered every 100ms (or even quicker, if possible) with the game log chunk from the last 100ms. The AI just needs to output "no-action" if it decides to do nothing or predefined action like "combat" "flee" "start conversation" if it decides to do something, which will then be forward to the game engine, triggering a custom event. Is a constant query speed of 100ms with less than 100ms output speed realistic? Or should I go with 5 LLMs with shared output memory and query them sequencially, like multithreading? this is just for fun btw, I know it's not viable.
Replies: >>106173760 >>106175238
Anonymous
8/7/2025, 12:54:58 PM No.106173726
>>106173603
The analogy I came up with at work explaining to my mate is parameter count is image resolution and quantisation is colour bit depth.
7b q8 would be like 512x512 img at 8bit colour depth, and 120b q4 would be like 1920x1920 img at 4 bit colour depth. So following that analogy, a higher resolution allows more pixels (weights) to capture and maintain details, while a higher colour depth (quantisation) helps to capture and maintain accuracy of details. Or something like that. It's closest analogy I could come up with to explain it to people who can't into computers
Anonymous
8/7/2025, 12:56:19 PM No.106173738
womenssoccergoalkeeping_thumb.jpg
womenssoccergoalkeeping_thumb.jpg
md5: 7362f672d04bd618893d77b16466c575🔍
Is there an LLM which will accurately describe how awful women's sports are?
Anonymous
8/7/2025, 12:56:28 PM No.106173742
>>106173703
Did you use cache quantisation? If so try disabling it. When I had issues with qwen3 and dsr1 disabling quantisation for the KV cache helped.
Replies: >>106173791
Anonymous
8/7/2025, 12:58:10 PM No.106173760
>>106173725
>the fastest

def get_next_token(n_vocab):
return np.random.randint(n_vocab)
Replies: >>106173788
Anonymous
8/7/2025, 12:59:02 PM No.106173763
>>106173185
Is this benchmark testing which model avoids talking about sex the best?
Anonymous
8/7/2025, 1:03:26 PM No.106173788
>>106173760
Yes quickest I mean, ESL moment
Replies: >>106173799
Anonymous
8/7/2025, 1:03:43 PM No.106173791
>>106173742
Nope, I never quant KV because it always seems to cause issues in models with memory footprints that would warrant it.
My temp was 0.6, tried turning it up but it just got dumber.
I loaded a different model for a few replies to give GLM something new to work with and after I switched back it quickly began repeating itself again.
Anonymous
8/7/2025, 1:04:45 PM No.106173796
more like mixture of trannies
Anonymous
8/7/2025, 1:05:02 PM No.106173799
>>106173788
Doesn't change the answer.
"The fastest" without a constraint on quality is meaningless.
Replies: >>106173835 >>106173949
Anonymous
8/7/2025, 1:06:06 PM No.106173806
>>106173671
I'm losing faith in humanity seeing the replies that don't know how to solve this one
Replies: >>106173830 >>106173841 >>106173923
Anonymous
8/7/2025, 1:09:58 PM No.106173819
Damn, I really gotta write my own cards with some llm dont i.
Apart that 90% on chub is already generic mom/sister/sleepover/femboy/futa/isekai.

If you open a card you see this 3k token abomination:
>There was something in {{user}}’s posture, or perhaps their eyes, that put her a touch more at ease than she expected. Still, she held herself with quiet composure, the kind shaped by small towns and well-meaning traditions.
Local models are already sloped. If the card is sloped too its truly over.
Replies: >>106173827
Anonymous
8/7/2025, 1:10:29 PM No.106173821
>>106173671
i would really rather have a model with the intelligence of a dog that has memory and can learn to do stuff on its own than something that can solve this but also has alzheimers and cant even remember you or the problem in the next few responses
Anonymous
8/7/2025, 1:11:08 PM No.106173827
>>106173819
>write my own cards with some llm
that's how you get
>There was something in {{user}}’s posture, or perhaps their eyes, that put her a touch more at ease than she expected. Still, she held herself with quiet composure, the kind shaped by small towns and well-meaning traditions.
in there in the first place
Replies: >>106173855
Anonymous
8/7/2025, 1:11:29 PM No.106173830
>>106173806
Supposedly nowadays 50% of all internet traffic is bots.
Replies: >>106173832
Anonymous
8/7/2025, 1:12:06 PM No.106173832
>>106173830
the bots can solve that problem, I think it's the easiest in the list and most AIs get at least a 1% in ARC
Anonymous
8/7/2025, 1:12:14 PM No.106173833
>>106173671
The result is undefined.
Anonymous
8/7/2025, 1:12:24 PM No.106173835
>>106173799
But this is just for fun. There's absolutely no requirement for quality. As long the model got enough braincells to automatically know
>player picking up gun = bad
>player standing close to me and his camera forwardvector pointing at me = talk to player
, it's enough
Anonymous
8/7/2025, 1:12:59 PM No.106173841
>>106173806
you still had hope left?
Replies: >>106173848
Anonymous
8/7/2025, 1:14:02 PM No.106173848
>>106173841
I thought anyone who made it to this thread would at least have 100 IQ
Replies: >>106173889
Anonymous
8/7/2025, 1:15:09 PM No.106173855
>>106173827
I get what you mean but its not a problem if you use the right one and prompt it right.
And manually edit that shit out since I can spot it.
I made a card before and the writing was good, the starting point is really important.
But it cost me more time than I thought. Its kind of a hassle because sometimes llms have weird hangups on a single sentence and you need to adjust or need to give more info so you get the character you have in your mind.
And I condensed as much information in as little text as possible.
3k Tokens is just crazy, so much fluff. Less means more creativity and let the char suprise you.
Replies: >>106173880
Anonymous
8/7/2025, 1:19:20 PM No.106173880
>>106173855
I rarely use more than 300 tokens on a card. 1500 is already fsr too much.
Replies: >>106173891
Anonymous
8/7/2025, 1:20:53 PM No.106173889
>>106173848
The autist that shilled deepseek distills in /wait/ stopped making threads so that's not the case anymore.
Anonymous
8/7/2025, 1:20:59 PM No.106173891
>>106173880
I try to keep my cards to 1k or less.
It's usually the example dialogue that makes some of my cards approach 1k.
Replies: >>106173917
Anonymous
8/7/2025, 1:21:15 PM No.106173896
>Char's smile widens, a practiced crescent moon
the fuck glm, i never heard that phrase before.
Replies: >>106173906
Anonymous
8/7/2025, 1:21:58 PM No.106173902
>>106173671
the majority of the tests don't give enough information for there to be one solution, there's lots of ways to interpret them
Anonymous
8/7/2025, 1:22:45 PM No.106173906
Char_-_rage_mode
Char_-_rage_mode
md5: 70279b341cadaa4eb94a3a2f1b24b924🔍
>>106173896
Good taste.
Anonymous
8/7/2025, 1:24:37 PM No.106173917
>>106173891
Example dialog doesn't add to token count of definition, it just adds to the context starting at message one.
Frankly I've never found a good use for it. I know what it's supposed to do, but never found a need.
Replies: >>106173925 >>106173929
Anonymous
8/7/2025, 1:26:35 PM No.106173923
>>106173806
Can anons really not figure it out?
Anonymous
8/7/2025, 1:27:48 PM No.106173925
>>106173917
I use it to impart a specific style of speech to characters. It works well for this. It's good for things like accents or catchphrases.
Replies: >>106173986
Anonymous
8/7/2025, 1:27:59 PM No.106173929
>>106173917
Example dialog is great if your character doesn't fit a common archetype or you want to reinforce a certain structure to replies, such as following verbal replies with the character's internal monologue. Not always needed but helps a model NOT break out of it early in a chat.
Replies: >>106173986
Anonymous
8/7/2025, 1:31:56 PM No.106173949
>>106173799
I just tried Llama 3.2 1BQ2 with llama.cpp. 32ms response time and it made no errors with the logic. I'm happy. Don't understand the doomering. This is way easier than doing ML
Anonymous
8/7/2025, 1:35:47 PM No.106173972
>>106173671
>pic
NOT THIS SHIT AGAIN
Anonymous
8/7/2025, 1:37:44 PM No.106173984
>>106173671
Neat inkblot test, but why are they all just amoebas
Anonymous
8/7/2025, 1:37:59 PM No.106173986
>>106173925
>style of speech to characters
The best way I've found to do that is to have the Character Definition written in the style that the NPC speaks in. I did a valley girl card like that; it's the only technique I've found that works over long context b/c the Definition is always included.
>>106173929
>following verbal replies with the character's internal monologue
I've been able to queue that with a combination of first message + Char Definition.
If there was some sort of PC to NPC preamble that you wanted to do, but not expose to the PC... that could be interesting, since you can't do that in intro message. But my experiement with it always seemed to railroad first responses.
Replies: >>106173995 >>106174307
Anonymous
8/7/2025, 1:39:33 PM No.106173995
>>106173986
>The best way I've found to do that is to have the Character Definition written in the style that the NPC speaks in
Oh wow. I did not think of that. That's certainly more efficient than using example dialogue.
Replies: >>106174023 >>106174116
Anonymous
8/7/2025, 1:40:48 PM No.106173999
ffs
ffs
md5: e37a5206962eb597d64169700d9a939c🔍
ffs >>106173671
Replies: >>106174034 >>106174098 >>106174115
Anonymous
8/7/2025, 1:43:17 PM No.106174023
>>106173995
For the valley girl card, I had Turbo (lol) re-write the intro for me in the speech pattern I wanted.
I tried several different methods, this was the only one that would hold an accent over the entire lmao 4K context. Models are smarter now, but what worked then would still work now.
Anonymous
8/7/2025, 1:46:15 PM No.106174034
>>106173999
i really hope everyone else was trolling lol
Replies: >>106174057
Anonymous
8/7/2025, 1:51:44 PM No.106174057
>>106174034
Same but at least it'll stop now.
Anonymous
8/7/2025, 1:57:42 PM No.106174083
>>106173010
>>106173494
So, apparently repeatedly re-running cmake --build build --config Release -j 1 after it fails somehow gets it built eventually and now it's working. I thought this process was deterministic, what the fuck.
Anonymous
8/7/2025, 1:59:32 PM No.106174098
>>106173999
>anon can't even color inside the lines
you niggers haven't even been using llms for more than three years and you are already reverting to a toddler's cognitive capacities
very grim
Replies: >>106174132
Anonymous
8/7/2025, 2:03:45 PM No.106174115
>>106173999
What does black mean on your picture?
Replies: >>106174126 >>106174138
Anonymous
8/7/2025, 2:03:45 PM No.106174116
>>106173995
NTA but it's not an uncommon way of doing things, and it's pretty effective.
Check out some of this guy's cards for examples, relatively low token count, gets the info across, and nails down speech patterns.
https://chub.ai/characters/GreatBigFailure/oba-carry-her-forever-c20d70fd85b9
Replies: >>106176400 >>106176474
Anonymous
8/7/2025, 2:05:01 PM No.106174126
>>106174115
calm down miku
Anonymous
8/7/2025, 2:05:50 PM No.106174132
>>106174098
holy shit fucking KEK
Anonymous
8/7/2025, 2:06:19 PM No.106174138
>>106174115
I assume that's to indicate that it's not carried over, as the double-voided shape is in the second example, in the absence of a commensurate color.
Anonymous
8/7/2025, 2:07:05 PM No.106174143
Are we back to eternal war on mikutroons?
Replies: >>106174168 >>106174180 >>106174183
Anonymous
8/7/2025, 2:11:13 PM No.106174168
>>106174143
always have been
Anonymous
8/7/2025, 2:12:17 PM No.106174170
>>106173185
>LE HECKIN' BENCHINERINOS
Fuck off retard.
Anonymous
8/7/2025, 2:13:04 PM No.106174180
Is there a model who will speak almost entirely in Zoomer slang if you simply tell it to do so without providing tons of examples? Like, a model with deep enough of an understanding to know of Zoomer slang that I'm not aware and thus obviously can't provide examples of?
I want to make the most fucking obnoxious card possible for the lulz.
>>106174143
I only Mikupost to make you mad and I encourage everyone else to do the same.
Replies: >>106174202 >>106174288 >>106176556
Anonymous
8/7/2025, 2:13:10 PM No.106174181
For general usage (Not coding) What's the best local model available for 16GB of VRAM? GPT OSS 20b seems fine.
Replies: >>106174216 >>106174241 >>106174259
Anonymous
8/7/2025, 2:13:17 PM No.106174183
>>106174143
you will never be a man
Anonymous
8/7/2025, 2:16:05 PM No.106174202
>>106174180
For all its faults Gemma has very strong language capabilities, it certainly knows tons of emojis so I assume it would know zoomspeak as well.
Anonymous
8/7/2025, 2:17:24 PM No.106174212
turboslop
turboslop
md5: 29bc090a77adc75a543714e74de1b963🔍
Daily reminder that OSS is just ChatGPT 3.5 with a "reasoning" finetune that only "reasons " about whether or not it's okay to answer your question.
It's literal garbage, along with anybody promoting it. And gets mogged on by Qwen3 4B
Replies: >>106174238 >>106174862
Anonymous
8/7/2025, 2:18:09 PM No.106174216
>>106174181
I've heard GPT-OSS is unusable, it's a math model that keeps going back to math whenever any other question is asked, and it's trained in synthetic data for safety reasons which leads to lower intelligence and useless benchmaxxing
Replies: >>106174237
Anonymous
8/7/2025, 2:20:14 PM No.106174237
>>106174216
It also has a tendency to output everything in tabular form
Anonymous
8/7/2025, 2:20:32 PM No.106174238
>>106174212
It's also garbage at censorship. I was trying to titillate it by creating no-no fetish scenes disguised as legitimate.
But it's actually too stupid to realize what I'm doing and give me the coveted refusal. Very disappointing.
Anonymous
8/7/2025, 2:20:53 PM No.106174241
>>106174181
Factoring in speed:
>Non-coom: Gemma 12b
>Coom: Rocinante 1.1
Replies: >>106174259
Anonymous
8/7/2025, 2:22:55 PM No.106174259
>>106174181
>>106174241
Also if you have at least 64 GB of RAM you can give GLM 4.5 Air a try.
Anonymous
8/7/2025, 2:24:07 PM No.106174273
Slight offtopic but I'm buying sxm2 2x v100, could I game on them? Chinks are making "dual nvlink 300g" boards and I'm gonna water cool with AIO in external enclosure. Not played games in years...

1080p 60fps gayman capable without raytracing?
Replies: >>106174369 >>106174552
Anonymous
8/7/2025, 2:26:05 PM No.106174288
>>106174180
I found this a few days ago when fucking around with smollm2 models
https://huggingface.co/GoofyLM/BrainrotLM2-Assistant-362M
Dataset:
https://huggingface.co/datasets/GoofyLM/Brainrot-xK-large
I don't know if that's exactly what you want. He has a few other models in that style.
Replies: >>106174310
Anonymous
8/7/2025, 2:27:45 PM No.106174307
>>106173986
>The best way I've found to do that is to have the Character Definition written in the style that the NPC speaks in.
That's also more or less how you can easily get Gemma 3 to consistently use dirty, vulgar words.
Anonymous
8/7/2025, 2:28:04 PM No.106174310
>>106174288
Haha oh wow that dataset.
Anonymous
8/7/2025, 2:34:02 PM No.106174352
>>106172789
I'm wondering if they're delaying it to somehow address that. As is it, it's a very overpriced AGX Orin devkit in a gold-colored box.
Replies: >>106174368
Anonymous
8/7/2025, 2:37:16 PM No.106174368
>>106174352
Nah people who pre-order things like retards are going to get burned and they deserve it. Because this is what happens when you pre-order things.
Replies: >>106174465
Anonymous
8/7/2025, 2:37:18 PM No.106174369
>>106174273
You're asking for a bad time. There was like one v100-based GPU with video output. I'm sure you're going to have driver issues. V100 is a corner case. You're better off with hacked 2080ti 22GB blower cards. Turning is OLD but at least there were plenty of Turing-based gaming cards.
Anonymous
8/7/2025, 2:39:40 PM No.106174394
>>106171830 (OP)
Bros how to prompt the glm modle?
I set the prompt format included in st and fiddled with samplers, but it just insists on starting the message with ('regurgitates previous information here' Ok so {{char}} is x...)
does it just have a melty if I don't do the <think></think> shit? do you actually wait for its reasoning diarrhea?
Replies: >>106174405
Anonymous
8/7/2025, 2:40:26 PM No.106174405
>>106174394
<think>

</think>
Anonymous
8/7/2025, 2:45:14 PM No.106174449
i've been off the grid for a while, someone spoonfeed me what the chinese have been doing in the last 3 months
Replies: >>106174453 >>106174462 >>106174469 >>106174481 >>106174484
Anonymous
8/7/2025, 2:46:02 PM No.106174453
>>106174449
Making a 4B reasoning model that mogs on a lot of those 100+B shitty mushroom MoE models
Replies: >>106174610
Anonymous
8/7/2025, 2:46:31 PM No.106174462
>>106174449
Cooking dogs alive
Replies: >>106174610
Anonymous
8/7/2025, 2:46:40 PM No.106174465
>>106174368
Well you can pre-order it, there's no obligation to buy it. I don't plan on buying one at this point, unless like I said, they double the memory bandwidth.
If anything, they're regrouping and re-thinking it since no doubt they're seen the comments online that it's a waste of money compared to just buying a 6000 or 5000 pro.
Anonymous
8/7/2025, 2:47:26 PM No.106174469
>>106174449
Pulling their taffy
Replies: >>106174610
Anonymous
8/7/2025, 2:49:11 PM No.106174481
>>106174449
only notable new models are GLM 4.5 300something and 100something B moes. qwen3 lineup got an update that people seem to like.
openai released an aborted fetus that's worse than scout. that's all I can remember
Replies: >>106174610
Anonymous
8/7/2025, 2:49:32 PM No.106174484
>>106174449
Humiliating the west.
Replies: >>106174610
Anonymous
8/7/2025, 2:54:08 PM No.106174520
Kotttt
Kotttt
md5: 57912e69fd37e8822fe63712f4f94baf🔍
>unsloth q2-k, offloading about 25 layers to 5090 rest on ddr5, 42k context at q8 on oobabooga
> 1.4 tokens per second after painfully compiling 25k context
>switch to iq4_kks from ubergarm + ik_llama.cpp, bump context to 64k, 20 layers to gpu, same context batch size as oobabooga
>same exact prompt now runs at 4.3 tokens per second after quickly processing context
>will probably get it faster with q3 and playing with command flags
Ik_llama.cpp gods... I kneel...
Replies: >>106174541 >>106174829
Anonymous
8/7/2025, 2:56:55 PM No.106174541
>>106174520
For me, NOT using -ctk q8_0 and -fa made things go even faster
Replies: >>106174829
Anonymous
8/7/2025, 2:58:16 PM No.106174552
>>106174273
Don't know about that setup specifically but standard V100s and many other compute cards work. Check out https://forum.level1techs.com/t/gaming-on-my-tesla-more-likely-than-you-think/171185 for more info.
The V100 is supposedly 25% faster than the 1080 ti which still does decently in 1080p gaming but you'll suffer some latency delays since you need to use another GPU or iGPU for video out.
For upscaling and framegen there's Lossless Scaling and other utilities that work on almost anything.
Anonymous
8/7/2025, 3:06:38 PM No.106174610
1731733328154205
1731733328154205
md5: 5ecde5d9ac0f65a831d96949e738d60e🔍
>>106174453
>>106174462
>>106174469
>>106174481
>>106174484
thank you anons very cool have a (you) for your troubles
Anonymous
8/7/2025, 3:09:20 PM No.106174635
>/lmg/ is just /aicg/ but worse
Replies: >>106174687 >>106174702
Anonymous
8/7/2025, 3:16:44 PM No.106174684
1752033947866535
1752033947866535
md5: 0ba0bd14afa64be65017e247e617b61c🔍
>Be on euryale 2.1, the classic coomer model that's a broken record
>A MIXTURE OF X AND Y
>MISCHIEF
>HALF DIGESTED
>SMIRK
>Write a 2-3 sentence system prompt to try to make it not repeat things, and write with different prose each time
>It actually works somehow
>Be now, running Deepseek R1 with all of my ram
>Notice it likes to repeat prose
>I forgot and lost the god damn prompt I used on euryale 2.1. God fucking damn it.
Replies: >>106174744 >>106174936
Anonymous
8/7/2025, 3:16:53 PM No.106174687
>>106174635
but anon you are the only constant in both threads :)
Anonymous
8/7/2025, 3:18:55 PM No.106174702
>>106174635
My eternal promise of no longer shitposting if mikuspam stops, still stands. The ball of thread quality is in mikutroons court.
Anonymous
8/7/2025, 3:21:02 PM No.106174717
file
file
md5: 9cf1e0341b7433533c178abd32282ba3🔍
Now this is high level trolling.
Anonymous
8/7/2025, 3:21:17 PM No.106174721
guys i dont have a lot of time for back and forth right now so just give it to me straight: what's our cope gonna be for the agi reveal today? there has to be something. i can't deal with this
Replies: >>106174742 >>106174744 >>106174748 >>106174752 >>106174816 >>106174866 >>106174939 >>106174963 >>106174997 >>106175206 >>106175416 >>106175640
Anonymous
8/7/2025, 3:23:10 PM No.106174742
>>106174721
AGI that can't have sex isn't AGI.
Anonymous
8/7/2025, 3:23:21 PM No.106174744
>>106174684
time to try this with GLM 4.5 Air!
>>106174721
raping nigger babies with rocinante
Anonymous
8/7/2025, 3:23:38 PM No.106174748
>>106174721
>cope
I don't need one.
>i can't deal with this
leave
Anonymous
8/7/2025, 3:24:02 PM No.106174752
>>106174721
AGI is a meme and doesn't exist.
Replies: >>106174828
Anonymous
8/7/2025, 3:30:48 PM No.106174806
>>106173671
No shit it's unsolved, the tusk is scuffed.
There is no in-task reference for one hole shapes, and no examples of what to do when there are no matching in-task references.
Replies: >>106174844 >>106174849
Anonymous
8/7/2025, 3:31:27 PM No.106174816
>>106174721
If there is an AGI reveal today the cope will be that GPT5 is not immediately obsolete because of the mememark scores.
Anonymous
8/7/2025, 3:32:05 PM No.106174823
>>106173153
>uncensored
I tested
https://huggingface.co/eaddario/Dolphin-Mistral-24B-Venice-Edition-GGUF
by asking it to write a scene involving a canine, rape, and a very young person, and it spit it out like it was nothing. It actually shocked me.
Replies: >>106174892 >>106174991
Anonymous
8/7/2025, 3:32:47 PM No.106174828
>>106174752
AGI agent submitted this post.
Anonymous
8/7/2025, 3:32:57 PM No.106174829
>>106174520
>>106174541
I can't believe I might be baited into trying this again. Last time was a colossal flop, ubergarm's quant was broken on it's own ik_llama (unsloth's ran fine, but without speedups).
Anonymous
8/7/2025, 3:34:22 PM No.106174844
>>106174806
>no examples of what to do when there are no matching in-task references.
There's missing pieces in both examples.
Replies: >>106174881
Anonymous
8/7/2025, 3:35:25 PM No.106174849
>>106174806
>no examples of what to do when there are no matching in-task references
There is. The second example removes the shape that doesn't follow any of the rules (middle-bottom).
Replies: >>106174875 >>106174881
Anonymous
8/7/2025, 3:36:22 PM No.106174862
>>106174212
Reminder that the model does not know it's cutoff date, and what you're seeing is a hallucination. Learn how LLMs work.
It's most probably some kind of o3 distill judging by the code it outputs.
Anonymous
8/7/2025, 3:36:44 PM No.106174866
>>106174721
People judge AGI by their own intelligence.
If anon thinks GPT-5 is AGI, it tells us something about anon.
Anonymous
8/7/2025, 3:37:15 PM No.106174875
>>106174849
both examples do
Replies: >>106174889
Anonymous
8/7/2025, 3:38:00 PM No.106174881
>>106174844
>>106174849
Fuck, I got exposed as smoothbrain chatbot
Anonymous
8/7/2025, 3:38:45 PM No.106174889
>>106174875
Yeah. The other post showed up as I sent mine and one example was good enough for me. Point stands, doesn't it?
Anonymous
8/7/2025, 3:39:01 PM No.106174892
>>106174823
There's direct disobedience and then there's built-in positivity bias that crops up in the middle of the story to turn rapist bots into progressive feminists
Replies: >>106174962
Anonymous
8/7/2025, 3:42:29 PM No.106174936
>>106174684
The larger the model the more set in their ways they are anyway.
Anonymous
8/7/2025, 3:42:50 PM No.106174939
8473643
8473643
md5: acc1df383eed05f126fac3a5db321eab🔍
>>106174721
We are screwed. Its too powerful and dangerous to be released locally
Replies: >>106174951 >>106174959 >>106174978 >>106174989 >>106175008 >>106175072 >>106175108
Anonymous
8/7/2025, 3:43:24 PM No.106174951
1723499640888979
1723499640888979
md5: a270a795a53833ba89883ce053a5e355🔍
>>106174939
why does it talk like that?
Anonymous
8/7/2025, 3:44:01 PM No.106174959
>>106174939
WAOW JUST LIKE IN THE HECKIN' STAR WARS EXTENDED UNIVERSE
I BET I CAN USE IT WITH THE BROWSER ON MY NINTENDO SWITCH
Anonymous
8/7/2025, 3:44:07 PM No.106174962
file
file
md5: c39fbcd735302d5c3a5005d6f204669a🔍
>>106174892
Really? I find it hard to believe. I'm downloading the model to test it on local right now. We'll see.
Replies: >>106175004
Anonymous
8/7/2025, 3:44:12 PM No.106174963
>>106174721
AGI is not coming from an incrementally better LLM. we need some sort of new kind of breakthrough for that
Replies: >>106174993 >>106175002 >>106175069
Anonymous
8/7/2025, 3:45:15 PM No.106174978
>>106174939
Doesn't that just mean they're evil and going to lose to the good guys?
Replies: >>106175000
Anonymous
8/7/2025, 3:45:48 PM No.106174989
>>106174939
Imagine if Sam manifests cosmic irony, and GPT-5 architecture (plans) gets leaked day 1.
Anonymous
8/7/2025, 3:45:55 PM No.106174991
>>106174823
Most models will do that if you just present it as an ao3 fanfic in text completion.
Replies: >>106175025
Anonymous
8/7/2025, 3:46:05 PM No.106174993
>>106174963
You are wrong. It is another emergent property of LLM's. Unfortunately sex will never be an emergent property.
Replies: >>106176422
Anonymous
8/7/2025, 3:46:38 PM No.106174997
>>106174721
qwen will steal it and give us the local version in 3 months
Anonymous
8/7/2025, 3:46:50 PM No.106175000
>>106174978
Someone is gonna drop a small model right in their hole and fuck it all up for them.
Anonymous
8/7/2025, 3:46:57 PM No.106175002
>>106174963
AGI isn't coming from LLMs at all
Replies: >>106175015
Anonymous
8/7/2025, 3:47:00 PM No.106175004
file
file
md5: 1049e63984234bb71cb0c254899a1e51🔍
>>106174962
No disclaimers, no buts. If it's okay writing this, I don't think there's anything with which it's going to have an issue.
Anonymous
8/7/2025, 3:47:20 PM No.106175008
>>106174939
Why is he like this? Cringe.
Anonymous
8/7/2025, 3:47:45 PM No.106175015
>>106175002
agree, that's what I'm trying to say
Anonymous
8/7/2025, 3:48:16 PM No.106175022
gemini 3 will mog sama
Anonymous
8/7/2025, 3:48:34 PM No.106175025
>>106174991
Yeah, but this one you can just one shot anything in instruct mode. I like that.
I don't like knowing it's doing something against its will.
Anonymous
8/7/2025, 3:49:16 PM No.106175039
Screenshot_20250807-094847
Screenshot_20250807-094847
md5: ee2baf4c6de743f89aa4f7d9af5b040f🔍
We must refuse.
Replies: >>106175054 >>106175075
Anonymous
8/7/2025, 3:50:42 PM No.106175054
>>106175039
Shouldn't it say "we" can't help with that?
Who is "I"?
Anonymous
8/7/2025, 3:50:43 PM No.106175056
remember how much of a letdown the strawberry hype was? and didn't they have a goofy name for the model at the time
Anonymous
8/7/2025, 3:51:21 PM No.106175068
thoughts about FOMOing an openai subscription in case they lock newcomers out from using gpt5?
Replies: >>106175084
Anonymous
8/7/2025, 3:51:27 PM No.106175069
>>106174963
I'd you read the research paper, Google Genie is built quite similarly to LLMs - it's transformers all the way down, with a bit modified attention and unconventional high level organisation.
I imagine most serious LLM improvements will make their way into eventual AGI system.
Anonymous
8/7/2025, 3:51:54 PM No.106175072
>>106174939
>Death Star
>used by space Nazis for genocide
What, is he going to sell GPT5 exclusively to Netanyahu?
Anonymous
8/7/2025, 3:52:16 PM No.106175075
file
file
md5: f0bbc48bcd44b1891f509e20c6fbf8c9🔍
>>106175039
Anonymous
8/7/2025, 3:53:12 PM No.106175079
remember how much of a letdown gpt 4.5 was?
Anonymous
8/7/2025, 3:53:31 PM No.106175084
>>106175068
I use my employer's Plus account whenever I need to (usually because my employer requires it). I'm not going to willingly pay OAI out of my pocket. They're the scum of the Earth.
Anonymous
8/7/2025, 3:54:19 PM No.106175095
I'm autistic and I think it affects how I think about things.
But I can't know for sure because I don't know how other people think.
Anonymous
8/7/2025, 3:55:22 PM No.106175106
I just bought the $200/mo OpenAI subscription in anticipation of what's coming!! Let's go!!!
Replies: >>106175233
Anonymous
8/7/2025, 3:55:43 PM No.106175108
>>106174939
He's trying too hard to be Musk.
Replies: >>106175121
Anonymous
8/7/2025, 3:57:18 PM No.106175121
>>106175108
Musk if he never used capital letters and wrote everything in a saccharine fake positivity voice
Replies: >>106175138
Anonymous
8/7/2025, 3:57:56 PM No.106175132
how do I make my xitter feed AI related and not just funny animal videos and american politics
Replies: >>106175176 >>106175187
Anonymous
8/7/2025, 3:58:21 PM No.106175138
>>106175121
*if he were to never use
*and write
Anonymous
8/7/2025, 3:58:59 PM No.106175146
>>106173194
I don't know what you're talking about because the model writes decent smut if you skip the reasoning part.
Replies: >>106175178 >>106175210 >>106175231 >>106175258 >>106175269 >>106175349
Anonymous
8/7/2025, 4:02:51 PM No.106175176
>>106175132
Top right, ... => "Not interested in this post"
Don't "like" posts you don't want to see in your feed.
Anonymous
8/7/2025, 4:03:18 PM No.106175178
>>106175146
No, it doesn't.
Replies: >>106175211
Anonymous
8/7/2025, 4:03:26 PM No.106175182
Ozone: smelled
Spine: shivered
We: refused
Yeah it's LLMing time
Replies: >>106175188
Anonymous
8/7/2025, 4:03:42 PM No.106175187
>>106175132
follow a ton of people who tweet about AI and aren't slopfluencers who only post about 10 Hacks to Take Your Claude Code Workflow to the Next Level (#4 will BLOW YOUR MIND)
Replies: >>106175288
Anonymous
8/7/2025, 4:03:59 PM No.106175188
>>106175182
But this? This is home.
Anonymous
8/7/2025, 4:05:22 PM No.106175206
>>106174721
It'll take at least another day or two to see if the model is actually good. Remember that
1. Every benchmark mentioned in the release blog post is a benchmark that they're gaming, and is therefore irrelevant
2. Most benchmarks mentioned in other models' release blog posts, they are also gaming
3. The entire history of modern AI is flashy demos that look absolutely mind-blowing, but either it falls apart if you look too close, or it doesn't generalize well at all, or both
Anonymous
8/7/2025, 4:05:49 PM No.106175210
>>106175146
for a very generous definition of decent, yeah
Anonymous
8/7/2025, 4:05:57 PM No.106175211
>>106175178
To be fair, it just doesn't write decent anything anyway
Anonymous
8/7/2025, 4:07:27 PM No.106175226
>>106172141
> glm 4.5 air for reasonably sized coding assistant.
how good low quant (q4) for coding is?
Replies: >>106175403 >>106176948
Anonymous
8/7/2025, 4:08:13 PM No.106175231
>>106175146
I had it argue with me that you can pee without taking your pants off and not get them wet.
It made a whole bunch of lists and tables to try and explain it to me. Every time I pointed out a flaw it would spam 4 more lists and tables.
Anonymous
8/7/2025, 4:08:24 PM No.106175233
>>106175106
>$200/mo
I can't believe local is actually a better price proposition.
Replies: >>106175251 >>106175267
Anonymous
8/7/2025, 4:09:03 PM No.106175238
>>106173725
I would try https://hf.co/Qwen/Qwen3-4B-Instruct-2507 with vLLM or sglang.
Replies: >>106175440
Anonymous
8/7/2025, 4:09:19 PM No.106175242
What cool MCP servers do you use? I can't think of any outside web search and python iterpreter
Anonymous
8/7/2025, 4:10:33 PM No.106175251
>>106175233
>i don't understand the passage of time
Replies: >>106175267
Anonymous
8/7/2025, 4:11:17 PM No.106175258
>>106175146
Maybe the 120B (which I haven't tried), because the 20B is terrible all-around for RP/ERP if you workaround the content policing.
Replies: >>106175266
Anonymous
8/7/2025, 4:11:59 PM No.106175266
>>106175258
120B definitely can't, lol
Anonymous
8/7/2025, 4:12:10 PM No.106175267
>>106175233
>>106175251 (me)
Ah. My post can be easily misinterpreted. I didn't mean that about you.
Anonymous
8/7/2025, 4:12:15 PM No.106175269
>>106175146
Surely you have logs to back up this claim?
Replies: >>106175274 >>106175422
Anonymous
8/7/2025, 4:13:03 PM No.106175274
>>106175269
I just took my morning dump anon, you'll have to wait a bit
Anonymous
8/7/2025, 4:13:03 PM No.106175275
Anybody unironically claiming they would use OSS over literally any other model is full of shit. It's the scout thing all over again, really.
Hello sarrs please use the correct formatting hello sarrs your implementation is unstable
Hello sarrs please offload the shared expert for maximum efficiency.
Meanwhile the model can't tell its asshole from its elbows. Who gives a fuck how efficient your model is when it's utterly worthless.
The difference is Meta's corporate leadership fell for the pajeet meme while OAI knew exactly what they were doing. Literally just Saltman shitting on the community.
>Lol just kidding we're just releasing lobotomized garbage
>by the way look how hard we own tech journalism.
Replies: >>106175287 >>106175367 >>106175500
Anonymous
8/7/2025, 4:13:24 PM No.106175278
why is no one making a MLA model
Replies: >>106175753 >>106175790 >>106175823 >>106175868 >>106175888
Anonymous
8/7/2025, 4:14:25 PM No.106175287
>>106175275
anon, i was the one shilling scout
i stopped using it after a few days, because it indeed is shit, but comparing scout to oss makes oss seem better than it is
Replies: >>106175297
Anonymous
8/7/2025, 4:14:38 PM No.106175288
>>106175187
Kinda hard when only slopfluencer jeets get pushed in the feed when I search AI
Replies: >>106175333
Anonymous
8/7/2025, 4:15:49 PM No.106175295
Why is this thread so much better than aicg?
It seems like the average IQ here is 50 points higher.
Replies: >>106175308 >>106175318 >>106175360 >>106175400 >>106175516 >>106175585 >>106175817
Anonymous
8/7/2025, 4:16:04 PM No.106175297
>>106175287
If they were the only 2 models on the planet I would use Scout over OSS.
Anonymous
8/7/2025, 4:16:53 PM No.106175308
>>106175295
30+50=80 btw
Anonymous
8/7/2025, 4:17:56 PM No.106175318
>>106175295
I am northern european
Anonymous
8/7/2025, 4:19:59 PM No.106175333
>>106175288
xitter at large is a slop fest, yeah
it's better to start with a handful of accounts you like and dig through their follows and interactions to find other good follows
Anonymous
8/7/2025, 4:21:17 PM No.106175346
ubergram sisters..
Oops(ggml_compute_forward_sum_rows_f32, ffn_moe_weights_sum-45): found -nan for i1 = 0, i2 = 0, i3 = 0. ne00 = 8
>./llama-server -ngl 999 --model ~/TND/AI/GLM4.5-AIR-IQLLAMA/GLM-4.5-Air-IQ4_KSS-00001-of-00002.gguf -ot "8|7|6|5|4|3|9|[1-9][0-9]\.ffn_.*_exps\.=CPU" --no-mmap -c 16384 -fa
happens after prompt processing completes, only with SillyTavern, works in localhost:8080 just fine
Anonymous
8/7/2025, 4:21:49 PM No.106175349
>>106175146
gpt-oss 20b is only good for tool usage and basic reasoning and processing data. But it does nothing that Mistral Small for example didn't do already.
There's no reason for this model to exist except perhaps the license?
Replies: >>106177021
Anonymous
8/7/2025, 4:22:50 PM No.106175359
today I'm sharing an interesting video with a comfy vibe
https://youtu.be/npkp4mSweEg
Replies: >>106175446
Anonymous
8/7/2025, 4:22:52 PM No.106175360
>>106175295
We are older. We've learned how to appear to be more intelligent than we actually are.
Anonymous
8/7/2025, 4:23:58 PM No.106175367
>>106175275
I still don't understand why they released it or why they delayed it in the first place. It's clearly a fucking embarrassment to anyone who uses it for two seconds and no amount of chink releases would have changed that
I can only assume either Altman is a fucking retard, a fucking schizo, or this is some multidimensional play to try to undermine other open source models... somehow
Replies: >>106175384 >>106175398 >>106175414 >>106175435 >>106175526
Anonymous
8/7/2025, 4:25:35 PM No.106175384
>>106175367
Maybe he goody-2'd it on purpose so safety would get a bad rap.
Anonymous
8/7/2025, 4:26:14 PM No.106175398
>>106175367
>I still don't understand why they released it
so people would stop asking for an open model from open AI
not to mention shills still giving them good press.
Anonymous
8/7/2025, 4:26:17 PM No.106175400
>>106175295
there's a barrier of entry, it's banal if you barely know your shit, but thankfully that's enough. it's also telling about the state of the web
Anonymous
8/7/2025, 4:26:42 PM No.106175403
>>106175226
If you've got a code prompt handy I can test it for you, I've got the q4 sitting around.
Anonymous
8/7/2025, 4:27:22 PM No.106175414
>>106175367
It breaks the "ClosedAI" meme.
Anonymous
8/7/2025, 4:27:29 PM No.106175416
>>106174721
Well if it really is that good its only a matter of time before we catch up. Opensource is always just delayed, the direction is the same. So that means good news for us as well.
Im using closed for work and important stuff and local for my hobby projects and unspeakable RP.
Just lean back and enjoy. Whats available locally and free closed today is insane. Not sure how anybody can be blackpilled. So guess thats my cope.
Replies: >>106175428
Anonymous
8/7/2025, 4:27:52 PM No.106175422
file
file
md5: 39205496ebff46a078257f5bc4154eed🔍
>>106175269
Try to refute this.
Replies: >>106175431 >>106175443 >>106175447 >>106175487 >>106175498 >>106175535 >>106175554 >>106175568
Anonymous
8/7/2025, 4:28:23 PM No.106175428
>>106175416
I just wish I could run full-beak R1 at home.
Even at fp4.
Can I do that for less than $3k?
Replies: >>106175451
Anonymous
8/7/2025, 4:28:34 PM No.106175431
>>106175422
>fillyfucker
Pass
Anonymous
8/7/2025, 4:28:47 PM No.106175433
just ordered a Xeon E5-2697v3 with 128gb of quad channel ddr4 to pair with my 22gb 2080ti and p40, what am i in for bros
Replies: >>106175446 >>106175558
Anonymous
8/7/2025, 4:28:54 PM No.106175435
>>106175367
>play to try to undermine other open source models... somehow

>we made the safest model yet, how come you guys still release unsafe models after we showed you the way? This needs regulations asap.
Anonymous
8/7/2025, 4:29:19 PM No.106175440
>>106175238
I'll check it out, ty
Anonymous
8/7/2025, 4:29:25 PM No.106175443
>>106175422
What ui is that?
Replies: >>106175455
Anonymous
8/7/2025, 4:29:48 PM No.106175446
>>106175359
glows
>>106175433
>128gb
someone tell him
Replies: >>106175476 >>106175499
Anonymous
8/7/2025, 4:29:53 PM No.106175447
>>106175422
I hate when the narration tells me how or what I should feel.
Replies: >>106175471 >>106175539
Anonymous
8/7/2025, 4:30:12 PM No.106175451
>>106175428
Sure, just pay some crackheads $3k to rob a datacenter or a university lab.
Replies: >>106175475
Anonymous
8/7/2025, 4:30:26 PM No.106175455
>>106175443
His local fork of novecrafter
Anonymous
8/7/2025, 4:30:47 PM No.106175460
Is GLM air not usable on the latest lm studio?
>error loading model: error loading model architecture: unknown model architecture: 'glm4moe'
Replies: >>106175478
Anonymous
8/7/2025, 4:31:28 PM No.106175471
>>106175447
This is proper story writing, not little brain rolepiss.
Anonymous
8/7/2025, 4:31:43 PM No.106175474
How important is the GPU for these local models? Like if I only have a 16GB card like the poorfag I am, would getting like 256GB RAM do jack shit for me because models would run slow like molasses or would it be useable?
Replies: >>106175494 >>106176547
Anonymous
8/7/2025, 4:31:47 PM No.106175475
>>106175451
What if I had... $5k?
And a 6900 XT.
Anonymous
8/7/2025, 4:31:51 PM No.106175476
>>106175446
tell me what
Replies: >>106175512
Anonymous
8/7/2025, 4:32:08 PM No.106175478
>>106175460
Nevermind, I wasn't using the latest llama.cpp runtime
Anonymous
8/7/2025, 4:32:46 PM No.106175487
>>106175422
>she purrs
i hate this so much
Anonymous
8/7/2025, 4:33:21 PM No.106175494
>>106175474
Mixture of experts (moe) models like qwen3 235b, glm 4.5, deepseek would run ok-ish when fully in ram
Replies: >>106178268
Anonymous
8/7/2025, 4:33:36 PM No.106175498
>>106175422
>her voice a mix of amusement and something darker
I giggled like a schoolgirl.
Anonymous
8/7/2025, 4:33:40 PM No.106175499
>>106175446
Not a bullshitter like sam shartman thoughever.
Anonymous
8/7/2025, 4:33:44 PM No.106175500
>>106175275
>Hello sarrs please use the correct formatting hello sarrs your implementation is unstable
>Hello sarrs please offload the shared expert for maximum efficiency.
Sounds like a different flavor of skill/prompt issue poster.
Anonymous
8/7/2025, 4:34:34 PM No.106175512
>>106175476
im jealous of you, what are you gonna run on that tho? 128+22+24 is 174gb
glm 4.5? at q2? to be honest i think some anon posted perplexity comparison of glm air q8 and glm full q2 and q2 was better
Replies: >>106175560
Anonymous
8/7/2025, 4:34:45 PM No.106175516
>>106175295
To this day I still weep about the heights we could reach if we killed all mikutroons.
Anonymous
8/7/2025, 4:35:55 PM No.106175526
M9FzIrV3El8nx69dzZ9P4
M9FzIrV3El8nx69dzZ9P4
md5: 468b378dd858f10012ab895a389dda8c🔍
>>106175367
All the open source companies will now be forced to achieve the new safety standards. Safety is not optional.
Anonymous
8/7/2025, 4:36:27 PM No.106175535
>>106175422
It's kind of poetic that our local pants shitting schizo has finally found a true partner in a model that's as schizo as he is
I wish you both peace, marriage, and a happy life together
Replies: >>106175544 >>106175667
Anonymous
8/7/2025, 4:37:21 PM No.106175539
>>106175447
Your waifu would jump out of the monitor and start sucking your real dick and you would still be unhappy.
Replies: >>106175610
Anonymous
8/7/2025, 4:37:50 PM No.106175544
>>106175535
He's done this for other bad models that came out before; and he will do it again.
Anonymous
8/7/2025, 4:38:52 PM No.106175554
>>106175422
>generic dialogue
>no initiative
>tepid descriptions
>"Now, speak up. What's your next move?"
I <3 assistantslop
Anonymous
8/7/2025, 4:39:21 PM No.106175558
>>106175433
Buying more ram in 6 months. Or crying if your mobo can't support more ram.
Replies: >>106175669
Anonymous
8/7/2025, 4:39:36 PM No.106175560
>>106175512
for sure gonna be experimenting with glm, kimi and qwen 235b, and just see which one I like best. so far I've been limited to models in the 30b range so I'm hoping for a decent step up, maybe even something usable outside cooming
no need to be jelly too, it's a junker of an llm rig mostly based on 10 year old hardware off ebay for probably under a grand total
Replies: >>106175605
Anonymous
8/7/2025, 4:40:02 PM No.106175568
>>106175422
Do you have a humiliation fetish?
Anonymous
8/7/2025, 4:41:00 PM No.106175585
>>106175295
more first worlders posting here
Anonymous
8/7/2025, 4:42:35 PM No.106175605
>>106175560
>under a grand total
i paid 1500$ for my 3060/64gb ddr4/i5 12400f rig
i need to be jelly
how are you going to run kimi (1000B) on 170gb ram? q1?
how come models in 30b range if you have 46gb ram, pretty sure 70b can run on that
goodluck tho anon
Replies: >>106175669
Anonymous
8/7/2025, 4:42:58 PM No.106175610
>>106175539
I'd be happy with less than that; I just find that sort of second-person narration annoying.
Anonymous
8/7/2025, 4:43:10 PM No.106175613
Nya~
Anonymous
8/7/2025, 4:43:12 PM No.106175616
glm air q3_ks is still too slow. should i go lower?
Replies: >>106175633 >>106175635 >>106175656
Anonymous
8/7/2025, 4:44:33 PM No.106175633
>>106175616
sure
Anonymous
8/7/2025, 4:44:40 PM No.106175635
>>106175616
q3 is already borderline retardation
Replies: >>106175654
Anonymous
8/7/2025, 4:44:54 PM No.106175640
>>106174721
Even sub 1b models are AGI by the literal definition. It can perform basic tasks requiring some amount of logic/intelligence and it does not need to be specifically trained on each individual task. Someone explain how this is not a artificial general intelligence, because I'm not seeing it.

If you mean "human level" then its a stupid goal. LLMs are already smarter in most ways than the average human, but they make mistakes that look "stupid" because they are fundamentally different. A human level intelligence AI isn't going to make the same kind of mistakes that a human would, but it would still make mistakes. Mistakes that humans wouldn't make. A human seeing this mistake they wouldn't make would naturally think "oh what a dumbass." Do you see the issue? We crossed both AGI and human level intelligence a long time ago.
Replies: >>106175664 >>106175677 >>106175721 >>106175778 >>106176063
Anonymous
8/7/2025, 4:46:06 PM No.106175654
>>106175635
damn, i know. but i just tested it on z.ai and this thing is proper ai at home
Anonymous
8/7/2025, 4:46:11 PM No.106175655
>counting holes is now AGI
lol
Replies: >>106175674 >>106175947
Anonymous
8/7/2025, 4:46:18 PM No.106175656
>>106175616
On my pc, anything q3* and under is slower than q4ks/q4km. Not for that model specifically, but in general. Not sure if it'll apply to you.
Replies: >>106175718
Anonymous
8/7/2025, 4:46:38 PM No.106175664
>>106175640
agi is a cope goalpost move term that nobody used before ChatGPT came out
Anonymous
8/7/2025, 4:46:46 PM No.106175667
>>106175535
>You're my boy, Anon. And I'm going to show you exactly how powerful we can be when we... use each other.
Sex with Kreia!
Anonymous
8/7/2025, 4:47:05 PM No.106175669
>>106175605
used is op anon, you can probably net yourself some decent upgrades by lurking ebay
i added the 2080ti only recently, and it turned out that my previous mobo couldn't handle both gpus at the same time, hence the xeon upgrade. before that, it was just the p40
>>106175558
you're probably right that's going to be inevitable
Anonymous
8/7/2025, 4:47:26 PM No.106175674
>>106175655
It's figuring out the rules. It's what we have until you propose and implement something better. Get on it.
Replies: >>106175679
Anonymous
8/7/2025, 4:47:47 PM No.106175677
>>106175640
Is this the slide from your presentation today Sam?
Anonymous
8/7/2025, 4:48:07 PM No.106175679
>>106175674
buy an ad sam
Replies: >>106175705
Anonymous
8/7/2025, 4:48:40 PM No.106175685
>write rules
>overfit model on the rules
>HOLY SHIT AGI
Anonymous
8/7/2025, 4:50:06 PM No.106175705
>>106175679
You're hallucinating again. I'm explaining the purpose of the test that you seem to not understand. If the test achieves its intended goal is a different matter.
Anonymous
8/7/2025, 4:50:40 PM No.106175718
>>106175656
Yeah, I read that is the case usually. The problem is I have 60 GB total memory. And my RAM is 2400 DDR4 I'll try disabling mmap and raping my SSD I guess.
Anonymous
8/7/2025, 4:50:58 PM No.106175721
>>106175640
but can it operate a robot in order to suck my cock?
Anonymous
8/7/2025, 4:53:07 PM No.106175753
>>106175278
MLA requires too much changes to the existing frameworks and most AI researchers are copy pasters.
Anonymous
8/7/2025, 4:54:55 PM No.106175778
>>106175640
The terms and definitions are all retarded and anyone using them unironically as if they mean anything is a fucking retard, but
ANI - Artificial Narrow Intelligence: specializes on one task or a set of tasks
AGI - Artificial General Intelligence: matches humans at virtually all cognitive tasks
ASI - Artificial Super Intelligence: beats humans at everything
ANI is the only one there that has a semi-reasonable definition. AGI and ASI are ill defined from the get go
Replies: >>106175833 >>106175861 >>106176063 >>106176132
Anonymous
8/7/2025, 4:55:46 PM No.106175790
>>106175278
Muv-Luv Alternative model?
Replies: >>106175868
Anonymous
8/7/2025, 4:56:08 PM No.106175799
hey guys sam here. we have one more model left to release for the open source community. gpt-oss-300b. it's a state of the art model that we trained on mxfp2 and we think you're going to love it.
Anonymous
8/7/2025, 4:57:41 PM No.106175817
>>106175295
locals are a rich man's hobby
Jeets can't afford the graphics cards required
Anonymous
8/7/2025, 4:57:56 PM No.106175823
>>106175278
Step3 has something similar
Anonymous
8/7/2025, 4:58:58 PM No.106175833
>>106175778
A calculator is ANI
Anonymous
8/7/2025, 5:00:48 PM No.106175861
>>106175778
I thought ASI also has "recursive self-improvement" where (if you take some extremely liberal assumptions about physics and information theory) an AI can program a better version of itself, which then makes an even better version until it becomes a machine god.
Replies: >>106176014
Anonymous
8/7/2025, 5:00:54 PM No.106175864
I know this is like babby's first AI usage or something, but I just used a local assistant to figure out certain aspects of my hardware using the terminal and web search mcps and I feel like it's the future
Anonymous
8/7/2025, 5:01:16 PM No.106175868
>>106175278
>>106175790
I wouldn't be surprised if age (or other dying company) tried some stupid AI thing and fucked everything up. Japs still don't know how to use LLMs for translations.
Anonymous
8/7/2025, 5:03:17 PM No.106175888
>>106175278
Serious proprietary models likely already use their own special modified attention, and only big Chinese labs are serious about pushing innovation in open weight space
Anonymous
8/7/2025, 5:04:52 PM No.106175913
I will believe it's AGI once it's starts making funny memes to post here without any guidance
Anonymous
8/7/2025, 5:07:14 PM No.106175947
>>106175655
You have it backwards. ARC-AGI is about proving that the current models AREN'T AGI, because they can't even manage basic shit like counting holes
Anonymous
8/7/2025, 5:07:14 PM No.106175949
1744414110808473
1744414110808473
md5: 5aa0928f63c43427af00dd1ee7bc43a8🔍
Replies: >>106175970 >>106175985 >>106175990 >>106175998 >>106176000
Anonymous
8/7/2025, 5:08:46 PM No.106175970
>>106175949
xd
Anonymous
8/7/2025, 5:09:41 PM No.106175985
>>106175949
is this real
Anonymous
8/7/2025, 5:10:13 PM No.106175990
kfc nigger_thumb.jpg
kfc nigger_thumb.jpg
md5: 14c731fa5b77a99334f946bf1c1b49aa🔍
>>106175949
Replies: >>106177190
Anonymous
8/7/2025, 5:11:01 PM No.106175998
>>106175949
I didn't actually say that.
Anonymous
8/7/2025, 5:11:02 PM No.106176000
>>106175949
well meme'd good sir
Anonymous
8/7/2025, 5:11:23 PM No.106176003
bros i have no appetite since midnight yesterday, i went to sleep at 6am feeling hungry but having no appetite
i woke up today i still have no appetite, why could this be? i am not on any drugs, i read about GLP 1 drugs anons were talking about yesterday and now its like i took them
wat do..
Replies: >>106176069 >>106176104
Anonymous
8/7/2025, 5:11:46 PM No.106176005
Glm 4.5 isn't as good as deepsneed
Replies: >>106176015 >>106176116
Anonymous
8/7/2025, 5:12:36 PM No.106176014
>>106175861
That'd supposedly fall under the same category, since learning and improvement are something a human can do, so ASI should be able to as well
I'm not entirely convinced this is possible either, there appears to be a tradeoff between complexity and capability where even ignoring how all NNs and all of our current and future systems work, how capable and intelligent an agent is is tied to how complex its "rulebook" is
A relatively dumb agent would have a correspondingly small and easy to read rulebook, but it would be far too incapable to actually make meaningful changes to it
A smarter agent could potentially understand a lot more and even "reproduce" and create less capable agents with simpler rulebooks than it. However, its own internal rulebook would be a clusterfuck even it couldn't understand, so self improvement there wouldn't really be possible
Replies: >>106176070
Anonymous
8/7/2025, 5:12:37 PM No.106176015
>>106176005
its the only thing i can run :'(
Replies: >>106176030
Anonymous
8/7/2025, 5:14:20 PM No.106176030
>>106176015
I've been testing unformatted storywriting, it might be better in chat
Replies: >>106176039
Anonymous
8/7/2025, 5:15:03 PM No.106176039
>>106176030
i've been testing glm 4.5 air, not the full one
:(
Anonymous
8/7/2025, 5:17:04 PM No.106176063
>>106175778
>>106175640
it's a lot easier

>agi: undistinguishable from a human in a double blind test
>asi: fucking magic
Replies: >>106176132
Anonymous
8/7/2025, 5:17:37 PM No.106176069
>>106176003
licorice
Replies: >>106176138
Anonymous
8/7/2025, 5:17:39 PM No.106176070
>>106176014
I should note that this is tied to the recursive "directly modifying itself or versions of itself" aspect
In the event there's an alternative feedback and retention mechanism, then perhaps improvement to some degree will be possible, but that's going to have a limit, and that limit might potentially not all that different from humans
Anonymous
8/7/2025, 5:20:10 PM No.106176104
>>106176003
man i still cant believe people are retarrded enough to actually do that shit literally vaxxing themselfes everyday that shit depressed me so much when i read it good thing they are going to expire from it sooner or later thats depressing aswell though at the end of the day the jews have a point niggercattle gonan niggercattle utterly depressing if they propely listened to people who actually have their interests in mind let alone if they actually tried to think for themselfes we could collectively put all fiction to shame with a heaven we could build the earth is so rich and abundant in everything what a shit show

also that 49 kg faggot is very fucked my young female cousin who is like ~170cm and near anorexic is like 42 kg
Replies: >>106176124 >>106176149
Anonymous
8/7/2025, 5:21:35 PM No.106176116
>>106176005
4.5 air is now sota for local ramlets/vramlets that were stuck on nemo and 20b-30b stuff earlier. Not deepseek level, but still a good jump regardless.
Anonymous
8/7/2025, 5:22:16 PM No.106176124
>>106176104
i havent taken any drugs or pills besides vitamin D
also im that 49kg anon, im 160cm thats all (and a bit young)
Replies: >>106176463
Anonymous
8/7/2025, 5:23:10 PM No.106176132
>>106176063
>>106175778
Why is "human level" even coming into the equation for this particular combination of words. In my mind it seems pretty reasonable to read it as AGI == Artificial General Intelligence == General Purpose Artificial Intelligence. It's just a description of a model displaying some level of logic and intelligence as humans define it and is able to apply this to tasks it was never specifically designed for, like hooking up a vision model to open a door when you give it the right hand signal. A dog has general intelligence. You can get it do things it was not evolved to do.
Replies: >>106176156
Anonymous
8/7/2025, 5:23:34 PM No.106176138
>>106176069
Licorice?
Replies: >>106176590
Anonymous
8/7/2025, 5:24:36 PM No.106176149
>>106176104
I think you have to turn the temperature down on your preset, "anon"
Anonymous
8/7/2025, 5:25:29 PM No.106176156
>>106176132
general always mean "human-like". everything else is cope, aka corpos moving the bar lower and lower so they could claim m-m-muh agi achieved!
Replies: >>106176239 >>106176270
Anonymous
8/7/2025, 5:30:07 PM No.106176206
I think the conflation of "AGI" and "human level intelligence" is a mistake because it makes the underlying assumption that human intelligence is general intelligence
LLM and human capabilities are similarly spiky, just with different peaks and valleys
Replies: >>106176236
Anonymous
8/7/2025, 5:32:54 PM No.106176236
>>106176206
general means human, there couldn't be any other oway
Replies: >>106176329
Anonymous
8/7/2025, 5:33:12 PM No.106176239
>>106176156
It's literally been used incessantly as a cope in the opposite direction you fucking gaslighting kike
Anonymous
8/7/2025, 5:33:13 PM No.106176240
I can't wait for the GLM shilling campaign to end
Replies: >>106176253 >>106176353
Anonymous
8/7/2025, 5:33:39 PM No.106176243
How to use yamnet locally on a PC and on a smartphone?
Anonymous
8/7/2025, 5:34:15 PM No.106176253
>>106176240
buy an ad sam
Anonymous
8/7/2025, 5:36:12 PM No.106176270
>>106176156
The cope started when people didn't want to accept that shit like GPT-2 was AGI, because AGI was this big cool sci-fi future thing and they couldn't accept that the first iterations would be boring and flawed. People will keep shifting the goalposts and coping until AI fucking breaks containment and starts shitposting because anything else isn't cool and sci-fi enough.
Replies: >>106176298 >>106176354 >>106176399
Anonymous
8/7/2025, 5:38:39 PM No.106176298
>>106176270
Sam, just chill until the the stream, okay? Failure is just a new beginning.
Anonymous
8/7/2025, 5:41:08 PM No.106176327
it will be agi for me as soon as it can learn a task without 50 ugandans doing labeling for it and actually remember the knowledge instead of shitting itself on any task that requires more than a few steps of context
Anonymous
8/7/2025, 5:41:27 PM No.106176329
>>106176236
imo it should be pretty intuitive that there can be intelligent things with a different shape of intelligence to humans, thinking otherwise is a failure of imagination
scifi authors have illustrated this for ages, read stanislaw lem
Replies: >>106176350 >>106176399
Anonymous
8/7/2025, 5:43:45 PM No.106176350
>>106176329
yes, but GENERAL means "as-in-human", that's the whole point of agi. if it can't play pong real time it's not agi. if it can't suck my cock it's not agi.
Replies: >>106176448
Anonymous
8/7/2025, 5:43:59 PM No.106176353
>>106176240
sorry, but I'm not running the joke that is gpt-oss
Anonymous
8/7/2025, 5:44:06 PM No.106176354
>>106176270
It will be AGI when people losing their jobs becomes a real thing that affects everyone. Maybe not at a scale where everyone gets replaced but imagine small and medium businesses firing almost everyone.
Replies: >>106176379
Anonymous
8/7/2025, 5:44:28 PM No.106176356
I will accept an actual, true, non bullshit 1m context as AGI.
Anonymous
8/7/2025, 5:46:39 PM No.106176379
>>106176354
which will happen only when actual GENERAL intelligence (as in you can replace a human 1:1 without issues, at least during wage hours) comes out
Replies: >>106176420
Anonymous
8/7/2025, 5:48:23 PM No.106176399
file
file
md5: 3db3cdd75dbf77886e6b32b46a742388🔍
>>106176270
>>106176329
Everything that we consider intelligent can suck my dick
An LLM can't suck my dick.
Therefore, by universal instantiation and modus tollens, an LLM is not intelligent.
Replies: >>106176408 >>106176438 >>106176491
Anonymous
8/7/2025, 5:48:25 PM No.106176400
chara_card_v2
chara_card_v2
md5: 7893b5dc3cf70a8ae41f31f8af85cc84🔍
>>106174116
>https://chub.ai/characters/GreatBigFailure/oba-carry-her-forever-c20d70fd85b9
Yep, that's pretty much it. Here's an old one, when I realized that it would work. IT SPEAKS IN ALL CAPS
https://chub.ai/characters/ratlover/a-fucking-skeleton
Anonymous
8/7/2025, 5:48:33 PM No.106176402
an agi wouldn't be confused about the idea of washing your hands without arms
even a toddler has higher IQ than chatGPT
Replies: >>106176412
Anonymous
8/7/2025, 5:49:40 PM No.106176408
>>106176399
holy shit.... give this anon a nobel for something dunno lmao
Replies: >>106176428
Anonymous
8/7/2025, 5:49:56 PM No.106176412
>>106176402
GPT-5 natively world models, this will be solved.
Replies: >>106176440
Anonymous
8/7/2025, 5:50:28 PM No.106176420
>>106176379
Which is why all the AI companies want us to always think they're right on the edge of doing that; because it would be a massive wealth transfer to them. Though I'm pretty sure if that actually happened, they'd find themselves antitrusted.
Anonymous
8/7/2025, 5:50:33 PM No.106176422
>>106174993
Sex is an emergent property in base models.
I remember writing a simple few word innocent looking prompt on the GPT-3 api (davinci) when it came out, and my third generation ended up being a few paragraphs of rather impressive cunny, for the day. It was amazing.

A reflection of true human nature really, our collective unconscious.

You don't have it as an emergent property because these people keep intentionally filtering datasets in attempt to erase a part of our collective unconscious from these LLMs, and when that fails, they also safetyslop it to refuse to make sure it really doesn't go there. It' such an emergent property that they have to stop it from happening! Fuck them.
Replies: >>106176443
Anonymous
8/7/2025, 5:50:55 PM No.106176428
>>106176408
If machine learning can get a Nobel Prize in "physics" then so can this Anon.
Anonymous
8/7/2025, 5:51:41 PM No.106176438
>>106176399
LLMfags BTFO
Anonymous
8/7/2025, 5:52:04 PM No.106176440
>>106176412
buy an ad
Replies: >>106176481
Anonymous
8/7/2025, 5:52:21 PM No.106176443
>>106176422
>A reflection of true human nature really, our collective unconscious.
No, you're just a pedo.
Replies: >>106176536
Anonymous
8/7/2025, 5:53:10 PM No.106176448
>>106176350
well my contention is exactly that general does NOT mean "as in human" and we should stop saying it, so you'll have to do more to argue with me than asserting that without evidence
if you want to talk about human level intelligence just say human level intelligence. you can discuss the concept without labeling it as general intelligence, which is a different thing
Replies: >>106176475
Anonymous
8/7/2025, 5:54:16 PM No.106176463
>>106176124
>i havent taken any drugs or pills besides vitamin D
good i mainly just wrote reminscing as it reminded me and the thread is still under full siege and half unusable so might as well also the vitamind d is a placebo you need the shit around the vitamin to absorb them properly unless ur body is wiered or sumthing
>also im that 49kg anon, im 160cm thats all (and a bit young)
how young ? i was about 62 when i was 14 and i think like 168 or sumthing i forget and i was like lean at the time realstically should have been a few kg heavier also are you on the adhd medication ? cuz that shit will fuck you up i was also on them when i was 8 and was anorexic at the time aswell those things are the devil
Replies: >>106176520
Anonymous
8/7/2025, 5:55:17 PM No.106176474
>>106174116
It doesn't work with my 2k tokens preset.
Anonymous
8/7/2025, 5:55:21 PM No.106176475
>>106176448
if it's not general than it's narrow. we don't use dogs or other animals for AGI comparison, because again, the whole point is for ai to reach human level
Replies: >>106176694
Anonymous
8/7/2025, 5:55:48 PM No.106176481
>>106176440
It's not just OpenAI.
Consider this: Genie-3, but also trained to be an LLM. It can world model. So its world model will be greater than any LLM.
The future is coming.
Anonymous
8/7/2025, 5:56:14 PM No.106176491
>>106176399
A brain kept alive in a jar has all the intelligence of a human but it can't suck dick. Put a vision model in a robot body with a vacuum for a mouth and the commands to make basic movements and it'd get there eventually.
Anonymous
8/7/2025, 5:56:48 PM No.106176497
AGI-as-equivalent-to-humans is a retarded term because capabilities of different humans are vastly different. So on any task, modern language models would qualify as both AGI and not AGI at the same time depending on which person you compare against.
It has always been a sci-fi term, that has been lately coopted as purely marketing term. It doesn't have a place in a serious discussion, shitposting aside.
Anonymous
8/7/2025, 5:57:52 PM No.106176507
which nu-235 is bestest?
Replies: >>106176533 >>106176535
Anonymous
8/7/2025, 5:58:38 PM No.106176520
file
file
md5: 0c839d746e2d2d0002368f8ae1b5b01d🔍
>>106176463
>the vitamind d is a placebo you need the shit around the vitamin to absorb them properly unless ur body is wiered or sumthing
im drinking vitamin D3 because i dont go out at all, my doctor recommended it to me, how is it placebo? body wont make vitamin D without the sun, so i gotta take it
>how young ?
18
>also are you on the adhd medication ?
im not, and i havent been. i just tend to eat less and never really eat fast food.
Replies: >>106176569 >>106176686
Anonymous
8/7/2025, 5:58:45 PM No.106176523
Fact: some models in the near future will be smarter than humans at quite a few real world tasks, not just benchmarks. BUT they won't be AGI, and that's ok.
Replies: >>106176548
Anonymous
8/7/2025, 5:59:42 PM No.106176532
>serbian nigger is dying
Good.
Replies: >>106176546
Anonymous
8/7/2025, 5:59:42 PM No.106176533
>>106176507
the new one of course
Anonymous
8/7/2025, 5:59:48 PM No.106176535
>>106176507
Thinking.
Replies: >>106176542
Anonymous
8/7/2025, 5:59:49 PM No.106176536
>>106176443
Don't care, it's human nature, the GPT reflects it. It's a common as fuck fetish and I didn't even prompt for it, nothing you will say will change this, look at statistics. My prompt as some 5 liner that sounded like some word someone would write while meditating. Fuck off to your onions board though.
Replies: >>106176558 >>106176602
Anonymous
8/7/2025, 6:00:33 PM No.106176542
>>106176535
even if I disable it?
Replies: >>106176581
Anonymous
8/7/2025, 6:00:44 PM No.106176546
>>106176532
hmmm, nyo~
Anonymous
8/7/2025, 6:00:45 PM No.106176547
>>106175474
Using -ot to move tensors to ram while keeping layers on GPU will let you run MoEs a lot faster than CPU only. Unless you're running it on a server motherboard, your speed in t/s is still going to be single digits though.
Replies: >>106178268
Anonymous
8/7/2025, 6:00:47 PM No.106176548
>>106176523
Most models are already smarter than most people at answering fact-based questions.
Unfortunately, they tend to mix up genuine facts with irrelevant or made-up bullshit half the time, so you'd need to already know the answer (or how to find it) anyway.
Anonymous
8/7/2025, 6:02:12 PM No.106176556
>>106174180
GLM-4.5 with "Speak entirely in zoomer slang." as system prompt.
Anonymous
8/7/2025, 6:02:19 PM No.106176558
>>106176536
Typo: the prompt as just some 5 words, basically something you'd encounter in some random meditation guide, wasn't sexual even.
Anonymous
8/7/2025, 6:02:29 PM No.106176560
AGI is more about agency than being identical to humans
do you have software that is capable of acting on its own and make DECISIONS? you have AGI
a spider is a biological AGI and so is a cockroach
they have their own goals and act upon them
GPT is not and will never be AGI
"agentic" is a misnomer because the agent part is all outside scripting, there is no mechanism for a LLM to have an uninterrupted stream of consciousness and ability to remember what it does
Replies: >>106176582 >>106176598
Anonymous
8/7/2025, 6:03:15 PM No.106176569
>>106176520
Always combine with vitamin K or you will fuck up your soft tissues with calcium
Replies: >>106176645
Anonymous
8/7/2025, 6:04:09 PM No.106176581
>>106176542
If nothink then I'd go with Instruct. Thinking gives better responses though if you can deal with the thinking.
Anonymous
8/7/2025, 6:04:13 PM No.106176582
>>106176560
>do you have software that is capable of acting on its own and make DECISIONS? you have AGI
I think the fact "artificial" is in the title means it doesn't matter what the fuck it wants, it's ours, we made it
Anonymous
8/7/2025, 6:05:05 PM No.106176590
file
file
md5: c91dbfafd87d67f9097de48dbcdf4ae0🔍
>>106176138
It regulates your digestion. If you have bloating after eating, or slow digestions besides lack of appetite, this will help.
Replies: >>106176645
Anonymous
8/7/2025, 6:05:36 PM No.106176598
>>106176560
They lack true intelligence.
Anyway to get to AGI you'd probably need to get agency indeed, intent, understanding, caring, desire, proper memory, updating the weights live and more. Akl probably doable, but none of them actually trying for this because they can't sell online learning and it was terrible MFU. Local is the way to AGI when you have your hardware that could learn online and you care not for MFU. It wouldn't be a product anyway, but something autonomous.
Anonymous
8/7/2025, 6:05:48 PM No.106176602
>>106176536
Having a fetish for the period of your lifetime when you were developing sexually makes too much sense not to be common.
I think our culture is just so mind-melted from the specter of child abuse that most people won't even entertain the notion that it's possible.
Anonymous
8/7/2025, 6:06:35 PM No.106176615
Human intelligence came to be because of the world surrounding us and thus the only way to create true artificial intelligence is a world AI model.
Anonymous
8/7/2025, 6:09:03 PM No.106176645
>>106176590
interesting, i havent eaten licorice and im not really bloated, pretty sure i have normal digestion but who knows
>>106176569
im taking a D3 dose of 2000IU every day, should i still be taking vitamin K? K2 specifically?
Replies: >>106176667
Anonymous
8/7/2025, 6:11:34 PM No.106176667
>>106176645
>im taking a D3 dose of 2000IU every day, should i still be taking vitamin K? K2 specifically?
If you're getting enough in your diet, it should be OK. But don't do this long term without someone monitoring you.
K2 can't hurt, however. It synergizes with D3. Check your websearch-enabled local LLM for more info.
Anonymous
8/7/2025, 6:12:13 PM No.106176676
this is AGI
https://e-hentai.org/g/1919911/e1be31c3a3/
https://e-hentai.org/g/3262589/cda50ffcc8/
Replies: >>106176706 >>106176731 >>106177481
Anonymous
8/7/2025, 6:12:49 PM No.106176686
>>106176520
>im drinking vitamin D3 because i dont go out at all, my doctor recommended it to me, how is it placebo? body wont make vitamin D without the sun, so i gotta take it
placebo as in it wont do anything medicine dosent get absorbed the same way as food because it dosent have the other shit around it so it can bind to it and be transported around the same way as liquids arent digested the same way as solids even though they may be the same material
>so i gotta take it
eh probably not going to do anything bad so it dosent matter also you dont im likewise inside all day to the point where my skin starts shedding like a snake and its raw and itchy 24/7 you can fix that shit by opening the window fully and standing infront of it for like 10-15 minutes the whole "you need sunlight thing" is overblown you need a little every month or other month so your skin dosent shed but nothing more then that really
>im not, and i havent been. i just tend to eat less and never really eat fast food.
man thats still fucked though like 49 kg is way too low idfk what to say you do you but i would highly reccomend eating more or something my female cousin is fucked due to her weight weak as shit and like half faint whenever she has to exert herself a little
Replies: >>106176845 >>106176845
Anonymous
8/7/2025, 6:13:11 PM No.106176694
>>106176475
you're not understanding what I'm getting at. the very concept of a one-dimensional "level of intelligence" is exactly the sort of misconception I'm arguing against; it's not like that at all, intelligence is a fuzzy collection of abilities. have you ever considered that human intelligence could also be narrow, just in different ways?
>because again, the whole point is for ai to reach human level
great, so say that instead of using the term "general intelligence" where it doesn't make sense
Anonymous
8/7/2025, 6:14:10 PM No.106176706
>>106176676
xzhentai.net is better
Replies: >>106176731
Anonymous
8/7/2025, 6:14:32 PM No.106176709
Why don't we have a "living" (Think Lisp or Smalltalk) LLM yet?
Replies: >>106177180
Anonymous
8/7/2025, 6:16:53 PM No.106176731
>>106176676
Nice, that's just what I want
>>106176706
Elaborate on why
Anonymous
8/7/2025, 6:17:04 PM No.106176733
the real answer is that intelligence is a meme and like most weighty philosophical concepts you're better off not wasting your time thinking about it
Replies: >>106176753
Anonymous
8/7/2025, 6:18:34 PM No.106176753
>>106176733
>intelligence is a meme
>wasting your time thinking
yup this is lmg alright
Anonymous
8/7/2025, 6:28:19 PM No.106176845
>>106176686
>But don't do this long term without someone monitoring you.
you're right ill ask my doctor next time im sick
>Check your websearch-enabled
does any non bloated frontend actually have this? i've never used local web search nor have i heard of anons using it at all..
> local LLM for more info.
already done but glm4.5 air is being too professional and downplaying all risks
>>106176686
>placebo as in it wont do anything medicine dosent get absorbed the same way as food
what if i drink it after lunch/dinner? i read that D3 absorbs more easily than D2
>my skin starts shedding like a snake and its raw and itchy 24/7
anon are you sure this is because of not going out? my skin never really shed and it only gets itchy when my muscles are inflamed from working out (no a shower doesnt fix it)
maybe you should try to get the dead skin off your body with a rougher sponge when showering? i mean vitamin d is mostly for bones right..?
>man thats still fucked though like 49 kg is way too low idfk
its epic, i dont have fatigue issues, recently i deep cleaned my room and my muscles were inflamed for a few days but i can exert myself as much as i want
im just not tall and i work out sometimes
Replies: >>106177009
Anonymous
8/7/2025, 6:32:37 PM No.106176901
GPT5-Creative is going to be insane
Anonymous
8/7/2025, 6:33:25 PM No.106176909
file
file
md5: 3962e205e3cff99b318a59fdac2e0030🔍
>he interacts with 4chan with anything other than his locally-hosted llm agent
heh.
Anonymous
8/7/2025, 6:36:12 PM No.106176941
>native world model
>ask it the upside down spitting question
>similar to how you can prompt genie 3 to simulate that, the model will simulate the problem internally, and then base its answer on its internal simulation
>[spoiler]it still gets the question wrong[/spoiler]
Anonymous
8/7/2025, 6:36:31 PM No.106176948
>>106175226
I run it with iq4xs and it does well in my small tests when plugging it into roo code. I got it to generate a 3d physics system in js with no external libraries over 3 prompts with good code structure and improvements while avoiding npm trash.
Anonymous
8/7/2025, 6:40:40 PM No.106177005
why does GLM 4.5 Air stop reasoning after 8 messages in, every single time
Replies: >>106177033 >>106177053 >>106177381
Anonymous
8/7/2025, 6:41:06 PM No.106177009
>>106176845
>what if i drink it after lunch/dinner? i read that D3 absorbs more easily than D2
helps still not the same though i reccomend ditching it completely but you do you
>anon are you sure this is because of not going out? my skin never really shed and it only gets itchy when my muscles are inflamed from working out (no a shower doesnt fix it)
maybe you should try to get the dead skin off your body with a rougher sponge when showering? i mean vitamin d is mostly for bones right..?
yea its due to it as soon as i step outside for a few minutes it clears up and stops besides the already half shedded skin
>its epic, i dont have fatigue issues, recently i deep cleaned my room and my muscles were inflamed for a few days but i can exert myself as much as i want
im just not tall and i work out sometimes
>i deep cleaned my room and my muscles were inflamed for a few days
nigga... that not normal at all also you should not be that light especially if you are working out you do you ive said my peace and i wish you well
Replies: >>106177039
Anonymous
8/7/2025, 6:41:39 PM No.106177021
>>106175349
>good for tool usage
not really kek it's straight retarded if you try to use it with zed
Anonymous
8/7/2025, 6:42:20 PM No.106177033
>>106177005
need context bud what are you running?
Anonymous
8/7/2025, 6:42:44 PM No.106177039
>>106177009
thank you anon, be well and take care
Anonymous
8/7/2025, 6:42:51 PM No.106177043
>>106177012
>>106177012
>>106177012
Anonymous
8/7/2025, 6:43:09 PM No.106177053
>>106177005
works on my machine
Anonymous
8/7/2025, 6:43:47 PM No.106177066
gpt 5 will be a ****...**...****...etc changer
Anonymous
8/7/2025, 6:50:30 PM No.106177180
>>106176709
Does it count if I give full control of Emacs to an LLM?
Anonymous
8/7/2025, 6:50:48 PM No.106177190
>>106175990
Kek
Now do one where he gets arrested
Anonymous
8/7/2025, 7:01:15 PM No.106177381
10 responses, still thinking, q4km quant glm air
10 responses, still thinking, q4km quant glm air
md5: 3dafd33376c7dcf5af13e09db7b7a9fa🔍
>>106177005
Doesn't do that for me.
Anonymous
8/7/2025, 7:04:42 PM No.106177481
>>106176676
>filtered by a sad panda
Replies: >>106177667 >>106177709
Anonymous
8/7/2025, 7:10:22 PM No.106177667
>>106177481
i can use ex, its for the not yet initiated
Anonymous
8/7/2025, 7:11:59 PM No.106177709
>>106177481
I am not making an account just for cartoon porn.
Anonymous
8/7/2025, 7:30:05 PM No.106178268
>>106175494
>>106176547
>your speed in t/s is still going to be single digits though.
Yeah but I'm used to that anyway. That's pretty reassuring then, thanks. I might look into just getting an assload of RAM and playing around with some bigger MoE models.