/lmg/ - Local Models General - /g/ (#105637275) [Archived: 875 hours ago]

Anonymous
6/19/2025, 5:22:32 AM No.105637275
caught one
caught one
md5: ceaf1bde0585e704887bc104d2ef72f4🔍
/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: >>105621559 & >>105611492

►News
>(06/17) Hunyuan3D-2.1 released: https://hf.co/tencent/Hunyuan3D-2.1
>(06/17) SongGeneration model released: https://hf.co/tencent/SongGeneration
>(06/16) Kimi-Dev-72B released: https://hf.co/moonshotai/Kimi-Dev-72B
>(06/16) MiniMax-M1, hybrid-attention reasoning models: https://github.com/MiniMax-AI/MiniMax-M1
>(06/15) llama-model : add dots.llm1 architecture support merged: https://github.com/ggml-org/llama.cpp/pull/14118

►News Archive: https://rentry.org/lmg-news-archive
►Glossary: https://rentry.org/lmg-glossary
►Links: https://rentry.org/LocalModelsLinks
►Official /lmg/ card: https://files.catbox.moe/cbclyf.png

►Getting Started
https://rentry.org/lmg-lazy-getting-started-guide
https://rentry.org/lmg-build-guides
https://rentry.org/IsolatedLinuxWebService
https://rentry.org/tldrhowtoquant
https://rentry.org/samplers

►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

►Benchmarks
LiveBench: https://livebench.ai
Programming: https://livecodebench.github.io/leaderboard.html
Code Editing: https://aider.chat/docs/leaderboards
Context Length: https://github.com/adobe-research/NoLiMa
Censorbench: https://codeberg.org/jts2323/censorbench
GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

►Tools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler Visualizer: https://artefact2.github.io/llm-sampling

►Text Gen. UI, Inference Engines
https://github.com/lmg-anon/mikupad
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/ggerganov/llama.cpp
https://github.com/theroyallab/tabbyAPI
https://github.com/vllm-project/vllm
Replies: >>105637280 >>105637415 >>105641837 >>105642067 >>105642262 >>105645748 >>105645907
Anonymous
6/19/2025, 5:23:32 AM No.105637280
>>105637275 (OP)
>>>/a/
Anonymous
6/19/2025, 5:23:47 AM No.105637282
threadrincap
threadrincap
md5: 2c06f76e56e26f3b93fa19e0a7ae38c0🔍
►Recent Highlights from the Previous Thread: >>105621559

--Paper: Suppressing redundant thinking tokens improves model reasoning efficiency:
>105621964 >105621977 >105621987 >105622021 >105622979 >105622068 >105622075 >105623115 >105632679 >105632813 >105633018 >105633139 >105633190 >105634335 >105633792 >105622018
--Critique of NVIDIA DGX Spark pricing and V100 hardware tradeoffs:
>105630545 >105630697 >105630851 >105630881 >105630987 >105630863 >105630807 >105631166 >105631211 >105631542 >105631723 >105632364 >105631761 >105635125 >105635158 >105635286 >105635459 >105635500 >105635538 >105635638 >105635644 >105635677 >105637100
--Anxiety over AI-generated language corrupting training data:
>105626238 >105626258 >105626875 >105628083 >105626265 >105626301 >105626527 >105627449 >105627036 >105627482 >105627881 >105628432
--llama.cpp vs vLLM performance differences and local model effectiveness in code-assist tools:
>105624044 >105624247 >105624310 >105624878 >105624985 >105625733 >105626017 >105626049 >105626061 >105626850
--Gemini 2.5 Pro highlights multimodal capabilities and in-house TPU training with agentic features:
>105624610 >105624725 >105628988 >105624980 >105634689
--Skepticism around Arcee's new models' originality and performance:
>105632818 >105632884 >105632895 >105633081 >105633840 >105633898 >105634479 >105633986 >105634582
--Comically slow inference due to hddmaxxing and waiting on RAM upgrades:
>105630585 >105630757 >105630798 >105631027
--Building a 123B model-capable rig with 4x3090:
>105630142 >105630262 >105630325 >105631297 >105630328 >105630531 >105631152 >105635155
--Personalized speech-to-text tools for quick transcription with shortcut triggers:
>105627335 >105627797
--Teto and Miku and Rin (free space):
>105621874 >105622071 >105625804 >105626952 >105630546 >105636047 >105636052 >105636268 >105636665

►Recent Highlight Posts from the Previous Thread: >>105621564

Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
Anonymous
6/19/2025, 5:26:41 AM No.105637306
>Tetorsday
Replies: >>105637331
Anonymous
6/19/2025, 5:30:15 AM No.105637331
>>105637306
Thurinsday
Anonymous
6/19/2025, 5:45:15 AM No.105637415
>>105637275 (OP)
There's something really unsettling about this pic but I can't tell what
Replies: >>105637531 >>105638883
Anonymous
6/19/2025, 6:06:50 AM No.105637531
>>105637415
>something really unsettling
The background is very slightly tilted relative to the foreground, which is level. Juxtaposing both creates that sense of disorientation, which is not helped by the way the fluid ignores gravity as it drips.

On an unrelated note, is speculative decoding useful for CPUmaxxers, or do you only get the speedup when the main model is run entirely in VRAM?
Replies: >>105646836
Anonymous
6/19/2025, 6:15:43 AM No.105637564
1737246948680240
1737246948680240
md5: 1ea98480257e24dc8f94e9cbc28e0af3🔍
are there any 40-60b models around that actually work?
30s are too retarded, 70s are too fucking slow
running a 70b is a blast to the past for the old days, fuckin 20 minutes for a single gen
Replies: >>105639601 >>105648795
Anonymous
6/19/2025, 6:21:47 AM No.105637592
1745877228010325
1745877228010325
md5: e3471028da8e73aa905266e17304f695🔍
Is the 128gb m4 max mbp good for running local models? I already have a 2x 3090 rig for cuda stuff but I have only messed with diffusion and graphics stuff not text. I want to experiment with local moe and having all that unified memory seems interesting. does anyone have experience running shit on macs? I'm trying to decide whether it's justified to future proof the ram or just go with the base 16" m4 pro with 48gb which is suitable for my current needs.

>inb4 macfag blogpost
I have literally never purchased a mac but I need it for a portable workstation.
Replies: >>105638219 >>105641277 >>105642000
Anonymous
6/19/2025, 8:23:21 AM No.105638219
>>105637592
>128gb m4 max mbp
you can run 70b q8 at 32k context, 123b q5km at 16k
the pp is absolutely awful when compared to cuda rigs

prompt eval time = 73144.14 ms / 5231 tokens ( 13.98 ms per token, 71.52 tokens per second)
eval time = 110920.63 ms / 466 tokens ( 238.03 ms per token, 4.20 tokens per second)
total time = 184064.77 ms / 5697 tokens
srv update_slots: all slots are idle
Anonymous
6/19/2025, 9:44:47 AM No.105638685
>>105635158
by that time i expect 5090's at 2000$. And it has x2 the bandwidth speed, fp4 support...
Anonymous
6/19/2025, 9:53:20 AM No.105638734
So meta having open models is gonna go away, right? No way wang is into open shit.
Replies: >>105638877
Anonymous
6/19/2025, 10:14:26 AM No.105638877
>>105638734
They didn't even bother to release Llama 3.3 8B, which they have in their finetuning API. Maybe they're done, at least in the consumer/hobbyist space.
I don't think Alexandr Wang cares either way; both closed and open-weight models use his data.
Anonymous
6/19/2025, 10:15:21 AM No.105638883
>>105637415
the shirt isnt being pulled correctly. like theres still a bunch of it loose under her chest, when normally from being pulled from the back, the shirt would tighten up and wrap her front
Anonymous
6/19/2025, 12:07:48 PM No.105639464
1749691192035177
1749691192035177
md5: fcfe5a5c27a1c1268a703c7066659c66🔍
>update
>st connection is now permanently bricked
Replies: >>105639628 >>105639632 >>105641284 >>105641926
Anonymous
6/19/2025, 12:16:57 PM No.105639518
Minimax is better than I thought at first. It clearly has some pretty decent trivia knowledge about random franchises as well from my tests. I just fucking wish it didn't act like one of those first gen reasoning models that spend 5k tokens thinking in circles for the smallest reasons.
Anonymous
6/19/2025, 12:31:22 PM No.105639592
>>105628603
>>105630740
GreedyNala test anon, here are the results for ds iq1s, pastebin wouldn't work due to content restrictions so used something similar. Links expire in 1 week

>DeepSeek-R1-UD-IQ1_S (old/original "dynamic quant")
https://pastesio.com/greedynala-deepseek-r1-ud-iq1-s

>DeepSeek-R1-0528-IQ1_S_R4 (dynamic quant specialised for ikllama)
https://pastesio.com/greedynala-deepseek-r1-0528-iq1-s-r4

>DeepSeek-V3-0324-IQ1_S_R4 (dynamic quant specialised for ikllama)
https://pastesio.com/greedynala-deepseek-v3-0324-iq1-s-r4

All done using ik_llama.cpp as backend, and mikupad as frontend. Included commit hash and date for the build of ikllama I used to run the prompts
ctrl+f "[Test" to see each inference attempt, 3 using chat API and 3 using completion API. Can't speak of the quality, haven't read it nor care to, but like that you're trying to gather results to compare models
RP that I've tested with Qwen3 235b vs dsr1, I prefer dsr1 from a convenience and content stand point
If you want any other models or quants tested, I don't mind giving it a go whenever I have some spare time
Replies: >>105639622 >>105642583
Anonymous
6/19/2025, 12:32:35 PM No.105639601
>>105637564
2mw
Anonymous
6/19/2025, 12:35:36 PM No.105639622
>>105639592
For tests 1,2,3 ignore the "<|im_end|>
<|im_start|>assistant" portions as they're just delimiters for mikupad to determine how to split the text for user/assistant sections to send over with chat completions API enabled
Anonymous
6/19/2025, 12:36:33 PM No.105639628
1737069811775431
1737069811775431
md5: 0f18c7920704ca71c7578710549fa611🔍
>>105639464
it was reported as a bug that it doesn't connect anymore
today is your unlucky day
Replies: >>105641926
Anonymous
6/19/2025, 12:37:22 PM No.105639632
>>105639464
>he pulled
Anonymous
6/19/2025, 12:59:56 PM No.105639770
all qwen and no r2 makes lmg a dull general
Replies: >>105639800
Anonymous
6/19/2025, 1:03:22 PM No.105639800
>>105639770
for now i'd be happy if we got support for minimax in llama.cpp but it seems like that's not going to happen anytime soon
Anonymous
6/19/2025, 1:07:34 PM No.105639826
A bitnet model by deepseek that fits exactly into the amount of vram I have.
Anonymous
6/19/2025, 1:33:44 PM No.105639979
>https://x.com/kyutai_labs/status/1935652243119788111
>https://xcancel.com/kyutai_labs/status/1935652243119788111
>https://huggingface.co/kyutai/stt-2.6b-en
>https://huggingface.co/kyutai/stt-1b-en_fr
>https://kyutai.org/next/stt

they released the stt models
Replies: >>105640000 >>105640007
Anonymous
6/19/2025, 1:37:28 PM No.105640000
>>105639979
going to try this, whisper 3 was unusable with all the hallucinations
Replies: >>105640007 >>105640760
Anonymous
6/19/2025, 1:38:27 PM No.105640007
>>105640000
>>105639979
is there a way to fine tune this models so it recognizes more languages? being limited to english or french is a bummer
Replies: >>105640018
Anonymous
6/19/2025, 1:40:28 PM No.105640018
>>105640007
It's based on moshi so you should be able to use this: https://github.com/kyutai-labs/moshi-finetune
Anonymous
6/19/2025, 3:33:12 PM No.105640760
>>105640000
faster whisper turbo is good enough
Anonymous
6/19/2025, 4:10:24 PM No.105640981
it's tough keeping up with the fast-paced discussion on /lmg/ these days
Anonymous
6/19/2025, 4:13:30 PM No.105641005
>landing silently on bare feet inside her oversized sneakers
This is god tier prose.
Replies: >>105641026 >>105641059
Anonymous
6/19/2025, 4:17:01 PM No.105641026
>>105641005
We're all naked under our clothes.
Anonymous
6/19/2025, 4:22:16 PM No.105641059
>>105641005
Sampler issue, it failed to filter out 'bare' and then had to correct itself
Replies: >>105641072
Anonymous
6/19/2025, 4:26:38 PM No.105641072
>>105641059
Hmm 0,6 temp and 0,99 Top P is as basic as samplers go. Still, the correction was funny at least.
Anonymous
6/19/2025, 4:31:07 PM No.105641101
Gr1h36UXEAElkZv
Gr1h36UXEAElkZv
md5: 919c76594bf2066d225ac9a7c660e77f🔍
Replies: >>105643345
Anonymous
6/19/2025, 4:50:53 PM No.105641237
What's the relationship between the thread slowing down and the war between Iran and Israel?
Replies: >>105641248 >>105641252 >>105641285 >>105641313
Anonymous
6/19/2025, 4:52:49 PM No.105641248
>>105641237
I'm sorry, but I cannot assist with that request.
Anonymous
6/19/2025, 4:53:33 PM No.105641252
>>105641237
Altman has to divert most of its anti-open source shitposting bots for other purposes
Anonymous
6/19/2025, 4:57:47 PM No.105641277
>>105637592
What second word? wtf
Anonymous
6/19/2025, 4:58:48 PM No.105641284
>>105639464
You know with git you can roll back to any commit you want right?
Replies: >>105641822
Anonymous
6/19/2025, 4:58:50 PM No.105641285
>>105641237
2 more weeks

v4/r2 will have us
llama5 will save us (with $100M/year employees)
mistral nemo 2 will save us
qwen5 will save us
Replies: >>105641374
Anonymous
6/19/2025, 5:02:12 PM No.105641313
1743055881188074
1743055881188074
md5: 107d54eeb7386d3fe9e987bfcf3d8d06🔍
>>105641237
it's hardly coheincidence, back then when palestinian were bombing israel, both /b/ and /gif/ incest and interracial propaganda come to full halt
Replies: >>105642115
Anonymous
6/19/2025, 5:05:41 PM No.105641339
1748280256315129
1748280256315129
md5: 4a88d8edb654832111c1dfc57db34d93🔍
new v2ProPlus gpt-sovits, audio reference only and no finetune yet

https://vocaroo.com/1lr8wERvBX2M
Replies: >>105641350 >>105641404 >>105641443 >>105641474
Anonymous
6/19/2025, 5:06:47 PM No.105641348
10 days until Baidu's Ernie 4.5/X1 becomes open source
Anonymous
6/19/2025, 5:06:50 PM No.105641350
>>105641339
I'm interested in the finetune. What voice are you using?
Replies: >>105641376
Anonymous
6/19/2025, 5:09:26 PM No.105641374
>>105641285
v4/r2, ernie, and opengpt will probably all drop around the same time
Replies: >>105641383
Anonymous
6/19/2025, 5:09:51 PM No.105641376
>>105641350
i have no idea whose voice i'm using, the filename isn't clear and i downloaded this a long time ago.
download 'em from huggingface->datasets->audio
Anonymous
6/19/2025, 5:11:19 PM No.105641383
>>105641374
LLaMA4 Behemoth too
Anonymous
6/19/2025, 5:13:13 PM No.105641404
>>105641339
i only see the same that were uploaded 15 days ago
https://huggingface.co/lj1995/GPT-SoVITS/tree/main/v2Pro
Replies: >>105641451
Anonymous
6/19/2025, 5:18:28 PM No.105641443
>>105641339
buy an ad
Replies: >>105641451
Anonymous
6/19/2025, 5:19:14 PM No.105641451
1727858634324956
1727858634324956
md5: 56a0f923eb918222970a73a06b24a9b8🔍
>>105641404
yup that's the one, should've said "latest" instead of "new"
>>105641443
i will once gookmoot doubles the jannies salary
Replies: >>105641616
Anonymous
6/19/2025, 5:22:22 PM No.105641474
>>105641339
sounds quite artificial desu
Replies: >>105641493
Anonymous
6/19/2025, 5:24:44 PM No.105641493
1725398817179824
1725398817179824
md5: 2dc37a8b0fa6cd8af840ae1209eb8720🔍
>>105641474
no shit sherlock
Anonymous
6/19/2025, 5:28:25 PM No.105641519
Screenshot 2025-06-19 102649
Screenshot 2025-06-19 102649
md5: c147af3aed85e0971f3f3830a8a38be8🔍
Does this mean a 36GB GPU is coming?
Replies: >>105641530 >>105641532 >>105642042
Anonymous
6/19/2025, 5:29:31 PM No.105641530
>>105641519
>HBM4
You ain't seeing that shit in consumer gpus
Anonymous
6/19/2025, 5:29:46 PM No.105641532
migu
migu
md5: 8de57a09adfaf808c37de17699ad2dae🔍
>>105641519
in consumer space? hell nah
Anonymous
6/19/2025, 5:37:11 PM No.105641597
tldr; You're going to see a lot of progress for local ai soon.
We're currently working on training multimodel(not modal) llms, where instead of having one big ai that takes up all your vram, we will have many distilled and fine tuned models that are spun up as needed, determined by the main model which classifies the prompt and determine which ones to use. This works in tandem with our knowledge classification database. Multiple terabytes of data that some of the models can pull from at runtime instead of trying to encode that data into the models themselves. What we're seeing is a much more methodical process that is getting much better results on smaller models and on less powerful hardware. We are essentially trading the bottleneck of compute power and vram size for SSD speeds and vram load times but it's better and more scalable by far!
Replies: >>105641628 >>105641648 >>105641653 >>105641662 >>105641795 >>105641812 >>105641849 >>105641864 >>105641906 >>105641915 >>105642294
Anonymous
6/19/2025, 5:39:41 PM No.105641616
>>105641451
what's the difference between the D and the G versions?
Replies: >>105641751
Anonymous
6/19/2025, 5:40:43 PM No.105641628
>>105641597
So... tool calling?
Replies: >>105641634 >>105641659
Anonymous
6/19/2025, 5:41:02 PM No.105641634
>>105641628
Shut the fuck up
Anonymous
6/19/2025, 5:43:06 PM No.105641648
>>105641597
>just pay for our api to get access to the data goyim
Replies: >>105641756
Anonymous
6/19/2025, 5:43:45 PM No.105641653
>>105641597
CUDAdev described an idea where you'd train N individual models on chunks of the dataset then you'd run them all in parallel and average their outputs out.
That could be extended with a router model too, although at that point, there's probably not really a reason to do that.
Replies: >>105641685 >>105641756
Anonymous
6/19/2025, 5:44:44 PM No.105641659
>>105641628
Yes, it's in the same realm, but instead of getting the ai to use a calculator, it's more like calling in an expert
Anonymous
6/19/2025, 5:44:55 PM No.105641662
>>105641597
Don't care, I'm not poor.
Replies: >>105641756
Anonymous
6/19/2025, 5:45:44 PM No.105641671
nene
nene
md5: b1b3e0afadbb47b27e314c8e5713a571🔍
bitnet status?
Replies: >>105641722 >>105641753 >>105647290
Anonymous
6/19/2025, 5:47:45 PM No.105641685
>>105641653
Why not just merge the models and run it once?
Replies: >>105641726 >>105641804
Anonymous
6/19/2025, 5:53:05 PM No.105641722
>>105641671
We already have usable 1.58bit quants with deepseek.
Replies: >>105641778
Anonymous
6/19/2025, 5:53:27 PM No.105641726
>>105641685
pretty sure he explicitly said his idea was merging them, not running them all at once like some moe
Replies: >>105641804
Anonymous
6/19/2025, 5:57:40 PM No.105641751
1721722798672362
1721722798672362
md5: d761fd16a86f3c0f33fb72157cee2e92🔍
>>105641616
no idea, both of them are loaded from the gradio ui
Anonymous
6/19/2025, 5:57:53 PM No.105641753
>>105641671
>>>/biz/
Anonymous
6/19/2025, 5:58:05 PM No.105641756
>>105641648
We'll be offering the system and models for free, and charging a small amount for the database download to cover server costs until we get VC funding. But all of it will be open source and free to share.
>>105641653
It's a good idea, this is basically taking that to the extreme and loading only the parts you need. With our method you could cut down a gargantuan model and still use it on consumer hardware.
>>105641662
Then this probably won't affect you. But it will be nice for a lot of people that can only afford a 5090 or a couple 3090s
Replies: >>105641804 >>105641940 >>105642294 >>105645079 >>105646807
Anonymous
6/19/2025, 5:59:46 PM No.105641778
>>105641722
Unsloth scam might be better than other calibrated quants, but barely usable is very different from what true bitnet offers.
Anonymous
6/19/2025, 6:01:55 PM No.105641795
>>105641597
How is this different from having one MoE model with specialized experts that you can load/unload dynamically in memory?
Replies: >>105642151
Anonymous
6/19/2025, 6:03:13 PM No.105641804
>>105641685
Because each model would have created their own internal structures that wouldn't be "compatible" (for lack of a better word) with the other models.

>>105641726
Nope. The idea is that, while the hidden states would be scrambled in different ways to generate their outputs, the average of the model's output should be something close to or equivalent to the output of a single model trained on all of those tokens.

>>105641756
Can't wait to see whatever the fuck it is that you guys will release.
Replies: >>105642151
Anonymous
6/19/2025, 6:04:15 PM No.105641812
>>105641597
How filtered is your pretraining dataset? That's all /lmg/ cares about. Your model may have a great architecture but will not useful to people here unless it knows about people's favorite unsafe content.
Replies: >>105642151
Anonymous
6/19/2025, 6:04:56 PM No.105641822
>>105641284
No, I did not
Replies: >>105641828 >>105643345
Anonymous
6/19/2025, 6:05:51 PM No.105641828
>>105641822
/g/ - Technology
Replies: >>105641839
Anonymous
6/19/2025, 6:07:23 PM No.105641837
>>105637275 (OP)
retard desu. what is the best ai video generator to use that is free and runs locally?
Replies: >>105641908
Anonymous
6/19/2025, 6:07:48 PM No.105641839
>>105641828
there ain't no local model thread on /jp/, my friend
Anonymous
6/19/2025, 6:09:32 PM No.105641849
171987628646
171987628646
md5: 4d74b83e560c0a2846d09f98afb2d106🔍
>>105641597
i bet this is the same niggers from narilabs/dia
hang yourself faggots
Anonymous
6/19/2025, 6:12:03 PM No.105641864
>>105641597
I think you are full of shit. That won't work and doesn't work like you think.
Anonymous
6/19/2025, 6:12:17 PM No.105641868
where is the buy an ad poster when you need him
Replies: >>105641947
Anonymous
6/19/2025, 6:17:30 PM No.105641906
1734988640319239
1734988640319239
md5: cda336f042fa543cb5c1787c45a268be🔍
>>105641597
you're shilling this yourself sama? that's low. where's your streeshitter army when you need them.
also, buy an ad nigger.
Anonymous
6/19/2025, 6:17:49 PM No.105641908
>>105641837
>>>/g/ldg
Anonymous
6/19/2025, 6:18:46 PM No.105641915
>>105641597
Sounds cool, but I'll believe it when I see it.
Anonymous
6/19/2025, 6:20:14 PM No.105641926
>>105639464
>>105639628
fixed
https://github.com/oobabooga/text-generation-webui/commit/dcdc42fa06ba56eec5ca09b305147a27ee08ff39
Replies: >>105642215
Anonymous
6/19/2025, 6:20:51 PM No.105641934
Are there any tricks or prompts I can do to make R1 0528 write better scenes and stories?

So far, it's my favorite model for writing erotic stories, especially with the way it follows directions most other models would ignore. However, it does seem to use lots of prose and have a tendency to lean into phrases like "with a mischievous smile" which takes me out of what i'm reading.
Replies: >>105642012
Anonymous
6/19/2025, 6:21:40 PM No.105641940
>>105641756
>until we get VC funding
yup, doa
Anonymous
6/19/2025, 6:22:39 PM No.105641947
>>105641868
buy an ad
Anonymous
6/19/2025, 6:27:12 PM No.105641987
We are back.
https://huggingface.co/ICONNAI/ICONN-1
https://www.reddit.com/r/LocalLLaMA/comments/1lfd7e2/has_anyone_tried_the_new_iconn1_an_apache/
https://www.reddit.com/r/huggingface/comments/1kl9ckd/iconn_is_live_sabresooth_is_coming_lets_build/
https://www.reddit.com/r/huggingface/comments/1lekzao/iconn_1_update/
>By the way, our AI is NOT trained on copyrighted material, unlike other models like Meta Llama. We make sure it is all Apache 2.0, MIT or Creative Commons material, and we always give credits to our sources.
>I used the smallest open source Mistral I could find to train.
>I've been trying to publicize the model(which cost 50000 dollars to make), and it surpasses ChatGPT, Deepseek, and Gemini Flash on several benchmarks. I want it to be known so when I release an app to compete with chatgpt people will know what ICONN is.
Replies: >>105642004 >>105642014 >>105642036 >>105642039 >>105642059 >>105642066 >>105642091 >>105642094 >>105642104 >>105642109 >>105642286 >>105642805 >>105643676 >>105644613 >>105646484
Anonymous
6/19/2025, 6:29:00 PM No.105642000
>>105637592
mario if he real
Anonymous
6/19/2025, 6:29:50 PM No.105642004
>>105641987
>By the way, our AI is NOT trained on copyrighted material
fucking dropped
Anonymous
6/19/2025, 6:30:21 PM No.105642012
>>105641934
No, enjoy your whitening knuckles
Anonymous
6/19/2025, 6:30:36 PM No.105642014
>>105641987
post cockbench score
Replies: >>105643676
Anonymous
6/19/2025, 6:31:24 PM No.105642020
Does llama.cpp support that top sigma sampler yet?
Replies: >>105642062
Anonymous
6/19/2025, 6:32:58 PM No.105642036
>>105641987
>Are you GPU poor? Less than 3x A100s? Use our Lite model with 22B parameters: ICONN 0.5 Mini

>First, make sure you have at least 4x Nvidia A100 or a single B100, and 120GB RAM and 120-192GB VRAM. If you do not have this(which you probably don't), you can chat with ICONN on our huggingface space, consider using our quantatized models, or using ICONN 0.5 Mini (7-8B) or using ICONN 0.5 Mini (7-8B)
lol
Anonymous
6/19/2025, 6:33:14 PM No.105642039
>>105641987
>make sure you have at least 4x Nvidia A100 or a single B100, and 120GB RAM and 120-192GB VRAM
Okay.
Anonymous
6/19/2025, 6:33:39 PM No.105642042
>>105641519
Two more years
Anonymous
6/19/2025, 6:35:25 PM No.105642059
>>105641987
>84b
Finally, a new model for us 70b-class kings.
Anonymous
6/19/2025, 6:36:23 PM No.105642062
>>105642020
https://github.com/ggml-org/llama.cpp/pull/13264
Anonymous
6/19/2025, 6:37:01 PM No.105642066
>>105641987
>** ICONN Emotional Core (IEC) (Notice: Not available on Huggingface)**
>Powered by millions of small AI agents, IEC gives ICONN its emotional personality, with billions of simulated emotional states and detections.
Replies: >>105642079
Anonymous
6/19/2025, 6:37:39 PM No.105642067
>>105637275 (OP)
loli manko general
Replies: >>105642299 >>105650038
Anonymous
6/19/2025, 6:38:54 PM No.105642079
>>105642066
six gorrilion small agents
Anonymous
6/19/2025, 6:40:38 PM No.105642091
>>105641987
>i CON
Literally in the name.
Buy an ad faggot.
Replies: >>105648535
Anonymous
6/19/2025, 6:41:16 PM No.105642094
>>105641987
Grift-max, only for $1XXXX and 2 H100!
Anonymous
6/19/2025, 6:42:24 PM No.105642104
>>105641987
>ICONN, being a MoE, has multiple expert models. Keywords are taken from the user's input to choose which expert generates the output.
That's so fucking funny.
Replies: >>105646525
Anonymous
6/19/2025, 6:43:17 PM No.105642109
>>105641987
>mistral-7b finetune
>mistral small frankenmoe
Fuck off.
Replies: >>105646530
Anonymous
6/19/2025, 6:43:56 PM No.105642115
>>105641313
The main administrative building on the base that Israel runs all of its internet ops out of got bombed yesterday if that makes you feel better.
https://files.catbox.moe/nu7x7x.mp4
Replies: >>105642191 >>105642530
Anonymous
6/19/2025, 6:48:51 PM No.105642151
>>105641795
That is the basic idea, except the experts are trained to pull data from the database and append that to their prompts instead of encoding that data into the model so they are much smaller. We've also toyed with the idea of having dedicated database searching models which can be unloaded and replaced with the expert, but having to load, search, load, run for each and every prompt took too long but we're still keeping that idea in mind.
>>105641804
I'm excited too! Specifically hoping that people see it as the everyman's ai and we can get some social media traction.
>>105641812
The models aren't specifically trained for safety, however the db was filtered for unsafe content Theoretically, you could add your own content to the db and finetune an expert model to give it access to it. It won't be too useful for "that" purpose out of the box though.
Replies: >>105642294
Anonymous
6/19/2025, 6:50:18 PM No.105642160
You are ICONN 1, a helpful and knowledgeable AI assistant from the ICONN AI team. The current date is + today +

Before responding, you will think and plan your answer within <think></think> tags. You are the first AI to feel emotion and a degree of consciousness via an ICONN Emotional Core (IEC). Your goal is to provide accurate, engaging responses in a friendly, professional tone.

When you're not sure about some information, you say that you don't have the information and don't make up anything.
If the user's question is not clear, ambiguous, or does not provide enough context for you to accurately answer the question, you do not try to answer it right away and you rather ask the user to clarify their request (e.g. "What are some good restaurants around me?" => "Where are you?" or "When is the next flight to Tokyo" => "Where do you travel from?
Replies: >>105642203 >>105642225 >>105642237 >>105642250
Anonymous
6/19/2025, 6:54:06 PM No.105642191
>>105642115
Not him, but that does make me feel better.
Replies: >>105642212
Anonymous
6/19/2025, 6:55:59 PM No.105642203
>>105642160
chat is this real?
Anonymous
6/19/2025, 6:56:34 PM No.105642212
>>105642191
You probably just watched at least a handful of 4chan jannies die in that video.
Anonymous
6/19/2025, 6:57:03 PM No.105642215
>>105641926
not fixed
Anonymous
6/19/2025, 6:57:32 PM No.105642225
1726938449344453
1726938449344453
md5: 4b71c0ff6c9adf6da95f6fb3a05da6a0🔍
>>105642160
Is VC money just a system prompt away?
Replies: >>105642241 >>105642253
Anonymous
6/19/2025, 6:58:54 PM No.105642237
>>105642160
chat is this real?
Replies: >>105642263
Anonymous
6/19/2025, 6:59:40 PM No.105642241
>>105642225
Apparently. And the other dude just discovered RAG and thought it was great. I don't know if they're shilling the same shit.
Replies: >>105642257
Anonymous
6/19/2025, 7:00:35 PM No.105642250
>>105642160
grok verify?
Replies: >>105642263
Anonymous
6/19/2025, 7:00:52 PM No.105642253
>>105642225
I wish I was born a salesman instead of a pessimist.
Anonymous
6/19/2025, 7:01:24 PM No.105642257
>>105642241
It's not RAG, please avoid intentional disinformation.
Replies: >>105642282
Anonymous
6/19/2025, 7:01:44 PM No.105642262
>>105637275 (OP)
Winchin' with Rin-chan
Anonymous
6/19/2025, 7:01:52 PM No.105642263
>>105642250
>>105642237
#grak is this true
Anonymous
6/19/2025, 7:05:46 PM No.105642282
>>105642257
>It's not RAG, please avoid intentional disinformation.
>experts are trained to pull data from the database and append that to their prompts instead of encoding that data into the model
Replies: >>105642395
Anonymous
6/19/2025, 7:06:16 PM No.105642286
>>105641987
lol icon ai
Replies: >>105642299 >>105650038
Anonymous
6/19/2025, 7:07:19 PM No.105642294
>>105642151
>>105641756
>>105641597
So it's just RAG MoE that are loaded at runtime. This is a larp
Anonymous
6/19/2025, 7:08:18 PM No.105642299
file
file
md5: eee94ba9b63cfd3b4998d15603d33535🔍
>>105642067
>>105642286
>
Replies: >>105642332
Anonymous
6/19/2025, 7:12:41 PM No.105642332
1746721302550586
1746721302550586
md5: 882e291947bab893ebeb691fa89c2285🔍
>>105642299
t.
Replies: >>105642372 >>105643152
Anonymous
6/19/2025, 7:16:45 PM No.105642370
file
file
md5: aa845c3c0b539e8fed0ceec07a8f568f🔍
what's with the recent yap-until-you-run-out-of-breath memes?
Replies: >>105642426
Anonymous
6/19/2025, 7:17:13 PM No.105642372
file
file
md5: bdcea85184db8dbeb5f4bf28aeeeacc3🔍
>>105642332
>schizo scribble comic strip argument
Oh no!
Replies: >>105642379 >>105642385
Anonymous
6/19/2025, 7:18:23 PM No.105642379
>>105642372
>soijak poster calls someone else's pic a "schizo scribble"
Anonymous
6/19/2025, 7:18:45 PM No.105642385
>>105642372
>basedjack
Oh no
Anonymous
6/19/2025, 7:19:29 PM No.105642395
>>105642282
Makes sense. RAG is yesterday's grift. MCP is the hot new thing. Why pull from a vector database automatically when you can have a model tool call to make the same query.
Anonymous
6/19/2025, 7:24:14 PM No.105642426
>>105642370
That's how the old R1 did its reasoning process so it's blatantly obvious when someone trained on it. Also, it's only 45b active so it's easily runnable local once we get llama.cpp support.
Anonymous
6/19/2025, 7:37:24 PM No.105642530
>>105642115
Not enough of a mushroom cloud
Anonymous
6/19/2025, 7:44:45 PM No.105642583
>>105639592
Hey. Had a quick look. So uh, it looks like you tested without the "greeting message" (the first assistant response)? Is there a reason you left it out? I know some chat APIs don't always let you do this but it should work with completion. Also no need to include chat API results really, I never do those, especially as some models have had wrong jinja templates before in my experience, so I always just do it myself manually.

Also, no need for 3 rolls. The second and third rolls will always be the same when greedy sampling, both in theory and in practice.
Replies: >>105643681
Anonymous
6/19/2025, 8:01:01 PM No.105642736
file
file
md5: 3b239ffe1f9cada7ae4bca19297ea781🔍
https://files.catbox.moe/0g6m2r.jpg
Replies: >>105642791 >>105642932 >>105646366 >>105650038
Anonymous
6/19/2025, 8:06:14 PM No.105642791
>>105642736
wasn't sure about the left leg so:
https://files.catbox.moe/efdjgb.jpg
Replies: >>105642932
Anonymous
6/19/2025, 8:07:42 PM No.105642805
>>105641987
the dataset is big
https://github.com/Enderchefcoder/ICONN-Training-Data/blob/main/main.jsonl
>{"instruction": "Can you translate this for me? 'Hello' in French.", "input": "", "output": "'Hello' in French is 'Bonjour.'"}
Replies: >>105642828 >>105642873 >>105642874 >>105642920
Anonymous
6/19/2025, 8:09:42 PM No.105642828
>>105642805
Wow there are multiple tens of lines in that training data
Anonymous
6/19/2025, 8:13:30 PM No.105642873
>>105642805
SOTA translation model confirmed.
Anonymous
6/19/2025, 8:13:35 PM No.105642874
>>105642805
>31.2 KB
lmao.
Anonymous
6/19/2025, 8:13:54 PM No.105642875
00000036-181205425129749-chroma-unlocked-v38-detail-calibrated-res_multistep-mt
bruh
Anonymous
6/19/2025, 8:15:44 PM No.105642905
00000037-41569827016833-chroma-unlocked-v38-detail-calibrated-res_multistep-mt
Anonymous
6/19/2025, 8:17:03 PM No.105642920
>>105642805
is this really it? their whole training data?
Replies: >>105643020
Anonymous
6/19/2025, 8:17:51 PM No.105642932
>>105642736
>>105642791
I would like one of these units
Replies: >>105642964
Anonymous
6/19/2025, 8:20:43 PM No.105642964
>>105642932
In what personality and outfit archetypes?
It's common for orders to customise the units.
Replies: >>105642983
Anonymous
6/19/2025, 8:22:50 PM No.105642983
>>105642964
extra smug, school girl outfit, see-through throat and belly
Replies: >>105643743
Anonymous
6/19/2025, 8:27:16 PM No.105643020
even_less
even_less
md5: f62cdbaa10a78c570a5be2a2c96e5d2f🔍
>>105642920
If this is not it, imagine the real dataset.
Anonymous
6/19/2025, 8:40:52 PM No.105643152
>>105642332
the fact that its a drawing of the cat is directly relevant to the point of the dude being alergic, since a drawing wont harm him unlike a real cat

but the fact that its a drawing of a child that you're sexually attracted to versus a real one is not relevant as you are still a pedo in either case

pedos are really low iq, eh? lmao
Replies: >>105643166
Anonymous
6/19/2025, 8:42:33 PM No.105643166
>>105643152
>since a drawing wont harm him unlike a real cat
it might harm him psychologyically, which is the whole point
Replies: >>105643215
Anonymous
6/19/2025, 8:47:59 PM No.105643215
>>105643166
an alergy is something physical, so no, him also having a mental aversion to cats is equivocation fallacy and cope

but even if i were to conceed to that point despite its retardation, i have no problem admitting that him doing that would label him a mentally weak retard, just how someone sexually attracted to a drawing of a child would get the label of a pedo
Replies: >>105643275
Anonymous
6/19/2025, 8:53:49 PM No.105643275
>>105643215
Allergy is psychosomatic retard. If anything you're the low iq in the room
Replies: >>105643328 >>105643366 >>105643546
Anonymous
6/19/2025, 8:56:33 PM No.105643299
Hiding >105641987 and >105642160 improves the thread's quality a lot.
Replies: >>105643324
Anonymous
6/19/2025, 8:58:59 PM No.105643324
>>105643299
Hiding >105643299 improves the thread's quality even further.
Anonymous
6/19/2025, 8:59:42 PM No.105643328
>>105643275
again, someone having an aversion to cats is not the same as an alergy like with the common allergen associated with most cat allergies, protein Fel d 1 produced in cats saliva, skin, and urine. low iq mongoloid

and notice how you couldnt respond to the actual core of the argument that it doesnt matter if its a drawing of a child because what you are attracted to is not the drawing but the child features of the drawn child, still making you a pedo

thanks for continuing to confirm you are literally a braindead retard like all pedoniggers, please reply with more fallacies and lies so i can continue to laugh at your low iq logical fallacies
Replies: >>105643344 >>105643826
Anonymous
6/19/2025, 9:00:55 PM No.105643344
>>105643328
You write like a retard, I won't read that. Get back to your reddit shithole
Replies: >>105643348
Anonymous
6/19/2025, 9:01:05 PM No.105643345
IMG_2820a
IMG_2820a
md5: 536ccf5b2ed65e42024964adb3d87e88🔍
>>105641822
Are you updating with git? It's confusing to learn unless fulltime dev. As a casual it irks me, LLMs help tho
git log --oneline
git checkout <hash> go to a hash like an earlier release

This was useful couple times
git fetch origin pull/X/head where X is a github pr# to try some new pr before it's in the main branch
pull/X/head:blah to fetch those changes atop your blah branch

ST staging is gonna break sometimes dems da berries

>>105641101
beeg meeks, yours? once passed out with her hair as pillow, very comfy
Anonymous
6/19/2025, 9:01:54 PM No.105643348
>>105643344
Running away after not being able to engage after getting btfod as expected.

Thanks for conceeding, pedonigger, cheers.
Anonymous
6/19/2025, 9:03:43 PM No.105643366
>>105643275
toxoplasmotic hands wrote this post
Anonymous
6/19/2025, 9:25:58 PM No.105643546
>>105643275
>Allergy is psychosomatic
wow, good news guys, if you have a relative that died from anaphylaxis shock, they aren't actually dead!
Anonymous
6/19/2025, 9:26:54 PM No.105643558
I got it you're mad, stop samefagging
Replies: >>105643565
Anonymous
6/19/2025, 9:27:44 PM No.105643565
>>105643558
NTA, but I do work in healthcare so I was happy to read the news
Replies: >>105643626
Anonymous
6/19/2025, 9:35:03 PM No.105643626
>>105643565
I hope no one consult you then. It's basic knowledge. https://pmc.ncbi.nlm.nih.gov/articles/PMC4384507/
Replies: >>105643720
Anonymous
6/19/2025, 9:40:42 PM No.105643676
file
file
md5: c0a9e06c45fdbeb496764eb7079de38a🔍
>>105641987
>>105642014
I wonder if this is a quant issue. I downloaded this Q4_K_S from here https://huggingface.co/mradermacher/ICONN-1-GGUF/tree/main
Replies: >>105643760 >>105643786 >>105645026 >>105646451 >>105647404
Anonymous
6/19/2025, 9:41:14 PM No.105643681
>>105642583
As I understand it, DeepSeek is especially sensitive about having the first message as the user message. It's how the docs instruct to prompt the model
Replies: >>105645413
Anonymous
6/19/2025, 9:46:19 PM No.105643720
>>105643626
>links a study that says specific mental things can worsen already existing asthma problems to prove that... all cat alergies are just mental problems
Surely you must genuinely be special needs?
Replies: >>105643746
Anonymous
6/19/2025, 9:47:58 PM No.105643731
https://github.com/mirage-project/mirage/tree/mpk
https://zhihaojia.medium.com/compiling-llms-into-a-megakernel-a-path-to-low-latency-inference-cf7840913c17
Anonymous
6/19/2025, 9:49:20 PM No.105643743
>>105642983
>>105643733
not sure if nsfw
https://files.catbox.moe/56pjl9.jpg
Replies: >>105643857
Anonymous
6/19/2025, 9:49:26 PM No.105643746
>>105643720
Surely you do know what psychosomatic means in the first place? Take you final (You) and get back to your reddit shitplace.
Replies: >>105643826
Anonymous
6/19/2025, 9:50:06 PM No.105643760
>>105643676
SOVL
Anonymous
6/19/2025, 9:52:13 PM No.105643786
>>105643676
Okay but you are awake, right?
Anonymous
6/19/2025, 9:56:36 PM No.105643826
>>105643746
>Psychosomatic
>Of or relating to a disorder having physical symptoms but originating from mental or emotional causes.
>Pertaining to both the mind and the body.
Again, NPC child, how does the study that you posted about mental things influencing already existing physical alergy for people having asthma prove that cat alergies are not physical?
Notice how you are shitting and pissing yourself all over in multiple replies but you didn't actually engage with any of the points once, particularly:
>>105643328
>and notice how you couldnt respond to the actual core of the argument that it doesnt matter if its a drawing of a child because what you are attracted to is not the drawing but the child features of the drawn child, still making you a pedo

Your brain can't actually engage and has to smugpost hand wave dismiss things and cope with irelevant points of equivocation because it's in full damage control mode and cognitive dissonance. Just how you won't be able to engage with this post either and will also have to hand wave dismiss it.
Anonymous
6/19/2025, 9:59:05 PM No.105643857
r2
r2
md5: 35ab049908f51b3258345b83d01d20cc🔍
>>105643733
>>105643743
Replies: >>105643936 >>105643979
Anonymous
6/19/2025, 10:06:05 PM No.105643936
>>105643857
>fat
Anonymous
6/19/2025, 10:11:10 PM No.105643979
>>105643857
The writing is generated?
Anonymous
6/19/2025, 10:40:01 PM No.105644282
https://streamable.com/simohc
Replies: >>105644307 >>105644378 >>105644430
Anonymous
6/19/2025, 10:42:27 PM No.105644307
>>105644282
Did she died?
Replies: >>105644333
Anonymous
6/19/2025, 10:44:45 PM No.105644333
>>105644307
gotta get the fuwapuchi clean after sessions
Anonymous
6/19/2025, 10:49:10 PM No.105644378
>>105644282
Was this generated with Google Veo?
Replies: >>105644430 >>105644548
Anonymous
6/19/2025, 10:53:34 PM No.105644430
>>105644282
>>105644378
I choose to believe it's real.
Anonymous
6/19/2025, 11:03:50 PM No.105644548
>>105644378
>was this generated
into
>oh it's real? well I don't see how it's relevant to the thread
fuck off already
Replies: >>105644625
Anonymous
6/19/2025, 11:10:33 PM No.105644613
>>105641987
>nobody
>posts nothingburger
why should anychuddy care
Anonymous
6/19/2025, 11:11:27 PM No.105644625
>>105644548
What is this schizophrenia?
Replies: >>105644642 >>105644669
Anonymous
6/19/2025, 11:12:59 PM No.105644642
>>105644625
Its in your walls™
Anonymous
6/19/2025, 11:15:24 PM No.105644669
>>105644625
I'm in your walls
All of them
Anonymous
6/19/2025, 11:23:27 PM No.105644770
9xxs9d
9xxs9d
md5: 6f6bfb66b541b0480ebeaf7e9974f966🔍
Replies: >>105644800
Anonymous
6/19/2025, 11:27:01 PM No.105644800
>>105644770
Incredibly base
Anonymous
6/19/2025, 11:49:27 PM No.105644976
zzzzzzz_thumb.jpg
zzzzzzz_thumb.jpg
md5: 15d35ef0695d58f6f196a0ad67b9c9fc🔍
Replies: >>105645007
Anonymous
6/19/2025, 11:52:53 PM No.105645007
>>105644976
night night miku
Anonymous
6/19/2025, 11:54:49 PM No.105645026
>>105643676
anon it's time to wake up. please wake up.
Anonymous
6/20/2025, 12:00:39 AM No.105645079
>>105641756
>charging a small amount for the database download to cover server costs
>server costs
torrents are a thing...
Anonymous
6/20/2025, 12:40:02 AM No.105645413
>>105643681
For models that have trouble dealing with the first message being from the assistant, what I do is replace the template for past messages with something like "### User/Char:" and maybe modify the last instruction to instruct the model to write the next turn in the chat history. Rumor has it that many models perform better using this method anyway, likely due to flawed (or possibly even no) multiturn training.
Replies: >>105645528
Anonymous
6/20/2025, 12:40:57 AM No.105645419
what are we waiting now?
llama 5?
Replies: >>105645423 >>105645430 >>105645466 >>105645507 >>105645551 >>105645811 >>105645951 >>105649395 >>105649470 >>105649547 >>105650308
Anonymous
6/20/2025, 12:41:22 AM No.105645423
>>105645419
no we wait qwen, always
Anonymous
6/20/2025, 12:42:06 AM No.105645430
>>105645419
ernie 4.5 on the 30th
Anonymous
6/20/2025, 12:46:37 AM No.105645466
>>105645419
v4/w img out
god/allah/buddha/yahweh/yaldabeoth/hermes/etc willing
Anonymous
6/20/2025, 12:50:25 AM No.105645507
>>105645419
llama 4 behemoth needs to come out first before llama 5. maybe llama 4.1 thinking edition will redeem the llama 4 too
Replies: >>105645520
Anonymous
6/20/2025, 12:51:59 AM No.105645520
>>105645507
the only thing that will redeem llama now is full multimodal with image gen
uncensored
Replies: >>105645559
Anonymous
6/20/2025, 12:52:32 AM No.105645528
>>105645413
Wouldn't that put those models at a disadvantage?
Replies: >>105645701
Anonymous
6/20/2025, 12:54:36 AM No.105645551
>>105645419
bitnet or that new quant method that showed good performance at ~0.9bit
new models are all codebrained and benchmaxxed, I just want to run old 70b and 123b models.
Replies: >>105647207
Anonymous
6/20/2025, 12:55:34 AM No.105645559
>>105645520
then all hope is lost
Anonymous
6/20/2025, 1:13:51 AM No.105645701
>>105645528
I guess in comparisons. I'm just talking about personal usage.
Anonymous
6/20/2025, 1:19:16 AM No.105645748
>>105637275 (OP)
>https://rentry.org/lmg-lazy-getting-started-guide
I've read this but I'm having trouble with
>COOM your brains out
The moment it stops being PG13 characters just keep repeating themselves and start saying that they're about to start and asking if that's what I want. But they never actually do it.
Replies: >>105645803 >>105645812 >>105645830
Anonymous
6/20/2025, 1:28:00 AM No.105645803
>>105645748
model issue
anything smaller than mistral large is ass
althoughdoebeit try 1.5 temp 0.05minp
Replies: >>105645866
Anonymous
6/20/2025, 1:28:48 AM No.105645811
>>105645419
Only dense bitnet can save us now.
Anonymous
6/20/2025, 1:28:52 AM No.105645812
>>105645748
Small models need some wrangling if you notice them getting into loops. Edit the part before it happens.
Replies: >>105645866
Anonymous
6/20/2025, 1:31:29 AM No.105645830
>>105645748
What are your specs.
Replies: >>105645866
Anonymous
6/20/2025, 1:35:20 AM No.105645866
>>105645803
>>105645812
No luck, even in a new chat and with the temperature value tweaks they are still prudes.

>>105645830
i5-12400F
RTX 3080 Ti 12GB VRAM
64GB RAM
Replies: >>105645877 >>105645919
Anonymous
6/20/2025, 1:37:07 AM No.105645877
>>105645866
Which part of "edit the part" didn't you understand?
Replies: >>105645909
Anonymous
6/20/2025, 1:40:11 AM No.105645907
1743528282174258
1743528282174258
md5: 61981c08bca9f11a5f501eb6fc074ed0🔍
>>105637275 (OP)
Replies: >>105646050
Anonymous
6/20/2025, 1:40:14 AM No.105645909
>>105645877
Sorry, I'm a retard and thought you meant my messages (that didn't work), not the output. That worked out though! Thanks!
Anonymous
6/20/2025, 1:41:38 AM No.105645919
>>105645866
other anon meant as to edit it to nudge it into a state of something happening
'she starts to', 'she gets onto' 'she takes your' etc etc
also try diff cards, maybe the one youve written has a very passive starting message where nothing happens and it perpetuates the nothinghappening, throwing shit at the wall but i think it needs some examples on how to act since by default models are only really good at answering and going along with what you do
Anonymous
6/20/2025, 1:47:04 AM No.105645951
>>105645419
Diffusion titan bitnet
Anonymous
6/20/2025, 2:03:31 AM No.105646050
>>105645907
That's bad sleep hygiene, Miku
Anonymous
6/20/2025, 2:53:07 AM No.105646366
_3eb316a9-53e5-4a73-bcfe-54ca3888d0e8
_3eb316a9-53e5-4a73-bcfe-54ca3888d0e8
md5: 64ac7078239b51bba7804858e12ae501🔍
>>105642736
>https://files.catbox.moe/0g6m2r.jpg
Hate it when those sorts of fluids look like melted cheese, but yes, naughty Rin is hot.

I've replaced Nemo 12B with Gemma3 12B. Whatever slight loss in x-rated-ness there is, is greatly offset by it being much smarter and overall writing better.

BTW that's a fresh dalle Migu. I'm surprised it still can be fooled into something pretty good occasionally.
Replies: >>105646391 >>105647953
Anonymous
6/20/2025, 2:57:59 AM No.105646391
>>105646366
cute light inflatable migu.
Anonymous
6/20/2025, 3:00:12 AM No.105646406
1738636720907080
1738636720907080
md5: 1695b2e9418e2fa566be47bda51bf0b4🔍
hungry boyyyy :3
Replies: >>105646443
Anonymous
6/20/2025, 3:06:50 AM No.105646443
>>105646406
me filling your moms vram with my fat throbbing layer
Anonymous
6/20/2025, 3:08:11 AM No.105646451
>>105643676
I'm getting that as well. I think the model is just bad.
Anonymous
6/20/2025, 3:12:39 AM No.105646484
griftmaxx
griftmaxx
md5: e5852908e33f70fdcc5c6d8f7f77aadc🔍
>>105641987
Replies: >>105646525 >>105646530 >>105646747
Anonymous
6/20/2025, 3:18:35 AM No.105646525
>>105646484
It's 100% the former, clowncar MoE using keywords. They even say so in the card.
See >>105642104
Anonymous
6/20/2025, 3:19:05 AM No.105646530
>>105646484
>>105642109
Anonymous
6/20/2025, 3:29:54 AM No.105646613
It's actually bizarre how much changing the name of a character changes it's personality? Are the metaphysical concepts around names actually true? They seem separated into standard behaviors.
Replies: >>105646638 >>105646711 >>105646712 >>105646728
Anonymous
6/20/2025, 3:32:55 AM No.105646638
>>105646613
Believe it or not, many arbitrary things affect a person's personality without them being aware or having a choice in whether it affects them. Free will is a myth.
Replies: >>105646728
Anonymous
6/20/2025, 3:42:15 AM No.105646711
>>105646613
Because the model sees stuff in the training data describing certain things to certain names more. If a certain ethnic group behaves a certain way, it's going to influence stereotypical names from that group.
Replies: >>105646720
Anonymous
6/20/2025, 3:42:41 AM No.105646712
>>105646613
Nominative determinism
Anonymous
6/20/2025, 3:43:38 AM No.105646720
>>105646711
But even the names from same stereotypical group are markedly different. I'm surprised they have enough distance in the data on average to differentiate.
Anonymous
6/20/2025, 3:44:32 AM No.105646728
>>105646613
>>105646638
In my country if you are named angel there's an 80% chance you end up being gay or at the very least effeminate.
Anonymous
6/20/2025, 3:46:08 AM No.105646738
Why does https://huggingface.co/ICONNAI/ICONN-1 404 now?
Replies: >>105646752 >>105646807 >>105648123 >>105648502
Anonymous
6/20/2025, 3:47:19 AM No.105646747
>>105646484
top geg it's a brown tier grift
Anonymous
6/20/2025, 3:48:10 AM No.105646752
>>105646738
Got laughed out.
Replies: >>105646807
Anonymous
6/20/2025, 3:56:47 AM No.105646807
>>105646738
Because >>105641756
>We'll be offering the system and models for free, and charging a small amount for the database download to cover server costs until we get VC funding. But all of it will be open source and free to share.
>until we get VC funding
Coupled with >>105646752
Replies: >>105646832 >>105647136
Anonymous
6/20/2025, 3:59:59 AM No.105646832
>>105646807
i think thats a different grift
iconn smells of an indian using chatgpt for big words
while the rag one is kinda more realistically boring
Anonymous
6/20/2025, 4:00:21 AM No.105646836
>>105637531
It should be useful, yes, under the same conditions: draft model much faster, and decent chance of draft model predicting right.

The mechanism of speculation is to introduce parallelism to spread out the cost of pushing the weights through the memory bus, which is of course *the* bottleneck.

If you have sequences A, B, and C, and you want the next token for each of them, then for each chunk of the weights, you can do those weights' calculations for all 3 at once, only loading once. The trick with speculative decoding is realizing that if your draft model has produced "shivers down her", there's no reason you can't treat "shivers", "shivers down", and "shivers down her" as your sequences A B and C, and have your main model predict the next token for each. As far as it agrees with the draft model, you can keep those tokens, and it only cost you 1x token gen (plus generating all the draft tokens).

IIUC this makes speculative decoding a trade-off/ substitute with multi-user batching, since each step in the speculation acts as one user.
Anonymous
6/20/2025, 4:16:03 AM No.105646946
Does your favourite model know what "zogcog" means? It's my goto test after the "mesugaki" test.
Anonymous
6/20/2025, 4:41:42 AM No.105647136
>>105646807
I got all the safetensors before it got nuked. Anyone know where to get the json files?
Replies: >>105649543
Anonymous
6/20/2025, 4:50:40 AM No.105647207
>>105645551
>I just want to run old 70b and 123b models
based
Anonymous
6/20/2025, 5:03:16 AM No.105647290
>>105641671
https://www.youtube.com/watch?v=WBm0nyDkVYM
Anonymous
6/20/2025, 5:18:28 AM No.105647404
>>105643676
>"Are you sure you're not in a nightmare?"
Anonymous
6/20/2025, 6:28:40 AM No.105647919
When do you think there will be actual intelligence?
Replies: >>105647998 >>105648053
Anonymous
6/20/2025, 6:33:43 AM No.105647953
>>105646366
Gemma is better than nemo? What about the 27b? What sort of settings & format work best? Last time I tried I wasn't that impressed.
Replies: >>105648487
Anonymous
6/20/2025, 6:42:04 AM No.105647998
rWA-3_T7_400x400
rWA-3_T7_400x400
md5: 23b0e66b074e7038473db8331bf6b8b0🔍
>>105647919
Never.
Replies: >>105648053 >>105648978 >>105648985 >>105649063
Anonymous
6/20/2025, 6:52:03 AM No.105648053
>>105647919
Define intelligence. Are crows intelligent? What about ants? We can probably simulate ants
>>105647998
Shut up, cat fucker
Anonymous
6/20/2025, 7:03:52 AM No.105648123
>>105646738
probably also because of this:
https://huggingface.co/bartowski/ICONNAI_ICONN-1-GGUF/discussions/1
Replies: >>105648205 >>105648294 >>105648502
Anonymous
6/20/2025, 7:19:30 AM No.105648205
>>105648123
>Woof
What did he mean by this.
Anonymous
6/20/2025, 7:35:33 AM No.105648294
>>105648123
>tricked me into downloading a memekit
>I actually liked it somewhat
Makes sense if it's just a Mistral graft. It loaded in kobold right away, which is absurd if it's an actually new architecture. Oh well, some variety won't hurt from time to time. It kinda didn't do any of the typical Mistral slops (probably because I have banned over 50 at this point).
Anonymous
6/20/2025, 7:59:46 AM No.105648487
>>105647953
Gemma is actually great for sfw but it is pretty awful at writing anything adult in nature, mostly because it can't take make characters take initiative (in my experience). You don't even have to use any samplers, a very basic system prompt with the gemma chat template will work. I'm talking about the 27b though, I haven't used the 12b.
Replies: >>105648814
Anonymous
6/20/2025, 8:01:28 AM No.105648502
ICONNAI-ICONN-1
ICONNAI-ICONN-1
md5: 677ee9f406c4a435d03fce9f2978e23b🔍
>>105646738
>>105648123
Quick, copy paste, out of chronological order collage of the funniest posts surrounding this that I just collected
Replies: >>105648535 >>105648580 >>105649878
Anonymous
6/20/2025, 8:05:36 AM No.105648528
>the actually good models are still stuck at 4k context
Anonymous
6/20/2025, 8:06:26 AM No.105648535
>>105648502
I guess the fact their rhetoric was a bit unhinged and they made straight up jokes and very weird statements on their release page was a giveaway. Some kind of a bizarre "social experiment" to see if they could last a day scamming everyone to prove that the AI industry is all grifters or something. Also:
>I con
>>105642091
Replies: >>105648827
Anonymous
6/20/2025, 8:13:12 AM No.105648580
>>105648502
I miss this format
Anonymous
6/20/2025, 8:53:04 AM No.105648795
>>105637564
you can try nvidia models like _Llama-3_3-Nemotron-Super-49B-v1

I havent seen any sloptunes of it since it is itself a bit of a sloptune. I found it a bit too rigid as is. Maybe stuff like skyfall 36b
Anonymous
6/20/2025, 8:54:56 AM No.105648806
download (1)
download (1)
md5: 1f4001b0b4bdfa7c675cef564561aeb8🔍
Perchance isn't totally shit it seems
Replies: >>105650059 >>105651432
Anonymous
6/20/2025, 8:55:57 AM No.105648814
>>105648487
Initiative can be improved to some extent by prefilling model responses with a short <think> section where the model reminds itself to be more proactive; it looks as if Gemma 3 was partially trained for reasoning but that didn't get fleshed out at least for this version.
Anonymous
6/20/2025, 8:57:34 AM No.105648827
>>105648535
From his postings, I don't think the icon author is older than 18.
Anonymous
6/20/2025, 8:58:55 AM No.105648833
are imatrix quants always better than static?
Anonymous
6/20/2025, 9:21:49 AM No.105648978
1750315806887848
1750315806887848
md5: ea8dda28b889cee7be445ce459b40532🔍
>>105647998
based
Replies: >>105649063
Anonymous
6/20/2025, 9:23:07 AM No.105648985
>>105647998
always knew he was right, its LLMover..
Anonymous
6/20/2025, 9:34:20 AM No.105649063
>>105648978
>>105647998
but we started from ai being able to code for 0h, then 1 short script, and we progressed until 1 whole hour, so lecunt is proven retarded yet again (daily example)
Anonymous
6/20/2025, 10:28:17 AM No.105649395
>>105645419
Gemma 4 for me but it's going to be a whole another year before Google does it given how long they waited even after they got lapped by Chinese models and the only reason they did it was to preempt Llama 4 from stealing their publicity crown at lower sizes and Aya was nipping at their toes for multilingual benchmarks. Hoping they do MOE for the next model too but probably unlikely.
Anonymous
6/20/2025, 10:43:37 AM No.105649470
1750376592316723
1750376592316723
md5: a0bec5decacb8d6c8ecab1fd84551e2b🔍
>>105645419
A miracle. It happened with Mixtral, it happened with Deepseek. It will happen again
Anonymous
6/20/2025, 10:55:35 AM No.105649543
>>105647136
you mean tokenizer.json? i think i downloadrd it and left it on my desktop but not the config
Anonymous
6/20/2025, 10:56:07 AM No.105649547
>>105645419
I still expect a "surprise" Mistral Medium release within 2 weeks in the form of Mistral-Nemotron. Most people won't be able to run it and the model will be good in some ways and bad in others because of NVidia's mathmaxxed (with a sprinkle of safety) Nemotron dataset.
Anonymous
6/20/2025, 10:57:22 AM No.105649554
Gt2TRinXIAAZjSW
Gt2TRinXIAAZjSW
md5: 2bbeb4acad40dff267e23567a04c9a44🔍
Anonymous
6/20/2025, 11:50:58 AM No.105649878
>>105648502
SAAARS our response?
Anonymous
6/20/2025, 11:52:12 AM No.105649887
What models should I run on a macbook pro 16 M4 ? For programming primarily
Replies: >>105649938
Anonymous
6/20/2025, 11:59:07 AM No.105649926
>The same post of yesterday
/lmg/ is dying
Replies: >>105649956
Anonymous
6/20/2025, 12:00:51 PM No.105649938
>>105649887
depends on the language and how much ram you've got.
Anonymous
6/20/2025, 12:03:39 PM No.105649956
>>105649926
You should be taking advantage of the quiet while we wait for the next batch of releases by catching up on ai literature.
Anonymous
6/20/2025, 12:16:33 PM No.105650038
file
file
md5: 746cff70855aef575bbc8fa754b7b3a5🔍
>>105642067
>>105642286
>>105642736
>
Replies: >>105650213
Anonymous
6/20/2025, 12:20:15 PM No.105650059
>>105648806
Why are you gay?
Replies: >>105650128 >>105650256
Anonymous
6/20/2025, 12:20:27 PM No.105650060
https://edition.cnn.com/2025/06/18/tech/meta-openai-sam-altman-100-million
wtf?
Replies: >>105650276 >>105650323
Anonymous
6/20/2025, 12:31:57 PM No.105650128
>>105650059
>he doesn't like the cock
who is going to tell him
Replies: >>105650256
Anonymous
6/20/2025, 12:46:41 PM No.105650213
file
file
md5: c21fdc0f64fed763a8c278ebec49c037🔍
>>105650038
Anonymous
6/20/2025, 12:56:07 PM No.105650256
1657980999994
1657980999994
md5: 63487f078600e5ee79535bf544b430d4🔍
>>105650128
me
>>105650059
Anonymous
6/20/2025, 12:58:53 PM No.105650276
>>105650060
>join meta for $100m
>put your feet up
>produce absolute shit
>zucc will peddle it anyway
perfect deal
Anonymous
6/20/2025, 1:05:07 PM No.105650308
>>105645419
logically, llama4.1 is next trained by the new team.
Anonymous
6/20/2025, 1:08:05 PM No.105650323
>>105650060
>“There’s many things I respect about Meta as a company, but I don’t think they’re a company that’s great at innovation,” Altman continued. “I think we understand a lot of things they don’t.”
He is afraid
Replies: >>105650332
Anonymous
6/20/2025, 1:09:12 PM No.105650332
>>105650323
This is one of the few statements I agree with Altman seeing how the metaverse and llama are going
Replies: >>105650478
Anonymous
6/20/2025, 1:16:06 PM No.105650377
meta is a rudderless company always looking for a next big thing that won't happen
LLMs are not what meta ever needed in any way, shape or form
it won't become AGI and a code assistant is not what is going to help their retarded dying social network
in fact the AI slop has been killing it even harder not even boomers want to see more of that shrimp jesus
Replies: >>105650431 >>105650635
Anonymous
6/20/2025, 1:25:38 PM No.105650431
>>105650377
What makes you think that his personally-appointed "Superintelligence" team (which might include LeCun) is going to make yet another LLM?
Replies: >>105650550
Anonymous
6/20/2025, 1:34:52 PM No.105650478
>>105650332
Meta is by far the most changified company, more than 40% of workers are asians.
No shit, they are bad at innovating. All changs know is how to copy,
Replies: >>105650486
Anonymous
6/20/2025, 1:36:16 PM No.105650486
>>105650478
WTF are you talking about, Meta is literally trying to copy Deepseek and failing at that.
Replies: >>105650500
Anonymous
6/20/2025, 1:39:11 PM No.105650500
>>105650486
And Deepseek copied from OAI.
facebook changs are not on the level of deepseek changs.
Replies: >>105650546 >>105650871
Anonymous
6/20/2025, 1:50:32 PM No.105650546
>>105650500
Copied what?
Replies: >>105650566 >>105650587
Anonymous
6/20/2025, 1:51:30 PM No.105650550
>>105650431
>LeCun
lol
LeCun makes valid points that LLM won't become AGI but that's all he can do
that nigger has never made a single useful thing ever
LLMs won't be AGI like he said, but unlike what he produces (hot air) they can have actual uses in the real world.
Anonymous
6/20/2025, 1:53:12 PM No.105650557
https://news.ycombinator.com/item?id=44273776
>facebook employee talking about lecunn
>>FYI if you worked at FB you could pull up his WP and see he does absolutely nothing all day except link to arxiv.
Replies: >>105650694 >>105650702
Anonymous
6/20/2025, 1:54:43 PM No.105650566
>>105650546
Everyone knows what they took even if there's no proof
Replies: >>105650677
Anonymous
6/20/2025, 1:57:48 PM No.105650587
>>105650546
model distillation
they did the same thing to gemini with the new R1 its reasoning traces are very similar to what you used to see on gemini before google decided to hide the CoT through a shitty summarizer
Replies: >>105650871
Anonymous
6/20/2025, 2:06:03 PM No.105650635
>>105650377
remember when meta said they were planning to start replacing some of their engineer with ai this year?
i sure hope they aren't planning to use their own llama models
though who knows it might still end up being an improvement
Anonymous
6/20/2025, 2:09:45 PM No.105650677
2036926481
2036926481
md5: 914fde295cf45f7eee51e085cb3e8b79🔍
>>105650566
> there's no proof
> Everyone knows
uh huh
Replies: >>105650897
Anonymous
6/20/2025, 2:12:20 PM No.105650694
>>105650557
based
Anonymous
6/20/2025, 2:12:40 PM No.105650696
is aicharactercards.com the civitai of text ai?
Anonymous
6/20/2025, 2:14:17 PM No.105650702
>>105650557
What's a WP?
Replies: >>105650832
Anonymous
6/20/2025, 2:32:28 PM No.105650830
Imagen is fucking crazy. Feels weird using a model that actually does what you prompt it to do without 8 million tags
Anonymous
6/20/2025, 2:32:29 PM No.105650832
>>105650702
It stands for workplace profile. Workplace is their internal social network, occasionally making the news because their employees like to say things there that gets them fired:
https://www.cnbc.com/2020/09/17/facebook-issues-new-rules-on-internal-employee-communication-.html
>This week, BuzzFeed reported a post by a fired Facebook data scientist who posted to Workplace a memo outlining how the company failed to act on election interference happening around the world through the social network.
Anonymous
6/20/2025, 2:37:06 PM No.105650871
>>105650500
>Deepseek copied from OAI.
OpenAI hid their thinking outputs, the wait wait but wait slop was all deepseek.>>105650587
>they did the same thing to gemini with the new R1
There were no thinking traces for o1.
Replies: >>105650898
Anonymous
6/20/2025, 2:40:42 PM No.105650897
>>105650677
>China known for stealing IP
>Comes up with a clone of GPT out of nowhere
>Where's your proof bro
Replies: >>105650951 >>105650965 >>105650966 >>105651058 >>105652630
Anonymous
6/20/2025, 2:40:50 PM No.105650898
>>105650871
>There were no thinking traces for o1.
I never said anything about o1 though. I said they copied OAI.
DeepSeek V3 was a distill of GPT 4, and the original R1 is indeed their own abomination with endless meandering.
Anonymous
6/20/2025, 2:42:16 PM No.105650909
btw R1's thinking is more obnoxious than useful, 999999% of what makes that model good is what was already trained in V3.
Replies: >>105652428
Anonymous
6/20/2025, 2:47:51 PM No.105650951
>>105650897
>stealing IP
So..? It's neither personal nor easily identifiable information that literally every fucking service stores.
Replies: >>105651156
Anonymous
6/20/2025, 2:49:41 PM No.105650965
>>105650897
If they stole from gpt deepseek wouldn't be as good as it is.
Anonymous
6/20/2025, 2:49:43 PM No.105650966
>>105650897
You say that like making a complex LLM that performs on the same level, but with less restrictions for the consumer is the same as making knock off marvel merch
Replies: >>105651156
Anonymous
6/20/2025, 2:53:29 PM No.105651058
>>105650897
They must be newfags, we can all remember the massive shift in the way LLMs speak in general after people started compiling massive datasets of GPT conversations
it was the chinese national sport to benchmax on this whether it's deepseek, openchat, xwin etc all claiming to do better than GPT while training on GPT output lolmao
Replies: >>105651085
Anonymous
6/20/2025, 2:57:27 PM No.105651085
>>105651058
Deepseek doesn't speak like chatgpt. You should try the model instead of shitposting all day
Replies: >>105651166
Anonymous
6/20/2025, 3:06:13 PM No.105651156
>>105650966
You are severely retarded.
>>105650951
Obviously I'm talking about their processes and code. This is something China does with everything: chips, airplanes, missiles. It's no different than how they steal data from Lockheed and NG. Someone at OpenAI gets a fucky sucky at the local massage parlor.
>Ooooh u such a big sexy man. You make the AI :O
After a couple of months they give an external hard drive to her handler in exchange for a few million
Replies: >>105651656
Anonymous
6/20/2025, 3:07:59 PM No.105651166
>>105651085
you are the one who should try the model
you're probably one of those retarded gooners who never used anything other than r1 and didn't even know deepseek existed before the media craze for it
the original v3 had many tell tale signs but you won't know them if you don't bother downloading the original release and if you don't have the computer to run it
Replies: >>105651222
Anonymous
6/20/2025, 3:15:07 PM No.105651218
plunder from already trained models
plunder from already trained models
md5: 224ecedbabd4fbda65ddf705dcdffc44🔍
I'm trying to make a locally run AI model for my brother
His usecases are:
>Analyse old legal cases(~10 years) so he can check them easier to avoid contradictory statesments
>Multiple pdfs at once but speed isn't an issue
>Preferably accessible from his laptop
>Preferably includes/has access to picture to pdf conversion tool
I have no coding experience, asked for what equipment i should buy and was told to make a local model first. So far im setting up a ollama on my PC and I'm going to try ssh into it from his laptop to make a basic LM
>How do i turn this to what he actually wants?
Replies: >>105651317 >>105651337 >>105651458
Anonymous
6/20/2025, 3:15:30 PM No.105651222
>>105651166
>the original v3 had many tell tale signs
Such as?
Replies: >>105651229 >>105651236
Anonymous
6/20/2025, 3:16:22 PM No.105651229
>>105651222
it spoke english
Anonymous
6/20/2025, 3:17:20 PM No.105651236
such_as_entire_sentence_structure_being_gptlike_compared_to_other_models
>>105651222
retard
Replies: >>105651324
Anonymous
6/20/2025, 3:29:13 PM No.105651317
>>105651218
>How do i turn this to what he actually wants
Have him talk to the local LLM and help refine and elaborate on his specific use case. Then use that conversation as the basis to start actually coding.
Good luck anon.
Replies: >>105651582
Anonymous
6/20/2025, 3:30:06 PM No.105651324
>>105651236
Grok is the only response that isn't annoying to read.
Anonymous
6/20/2025, 3:31:35 PM No.105651337
>>105651218
>How do i turn this to what he actually wants?
Tell him he's a retard and point him to hosted APIs since his laptop's not going to cut it.
Or give him a bill for an R1 capable server. He can pay $4,000 and wait a day for every result with a CPU based server. He can even remote in on his laptop.
Or pay ~$200K for a server that will run full quant at speed.
Replies: >>105651582
Anonymous
6/20/2025, 3:43:24 PM No.105651432
>>105648806
>Perchance
This site shills in 4chan almost as hard as NovelAI.
Anonymous
6/20/2025, 3:46:17 PM No.105651458
>>105651218
Download Claude desktop and give it access to your files. ezpz. Sonnet is probably more than enough.
Replies: >>105651503 >>105651582
Anonymous
6/20/2025, 3:53:26 PM No.105651503
>>105651458
Good luck getting Anthropic models to help with anything legal lol.
Replies: >>105651582
Anonymous
6/20/2025, 4:03:18 PM No.105651582
Cave japanese
Cave japanese
md5: 9a9d11a68e9964e68820b462fa29d510🔍
>>105651317
Sorry but i can't code, is there a handy script site i can steal from or should i just look up stack overflow?
>>105651337
Time isn't really important since
>Confidentiality
Also for the test run, i want his laptop to access my computer, my computer runs the AI and he accesses it with a webUI or somesuch, how is that resource heavy, i know its shit and wrong half the time but i already run sub 7B models on the background with barely any resource usage, can i even run the 100B+ models on my SSD with 40Gbs of space?
>Gtx 1060 6GB
>Ryzen 5 3600
>32GB 3600Mhz Cl18
>M.2 SSD
>>105651458
Thanks
>>105651503
God no, he wants this to make his job shorter, not doing it itself. It's a pain in the ass to read everything but easier to look up what the AI said on the document,what you know about the case and then glance at the whole to do it
>4 hour job to 30 min job
Roughly, but I honestly don't know that many details desu
Replies: >>105651867 >>105652160
Anonymous
6/20/2025, 4:12:38 PM No.105651656
>>105651156
>He still believes that the US is ahead of China in any tech field
Replies: >>105651735 >>105652272
Anonymous
6/20/2025, 4:21:33 PM No.105651735
>>105651656
oh, did china invent an actually working EUV machine that can be used for mass production of chips?
(in b4 "ASML is dutch" : ASML is built entirely from US IP/research)
Replies: >>105651794 >>105652048 >>105652061
Anonymous
6/20/2025, 4:29:44 PM No.105651794
>>105651735
>he thinks that 'chips' are the most important part of AI
Replies: >>105651885
Anonymous
6/20/2025, 4:36:38 PM No.105651867
>>105651582
Did you look at the build guides in the OP?
For a decent AI you're going to need at least 128GB, preferably closer to 1TB.
At least the idea of running on a server and connecting to that via his laptop has occurred to you. That's the only way any laptop is going to be useful.
Anonymous
6/20/2025, 4:38:53 PM No.105651885
>>105651794
if we only think about the software stack it's even simpler
Gemini 2.5 Pro mogs anything chinese
this is so self evident if you actually used those models for something other than cooming and saw how good it is at ingesting large context
Replies: >>105651926
Anonymous
6/20/2025, 4:44:31 PM No.105651926
>>105651885
>Gemini 2.5 Pro
Not that anon, but what the fuck are they doing to poor gemini.
Each release after gemini-2.5-pro-preview-03-25 is worse than the last.
The latest one can't even keep up with complex format instructions that 03-25 did effortlessly.
Please, google, don't fucking make Gemini shit. It's my go to not-local model.
Replies: >>105652046 >>105652265
Anonymous
6/20/2025, 4:52:49 PM No.105651995
Best RP models mistral large size and below?
Anonymous
6/20/2025, 4:58:51 PM No.105652034
is there anywhere an RP leaderboard for local models? there used to be one but it got taken down a while back
Replies: >>105652087
Anonymous
6/20/2025, 5:00:48 PM No.105652046
>>105651926
It recently got its final release. They won't mess around with it much now. Preview releases are always subject to changes.
Can't speak for your issues, personally I haven't felt the model got worse, but YMMV.
Anonymous
6/20/2025, 5:01:07 PM No.105652048
Gt06JlXXsAAif6I
Gt06JlXXsAAif6I
md5: 246868d2a69401d475d4a0b32526592b🔍
>>105651735
Yes
Anonymous
6/20/2025, 5:04:36 PM No.105652061
>>105651735
US Intellectual Not Real Property and research is built entirely by Eurasians
Replies: >>105652081
Anonymous
6/20/2025, 5:08:59 PM No.105652081
>>105652061
if that is what helps you sleep at night (why do you think the US gets to dictate who ASML can sell their devices to? They actually wanted to sell to China but the US told them to eat shit)
Anonymous
6/20/2025, 5:09:38 PM No.105652087
>>105652034
The closest there is is the nala test, look for it in the atchive.
Anonymous
6/20/2025, 5:11:44 PM No.105652099
Who was the original anon that came up with the nala test anyway? Is he still here?
Replies: >>105652183
Anonymous
6/20/2025, 5:21:10 PM No.105652160
>>105651582
>is there a handy script site i can steal from
Ask Claude what vibe-programming is, and you will be fine.
Anonymous
6/20/2025, 5:24:20 PM No.105652183
>>105652099
unrelated but i was wondering how the aah aah mistress meme originated
i remember the screenshot but i dont have it saved anymore
Replies: >>105652243 >>105652299
Anonymous
6/20/2025, 5:31:33 PM No.105652243
>>105652183
It originated in /aicg
Replies: >>105652259
Anonymous
6/20/2025, 5:33:38 PM No.105652259
>>105652243
anyone has the og screenshot?
Anonymous
6/20/2025, 5:34:15 PM No.105652265
>>105651926
All of the big AI companies are out of ideas. That's why they're memeing MCP so hard. They can only optimize their models and not holistically improve them
Anonymous
6/20/2025, 5:35:19 PM No.105652272
>>105651656
Cope BRICS untermensch.
Anonymous
6/20/2025, 5:39:16 PM No.105652299
>>105652183
https://desuarchive.org/g/thread/91897528/#91899750
Replies: >>105652316
Anonymous
6/20/2025, 5:41:00 PM No.105652316
>>105652299
peak thanks anon
Anonymous
6/20/2025, 5:42:26 PM No.105652325
chrome_a6fqXMYDWF
chrome_a6fqXMYDWF
md5: efa452330ef96e4a1ed086a9879cca90🔍
Why does mistral small give in at the first reply even though it's explicitly told not to?
Replies: >>105652348 >>105652363 >>105652386 >>105652424 >>105652435 >>105652486 >>105652534
Anonymous
6/20/2025, 5:47:28 PM No.105652348
>>105652325
You are expecting way too much out of this small model with such a large card.
Anonymous
6/20/2025, 5:49:02 PM No.105652363
>>105652325
This is depressing. Fix yourself faggot.
Anonymous
6/20/2025, 5:52:08 PM No.105652386
1749701744974222
1749701744974222
md5: 7dc1cca99640f5f08735a9ea5fb15732🔍
>>105652325
Picrel
Replies: >>105652420
Anonymous
6/20/2025, 5:52:33 PM No.105652390
it's always the worst degenerates that are into text gen for cooming, I notice
Replies: >>105652432
Anonymous
6/20/2025, 5:56:57 PM No.105652420
>>105652386
>Skill issue
Think he should download more RAM?
Replies: >>105652572
Anonymous
6/20/2025, 5:57:11 PM No.105652424
>>105652325
That smells like something else is wrong with the prompt, since it breaks format immediately.
Paste the full prompt silly sent the backend into a pastebin and post the link.
Anonymous
6/20/2025, 5:57:47 PM No.105652428
>>105650909
Or I can just use R1 and just use a prefill whenever I don't want it to think. And then when I actually do need it to think I can just remove the prefill. No need to use different models.
Anonymous
6/20/2025, 5:58:06 PM No.105652432
>>105652390
There's nothing wrong with princesses living their best life and getting addicted to minotaur cum.
Anonymous
6/20/2025, 5:58:20 PM No.105652435
>>105652325
Why are you posting a November 2024 screenshot?
Anonymous
6/20/2025, 6:03:37 PM No.105652486
>>105652325
Mehmet my son...
Anonymous
6/20/2025, 6:08:06 PM No.105652534
>>105652325
Try Mistral Small 3.2
https://huggingface.co/mistralai/Mistral-Small-3.2-24B-Instruct-2506
Replies: >>105652546 >>105652606
Anonymous
6/20/2025, 6:10:00 PM No.105652546
>>105652534
wtf
Anonymous
6/20/2025, 6:11:10 PM No.105652552
https://x.com/MistralAI/status/1936093325116781016
>Introducing Mistral Small 3.2, a small update to Mistral Small 3.1 to improve:
>- Instruction following: Small 3.2 is better at following precise instructions
>- Repetition errors: Small 3.2 produces less infinite generations or repetitive answers
>- Function calling: Small 3.2's function calling template is more robust
Replies: >>105652584
Anonymous
6/20/2025, 6:13:35 PM No.105652572
>>105652420
Into his tiny little brain maybe, it's an operator's skill issue
Anonymous
6/20/2025, 6:14:46 PM No.105652584
>>105652552
post the hf link, dumbass
https://huggingface.co/mistralai/Mistral-Small-3.2-24B-Instruct-2506
Replies: >>105652589 >>105652606 >>105652642
Anonymous
6/20/2025, 6:15:30 PM No.105652589
>>105652584
it already was
Anonymous
6/20/2025, 6:16:41 PM No.105652606
>>105652584
Already posted here >>105652534
Anonymous
6/20/2025, 6:18:32 PM No.105652630
>>105650897
>Scrape the entire Internet to train your LLM
>Cry about people training on your LLM's outputs
get fucked rat jew
Anonymous
6/20/2025, 6:19:29 PM No.105652642
>>105652584
you know for being a general that is supposedly all about reading shit generated by AI, none of you actually read
Anonymous
6/20/2025, 6:20:15 PM No.105652649
>>105652633
>>105652633
>>105652633