Search Results
7/26/2025, 5:21:33 PM
►Recent Highlights from the Previous Thread: >>106022725
--Confidential computing may protect against providers but not against corporations or determined attackers:
>106032051 >106032109 >106032125 >106032847 >106032906 >106032943 >106033227
--Pretraining data augmentation and synthetic recycling as emerging industry standards:
>106032565 >106032668 >106032775 >106033083 >106033222 >106033274
--Optimizing 24B model inference on 16GB GPU via tensor offloading and quantization:
>106030678 >106030699 >106030726 >106030746 >106030888 >106030911 >106030936 >106030952 >106031007 >106031309 >106031127 >106031129
--Meta AI's leaked system prompt reveals poor prompt engineering with excessive negative directives:
>106030482 >106033266 >106033329 >106033349 >106033368 >106033422
--LLMs as unreliable standalone tools but useful when integrated with traditional systems:
>106025198 >106025285 >106025482 >106025526 >106025721 >106025949 >106026069 >106028082
--ST fails to remove thinking blocks starting with parentheses from context:
>106028030 >106028084 >106028130 >106029199 >106029213 >106029266
--Anon builds minimal terminal interface for LLMs:
>106024282 >106024421 >106024944 >106024758
--Intern-S1 is a Qwen3 fine-tune with 235B MoE architecture and multimodal support:
>106031261 >106031267 >106031296 >106031307 >106031358 >106031277 >106031280
--Small 350M model shows strong coherence with custom tokenizer:
>106034024 >106034064 >106034083 >106034143 >106034205
--Challenges in building persistent, stat-aware LLM-driven RPGs locally:
>106033668 >106033772 >106033817
--Anon shares thoughts on NuQwen Thinker for RP:
>106024560 >106024601 >106024668
--Links:
>106033095 >106023407 >106029240 >106023197
--Miku (free space):
>106022834 >106022983 >106023491 >106026316 >106029061 >106029200 >106030177 >106031397
►Recent Highlight Posts from the Previous Thread: >>106022743
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
--Confidential computing may protect against providers but not against corporations or determined attackers:
>106032051 >106032109 >106032125 >106032847 >106032906 >106032943 >106033227
--Pretraining data augmentation and synthetic recycling as emerging industry standards:
>106032565 >106032668 >106032775 >106033083 >106033222 >106033274
--Optimizing 24B model inference on 16GB GPU via tensor offloading and quantization:
>106030678 >106030699 >106030726 >106030746 >106030888 >106030911 >106030936 >106030952 >106031007 >106031309 >106031127 >106031129
--Meta AI's leaked system prompt reveals poor prompt engineering with excessive negative directives:
>106030482 >106033266 >106033329 >106033349 >106033368 >106033422
--LLMs as unreliable standalone tools but useful when integrated with traditional systems:
>106025198 >106025285 >106025482 >106025526 >106025721 >106025949 >106026069 >106028082
--ST fails to remove thinking blocks starting with parentheses from context:
>106028030 >106028084 >106028130 >106029199 >106029213 >106029266
--Anon builds minimal terminal interface for LLMs:
>106024282 >106024421 >106024944 >106024758
--Intern-S1 is a Qwen3 fine-tune with 235B MoE architecture and multimodal support:
>106031261 >106031267 >106031296 >106031307 >106031358 >106031277 >106031280
--Small 350M model shows strong coherence with custom tokenizer:
>106034024 >106034064 >106034083 >106034143 >106034205
--Challenges in building persistent, stat-aware LLM-driven RPGs locally:
>106033668 >106033772 >106033817
--Anon shares thoughts on NuQwen Thinker for RP:
>106024560 >106024601 >106024668
--Links:
>106033095 >106023407 >106029240 >106023197
--Miku (free space):
>106022834 >106022983 >106023491 >106026316 >106029061 >106029200 >106030177 >106031397
►Recent Highlight Posts from the Previous Thread: >>106022743
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
Page 1