Search Results
6/19/2025, 5:23:47 AM
►Recent Highlights from the Previous Thread: >>105621559
--Paper: Suppressing redundant thinking tokens improves model reasoning efficiency:
>105621964 >105621977 >105621987 >105622021 >105622979 >105622068 >105622075 >105623115 >105632679 >105632813 >105633018 >105633139 >105633190 >105634335 >105633792 >105622018
--Critique of NVIDIA DGX Spark pricing and V100 hardware tradeoffs:
>105630545 >105630697 >105630851 >105630881 >105630987 >105630863 >105630807 >105631166 >105631211 >105631542 >105631723 >105632364 >105631761 >105635125 >105635158 >105635286 >105635459 >105635500 >105635538 >105635638 >105635644 >105635677 >105637100
--Anxiety over AI-generated language corrupting training data:
>105626238 >105626258 >105626875 >105628083 >105626265 >105626301 >105626527 >105627449 >105627036 >105627482 >105627881 >105628432
--llama.cpp vs vLLM performance differences and local model effectiveness in code-assist tools:
>105624044 >105624247 >105624310 >105624878 >105624985 >105625733 >105626017 >105626049 >105626061 >105626850
--Gemini 2.5 Pro highlights multimodal capabilities and in-house TPU training with agentic features:
>105624610 >105624725 >105628988 >105624980 >105634689
--Skepticism around Arcee's new models' originality and performance:
>105632818 >105632884 >105632895 >105633081 >105633840 >105633898 >105634479 >105633986 >105634582
--Comically slow inference due to hddmaxxing and waiting on RAM upgrades:
>105630585 >105630757 >105630798 >105631027
--Building a 123B model-capable rig with 4x3090:
>105630142 >105630262 >105630325 >105631297 >105630328 >105630531 >105631152 >105635155
--Personalized speech-to-text tools for quick transcription with shortcut triggers:
>105627335 >105627797
--Teto and Miku and Rin (free space):
>105621874 >105622071 >105625804 >105626952 >105630546 >105636047 >105636052 >105636268 >105636665
►Recent Highlight Posts from the Previous Thread: >>105621564
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
--Paper: Suppressing redundant thinking tokens improves model reasoning efficiency:
>105621964 >105621977 >105621987 >105622021 >105622979 >105622068 >105622075 >105623115 >105632679 >105632813 >105633018 >105633139 >105633190 >105634335 >105633792 >105622018
--Critique of NVIDIA DGX Spark pricing and V100 hardware tradeoffs:
>105630545 >105630697 >105630851 >105630881 >105630987 >105630863 >105630807 >105631166 >105631211 >105631542 >105631723 >105632364 >105631761 >105635125 >105635158 >105635286 >105635459 >105635500 >105635538 >105635638 >105635644 >105635677 >105637100
--Anxiety over AI-generated language corrupting training data:
>105626238 >105626258 >105626875 >105628083 >105626265 >105626301 >105626527 >105627449 >105627036 >105627482 >105627881 >105628432
--llama.cpp vs vLLM performance differences and local model effectiveness in code-assist tools:
>105624044 >105624247 >105624310 >105624878 >105624985 >105625733 >105626017 >105626049 >105626061 >105626850
--Gemini 2.5 Pro highlights multimodal capabilities and in-house TPU training with agentic features:
>105624610 >105624725 >105628988 >105624980 >105634689
--Skepticism around Arcee's new models' originality and performance:
>105632818 >105632884 >105632895 >105633081 >105633840 >105633898 >105634479 >105633986 >105634582
--Comically slow inference due to hddmaxxing and waiting on RAM upgrades:
>105630585 >105630757 >105630798 >105631027
--Building a 123B model-capable rig with 4x3090:
>105630142 >105630262 >105630325 >105631297 >105630328 >105630531 >105631152 >105635155
--Personalized speech-to-text tools for quick transcription with shortcut triggers:
>105627335 >105627797
--Teto and Miku and Rin (free space):
>105621874 >105622071 >105625804 >105626952 >105630546 >105636047 >105636052 >105636268 >105636665
►Recent Highlight Posts from the Previous Thread: >>105621564
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
Page 1