Search Results
6/21/2025, 4:15:02 PM
►Recent Highlights from the Previous Thread: >>105652633
(1/2)
--Papers:
>105654168 >105660467
--GRPO-CARE improves reasoning consistency in multimodal models on SEED-Bench-R1:
>105652873
--Critique of Musk's plan to rewrite training data with Grok and its implications for bias, accuracy, and ideological filtering:
>105660294 >105660320 >105660328 >105660337 >105660341 >105660343 >105660346 >105660373 >105660347 >105660459 >105660482 >105660515 >105660526 >105660532 >105660568 >105660559 >105660571 >105660632 >105660667 >105660680 >105660726 >105660725 >105660759
--Prompting techniques and philosophical debates around LLM self-awareness and AGI consciousness:
>105654616 >105655666 >105655670 >105655699 >105655705 >105655838 >105656283 >105656425 >105656438 >105661576 >105656770 >105656873 >105656965 >105657043 >105657220 >105657504 >105657611 >105657622 >105657813 >105657859 >105657863 >105658064 >105658111 >105658190 >105658268 >105660409
--Debating the optimal sampler chain for creative writing: XTC -> top nsigma -> temp:
>105656828 >105656912 >105656987 >105657078 >105657213 >105658087 >105658212 >105658389 >105658469 >105658613 >105658721 >105658846 >105658905 >105659069 >105659183 >105659447
--Debating AGI claims amid LLM consciousness and capability limitations:
>105652855 >105653377 >105653577 >105655036 >105655147 >105655182 >105655232 >105655345 >105655453 >105655493 >105655500 >105658428
--Mistral Small 3.2 shows measurable improvements in repetition and instruction following, prompting speculation on training methodologies:
>105658424 >105658467 >105658488 >105658665 >105658676 >105658776 >105658696 >105658710 >105658525 >105658938 >105658951
--LongWriter-Zero-32B excels in ultra-long text generation via reinforcement learning:
>105661432 >105661490 >105661519
►Recent Highlight Posts from the Previous Thread: >>105652637
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
(1/2)
--Papers:
>105654168 >105660467
--GRPO-CARE improves reasoning consistency in multimodal models on SEED-Bench-R1:
>105652873
--Critique of Musk's plan to rewrite training data with Grok and its implications for bias, accuracy, and ideological filtering:
>105660294 >105660320 >105660328 >105660337 >105660341 >105660343 >105660346 >105660373 >105660347 >105660459 >105660482 >105660515 >105660526 >105660532 >105660568 >105660559 >105660571 >105660632 >105660667 >105660680 >105660726 >105660725 >105660759
--Prompting techniques and philosophical debates around LLM self-awareness and AGI consciousness:
>105654616 >105655666 >105655670 >105655699 >105655705 >105655838 >105656283 >105656425 >105656438 >105661576 >105656770 >105656873 >105656965 >105657043 >105657220 >105657504 >105657611 >105657622 >105657813 >105657859 >105657863 >105658064 >105658111 >105658190 >105658268 >105660409
--Debating the optimal sampler chain for creative writing: XTC -> top nsigma -> temp:
>105656828 >105656912 >105656987 >105657078 >105657213 >105658087 >105658212 >105658389 >105658469 >105658613 >105658721 >105658846 >105658905 >105659069 >105659183 >105659447
--Debating AGI claims amid LLM consciousness and capability limitations:
>105652855 >105653377 >105653577 >105655036 >105655147 >105655182 >105655232 >105655345 >105655453 >105655493 >105655500 >105658428
--Mistral Small 3.2 shows measurable improvements in repetition and instruction following, prompting speculation on training methodologies:
>105658424 >105658467 >105658488 >105658665 >105658676 >105658776 >105658696 >105658710 >105658525 >105658938 >105658951
--LongWriter-Zero-32B excels in ultra-long text generation via reinforcement learning:
>105661432 >105661490 >105661519
►Recent Highlight Posts from the Previous Thread: >>105652637
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
Page 1