Search Results
8/2/2025, 11:18:14 AM
►Recent Highlights from the Previous Thread: >>106108045
--Paper: IndexTTS2: Breakthrough open-source zero-shot expressive TTS model soon to be released by Bilibili:
>106108381 >106108506
--Running >200B models via extreme quantization and offloading on consumer hardware:
>106110419 >106110426 >106110445 >106110633 >106110681 >106111115
--Qwen Thinker favored for roleplay depth despite slower speed vs Instruct:
>106109987 >106110063 >106110103 >106110130 >106110110 >106110097
--GLM-4.5-GGUF release with concerns over imatrix calibration quality:
>106108341 >106108382
--Community reaction to GLM 4.5 support attempt in llama.cpp:
>106110329 >106110516 >106110547 >106110582 >106110838 >106110850 >106110970 >106111121 >106111137 >106110940 >106110963
--Improving vision model accuracy through synthetic data and finetuning:
>106111679 >106111750
--New --cpu-moe flag in llamacpp for simplified MoE expert offloading:
>106111142
--AI generates "smirulakte" due to sampler settings and model instability:
>106111497 >106111502 >106111539 >106111504 >106111520 >106111529 >106111500 >106111715
--Speculation on Horizon Alpha and stealth model parameter counts and performance:
>106110634 >106111055 >106111085 >106111138
--Seeking capable uncensored ERP models despite hardware and naming absurdity:
>106110567 >106110647 >106110726 >106110750 >106110762
--Qwen 235B shows major improvement in conversational quality and expressiveness:
>106110230 >106110272 >106110288 >106110308
--Running Qwen-code on CPU with high token processing overhead:
>106111566 >106111718 >106111785
--Horizon Beta feels less assistant-like, raising hopes it's not just a cloud model:
>106111772
--Disappointment over safety restrictions in deepseek-671B-MoE finetune:
>106109775
--Miku (free space):
>106110757 >106113331
►Recent Highlight Posts from the Previous Thread: >>106108052
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
--Paper: IndexTTS2: Breakthrough open-source zero-shot expressive TTS model soon to be released by Bilibili:
>106108381 >106108506
--Running >200B models via extreme quantization and offloading on consumer hardware:
>106110419 >106110426 >106110445 >106110633 >106110681 >106111115
--Qwen Thinker favored for roleplay depth despite slower speed vs Instruct:
>106109987 >106110063 >106110103 >106110130 >106110110 >106110097
--GLM-4.5-GGUF release with concerns over imatrix calibration quality:
>106108341 >106108382
--Community reaction to GLM 4.5 support attempt in llama.cpp:
>106110329 >106110516 >106110547 >106110582 >106110838 >106110850 >106110970 >106111121 >106111137 >106110940 >106110963
--Improving vision model accuracy through synthetic data and finetuning:
>106111679 >106111750
--New --cpu-moe flag in llamacpp for simplified MoE expert offloading:
>106111142
--AI generates "smirulakte" due to sampler settings and model instability:
>106111497 >106111502 >106111539 >106111504 >106111520 >106111529 >106111500 >106111715
--Speculation on Horizon Alpha and stealth model parameter counts and performance:
>106110634 >106111055 >106111085 >106111138
--Seeking capable uncensored ERP models despite hardware and naming absurdity:
>106110567 >106110647 >106110726 >106110750 >106110762
--Qwen 235B shows major improvement in conversational quality and expressiveness:
>106110230 >106110272 >106110288 >106110308
--Running Qwen-code on CPU with high token processing overhead:
>106111566 >106111718 >106111785
--Horizon Beta feels less assistant-like, raising hopes it's not just a cloud model:
>106111772
--Disappointment over safety restrictions in deepseek-671B-MoE finetune:
>106109775
--Miku (free space):
>106110757 >106113331
►Recent Highlight Posts from the Previous Thread: >>106108052
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
8/2/2025, 11:02:33 AM
►Recent Highlights from the Previous Thread: >>106108045
--Paper: IndexTTS2: Breakthrough open-source zero-shot expressive TTS model soon to be released by Bilibili:
>106108381 >106108506
--Running >200B models via extreme quantization and offloading on consumer hardware:
>106110419 >106110426 >106110445 >106110633 >106110681 >106111115
--Qwen Thinker favored for roleplay depth despite slower speed vs Instruct:
>106109987 >106110063 >106110103 >106110130 >106110110 >106110097
--GLM-4.5-GGUF release with concerns over imatrix calibration quality:
>106108341 >106108382
--Community reaction to GLM 4.5 support attempt in llama.cpp:
>106110329 >106110516 >106110547 >106110582 >106110838 >106110850 >106110970 >106111121 >106111137 >106110940 >106110963
--Improving vision model accuracy through synthetic data and finetuning:
>106111679 >106111750
--New --cpu-moe flag in llamacpp for simplified MoE expert offloading:
>106111142
--:
>106111497 >106111502 >106111539 >106111504 >106111520 >106111529 >106111500 >106111715
--Speculation on Horizon Alpha and stealth model parameter counts and performance:
>106110634 >106111055 >106111085 >106111138
--Seeking capable uncensored ERP models despite hardware and naming absurdity:
>106110567 >106110647 >106110726 >106110750 >106110762
--Qwen 235B shows major improvement in conversational quality and expressiveness:
>106110230 >106110272 >106110288 >106110308
--Running Qwen-code on CPU with high token processing overhead:
>106111566 >106111718 >106111785
--Horizon Beta feels less assistant-like, raising hopes it's not just a cloud model:
>106111772
--Disappointment over safety restrictions in deepseek-671B-MoE finetune:
>106109775
--Miku (free space):
>106110757
►Recent Highlight Posts from the Previous Thread: >>106108052
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
--Paper: IndexTTS2: Breakthrough open-source zero-shot expressive TTS model soon to be released by Bilibili:
>106108381 >106108506
--Running >200B models via extreme quantization and offloading on consumer hardware:
>106110419 >106110426 >106110445 >106110633 >106110681 >106111115
--Qwen Thinker favored for roleplay depth despite slower speed vs Instruct:
>106109987 >106110063 >106110103 >106110130 >106110110 >106110097
--GLM-4.5-GGUF release with concerns over imatrix calibration quality:
>106108341 >106108382
--Community reaction to GLM 4.5 support attempt in llama.cpp:
>106110329 >106110516 >106110547 >106110582 >106110838 >106110850 >106110970 >106111121 >106111137 >106110940 >106110963
--Improving vision model accuracy through synthetic data and finetuning:
>106111679 >106111750
--New --cpu-moe flag in llamacpp for simplified MoE expert offloading:
>106111142
--:
>106111497 >106111502 >106111539 >106111504 >106111520 >106111529 >106111500 >106111715
--Speculation on Horizon Alpha and stealth model parameter counts and performance:
>106110634 >106111055 >106111085 >106111138
--Seeking capable uncensored ERP models despite hardware and naming absurdity:
>106110567 >106110647 >106110726 >106110750 >106110762
--Qwen 235B shows major improvement in conversational quality and expressiveness:
>106110230 >106110272 >106110288 >106110308
--Running Qwen-code on CPU with high token processing overhead:
>106111566 >106111718 >106111785
--Horizon Beta feels less assistant-like, raising hopes it's not just a cloud model:
>106111772
--Disappointment over safety restrictions in deepseek-671B-MoE finetune:
>106109775
--Miku (free space):
>106110757
►Recent Highlight Posts from the Previous Thread: >>106108052
Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
Page 1