Anonymous
8/5/2025, 5:23:43 PM
No.106150574
For some reason my local GLM instance doesn't do <thinking></thinking> tags to encompass its reasoning process. I'm running GLM 4.5 Q4_XL using chat completion.
Do I need to do something else to enable reasoning mode with the llama.cpp implementation?
Do I need to do something else to enable reasoning mode with the llama.cpp implementation?