>>105885211
>What idiot thought this was the right default behavior?
The model's config says to add the BOS token. llama.cpp adds the BOS token. The chat template says to add a BOS token. You add the BOS token.
The issue is not llama.cpp. Not on this one.