>Wake up
>Still no Hunyuan-A13B support
>Still no gguf
>https://github.com/ggml-org/llama.cpp/pull/14425#issuecomment-3015113465
>"They probably fucked up"
>Can't even run it local unquant with 64GB VRAM because fuck you