Good news, llama.cpp can somehow start GLM-Air gguf on my toaster.
Bad news, I get, like, 0.5 tokens per second or so.