>>106879668
lol I preordered this thing so long ago. Is it even remotely useful for local LLMs, diffusion, or video?
I already built an AI server with 512 GB of RAM and 6 3090s