>>513498035Same exact models as on an NVidia. Can even run it ALL on your CPU & regular ram, it's just slower. You can do that with KoboldCPP if I recall correctly, When you launch KoboldCPP just set the "Backed:" to "Use Vulkan" for Radeon or "Use CPU" if that doesn't work. I've totally loaded 4B multimodal models and entirely run them from regular ram because the speed difference for a model that small isn't a terribly noticeable difference between VRAM and RAM.