llama.cpp with native AMD RDNA4 (gfx1201) ROCm 7.11 support - 98.97 tok/s AI inference, competitive with RTX 4070 Ti, 32GB VRAM
-
Updated
Jan 3, 2026 - C++
llama.cpp with native AMD RDNA4 (gfx1201) ROCm 7.11 support - 98.97 tok/s AI inference, competitive with RTX 4070 Ti, 32GB VRAM
Add a description, image, and links to the gfx1201 topic page so that developers can more easily learn about it.
To associate your repository with the gfx1201 topic, visit your repo's landing page and select "manage topics."