AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.
- ROCm 7.2, gfx1100 (corrected from gfx1101), kernel 6.17 - llama-server built with HIP from llama.cpp 11c325c - Gemma3-4B baseline: 109 tok/s decode, 396 tok/s prefill - Critical: iGPU crash requires HIP_VISIBLE_DEVICES=0 - All Phase 0 tasks marked done Co-Authored-By: Virgil <virgil@lethean.io> |
||
|---|---|---|
| backend.go | ||
| CLAUDE.md | ||
| FINDINGS.md | ||
| go.mod | ||
| README.md | ||
| register_rocm.go | ||
| rocm.go | ||
| rocm_stub.go | ||
| TODO.md | ||
go-rocm
AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.