AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.
Implements inference.Backend via llama-server subprocess (llama.cpp + HIP/ROCm). Targets RX 7800 XT (gfx1101, RDNA 3, 16GB VRAM). Includes: - Backend registration with build tags (linux/amd64) - Stub backend.go with llama-server lifecycle outline - CLAUDE.md with build instructions for llama.cpp + ROCm - TODO.md with 5-phase task queue - FINDINGS.md with hardware specs, VRAM budget, design rationale Co-Authored-By: Virgil <virgil@lethean.io> |
||
|---|---|---|
| backend.go | ||
| CLAUDE.md | ||
| FINDINGS.md | ||
| go.mod | ||
| README.md | ||
| register_rocm.go | ||
| rocm.go | ||
| rocm_stub.go | ||
| TODO.md | ||
go-rocm
AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.