AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.
Adds server.go with the process lifecycle layer that manages spawning llama-server, waiting for readiness, and graceful shutdown. Includes three helper functions (findLlamaServer, freePort, serverEnv) and the full startServer/waitReady/stop lifecycle. The serverEnv function critically filters HIP_VISIBLE_DEVICES to mask the Ryzen 9 iGPU which crashes llama-server if not excluded. Co-Authored-By: Virgil <virgil@lethean.io> Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> |
||
|---|---|---|
| docs/plans | ||
| internal/llamacpp | ||
| backend.go | ||
| CLAUDE.md | ||
| FINDINGS.md | ||
| go.mod | ||
| go.sum | ||
| README.md | ||
| register_rocm.go | ||
| rocm.go | ||
| rocm_stub.go | ||
| server.go | ||
| server_test.go | ||
| TODO.md | ||
go-rocm
AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.