Implements inference.Backend via llama-server subprocess (llama.cpp + HIP/ROCm). Targets RX 7800 XT (gfx1101, RDNA 3, 16GB VRAM). Includes: - Backend registration with build tags (linux/amd64) - Stub backend.go with llama-server lifecycle outline - CLAUDE.md with build instructions for llama.cpp + ROCm - TODO.md with 5-phase task queue - FINDINGS.md with hardware specs, VRAM budget, design rationale Co-Authored-By: Virgil <virgil@lethean.io>
7 lines
206 B
Go
7 lines
206 B
Go
//go:build !linux || !amd64
|
|
|
|
package rocm
|
|
|
|
// ROCmAvailable reports whether ROCm GPU inference is available.
|
|
// Returns false on non-Linux or non-amd64 platforms.
|
|
func ROCmAvailable() bool { return false }
|