Implements inference.Backend via llama-server subprocess (llama.cpp + HIP/ROCm). Targets RX 7800 XT (gfx1101, RDNA 3, 16GB VRAM). Includes: - Backend registration with build tags (linux/amd64) - Stub backend.go with llama-server lifecycle outline - CLAUDE.md with build instructions for llama.cpp + ROCm - TODO.md with 5-phase task queue - FINDINGS.md with hardware specs, VRAM budget, design rationale Co-Authored-By: Virgil <virgil@lethean.io>
7 lines
153 B
Modula-2
7 lines
153 B
Modula-2
module forge.lthn.ai/core/go-rocm
|
|
|
|
go 1.25.5
|
|
|
|
require forge.lthn.ai/core/go-inference v0.0.0
|
|
|
|
replace forge.lthn.ai/core/go-inference => ../go-inference
|