AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.
Add ChatComplete() and Complete() methods to the llamacpp Client, backed by a shared parseSSE() line parser. Types include ChatMessage, ChatRequest, CompletionRequest and their chunked response structs. Tests cover multi-chunk streaming, empty responses, HTTP errors, and context cancellation — all using httptest SSE servers. Co-Authored-By: Virgil <virgil@lethean.io> Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> |
||
|---|---|---|
| docs/plans | ||
| internal/llamacpp | ||
| backend.go | ||
| CLAUDE.md | ||
| FINDINGS.md | ||
| go.mod | ||
| go.sum | ||
| README.md | ||
| register_rocm.go | ||
| rocm.go | ||
| rocm_stub.go | ||
| TODO.md | ||
go-rocm
AMD ROCm GPU inference for Linux. llama.cpp + HIP backend for RDNA 3.