Snider
|
c3c2c14dba
|
feat(backend): add HTTP/Llama TextModel wrappers + verify downstream
Phase 1 Steps 1.4 and 1.5 complete:
- HTTPTextModel wraps HTTPBackend as inference.TextModel (Generate/Chat
yield entire response as single Token, Classify unsupported,
BatchGenerate sequential)
- LlamaTextModel embeds HTTPTextModel, overrides ModelType -> "llama"
and Close -> llama.Stop()
- 19 new tests (17 HTTPTextModel + 2 LlamaTextModel), all passing
- Verified service.go and judge.go downstream consumers unchanged
- Updated CLAUDE.md: backend_mlx.go DONE, architecture table current,
critical context reflects Phase 1 complete
- Updated TODO.md: Steps 1.4 and 1.5 marked done
Co-Authored-By: Virgil <virgil@lethean.io>
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
|
2026-02-20 01:23:34 +00:00 |
|