From 861e44fedd19cdaf753c4d4b6a47c8d0e46c81f6 Mon Sep 17 00:00:00 2001 From: Snider Date: Fri, 20 Feb 2026 15:11:19 +0000 Subject: [PATCH] docs: add README with quick start and docs links Co-Authored-By: Virgil --- README.md | 40 ++++++++++++++++++++++++++++++++++++++++ 1 file changed, 40 insertions(+) create mode 100644 README.md diff --git a/README.md b/README.md new file mode 100644 index 0000000..a2825b1 --- /dev/null +++ b/README.md @@ -0,0 +1,40 @@ +# go-ml + +ML inference backends, multi-suite scoring engine, and agent orchestrator for the Lethean AI stack. The package provides pluggable backends (Apple Metal via go-mlx, managed llama-server subprocesses, and OpenAI-compatible HTTP APIs), a concurrent scoring engine that evaluates model responses across heuristic, semantic, content, and standard benchmark suites, 23 capability probes, GGUF model management, and an SSH-based agent orchestrator that streams checkpoint evaluation results to InfluxDB and DuckDB. + +**Module**: `forge.lthn.ai/core/go-ml` +**Licence**: EUPL-1.2 +**Language**: Go 1.25 + +## Quick Start + +```go +import "forge.lthn.ai/core/go-ml" + +// HTTP backend (Ollama, LM Studio, any OpenAI-compatible endpoint) +backend := ml.NewHTTPBackend("http://localhost:11434", "qwen3:8b") +resp, err := backend.Generate(ctx, "Hello", ml.GenOpts{MaxTokens: 256}) + +// Scoring engine +engine := ml.NewEngine(backend, ml.Options{Suites: "heuristic,semantic", Concurrency: 4}) +scores := engine.ScoreAll(responses) +``` + +## Documentation + +- [Architecture](docs/architecture.md) — backends, scoring engine, agent orchestrator, data pipeline +- [Development Guide](docs/development.md) — building, testing, contributing +- [Project History](docs/history.md) — completed phases and known limitations + +## Build & Test + +```bash +go test ./... +go test -race ./... +go test -bench=. ./... +go build ./... +``` + +## Licence + +European Union Public Licence 1.2 — see [LICENCE](LICENCE) for details.