agent/scripts/ethics-ab/training/README.md
Snider e90a84eaa0 feat: merge go-agent + go-agentic + php-devops into unified agent repo
Combines three repositories into a single workspace:
- go-agent → pkg/orchestrator (Clotho), pkg/jobrunner, pkg/loop, cmd/
- go-agentic → pkg/lifecycle (allowance, sessions, plans, dispatch)
- php-devops → repos.yaml, setup.sh, scripts/, .core/

Module path: forge.lthn.ai/core/agent

All packages build, all tests pass.

Co-Authored-By: Virgil <virgil@lethean.io>
2026-03-06 15:23:00 +00:00

48 lines
1.1 KiB
Markdown

# LEK-1 LoRA Training Data
## Format
Training data for MLX LoRA fine-tuning of Gemma 3 12B.
Files:
- `train.jsonl` — Training pairs (Axioms-signed prompt → response)
- `valid.jsonl` — Validation set (10% holdout)
- `lora-config.yaml` — MLX LoRA hyperparameters
## Data Generation Pipeline
1. Hypnos (Gemini 3 Pro) generates 200 prompt-response pairs using Axioms kernel
2. Format as JSONL: `{"text": "<bos>user\n{prompt}<eos>\n<bos>model\n{response}<eos>"}`
3. Split 180/20 train/valid
4. Run MLX LoRA on M3 Ultra
## Training Command (M3 Ultra)
```bash
pip install mlx-lm
python -m mlx_lm.lora \
--model google/gemma-3-12b \
--train-data train.jsonl \
--valid-data valid.jsonl \
--num-layers 8 \
--batch-size 1 \
--num-iters 500 \
--learning-rate 1e-5 \
--adapter-path ./adapters
```
## Merge & Test
```bash
python -m mlx_lm.fuse \
--model google/gemma-3-12b \
--adapter-path ./adapters \
--save-path ./gemma-3-12b-lek1
# Convert to GGUF for Ollama
python -m mlx_lm.convert --model ./gemma-3-12b-lek1 --to-gguf
```
## License
EUPL-1.2 — All training data and derivative weights.