No description
LoRA: low-rank adaptation with trainable A/B matrices, Kaiming normal init, safetensors save/load. AdamW: decoupled weight decay optimizer with positional moment tracking for gradient-replaced params. 14 tests passing including end-to-end LoRA+AdamW training loop. Co-Authored-By: Virgil <virgil@lethean.io> |
||
|---|---|---|
| agentic | ||
| ai | ||
| mcp | ||
| ml | ||
| mlx | ||
| rag | ||
| go.mod | ||
| go.sum | ||
| test-mlx.go | ||
| TEST-RESULTS.md | ||