-
Notifications
You must be signed in to change notification settings - Fork 440
Open
Labels
enhancementNew feature or requestNew feature or request
Description
Overview
Implement the full training pipeline from ADR-129 to retrain RuvLTRA models with TurboQuant KV-cache profiling on Google Cloud.
Phases
Phase 1: imatrix Recalibration + TurboQuant KV Profiling (Week 1)
- Build
gcr.io/ruv-dev/ruvltra-training:latestDocker image - Run imatrix recalibration with code-focused calibration data
- Generate
.turboquant.jsonsidecar profiles per model - Benchmark recalibrated GGUFs vs baseline (ablation run B)
Phase 2: WET-Augmented LoRA Fine-Tuning (Week 2-3)
- Export brain memories + WET data as training corpus
- Run eval contamination check (13-gram overlap)
- Validate dataset governance (schema, dedup, quality scores)
- Run LoRA SFT on Vertex AI A100 (ablation run C)
- Run DPO training (ablation run D)
Phase 3: Benchmarking & Validation (Week 3-4)
- Run full ablation matrix (runs A-E)
- Evaluate all 7 release gates (G1-G7)
- Produce contamination report + ablation report
- Automate via
scripts/training/release_gate.py
Phase 4: Publishing (Week 4)
- Produce GGUF variants +
.turboquant.jsonsidecars - Publish to HuggingFace (all 4 models)
- Update model cards with benchmark results
- Update
ruvllmregistry with checksums - Publish
ruvllmand@ruvector/ruvllmwith sidecar loading - Set up weekly benchmark scheduler job
Release Gates (G1-G7)
| Gate | Criterion |
|---|---|
| G1 | HumanEval pass@1 ≥ 45% (0.5B) / ≥ 55% (3B) |
| G2 | Routing accuracy ≥ 80% (no regression) |
| G3 | Wikitext-2 PPL increase < 5% |
| G4 | TurboQuant ≥ 8x compression, PPL delta < 1% |
| G5 | Long context PPL < 20 at 16K tokens |
| G6 | Zero eval contamination |
| G7 | Inference ≥ 80 tok/s (0.5B) / ≥ 40 (3B) |
Infrastructure
- Compute: L4 GPU (Cloud Run Jobs) + A100-80GB (Vertex AI)
- Data: Brain memories (3,870+), WET corpus, Claude Flow routing (2,700+), ADR corpus (129 docs)
- Estimated cost: ~$70-210 (experimental compute)
Files Created
scripts/training/release_gate.py— Automated ship/no-ship checkerscripts/training/export_training_data.py— Dataset export with governancescripts/training/contamination_check.py— Eval contamination detectionscripts/training/Dockerfile— Training imagescripts/training/deploy_training.sh— Cloud Run job creationscripts/training/run_calibration.py— Phase 1 entry pointscripts/training/run_sft.py— Phase 2 entry pointcrates/ruvllm/src/quantize/turboquant_profile.rs— Sidecar config loading
Related
- ADR-129: docs/adr/ADR-129-ruvltra-gcloud-training-turboquant.md
- TurboQuant:
crates/ruvllm/src/quantize/turbo_quant.rs - Models: ruv/ruvltra-claude-code (7,615 downloads)
🤖 Generated with claude-flow
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or request