Skip to content

raajmandale/mos-parameter-golf

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 

Repository files navigation

M-OS — Parameter Golf (CRS-LM)

Context Reconstruction × Pattern Runtime
Small-model intelligence through structured context control

status runtime ai track


🧠 Core Idea

Instead of scaling models endlessly:

Control the context, not the parameters


⚡ Concept Flow

Raw Context → CRS Engine → Smart Context → TinyLM


✨ What CRS Does

  • ✂️ Removes irrelevant noise
  • 📉 Compresses token space
  • 🔄 Reconstructs missing structure
  • 🧠 Preserves reasoning signal

🧬 Architecture


⚙️ Pipeline

Input Text ↓ Tokenizer ↓ CRS Filter Engine (SACR) ↓ Compressed Context ↓ TinyLM ↓ Prediction


📊 Benchmark (Visual)


🎬 Demo (Live Simulation)


📈 Results Snapshot

Mode Tokens Loss Speed
Baseline 81 0.1873 0.44s
CRS-LM 76 0.1824 0.40s

⚠️ Reality Check

  • ✅ ~6–40% token reduction (config dependent)
  • ⚠️ Aggressive filtering reduces quality
  • ❌ Not production-ready
  • ✔️ Strong research direction

🧪 Why This Matters

Traditional LLM CRS-LM
Uses full context Uses filtered context
Token-heavy Token-efficient
No structure awareness Structure-aware
Linear reasoning Reconstructed reasoning

🔗 Key Components

  • 🧠 CRS Engine → context filtering + compression
  • ⚙️ SACR → structure-aware reduction logic
  • 🤖 TinyLM → lightweight reasoning model
  • 📊 Benchmark Layer → token vs loss tradeoff

📁 Project Structure

mos-parameter-golf/ │ ├── crs-lm/ │ ├── banner.svg │ ├── architecture.svg │ ├── benchmark.svg │ ├── demo.svg │ ├── README.md │ ├── model/ │ ├── tokenizer/ │ ├── crs/ │ ├── train.py │ ├── infer.py │ └── eval.py │ ├── benchmarks/ ├── results/ └── README.md


⚙️ Quick Start

git clone https://github.com/raajmandale/mos-parameter-golf
cd mos-parameter-golf/crs-lm

pip install -r requirements.txt

python train.py
python infer.py
python eval.py
🧬 Future Direction
🔗 CRS + Deterministic Fragment Graph (DFG)
🧠 AI Memory Layer (XLifelineAI)
⚙️ M-OS runtime integration
🤖 Agent memory optimization
📌 Status
🧪 Research Prototype
⚠️ Experimental System
🚀 High Potential Direction
👨‍💻 Author

Raaj Mandale
Systems Architect • AI Infrastructure • M-OS • QBAIX

⭐ Support

If this work resonates:

⭐ Star the repo
🍴 Fork it
🚀 Share it
🧠 Final Thought

LLMs don’t need more tokens.
They need better context.

Releases

No releases published

Packages

 
 
 

Contributors

Languages