Industrial-grade fine-tuning system implementing the Atom of Thoughts (AoT) reasoning framework. Designed to run on limited-resource devices using PEFT and 4-bit quantization.
Standard Chain-of-Thought (CoT) prompting is linear and memory-heavy. AoT treats reasoning as a Markovian process:
- Decomposition — breaks a complex problem into independent atomic states
- Atomic Reasoning — solves each state in isolation, preventing history interference
- Contraction — merges atomic solutions into a final, verifiable answer
This reduces token bloat and improves reasoning accuracy on complex tasks.
src/training/aot_engine.py— Core Decompose-Solve-Contract loopsrc/model/model_manager.py— Model loading with MXFP4 and NF4 4-bit quantizationsrc/training/trainer.py— Unified training pipeline for multi-scale models (0.5B to 20B)local_lite/— Optimized sub-system for CPU-only and 16GB RAM environments
pip install -r requirements.txtpython src/training/trainer.py --model qwen_7b --template aot --dataset ./data/aot_research_data.jsonSet WANDB_MODE=disabled if you are not using Weights and Biases:
WANDB_MODE=disabled python src/training/trainer.py --model qwen_7b --template aot --dataset ./data/aot_research_data.jsonTo push local changes to the repository, run the sync script from the project root:
./scripts/sync_to_git.ps1This script stages all modified files, commits with a timestamp message, and pushes to the main branch.
- Smoke tests confirmed successful inference on Qwen2.5-0.5B with 4-bit NF4 quantization on CPU
- Training loop runs end-to-end with AoT decompose-solve-contract prompt structure
- Full evaluation on a held-out reasoning benchmark is in progress
Python · PyTorch · Transformers · PEFT · BitsAndBytes · Qwen2.5 · WandB
MIT