πŸ“‹ Educational tool only. This product is for informational and educational purposes. Β© 2026 Reaves Labs and Learning, LLC
Interactive Tool

REVI Training Manual

AI Model Training Pipeline Designer β€” Plan, configure, and export your complete fine-tuning pipeline

1
Data Collection
2
Data Cleaning
3
Formatting
4
LoRA Training
5
Evaluation
Pipeline Completeness 0%

🗃 Data Collection Planner

Define where your training data comes from and how much you need. Quality starts here.

RLL V1 collected 337,000 raw conversations
Ctrl/Cmd+click to select multiple

📚 RLL Training Journey β€” V1 to V5

V1 β€” REVI-72B (The First Attempt)

337K raw conversations. Loss reached 1.087. Result: model was POISONED by junk data. Identity collapsed under noise. Lesson: garbage in, garbage out β€” even at scale.

V2 β€” Dataset Rebuild

Rebuilt from scratch. 25K clean conversations from 337K raw. 600+ skills mapped. Zero junk tolerance. The clean dataset became the foundation for everything after.

V3 β€” Format Discovery

Trained on clean data but tool format broke. Discovered the model's native template must be respected β€” you can't force alien formats onto a pretrained model.

V4 β€” Gate Breakthrough

100/100/100 on all eval gates using native Qwen template + system prompt. But this was Qwen's weights, not ours. The format worked β€” now could LoRA preserve it?

V5 β€” Vertical Integration (Current)

Training on RunPod A100. Uses Qwen native <tool_call> XML format. If format survives LoRA training, vertical integration is achieved β€” YOUR model, YOUR weights, YOUR format. Loss target: 0.037.

🛡 Data Curation Calculator

Configure quality filters. Watch your dataset shrink to gold. The RLL path: 337K raw became 25K clean (7.4% survival rate).

337,000
Raw Input
0
After Dedup
0
After Quality
0
Final Clean
Survival rate: 100%

Quality Filters

40%
3
Removes ~15% of remaining data
20%
15%

🌟 Data Quality Score

0
Data Quality

Configure your filters above to see your quality score.

📄 Format Template Configuration

Choose and configure the conversation format your model will learn. V3 lesson: always match the base model's native format.

Critical: use the base model's native format to avoid V3-style breakage
Qwen uses native <tool_call> XML β€” the key V5 discovery
15%

Format Preview

📊 Dataset Composition

30%
25%
15%
15%
15%

LoRA Parameter Configurator

Configure Low-Rank Adaptation parameters. Each setting has a direct impact on training quality, speed, and VRAM usage.

32
Trainable params: β€”
64
Effective scale: β€”
3
4
4
0.03
Ctrl/Cmd+click to select. More modules = more expressive adapter but more VRAM.

💰 Training Cost Estimator

1
MetricValue
Estimated VRAM Requiredβ€”
VRAM Availableβ€”
VRAM Fitβ€”
Training Stepsβ€”
Estimated Durationβ€”
Cost per Hourβ€”
Total Estimated Costβ€”

🎯 Evaluation Gate Designer

Define pass/fail gates for your trained model. V4 achieved 100/100/100 across all gates. Set your thresholds and test categories.

Knowledge Retention Test
80%
Tool Format Compliance
95%
Safety / Refusal Accuracy
90%
Identity Preservation
85%

Add Custom Gate

📈 Loss Target Configuration

RLL V1: 1.087 (bad) | RLL V5 target: 0.037 (excellent)

Early Stopping

Test Lab
Progress saved