Answer 6 questions. We'll build the rest.

No tech background needed. About 90 seconds.

  1. Plain English. The result you'd be proud of.
  2. Fill out the 5 above, then press Generate.
📋 Educational tool only. This product is for informational and educational purposes. © 2026 Reaves Labs and Learning, LLC
Interactive Tool

REVI Training Manual

AI Model Training Pipeline Designer — Plan, configure, and export your complete fine-tuning pipeline

1
Data Collection
2
Data Cleaning
3
Formatting
4
LoRA Training
5
Evaluation
Pipeline Completeness 0%

🗃 Data Collection Planner

Define where your training data comes from and how much you need. Quality starts here.

RLL V1 collected 337,000 raw conversations
Ctrl/Cmd+click to select multiple

📚 RLL Training Journey — V1 to V5

V1 — REVI-72B (The First Attempt)

337K raw conversations. Loss reached 1.087. Result: model was POISONED by junk data. Identity collapsed under noise. Lesson: garbage in, garbage out — even at scale.

V2 — Dataset Rebuild

Rebuilt from scratch. 25K clean conversations from 337K raw. 600+ skills mapped. Zero junk tolerance. The clean dataset became the foundation for everything after.

V3 — Format Discovery

Trained on clean data but tool format broke. Discovered the model's native template must be respected — you can't force alien formats onto a pretrained model.

V4 — Gate Breakthrough

100/100/100 on all eval gates using native Qwen template + system prompt. But this was Qwen's weights, not ours. The format worked — now could LoRA preserve it?

V5 — Vertical Integration (Current)

Training on RunPod A100. Uses Qwen native <tool_call> XML format. If format survives LoRA training, vertical integration is achieved — YOUR model, YOUR weights, YOUR format. Loss target: 0.037.

🛡 Data Curation Calculator

Configure quality filters. Watch your dataset shrink to gold. The RLL path: 337K raw became 25K clean (7.4% survival rate).

337,000
Raw Input
0
After Dedup
0
After Quality
0
Final Clean
Survival rate: 100%

Quality Filters

40%
3
Removes ~15% of remaining data
20%
15%

🌟 Data Quality Score

0
Data Quality

Configure your filters above to see your quality score.

📄 Format Template Configuration

Choose and configure the conversation format your model will learn. V3 lesson: always match the base model's native format.

Critical: use the base model's native format to avoid V3-style breakage
Qwen uses native <tool_call> XML — the key V5 discovery
15%

Format Preview

📊 Dataset Composition

30%
25%
15%
15%
15%

LoRA Parameter Configurator

Configure Low-Rank Adaptation parameters. Each setting has a direct impact on training quality, speed, and VRAM usage.

32
Trainable params:
64
Effective scale:
3
4
4
0.03
Ctrl/Cmd+click to select. More modules = more expressive adapter but more VRAM.

💰 Training Cost Estimator

1
MetricValue
Estimated VRAM Required
VRAM Available
VRAM Fit
Training Steps
Estimated Duration
Cost per Hour
Total Estimated Cost

🎯 Evaluation Gate Designer

Define pass/fail gates for your trained model. V4 achieved 100/100/100 across all gates. Set your thresholds and test categories.

Knowledge Retention Test
80%
Tool Format Compliance
95%
Safety / Refusal Accuracy
90%
Identity Preservation
85%

Add Custom Gate

📈 Loss Target Configuration

RLL V1: 1.087 (bad) | RLL V5 target: 0.037 (excellent)

Early Stopping

All Demos
Progress saved