VibeThinker-1.5B-SCU: Automated Information-Theoretic Early Stopping

A Scientifically Validated "Self-Regulating" Model

Patent Pending GitHub

This model is a proof-of-concept for the Shannon Control Unit (SCU), an automated training framework that acts as a "Safety Brake" against overfitting. It was trained on the WeiboAI/VibeThinker-1.5B model using the FineWeb-Edu dataset.

πŸš€ The "Transistor" Moment in AI Training

Traditional training is like a vacuum tube: it amplifies learning blindly until manual intervention (early stopping) cuts the power. SCU acts as a transistor, modulating regularization in real-time based on the model's internal Information Ratio.

Scientific Discovery: Step 386

In this experiment, SCU detected that the 1.5B model had saturated its learning capacity after only 16 Million tokens (Step 386).

  • Automatic Reaction: The controller saturated regularization ($\lambda \to 2.0$), effectively freezing the weights to prevent the model from memorizing noise.
  • The Result: Optimal performance (6.14 BPT) matching the best manual baseline, but with guaranteed safety against the overfitting "crash" seen in unregulated runs.

This suggests that for highly optimized models like VibeThinker, ~90% of standard training compute may be wasted on overfitting, which SCU identifies and prevents automatically.

πŸ“Š Benchmark Results

Metric Base Model Baseline (Manual) SCU V3 (Auto-Brake) SCU V4 (Unregulated)
BPT Score 9.92 6.13 6.14 6.77
Status Untrained Risky Safe & Optimal Crashed

πŸ› οΈ Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch

# 1. Load Base Model
base_id = "WeiboAI/VibeThinker-1.5B"
model = AutoModelForCausalLM.from_pretrained(
    base_id, 
    torch_dtype=torch.bfloat16, 
    device_map="auto", 
    trust_remote_code=True
)

# 2. Load SCU Adapter
adapter_id = "hunterbown/VibeThinker-1.5B-SCU" 
model = PeftModel.from_pretrained(model, adapter_id)

# 3. Inference
tokenizer = AutoTokenizer.from_pretrained(base_id, trust_remote_code=True)
prompt = "Explain the concept of information capacity."
# ... standard generation ...

πŸ“œ License & Citation

If you use this in research, please cite:

@misc{bown2025scu,
  author = {Bown, Hunter},
  title = {Shannon Control Unit: Automated Information-Theoretic Early Stopping},
  year = {2025},
  url = {https://github.com/Shannon-Labs/shannon-control-unit}
}
Downloads last month
2
Video Preview
loading

Model tree for hunterbown/VibeThinker-1.5B-SCU

Base model

Qwen/Qwen2.5-1.5B
Adapter
(2)
this model

Evaluation results