Embarrassingly Simple Self-Distillation Improves Code Generation Paper • 2604.01193 • Published 2 days ago • 15
Multiplex Thinking: Reasoning via Token-wise Branch-and-Merge Paper • 2601.08808 • Published Jan 13 • 39
GAD-Models Collection Model checkpoints of Black-Box On-Policy Distillation of Large Language Models • 5 items • Updated Nov 17, 2025 • 6
Black-Box On-Policy Distillation of Large Language Models Paper • 2511.10643 • Published Nov 13, 2025 • 52
Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model? Paper • 2504.13837 • Published Apr 18, 2025 • 141
4D LangSplat: 4D Language Gaussian Splatting via Multimodal Large Language Models Paper • 2503.10437 • Published Mar 13, 2025 • 34
DiffCLIP Collection Official models for DiffCLIP: Differential Attention Meets CLIP • 4 items • Updated Mar 9, 2025 • 3
Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach Paper • 2502.05171 • Published Feb 7, 2025 • 154