UniSD: Towards a Unified Self-Distillation Framework for Large Language Models Paper • 2605.06597 • Published 7 days ago • 12
Self-Distillation for Model Stacking Unlocks Cross-Lingual NLU in 200+ Languages Paper • 2406.12739 • Published Jun 18, 2024 • 2
Self-Distillation Zero: Self-Revision Turns Binary Rewards into Dense Supervision Paper • 2604.12002 • Published about 1 month ago • 11
Unifying Group-Relative and Self-Distillation Policy Optimization via Sample Routing Paper • 2604.02288 • Published Apr 2 • 33
Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models Paper • 2601.18734 • Published Jan 26 • 7
Why Does Self-Distillation (Sometimes) Degrade the Reasoning Capability of LLMs? Paper • 2603.24472 • Published Mar 25 • 55
DataFlex: A Unified Framework for Data-Centric Dynamic Training of Large Language Models Paper • 2603.26164 • Published Mar 27 • 364
Goedel-Prover-V2: Scaling Formal Theorem Proving with Scaffolded Data Synthesis and Self-Correction Paper • 2508.03613 • Published Aug 5, 2025 • 16
DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition Paper • 2504.21801 • Published Apr 30, 2025 • 5