1. 📘 Topic and Domain: The paper investigates self-distillation in large language models (LLMs) for mathematical reasoning tasks, focusing on how post-training methods affect reasoning capability.
2. 💡 Previous Research and New Ideas: Based on prior work showing self-distillation improves performance in domains like agentic environments and scientific reasoning, the paper introduces a new hypothesis that performance degradation in math reasoning stems from suppression of epistemic verbalization—the model's expression of uncertainty during reasoning.
3. ❓ Problem: The paper addresses why self-distillation, while effective in some domains, can degrade reasoning performance in mathematical tasks despite guiding models toward correct answers.
4. 🛠️ Methods: The authors use controlled experiments varying information richness in teacher conditioning and task coverage, analyzing how different conditioning contexts (unguided vs. solution-guided generation) affect epistemic token usage and out-of-distribution performance across multiple models (Qwen3-8B, DeepSeek-Distill-Qwen-7B, and Olmo3-7B-Instruct).
5. 📊 Results and Evaluation: Self-distillation with rich conditioning contexts reduces epistemic verbalization and response length, enabling rapid in-domain optimization with limited task coverage but causing up to 40% performance degradation on OOD benchmarks (AIME24, AMC23); GRPO maintains or improves performance while SDPO degrades it, and performance drops correlate with reduced uncertainty expression.
1. 📘 主题与领域: 该论文研究大型语言模型(LLM)中的自蒸馏方法在数学推理任务中的应用,重点关注后训练方法如何影响推理能力。
2. 💡 先前研究与新思路: 基于先前在智能体环境和科学推理等领域证明自蒸馏有效的研�,本文提出一个新假设:数学推理性能下降的原因在于抑制了认知语言化(epistemic verbalization)——即模型在推理过程中表达不确定性的行为。
3. ❓ 问题: 该论文探讨为什么自蒸馏在某些领域有效,却在数学任务中可能导致推理性能下降,尽管该方法引导模型朝向正确答案。
4. 🛠️ 方法: 作者通过控制实验,改变教师条件设置中的信息丰富度和任务覆盖率,分析不同条件上下文(无引导生成 vs. 解题方案引导生成)如何影响认知标记使用和分布外性能,实验涵盖多个模型(Qwen3-8B、DeepSeek-Distill-Qwen-7B 和 Olmo3-7B-Instruct)。
5. 📊 结果与评估: 使用丰富条件上下文的自蒸馏减少了认知语言化和响应长度,在有限任务覆盖下实现快速领域内优化,但在分布外基准测试(AIME24、AMC23)上导致高达40%的性能下降;GRPO保持或提升性能而SDPO导致性能下降,性能下降与不确定性表达的减少相关。