Self-Distillation Can Hurt LLM Reasoning by Silencing Useful Doubt

Microsoft Research finds that self-distillation suppresses models’ expressions of uncertainty during reasoning, causing up to 40% accuracy drops on math benchmarks.
artificial-intelligence
Author

Kabui, Charles

Published

2026-04-02

Keywords

self-distillation, llm-reasoning, epistemic-verbalization, post-training, uncertainty-expression