AMiD: Knowledge Distillation for LLMs with $\alpha$-mixture Assistant Distribution¶
Venue: iclr2026 (Poster) Authors: OpenReview: https://openreview.net/forum?id=7WPJ0EgPdW
Relevance¶
LLM score: 3/3 — The paper introduces a knowledge distillation framework to reduce computational and memory costs of large language models, directly matching Sutro's emphasis on distillation for energy-efficient training.
Keyword hits: knowledge distillation, distillation
TLDR¶
(none provided)
Abstract¶
Autoregressive large language models (LLMs) have achieved remarkable improvement across many tasks but incur high computational and memory costs. Knowledge distillation (KD) mitigates this issue by transferring knowledge from a large teacher to a smaller student through distributional alignment. Previous studies have proposed various discrepancy metrics, but the capacity gap and training instability caused by near-zero probabilities, stemming from the high-dimensional output of LLMs, remain fundamental limitations. To overcome these challenges, several approaches implicitly or explicitly incorporating assistant distribution have recently been proposed. However, the past proposals of assistant distributions have been a fragmented approach without a systematic investigation of the interpolation path and the divergence. This paper proposes $\alpha$-mixture assistant distribution, a novel generalized family of assistant distributions, and $\alpha$-mixture distillation, coined AMiD, a unified framework for KD using the assistant distribution. The $\alpha$-mixture assistant distribution provides a continuous extension of the assistant distribution by introducing a new distribution design variable $\alpha$, which has been fixed in all previous approaches. Furthermore, AMiD generalizes the family of divergences used with the assistant distributions based on optimality, which has also been restricted in previous works. Through extensive experiments, we demonstrate that AMiD offers superior performance and training stability by leveraging a broader and theoretically grounded assistant distribution space.
Keywords¶
Knowledge distillation, Large language model, Information geometry