Skip to content

KDP: Simplifying Representation Dynamics in Kernel Space

Venue: iclr2026 (Poster) Authors: OpenReview: https://openreview.net/forum?id=262LUKGdQn

Relevance

LLM score: 1/3 — The paper proposes a model compression method via layer pruning, which yields inference efficiency gains but is not directly about energy-efficient training, data movement, sparsity during training, or other Sutro Group priorities. Keyword hits: model compression, pruning, kernel

TLDR

(none provided)

Abstract

This paper proposes Kernelized Dynamics Pruning (KDP), a novel layer pruning method from the perspective of simplifying representation dynamics within large language models (LLMs). Motivated by the high similarity between consecutive layer representations, we view the LLM's forward pass as a discrete-time dynamical system. We speculate that this phenomenon indicates the model's internal dynamics have entered a ``slow manifold'', which exhibits computational redundancy. Based on this insight, we project the representations into a kernel space where the complex, non-linear transformation between them is simplified to an approximately linear one. Then, a simple network learns the inverse kernel transformation, thereby enabling the pruning of the entire layer block. Both theoretical analysis and extensive experiments validate the effectiveness of KDP, demonstrating its superiority over existing pruning baselines. Code is available at https://anonymous.4open.science/r/draft-123abc.

Keywords

Large Language Models, Model Compression, Structured Pruning, Kernel Space