Authored by the Kimi Team at Moonshot AI (including Su Jianlin), this paper introduces a fundamental rethink of the "residual connection" by replacing fixed accumulation with learned softmax attention over the network's depth.

The Core Innovation: Standard Transformers sum layer outputs equally (hl = hl-1 + fl(hl-1)). Attention Residuals (AttnRes) instead uses learned "pseudo-queries" to let each layer retrieve specific information from earlier clean signals, effectively solving the "PreNorm Dilution" problem where early signals are drowned out by later noise. The Block AttnRes variant grouping layers makes this practical for massive models.

Significance and Impact: The paper demonstrates a 1.25x compute efficiency gain, allowing models to match standard performance with 25% less training budget. Massive reasoning improvements were observed, including +7.5 points on GPQA-Diamond. Su Jianlin frames this as "Sequence-Depth Duality": bringing the learned routing of the sequence dimension to the previously "fixed plumbing" of the depth dimension. The community has hailed it as the "final piece of the Transformer puzzle," with the official repository seeing immediate viral adoption.

Outputs 2

Attention Residuals Paper

paper

arXiv: 2603.15031

Attention Residuals Code

library

GitHub Repository

architecturescaling

More Links