Pion: A Spectrum-Preserving Optimizer via Orthogonal Equivalence Transformation
Abstract
We introduce Pion, a spectrum-preserving optimizer for large language model (LLM) training based on orthogonal equivalence transformation. Unlike additive optimizers such as Adam and Muon, Pion updates each weight matrix through left and right orthogonal transformations, preserving its singular values throughout training. This yields an optimization mechanism that modulates the geometry of weight matrices while keeping their spectral norm fixed. We derive the Pion update rule, systematically exa...
Description / Details
We introduce Pion, a spectrum-preserving optimizer for large language model (LLM) training based on orthogonal equivalence transformation. Unlike additive optimizers such as Adam and Muon, Pion updates each weight matrix through left and right orthogonal transformations, preserving its singular values throughout training. This yields an optimization mechanism that modulates the geometry of weight matrices while keeping their spectral norm fixed. We derive the Pion update rule, systematically examine its design choices, and analyze its convergence behavior along with several key properties. Empirical results show that Pion offers a stable and competitive alternative to standard optimizers for both LLM pretraining and finetuning.
Source: arXiv:2605.12492v1 - http://arxiv.org/abs/2605.12492v1 PDF: https://arxiv.org/pdf/2605.12492v1 Original Link: http://arxiv.org/abs/2605.12492v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
May 13, 2026
Data Science
Machine Learning
0