LoRA Meets Second-Order Optimization: Towards Optimal Low-Rank Updates

ICLR 2026 Conference Submission17330 Authors

19 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Fine-tuning, Low-rank matrix weights, Shampoo, Adaptive metric
TL;DR: We design a better low-rank approximation to the full fine-tuning gradient based on an adaptive metric, and effectively improve and acclerate the low-rank fine-tuning.
Abstract: Low-rank fine-tuning is widely applied for the effective adaptation of large models. Most existing methods rely on low-rank matrix factorization, whose performance is limited by the condition number of the associated Jacobi operator. Although these methods are computationally efficient, their performance still falls short compared to full fine-tuning. To address this, we propose SoLoRA, which leverages an adaptive metric to find a low-rank approximation of the full fine-tuning gradient. This low-rank approximation can be viewed as an approximation of Hessian, effectively incorporating second-order information to achieve faster convergence and higher optimization efficiency. Furthermore, the low-rank approximation in SoLoRA is computationally simple and easy to implement, achieving a close approximation to the performance of full fine-tuning with almost no additional computational overhead. We conduct fine-tuning experiments on large language models and diffusion models, and the results consistently demonstrate that SoLoRA achieves superior performance advantages over state-of-the-art low-rank fine-tuning methods.
Supplementary Material: zip
Primary Area: optimization
Submission Number: 17330
Loading