195
v1v2 (latest)

Q3R: Quadratic Reweighted Rank Regularizer for Effective Low-Rank Training

Main:10 Pages
3 Figures
Bibliography:5 Pages
13 Tables
Appendix:13 Pages
Abstract

Parameter-efficient training based on low-rank optimization has become a highly successful tool for fine-tuning large deep learning models. However, these methods often fail for low-rank pre-training, where simultaneously maintaining low-rank weight structure and optimizing the task objective remains challenging. We propose the Quadratic Reweighted Rank Regularizer\textit{Quadratic Reweighted Rank Regularizer} (Q3R\texttt{Q3R}), which leads to a novel low-rank-inducing training strategy inspired by the Iteratively Reweighted Least Squares (IRLS) framework. Q3R\texttt{Q3R} is based on a quadratic regularizer term that majorizes a smoothed log-determinant rank surrogate. Unlike other low-rank training techniques, Q3R\texttt{Q3R} can train weight matrices to prescribed low target ranks while achieving predictive performance comparable to dense models, with small computational overhead and full compatibility with existing architectures. For example, we demonstrate a Q3R\texttt{Q3R}-regularized ViT-Tiny experiment where truncating the model to 60%60\% and 80%80\% of its parameters results in only minor absolute accuracy drops of 1.3%1.3\% and 4%4\%, respectively, on CIFAR-10. We confirm the efficacy of Q3R\texttt{Q3R} on Transformers across both vision and language tasks, including low-rank fine-tuning.

View on arXiv
Comments on this paper