Skip to content

PyTorch code for our paper "AdaSVD: Adaptive Singular Value Decomposition for Large Language Models"

License

Notifications You must be signed in to change notification settings

ZHITENGLI/AdaSVD

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

AdaSVD: Adaptive Singular Value Decomposition for Large Language Models

Zhiteng Li, Mingyuan Xia, Jingyuan Zhang, Zheng Hui, Linghe Kong, Yulun Zhang, and Xiaokang Yang, "AdaSVD: Adaptive Singular Value Decomposition for Large Language Models", arxiv, 2025

[arXiv] [supplementary material]

🔥🔥🔥 News

  • 2025-03-09: This repo is released.

Abstract: Large language models (LLMs) have achieved remarkable success in natural language processing (NLP) tasks, yet their substantial memory requirements present significant challenges for deployment on resource-constrained devices. Singular Value Decomposition (SVD) has emerged as a promising compression technique for LLMs, offering considerable reductions in memory overhead. However, existing SVD-based methods often struggle to effectively mitigate the errors introduced by SVD truncation, leading to a noticeable performance gap when compared to the original models. Furthermore, applying a uniform compression ratio across all transformer layers fails to account for the varying importance of different layers. To address these challenges, we propose AdaSVD, an adaptive SVD-based LLM compression approach. Specifically, AdaSVD introduces adaComp, which adaptively compensates for SVD truncation errors by alternately updating the singular matrices $\mathcal{U}$ and $\mathcal{V}^\top$. Additionally, AdaSVD introduces adaCR, which adaptively assigns layer-specific compression ratios based on the relative importance of each layer. Extensive experiments across multiple LLM/VLM families and evaluation metrics demonstrate that AdaSVD consistently outperforms state-of-the-art (SOTA) SVD-based methods, achieving superior performance with significantly reduced memory requirements. Code and models of AdaSVD will be available at https://github.com/ZHITENGLI/AdaSVD.

⚒️ TODO

  • Complete this repository

🔗 Contents

🔎 Results

AdaSVD outperforms previous state-of-the-art SVD-based LLM compression methods.
  • LLaMA2 7B

  • OPT-6.7B, LLaMA2-7B, Mistral-7B, and Vicuna-7B

AdaSVD can also be applied to VLM models like LLaVA.
  • Image captioning on COCO dataset

Citation

If you find the code helpful in your research or work, please cite the following paper.

@article{li2025adasvd,
  title={AdaSVD: Adaptive Singular Value Decomposition for Large Language Models},
  author={Li, Zhiteng and Xia, Mingyuan and Zhang, Jingyuan and Hui, Zheng and Kong, Linghe and Zhang, Yulun and Yang, Xiaokang},
  journal={arXiv e-prints},
  pages={arXiv--2502},
  year={2025}
}

💡 Acknowledgements

This work is released under the Apache 2.0 license. The codes are based on SVD-LLM. Please also follow their licenses. Thanks for their awesome works.

About

PyTorch code for our paper "AdaSVD: Adaptive Singular Value Decomposition for Large Language Models"

Resources

License

Stars

Watchers

Forks

Packages

No packages published