Low-Rank Adapters Meet Neural Architecture Search for LLM Compression
Abstract
The rapid expansion of Large Language Models (LLMs) has posed significant challenges regarding the computational resources required for fine-tuning and deployment. Recent advancements in low-rank adapters have demonstrated their efficacy in parameter-efficient fine-tuning (PEFT) of these models. This retrospective paper comprehensively discusses innovative approaches that synergize low-rank representations with Neural Architecture Search (NAS) techniques, particularly weight-sharing super-networks. Robust solutions for compressing and fine-tuning large pre-trained models are developed by integrating these methodologies. Our analysis highlights the potential of these combined strategies to democratize the use of LLMs, making them more accessible for deployment in resource-constrained environments. The resulting models exhibit reduced memory footprints and faster inference times, paving the way for more practical and scalable applications of LLMs. Models and code are available at https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Gradient Weight-normalized Low-rank Projection for Efficient LLM Training (2024)
- SaLoRA: Safety-Alignment Preserved Low-Rank Adaptation (2025)
- Efficient Deployment of Large Language Models on Resource-constrained Devices (2025)
- FlexiGPT: Pruning and Extending Large Language Models with Low-Rank Weight Sharing (2025)
- ASLoRA: Adaptive Sharing Low-Rank Adaptation Across Layers (2024)
- LoRS: Efficient Low-Rank Adaptation for Sparse Large Language Model (2025)
- Low-Rank Adaptation for Foundation Models: A Comprehensive Review (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper