Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator

In this paper, we present a Low-Rank Gradient Estimator (LoGE) to accelerate the finetune-time computation of transformers, especially large language models (LLMs). Unlike Parameter-Efficient Fine-Tuning (PEFT) methods, which primarily aim to minimize the number of fine-tuning parameters, LoGE also...

Full description

Saved in:
Bibliographic Details
Main Authors: Luoming Zhang, Zhenyu Lou, Yangwei Ying, Cheng Yang, Hong Zhou
Format: Article
Language:English
Published: MDPI AG 2024-12-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/15/1/82
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841549427137839104
author Luoming Zhang
Zhenyu Lou
Yangwei Ying
Cheng Yang
Hong Zhou
author_facet Luoming Zhang
Zhenyu Lou
Yangwei Ying
Cheng Yang
Hong Zhou
author_sort Luoming Zhang
collection DOAJ
description In this paper, we present a Low-Rank Gradient Estimator (LoGE) to accelerate the finetune-time computation of transformers, especially large language models (LLMs). Unlike Parameter-Efficient Fine-Tuning (PEFT) methods, which primarily aim to minimize the number of fine-tuning parameters, LoGE also significantly reduces the computational load of activation gradient calculations by decomposing pre-trained weights and utilizing low-rank matrices during the backward pass. Our approach includes an effective solution for identifying sensitive and important latent subspaces in large models before training with downstream datasets. As LoGE does not alter the network structure, it can be conveniently integrated into existing models. We validated LoGE’s efficacy through comprehensive experiments across various models on various tasks. For the widely used LLaMA model equipped with LoRA, LoGE achieves up to a 1.3× speedup while maintaining graceful accuracy.
format Article
id doaj-art-1beee20bd63a4602bfe585bffe9f8f12
institution Kabale University
issn 2076-3417
language English
publishDate 2024-12-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj-art-1beee20bd63a4602bfe585bffe9f8f122025-01-10T13:14:22ZengMDPI AGApplied Sciences2076-34172024-12-011518210.3390/app15010082Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient EstimatorLuoming Zhang0Zhenyu Lou1Yangwei Ying2Cheng Yang3Hong Zhou4The College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310027, ChinaThe College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310027, ChinaThe College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310027, ChinaThe College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310027, ChinaThe College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310027, ChinaIn this paper, we present a Low-Rank Gradient Estimator (LoGE) to accelerate the finetune-time computation of transformers, especially large language models (LLMs). Unlike Parameter-Efficient Fine-Tuning (PEFT) methods, which primarily aim to minimize the number of fine-tuning parameters, LoGE also significantly reduces the computational load of activation gradient calculations by decomposing pre-trained weights and utilizing low-rank matrices during the backward pass. Our approach includes an effective solution for identifying sensitive and important latent subspaces in large models before training with downstream datasets. As LoGE does not alter the network structure, it can be conveniently integrated into existing models. We validated LoGE’s efficacy through comprehensive experiments across various models on various tasks. For the widely used LLaMA model equipped with LoRA, LoGE achieves up to a 1.3× speedup while maintaining graceful accuracy.https://www.mdpi.com/2076-3417/15/1/82low rankgradient estimatorefficient fine-tuningefficient AI
spellingShingle Luoming Zhang
Zhenyu Lou
Yangwei Ying
Cheng Yang
Hong Zhou
Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
Applied Sciences
low rank
gradient estimator
efficient fine-tuning
efficient AI
title Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
title_full Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
title_fullStr Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
title_full_unstemmed Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
title_short Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
title_sort efficient fine tuning of large language models via a low rank gradient estimator
topic low rank
gradient estimator
efficient fine-tuning
efficient AI
url https://www.mdpi.com/2076-3417/15/1/82
work_keys_str_mv AT luomingzhang efficientfinetuningoflargelanguagemodelsviaalowrankgradientestimator
AT zhenyulou efficientfinetuningoflargelanguagemodelsviaalowrankgradientestimator
AT yangweiying efficientfinetuningoflargelanguagemodelsviaalowrankgradientestimator
AT chengyang efficientfinetuningoflargelanguagemodelsviaalowrankgradientestimator
AT hongzhou efficientfinetuningoflargelanguagemodelsviaalowrankgradientestimator