FedKRSO: Communication and Memory Efficient Federated Fine-Tuning of Large Language Models

FedKRSO: Communication and Memory Efficient Federated Fine-Tuning of Large Language Models
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

Fine-tuning is essential to adapt general-purpose large language models (LLMs) to domain-specific tasks. As a privacy-preserving framework to leverage decentralized data for collaborative model training, Federated Learning (FL) is gaining popularity in LLM fine-tuning, but remains challenging due to the high cost of transmitting full model parameters and computing full gradients on resource-constrained clients. While Parameter-Efficient Fine-Tuning (PEFT) methods are widely used in FL to reduce communication and memory costs, they often sacrifice model performance compared to FFT. This paper proposes FedKRSO (Federated $K$-Seed Random Subspace Optimization), a novel method that enables communication and memory efficient FFT of LLMs in federated settings. In FedKRSO, clients update the model within a shared set of random low-dimension subspaces generated by the server to save memory usage. Furthermore, instead of transmitting full model parameters in each FL round, clients send only the model update accumulators along the subspaces to the server, enabling efficient global model aggregation and dissemination. By using these strategies, FedKRSO can substantially reduce communication and memory overhead while overcoming the performance limitations of PEFT, closely approximating the performance of federated FFT. The convergence properties of FedKRSO are analyzed rigorously under general FL settings. Extensive experiments on the GLUE benchmark across diverse FL scenarios demonstrate that FedKRSO achieves both superior performance and low communication and memory overhead, paving the way towards on federated LLM fine-tuning at the resource-constrained edge.


💡 Research Summary

FedKRSO (Federated K‑Seed Random Subspace Optimization) addresses the prohibitive memory and communication costs of full‑parameter fine‑tuning (FFT) for large language models (LLMs) in federated learning (FL). The method introduces a server‑maintained pool of K random seeds. Each seed defines a low‑dimensional random projection matrix (P_k \in \mathbb{R}^{r \times d_n}) with (r \ll d_n) and the property (\mathbb{E}


Comments & Academic Discussion

Loading comments...

Leave a Comment