menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Mitigating...
source image

Arxiv

3d

read

47

img
dot

Image Credit: Arxiv

Mitigating Catastrophic Forgetting with Adaptive Transformer Block Expansion in Federated Fine-Tuning

  • Federated fine-tuning (FedFT) is used for adapting large language models (LLMs) to distributed data environments while preserving data privacy.
  • Existing FedFT methods primarily employ parameter-efficient fine-tuning techniques but struggle with catastrophic forgetting in distributed environments.
  • To address this issue, the proposed FedBE framework integrates an adaptive transformer block expansion mechanism with dynamic trainable-block allocation.
  • FedBE shows improved accuracy retention (12-74%) on general tasks after fine-tuning and accelerates model convergence (1.9-3.1x) without compromising downstream task accuracy.

Read Full Article

like

2 Likes

For uninterrupted reading, download the app