menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

On the Imp...
source image

Arxiv

2d

read

260

img
dot

Image Credit: Arxiv

On the Implicit Relation Between Low-Rank Adaptation and Differential Privacy

  • Low-rank adaptation methods in natural language processing, such as LoRA and FLoRA, involve keeping pre-trained model weights fixed and incorporating trainable low-rank decomposition matrices into some layers of the transformer architecture, called adapters.
  • Researchers have found that the low-rank adaptation used in LoRA and FLoRA introduces random noise into the batch gradients with respect to the adapter parameters, leading to a variance in the injected noise that increases as the adaptation rank decreases.
  • The study establishes a relationship between low-rank adaptation and differential privacy, showing that the dynamics of low-rank adaptation is similar to differentially private fine-tuning of the adapters.
  • The researchers suggest that low-rank adaptation offers privacy protection without the high space complexity of differentially private stochastic gradient descent (DPSGD), providing an efficient alternative for privacy-preserving fine-tuning in NLP models.

Read Full Article

like

15 Likes

For uninterrupted reading, download the app