menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Compressin...
source image

Arxiv

3d

read

63

img
dot

Image Credit: Arxiv

Compressing Sine-Activated Low-Rank Adapters through Post-Training Quantization

  • Low-Rank Adaptation (LoRA) method is effective for parameter-efficient fine-tuning by reducing trainable parameters through low-rank matrices.
  • Recent research applied sinusoidal transformation to low-rank adapters to increase stable rank without adding parameters, aiming to enhance performance compared to full-rank fine-tuning.
  • This study explores the application of the sine-activated technique in Post-Training Quantization to preserve benefits during model compression.
  • The results show that using sinusoidal non-linearity even after quantization leads to highly compressed adapters with minimal performance loss in various fine-tuning tasks.

Read Full Article

like

3 Likes

For uninterrupted reading, download the app