menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

SVD-Free L...
source image

Arxiv

1d

read

81

img
dot

Image Credit: Arxiv

SVD-Free Low-Rank Adaptive Gradient Optimization for Large Language Models

  • Low-rank optimization is being used in training large language models to reduce memory usage of adaptive optimizers by restricting learning to a lower-dimensional space.
  • A new two-step procedure is proposed to approximate SVD-based gradient projections efficiently in large models.
  • The procedure involves constructing an orthogonal basis using Discrete Cosine Transform matrices and adaptively selecting basis columns aligned with each layer's gradient.
  • The method achieves optimal low-rank projections, matching SVD-based methods' performance while being computationally efficient, faster, and reducing memory usage.

Read Full Article

like

4 Likes

For uninterrupted reading, download the app