menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

DenoiseRot...
source image

Arxiv

3d

read

291

img
dot

Image Credit: Arxiv

DenoiseRotator: Enhance Pruning Robustness for LLMs via Importance Concentration

  • Pruning is a common technique to compress large language models by removing unimportant weights, but it often leads to performance degradation, especially under semi-structured sparsity constraints.
  • A new approach called DenoiseRotator is proposed to enhance pruning robustness by redistributing parameter importance to make the model more amenable to pruning.
  • DenoiseRotator minimizes the information entropy of normalized importance scores, concentrating importance onto a smaller subset of weights, thus improving pruning effectiveness.
  • Evaluation on various models shows that DenoiseRotator consistently enhances perplexity and zero-shot accuracy when compared to existing pruning techniques like Magnitude, SparseGPT, and Wanda.

Read Full Article

like

17 Likes

For uninterrupted reading, download the app