menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Outlier We...
source image

Arxiv

3d

read

108

img
dot

Image Credit: Arxiv

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

  • Large Language Models (LLMs) are known for their high performance but face challenges in practical deployment due to their large size.
  • Efforts have been made to apply traditional network pruning techniques to LLMs to reduce their size without impacting performance.
  • A new pruning methodology called Outlier Weighed Layerwise sparsity (OWL) has been introduced, which considers non-uniform layerwise sparsity ratios based on outlier ratios within each layer.
  • Empirical evaluations show that OWL outperforms previous methods, achieving significant performance gains and faster inference speeds at high sparsity levels.

Read Full Article

like

6 Likes

For uninterrupted reading, download the app