Researchers have introduced Generalized Fisher-Weighted SVD (GFWSVD) as a new post-training technique for compressing large language models (LLMs).
GFWSVD considers both diagonal and off-diagonal elements of the Fisher information matrix to better reflect parameter importance, addressing the limitations of simple diagonal approximations used in previous methods.
The method utilizes a scalable adaptation of the Kronecker-factored approximation algorithm for the observed Fisher information, improving compression performance.
In experiments on LLM compression, GFWSVD outperformed existing compression techniques like FWSVD, SVD-LLM, and ASVD at various compression rates, showcasing improved effectiveness.