Large Language Models (LLMs) are becoming increasingly large, limiting their use in computationally constrained environments.
Researchers have proposed a novel approach to extract task-specific circuits from LLMs for faster inference.
The extracted subset of the LLM can perform a targeted task without additional training and with a small amount of data samples.
The resulting models are considerably smaller, reducing the number of parameters up to 82.77% and more interpretable using Mechanistic Interpretability (MI) techniques.