Parameter-Efficient Fine-Tuning (PEFT) methods have become popular for adapting pre-trained Large Language Models (LLMs) to downstream tasks efficiently.
Existing PEFT techniques usually apply LoRA adapters uniformly across all layers, leading to redundant parameter allocation and suboptimal adaptation efficiency.
To address these issues, FLoE is introduced as a PEFT framework that utilizes Fisher information and Bayesian optimization for dynamic layer selection and optimal LoRA rank allocation, resulting in impressive efficiency-accuracy trade-offs.
FLoE is particularly beneficial in resource-constrained environments where rapid adaptation is crucial.