Current parameter-efficient fine-tuning methods for adapting pre-trained language models to downstream tasks are susceptible to interference from noisy data.
Proposed a noise-robust adaptation method via asymmetric LoRA poisoning experts (LoPE) framework that enhances model robustness with generated noisy data.
LoPE strategically integrates a dedicated poisoning expert in an asymmetric LoRA configuration to enhance noise discrimination and processing ability.
Extensive experiments demonstrate that LoPE achieves strong performance and robustness purely through low-cost noise injection, eliminating the need for data cleaning.