Multi-task model merging techniques aim to combine knowledge from task-specific experts into a unified model efficiently.
A new approach called Layer-wise Optimal Task Vector Merging (LOT Merging) is introduced to minimize feature drift during model merging.
LOT Merging minimizes feature differences between task-specific experts and the unified model in a layer-by-layer manner, enhancing performance.
Extensive experiments show that LOT Merging outperforms existing methods, achieving improvements of up to 4.4% on vision and vision-language benchmarks.