Researchers at Sakana AI have developed a resource-efficient framework called CycleQD that combines the skills of different language models without expensive training processes.
CycleQD creates swarms of task-specific models, offering a sustainable alternative to increasing model size.
The technique incorporates quality diversity (QD), using evolutionary algorithms to create populations of models with different skill domains.
CycleQD outperforms traditional fine-tuning methods, demonstrating its effectiveness in training models to excel across multiple skills.