Combinatorial optimization problems are difficult due to their discrete structure and large solution space.
Deep reinforcement learning (DRL) has shown the ability to learn heuristics from data but can struggle with limited exploration and local optima.
Genetic Algorithms (GAs) excel in global exploration but are sample inefficient and computationally intensive.
A new framework called Evolutionary Augmentation Mechanism (EAM) combines DRL efficiency with GA's global search power by refining solutions through genetic operations.
EAM enhances exploration and speeds up convergence by integrating evolved solutions back into the policy training loop.
Theoretical analysis ensures stable policy updates by establishing an upper bound on the KL divergence between evolved and policy distributions.
EAM is versatile and can be used with various DRL solvers like Attention Model, POMO, and SymNCO.
Extensive testing on benchmark problems like TSP, CVRP, PCTSP, and OP shows EAM improves solution quality and training efficiency compared to baselines.