Deep Reinforcement Learning (RL) models often fail to generalize when changes occur in the environment's observations or task requirements.
This paper proposes a zero-shot method for mapping between latent spaces across different agents trained on different visual and task variations.
The approach learns a transformation that maps embeddings from one agent's encoder to another agent's encoder without further fine-tuning.
The framework preserves high performance under visual and task domain shifts, allowing for more robust reinforcement learning in dynamically changing environments.