Recent advancements in imitation learning have led to transformer-based behavior foundation models (BFMs) that enable multi-modal, human-like control for humanoid agents.
The introduction of "Task Tokens" provides a method to tailor BFMs to specific tasks while maintaining flexibility.
Task Tokens leverage the transformer architecture of BFMs to learn a task-specific encoder through reinforcement learning, allowing the incorporation of user-defined priors and balancing reward design.
Task Tokens demonstrate efficacy in various tasks, including out-of-distribution scenarios, and are compatible with other prompting modalities.