Continuous time systems are typically modeled using discrete time dynamics, requiring a small simulation step for accuracy.
Proposed approach involves using temporally-extended actions to control continuous decision timescale, allowing for deep horizon search with shallow planner depth.
This method speeds up trajectory simulation, reduces errors in model-based reinforcement learning, and improves training time for models.
Experimental evaluation demonstrates that the approach results in faster planning, better solutions, and solves problems not addressed in the standard formulation.