This work proposes an approach that integrates reinforcement learning and model predictive control (MPC) to solve finite-horizon optimal control problems in mixed-logical dynamical systems efficiently.
The approach aims to mitigate the curse of dimensionality by decoupling the decision on the discrete variables from the decision on the continuous variables.
Reinforcement learning determines the discrete decision variables, simplifying the online optimization problem of the MPC controller and reducing computational time.
Simulation experiments on a microgrid system demonstrate that the proposed method substantially reduces the online computation time of MPC while maintaining high feasibility and low suboptimality.