A learning-based approach is proposed to synthesize policies for huge parameterized Markov decision processes (MDPs).
The method generalizes optimal policies obtained from model-checking small instances to larger ones using decision-tree learning.
By bypassing the need for explicit state-space exploration of large models, the method provides a practical solution to the state-space explosion problem.
Experimental results show that the policies perform well even for models beyond the reach of state-of-the-art analysis tools.