1–2–3–Go! Policy Synthesis for Parameterized Markov Decision Processes via Decision-Tree Learning and Generalization

Muqsit Azeem, Debraj Chakraborty, Sudeep Kanav, Jan Křetínský, Mohammadsadegh Mohagheghi, Stefanie Mohr, Maximilian Weininger

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Despite the advances in probabilistic model checking, the scalability of the verification methods remains limited. In particular, the state space often becomes extremely large when instantiating parameterized Markov decision processes (MDPs) even with moderate values. Synthesizing policies for such huge MDPs is beyond the reach of available tools. We propose a learning-based approach to obtain a reasonable policy for such huge MDPs. The idea is to generalize optimal policies obtained by model-checking small instances to larger ones using decision-tree learning. Consequently, our method bypasses the need for explicit state-space exploration of large models, providing a practical solution to the state-space explosion problem. We demonstrate the efficacy of our approach by performing extensive experimentation on the relevant models from the quantitative verification benchmark set. The experimental results indicate that our policies perform well, even when the size of the model is orders of magnitude beyond the reach of state-of-the-art analysis tools.

Original languageEnglish
Title of host publicationVerification, Model Checking, and Abstract Interpretation - 26th International Conference, VMCAI 2025, Proceedings
EditorsKrishna Shankaranarayanan, Sriram Sankaranarayanan, Ashutosh Trivedi
PublisherSpringer Science and Business Media Deutschland GmbH
Pages97-120
Number of pages24
ISBN (Print)9783031827020
DOIs
StatePublished - 2025
Event26th International Conference on Verification, Model Checking, and Abstract Interpretation, VMCAI 2025 - Denver, United States
Duration: 20 Jan 202521 Jan 2025

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume15530 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference26th International Conference on Verification, Model Checking, and Abstract Interpretation, VMCAI 2025
Country/TerritoryUnited States
CityDenver
Period20/01/2521/01/25

Keywords

  • Markov decision process
  • model checking
  • policy synthesis
  • probabilistic verification

Fingerprint

Dive into the research topics of '1–2–3–Go! Policy Synthesis for Parameterized Markov Decision Processes via Decision-Tree Learning and Generalization'. Together they form a unique fingerprint.

Cite this