Detail výsledku
Small Decision Trees for MDPs with Deductive Synthesis
Češka Milan, doc. RNDr., Ph.D., UITS (FIT)
Junges Sebastian
Macák Filip, Ing., UITS (FIT)
Markov decision processes (MDPs) describe decision making subject to probabilistic uncertainty. A classical problem on MDPs is to compute a policy, selecting actions in every state, that maximizes the probability of reaching a dedicated set of target states. Computing such policies in tabular form is efficiently possible via standard algorithms. However, for further processing by either humans or machines, policies should be represented concisely, e.g., as a decision tree. This paper considers finding (almost) optimal decision trees of minimal depth and contributes a deductive synthesis approach. Technically, we combine pruning the space of concise policies with an abstraction-refinement loop with an SMT-encoding that maps candidate policies into decision trees. Our experiments show that this approach beats the state-of-the-art solver using an MILP encoding by orders of magnitude. The approach also pairs well with heuristic approaches that map a fixed policy into a decision tree: for an MDP with 1.5M states, our approach reduces the size of the given tree by 90%, while sacrificing only 1% of the optimal performance.
Markov Decision Processes; Decision trees; Deductive synthesis
@inproceedings{BUT198879,
author="Roman {Andriushchenko} and Milan {Češka} and {} and Filip {Macák}",
title="Small Decision Trees for MDPs with Deductive Synthesis",
booktitle="Computer Aided Verification",
year="2025",
pages="169--192",
publisher="Springer Cham",
doi="10.1007/978-3-031-98679-6\{_}8",
isbn="978-3-031-98678-9",
url="https://doi.org/10.1007/978-3-031-98679-6_8"
}
Verification and Analysis for Safety and Security of Applications in Life, EU, HORIZON EUROPE, SEP-210979090, zahájení: 2024-06-01, ukončení: 2027-05-31, řešení
VESCAA: Verifikovatelná a efektivní syntéza kontrolerů, GAČR, Standardní projekty, GA23-06963S, zahájení: 2023-03-01, ukončení: 2025-12-31, řešení