Robust decomposable Markov decision processes motivated by allocating school budgets |
| |
Authors: | Nedialko B Dimitrov Stanko Dimitrov Stefanka Chukova |
| |
Institution: | 1. Graduate Program in Operations Research & Industrial Engineering, The University of Texas at Austin, USA;2. Management Sciences, University of Waterloo, Canada;3. School of Mathematics, Statistics and Operations Research, Victoria University of Wellington, New Zealand |
| |
Abstract: | Motivated by an application to school funding, we introduce the notion of a robust decomposable Markov decision process (MDP). A robust decomposable MDP model applies to situations where several MDPs, with the transition probabilities in each only known through an uncertainty set, are coupled together by joint resource constraints. Robust decomposable MDPs are different than both decomposable MDPs, and robust MDPs and cannot be solved by a direct application of the solution methods from either of those areas. In fact, to the best of our knowledge, there is no known method to tractably compute optimal policies in robust, decomposable MDPs. We show how to tractably compute good policies for this model, and apply the derived method to a stylized school funding example. |
| |
Keywords: | Markov processes Dynamic programming-optimal control School funding |
本文献已被 ScienceDirect 等数据库收录! |
|