๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Bounded-parameter Markov decision processes

โœ Scribed by Robert Givan; Sonia Leach; Thomas Dean


Publisher
Elsevier Science
Year
2000
Tongue
English
Weight
375 KB
Volume
122
Category
Article
ISSN
0004-3702

No coin nor oath required. For personal study only.

โœฆ Synopsis


In this paper, we introduce the notion of a bounded-parameter Markov decision process (BMDP) as a generalization of the familiar exact MDP. A bounded-parameter MDP is a set of exact MDPs specified by giving upper and lower bounds on transition probabilities and rewards (all the MDPs in the set share the same state and action space). BMDPs form an efficiently solvable special case of the already known class of MDPs with imprecise parameters (MDPIPs). Bounded-parameter MDPs can be used to represent variation or uncertainty concerning the parameters of sequential decision problems in cases where no prior probabilities on the parameter values are available. Boundedparameter MDPs can also be used in aggregation schemes to represent the variation in the transition probabilities for different base states aggregated together in the same aggregate state.

We introduce interval value functions as a natural extension of traditional value functions. An interval value function assigns a closed real interval to each state, representing the assertion that the value of that state falls within that interval. An interval value function can be used to bound the performance of a policy over the set of exact MDPs associated with a given bounded-parameter MDP. We describe an iterative dynamic programming algorithm called interval policy evaluation that computes an interval value function for a given BMDP and specified policy. Interval policy evaluation on a policy ฯ€ computes the most restrictive interval value function that is sound, i.e., that bounds the value function for ฯ€ in every exact MDP in the set defined by the bounded-parameter MDP. We define optimistic and pessimistic criteria for optimality, and provide a variant of value iteration (Bellman, 1957) that we call interval value iteration that computes policies for a BMDP that are optimal with respect to these criteria. We show that each algorithm we present converges to the desired values in a polynomial number of iterations given a fixed discount factor.


๐Ÿ“œ SIMILAR VOLUMES


Bounding reward measures of Markov model
โœ Peter Buchholz ๐Ÿ“‚ Article ๐Ÿ“… 2011 ๐Ÿ› John Wiley and Sons ๐ŸŒ English โš– 388 KB

## SUMMARY For a Markov reward process, where upper and lower bounds for the transition rates and rewards are known, a new approach to bound the expected reward is presented. Based on a previous paper where sharp bounds have been defined for the problem, but only an inefficient and unstable algorit

Markov ratio decision processes
โœ V. Aggarwal; R. Chandrasekaran; K. P. K. Nair ๐Ÿ“‚ Article ๐Ÿ“… 1977 ๐Ÿ› Springer ๐ŸŒ English โš– 466 KB
Semi-infinite Markov decision processes
โœ Ming Chen; Jerzy A. Filar; Ke Liu ๐Ÿ“‚ Article ๐Ÿ“… 2000 ๐Ÿ› Springer ๐ŸŒ English โš– 184 KB