Please use this identifier to cite or link to this item: http://hdl.handle.net/11718/1476
Title: Mean variance optimality criteria for discounted Markov decision processes
Authors: Satia, J. K.
Keywords: Optimality;Markov decision processes;Mean variance optimality criteria
Issue Date: 22-Mar-2010
Series/Report no.: WP;1978/246
Abstract: The criteria of maximizing expected rewards has been widely used in Markov decision processes following Howard [2]. Recently considerations related to higher moments of rewards have also been incorporated by Jaquette [4] and Goldwerger [1]. This paper considers mean variance criteria for discounted Markov decision processes. Variability in rewards arising both out of variability of rewards during each period and due to stochastic nature of transitions is considered. It is shown that randomized policies need not be considered when a function of mean and variance ( - a ) is to be optimized. However an example illustrates that policies which will simultaneously minimize variances for all states may not exist. We, therefore, provide a dynamic programming formulation for optimizing i - a i for each state i. An example is given to illustrate the procedure.
URI: http://hdl.handle.net/11718/1476
Appears in Collections:Working Papers

Files in This Item:
File Description SizeFormat 
WP 1978_246.pdf533.99 kBAdobe PDFView/Open


Items in IIMA Institutional Repository are protected by copyright, with all rights reserved, unless otherwise indicated.