A Short Note on Combining Multiple Policies in Risk-Sensitive Exponential Average Reward Markov Decision Processes
dc.contributor.author | Chang, Hyeong Soo | |
dc.date.accessioned | 2009-09-30T13:36:39Z | |
dc.date.available | 2009-09-30T13:36:39Z | |
dc.date.issued | 2009 | |
dc.description | This work was done while he was a visiting associate professor at ISR, University of Maryland, College Park. | en |
dc.description.abstract | This short note presents a method of combining multiple policies in a given policy set such that the resulting policy improves all policies in the set for risk-sensitive exponential average reward Markov decision processes (MDPs), extending the work of Howard and Matheson for the singleton policy set case. Some applications of the method in solving risk-sensitive MDPs are also discussed. | en |
dc.description.sponsorship | Steven I. Marcus and Michael C. Fu | en |
dc.format.extent | 260067 bytes | |
dc.format.mimetype | application/pdf | |
dc.identifier.uri | http://hdl.handle.net/1903/9433 | |
dc.language.iso | en | en |
dc.relation.isAvailableAt | Institute for Systems Research | en_us |
dc.relation.isAvailableAt | Digital Repository at the University of Maryland | en_us |
dc.relation.isAvailableAt | University of Maryland (College Park, MD) | en_us |
dc.relation.ispartofseries | TR_2009-14 | en |
dc.subject | Risk-sensitive Markov decision process | en |
dc.subject | policy improvement | en |
dc.title | A Short Note on Combining Multiple Policies in Risk-Sensitive Exponential Average Reward Markov Decision Processes | en |
dc.type | Technical Report | en |
Files
Original bundle
1 - 1 of 1