Approximate Receding Horizon Approach for Markov Decision Processes: Average Reward Case
dc.contributor.author | Chang, Hyeong Soo | en_US |
dc.contributor.author | Marcus, Steven I. | en_US |
dc.contributor.department | ISR | en_US |
dc.date.accessioned | 2007-05-23T10:10:51Z | |
dc.date.available | 2007-05-23T10:10:51Z | |
dc.date.issued | 2001 | en_US |
dc.description.abstract | Building on the receding horizon approach by Hernandez-Lerma andLasserre in solving Markov decision processes (MDPs),this paper first analyzes the performance of the (approximate) receding horizon approach in terms of infinite horizon average reward. <p>In this approach, we fix a finite horizon and at each decision time, we solve the given MDP with the finite horizon for an approximately optimal current action and take the action to control the MDP.<p>We then analyze recently proposed on-line policy improvementscheme, "roll-out," by Bertsekas and Castanon, and a generalization of the rollout algorithm, "parallel rollout" by Chang et al., in terms of the infinite horizon average reward in the framework of the (approximate) receding horizon control.<p>We finally discuss practical implementations of these schemes via simulation. | en_US |
dc.format.extent | 188249 bytes | |
dc.format.mimetype | application/pdf | |
dc.identifier.uri | http://hdl.handle.net/1903/6208 | |
dc.language.iso | en_US | en_US |
dc.relation.ispartofseries | ISR; TR 2001-46 | en_US |
dc.subject | Global Communication Systems | en_US |
dc.title | Approximate Receding Horizon Approach for Markov Decision Processes: Average Reward Case | en_US |
dc.type | Technical Report | en_US |
Files
Original bundle
1 - 1 of 1