A Distributed Algorithm for Solving a Class of Multi-agent Markov Decision Problems

dc.contributor.advisorFu, Michael C.en_US
dc.contributor.authorChang, Hyeong Sooen_US
dc.contributor.authorFu, Michael C.en_US
dc.contributor.departmentISRen_US
dc.date.accessioned2007-05-23T10:13:46Z
dc.date.available2007-05-23T10:13:46Z
dc.date.issued2003en_US
dc.description.abstractWe consider a class of infinite horizon Markov decision processes (MDPs) with multiple decision makers, called agents,and a general joint reward structure, but a special decomposable state/action structure such that each individual agent's actions affect the system's state transitions independently from the actions of all other agents. We introduce the concept of ``localization," where each agent need only consider a ``local" MDP defined on its own state and action spaces. Based on this localization concept, we propose an iterative distributed algorithm that emulates gradient ascent and which converges to a locally optimal solution for the average reward case. The solution is an ``autonomous" joint policy such that each agent's action is based on only its local state. Finally, we discuss the implication of the localization concept for discounted reward problems.en_US
dc.format.extent195556 bytes
dc.format.mimetypeapplication/pdf
dc.identifier.urihttp://hdl.handle.net/1903/6357
dc.language.isoen_USen_US
dc.relation.ispartofseriesISR; TR 2003-25en_US
dc.subjectNULLen_US
dc.titleA Distributed Algorithm for Solving a Class of Multi-agent Markov Decision Problemsen_US
dc.typeTechnical Reporten_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
TR_2003-25.pdf
Size:
190.97 KB
Format:
Adobe Portable Document Format