Representation Learning for Reinforcement Learning: Modeling Non-Gaussian Transition Probabilities with a Wasserstein Critic

dc.contributor.advisorZhang, Kaiqingen_US
dc.contributor.authorTse, Ryanen_US
dc.contributor.departmentElectrical Engineeringen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2025-01-29T06:46:40Z
dc.date.available2025-01-29T06:46:40Z
dc.date.issued2024en_US
dc.description.abstractReinforcement learning algorithms depend on effective state representations when solving complex, high-dimensional environments. Recent methods learn state representations using auxiliary objectives that aim to capture relationships between states that are behaviorally similar, meaning states that lead to similar future outcomes under optimal policies. These methods learn explicit probabilistic state transition models and compute distributional distances between state transition probabilities as part of their measure of behavioral similarity. This thesis presents a novel extension to several of these methods that directly learns the 1-Wasserstein distance between state transition distributions by exploiting the Kantorovich-Rubenstein duality. This method eliminates parametric assumptions about the state transition probabilities while providing a smoother estimator of distributional distances. Empirical evaluation demonstrates improved sample efficiency over some of the original methods and a modest increase in computational cost per sample. The results establish that relaxing theoretical assumptions about state transition modeling leads to more flexible and robust representation learning while maintaining strong performance characteristics.xen_US
dc.identifierhttps://doi.org/10.13016/fhth-mcku
dc.identifier.urihttp://hdl.handle.net/1903/33716
dc.language.isoenen_US
dc.subject.pqcontrolledArtificial intelligenceen_US
dc.subject.pqcontrolledComputer scienceen_US
dc.subject.pqcontrolledApplied mathematicsen_US
dc.subject.pquncontrolledreinforcement learningen_US
dc.subject.pquncontrolledrepresentation learninen_US
dc.subject.pquncontrolledWassersteinen_US
dc.titleRepresentation Learning for Reinforcement Learning: Modeling Non-Gaussian Transition Probabilities with a Wasserstein Criticen_US
dc.typeThesisen_US

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
Tse_umd_0117N_24835.pdf
Size:
17.62 MB
Format:
Adobe Portable Document Format