Description
This work explores combining state-of-the-art \gls{mbrl} algorithms focused on learning complex policies with large state-spaces and augmenting them with distributional reward perspective on \gls{rl} algorithms. Distributional \gls{rl} provides a probabilistic reward formulation as opposed to the classic \gls{rl} formulation which models the estimation of this distributional return. These probabilistic reward formulations help the agent choose highly risk-averse actions, which in turn makes the learning more stable. To evaluate this idea, I experiment in simulation on complex high-dimensional environments when subject under different noisy conditions.
Details
Title
- Learning Policies for Model-Based Reinforcement Learning Using Distributed Reward Formulation
Contributors
- Agarwal, Nikhil (Author)
- Ben Amor, Heni (Thesis advisor)
- Phielipp, Mariano (Committee member)
- DV, Hemanth (Committee member)
- Arizona State University (Publisher)
Date Created
The date the item was original created (prior to any relationship with the ASU Digital Repositories.)
2021
Subjects
Resource Type
Collections this item is in
Note
- Partial requirement for: M.S., Arizona State University, 2021
- Field of study: Computer Science