dc.contributor.author | Guberina, Marko | |
dc.contributor.author | Desta, Betelhem Dejene | |
dc.date.accessioned | 2022-10-14T07:40:59Z | |
dc.date.available | 2022-10-14T07:40:59Z | |
dc.date.issued | 2022-10-14 | |
dc.identifier.uri | https://hdl.handle.net/2077/73890 | |
dc.description.abstract | Reinforcement learning struggles to solve control tasks on directly on images. Performance
on identical tasks with access to the underlying states is much better. One
avenue to bridge the gap between the two is to leverage unsupervised learning as a
means of learning state representations from images, thereby resulting in a better
conditioned reinforcement learning problem. Through investigation of related work,
characteristics of successful integration of unsupervised learning and reinforcement
learning are identified. We hypothesize that joint training of state representations
and policies result in higher sample-efficiency if adequate regularization is provided.
We further hypothesize that representations which correlate more strongly with the
underlying Markov decision process result in additional sample-efficiency. These hypotheses
are tested through a simple deterministic generative representation learning
model (autoencoder) trained with image reconstruction loss and additional forward
and inverse auxiliary losses. While our algorithm does not reach state-of-the-art
performance, its modular implementation integrated in the reinforcement learning
library Tianshou enables easy use to reinforcement learning practitioners, and thus
also accelerates further research. We also identify which aspects of our solution are
most important and use them to formulate promising research directions. In our
tests we limited ourselves to Atari environments and primarily used Rainbow as the
underlying reinforcement learning algorithm. | en_US |
dc.language.iso | eng | en_US |
dc.subject | sample-efficient reinforcement learning | en_US |
dc.subject | state representation learning | en_US |
dc.subject | unsupervised learning | en_US |
dc.subject | autoencoder | en_US |
dc.title | Improving sample-efficiency of model-free reinforcement learning algorithms on image inputs with representation learning | en_US |
dc.type | text | |
dc.setspec.uppsok | Technology | |
dc.type.uppsok | H2 | |
dc.contributor.department | Göteborgs universitet/Institutionen för data- och informationsteknik | swe |
dc.contributor.department | University of Gothenburg/Department of Computer Science and Engineering | eng |
dc.type.degree | Student essay | |