Publications

Detailed Information

Deep Reinforcement Learning with LSTM-based Exploration Bonus

DC Field Value Language
dc.contributor.advisor유석인-
dc.contributor.author양진호-
dc.date.accessioned2017-07-14T02:36:19Z-
dc.date.available2017-07-14T02:36:19Z-
dc.date.issued2017-02-
dc.identifier.other000000140903-
dc.identifier.urihttps://hdl.handle.net/10371/122687-
dc.description학위논문 (석사)-- 서울대학교 대학원 : 컴퓨터공학부, 2017. 2. 유석인.-
dc.description.abstractDeep learning is dominant method in recent machine learning community. Deep learning outperforms traditional methods in various tasks such as image classification, object recognition, speech recognition, and natural language processing. However, most deep learning algorithm is focused on supervised learning task. Supervised learning only considers static environment therefore it is not proper to dynamic environment.

To solve this problem, reinforcement learning method combined with deep learning called deep reinforcement learning is proposed. Deep reinforcement learning is composed of two parts. First part is feature extraction using deep learning method. Second part is learning proper action for agent using reinforcement learning through trial and error.

However, reinforcement learning algorithm has exploration and exploration trade-off problem. It is hard to find optimal exploration and exploitation ratio for reinforcement learning. To solve this problem, we propose novel deep reinforcement learning algorithm with LSTM-based exploration bonus. LSTM-based exploration bonus uses Long-Short Term Memory (LSTM) networks as predictor and makes exploration bonus based on prediction error. LSTM-based exploration bonus guides agent play more daring action. As a result, agent could find optimal solution in more short time.

We test our method playing various Atari games. Experimental results shows our method outperforms plain deep reinforcement learning method. This shows the effectiveness of our method.
-
dc.description.tableofcontentsIntroduction 1
Related Works 3
Deep Reinforcement Learning 8
Experiment 23
Conclusion 27
Bibliography 28
요약 33
-
dc.formatapplication/pdf-
dc.format.extent3028176 bytes-
dc.format.mediumapplication/pdf-
dc.language.isoen-
dc.publisher서울대학교 대학원-
dc.subjectDeep Learning-
dc.subjectDeep Reinforcement Learning-
dc.subjectLong-Short Term Memory Networks-
dc.subjectExploration and Exploitation Trade-O ff-
dc.subject.ddc621-
dc.titleDeep Reinforcement Learning with LSTM-based Exploration Bonus-
dc.typeThesis-
dc.description.degreeMaster-
dc.citation.pages33-
dc.contributor.affiliation공과대학 컴퓨터공학부-
dc.date.awarded2017-02-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share