Publications

Detailed Information

Deep Bayesian Neural Networks for Continual Learning : 지속학습을 위한 심층 베이지안 신경망

DC Field Value Language
dc.contributor.advisor장병탁-
dc.contributor.author손성호-
dc.date.accessioned2018-05-29T03:31:47Z-
dc.date.available2018-05-29T03:31:47Z-
dc.date.issued2018-02-
dc.identifier.other000000149814-
dc.identifier.urihttps://hdl.handle.net/10371/141547-
dc.description학위논문 (석사)-- 서울대학교 대학원 : 공과대학 컴퓨터공학부, 2018. 2. 장병탁.-
dc.description.abstractIn the problem of continual learning, tasks are given in sequence and the goal of the machine is to learn new tasks while retaining previously learned tasks' performances. Even though deep neural networks have become prevalent among machine learning techniques recently, they may fail to deal with continual learning environment. As the model learns new tasks, parameters contributing largely to previous tasks' performances may change and result in poor performance. This phenomenon is called catastrophic forgetting, and researchers tackled this problem using regularizers and structure optimization techniques.

We aim to show that applying Bayesian modelling to deep neural network is beneficial for continual learning. Not only does Bayesian framework provide systematic way of performing online learning, it also provides uncertainty estimates for measuring each parameter's contribution to previously learned tasks' performances. By rescaling gradients applied to parameters with large performance contribution, the model can retain performances to previous tasks longer. This thesis shows how deep Bayesian neural networks utilize model uncertainty, which leads to alleviation of catastrophic forgetting in continual learning environment.
-
dc.description.tableofcontentsChapter 1 Introduction 1

Chapter 2 Preliminaries 5
2.1 Approaches on Catastrophic Forgetting 5
2.2 Deep Bayesian Neural Networks 7

Chapter 3 Bayes by Backprop 9
3.1 Variational Approximation to the Posterior 10
3.2 Monte Carlo Gradients 11
3.3 Gaussian Variational Posterior 11

Chapter 4 Proposed Method 14
4.1 Bayes by Backprop for Online Learning 15
4.1.1 Recursive Bayesian Estimation 15
4.1.2 Mode-fitting with Posterior Uncertainty Reset 16
4.2 Rescaled Gradient Method 19

Chapter 5 Experiments 23
5.1 Configuration 23
5.2 Disjoint MNIST Problem 24
5.3 Permuted MNIST Problem 28

Chapter 6 Conclusion 41

Bibliography 45

국문초록 49
-
dc.formatapplication/pdf-
dc.format.extent4845684 bytes-
dc.format.mediumapplication/pdf-
dc.language.isoen-
dc.publisher서울대학교 대학원-
dc.subjectdeep learning-
dc.subjectbayesian neural network-
dc.subjectcontinual learning-
dc.subjectcatastrophic forgetting-
dc.subjectuncertainty estimation-
dc.subject.ddc621.39-
dc.titleDeep Bayesian Neural Networks for Continual Learning-
dc.title.alternative지속학습을 위한 심층 베이지안 신경망-
dc.typeThesis-
dc.contributor.AlternativeAuthorSeongho Son-
dc.description.degreeMaster-
dc.contributor.affiliation공과대학 컴퓨터공학부-
dc.date.awarded2018-02-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share