Publications

Detailed Information

Neural Tangent Kernel Analysis of Deep Narrow Neural Networks : 좁고 깊은 심층신경망의 뉴럴 탄젠트 커널 분석

DC Field Value Language
dc.contributor.advisorErnest K. Ryu-
dc.contributor.author이종민-
dc.date.accessioned2023-11-20T04:49:35Z-
dc.date.available2023-11-20T04:49:35Z-
dc.date.issued2023-
dc.identifier.other000000179443-
dc.identifier.urihttps://hdl.handle.net/10371/197304-
dc.identifier.urihttps://dcollection.snu.ac.kr/common/orgView/000000179443ko_KR
dc.description학위논문(석사) -- 서울대학교대학원 : 자연과학대학 수리과학부, 2023. 8. Ernest K. Ryu.-
dc.description.abstractThe tremendous recent progress in analyzing the training dynamics of over parameterized neural networks has primarily focused on wide networks and therefore does not sufficiently address the role of depth in deep learning. In this work, we present the first trainability guarantee of infinitely deep but narrow neural networks. We study the infinite-depth limit of a multilayer perceptron (MLP) with a specific initialization and establish a trainability guarantee using the NTK theory. We then extend the analysis to an infinitely deep convolutional neural network (CNN) and perform brief experiments.-
dc.description.abstract과매개화된 신경망의 훈련 역학을 분석하는 최근의 엄청난 발전은 주로 넓은 네트워크에 초점을 맞추었기 때문에 딥 러닝에서 깊이의 역할을 충분히 다루지 못 한다. 이 논문에서 우리는 무한히 깊지만 좁은 신경망의 훈련 가능성을 처음으로 보인다. 우리는 특정 초기화하에서 무한한 깊이의 다층 신경망을 연구하고 뉴럴 탄젠트커널 이론을 사용하여 학습 가능성울 보장한다. 그런 다음 분석을 무한히 깊은 합성곱 신경망으로 확장하고 간단한 실험을 수행한다.-
dc.description.tableofcontents1 Introduction 2
1.1 Prior works 3
2 Preliminaries and Notations 5
2.1 Kernel gradient flow 6
2.2 Neural tangent kernel 7
3 NTK analysis of infinitely deep MLP 9
3.1 Initialization 10
3.1.1 Gradient flow and neural tangent kernel 11
3.2 Convergence in infinite-depth limit 13
3.3 Proof outline 15
4 NTK analysis of infinitely deep CNN 19
4.0.1 Initialization 20
4.0.2 Convergence in infinite-depth limit 22
5 Experiments 24
5.1 Convergence of the scaled NTK 24
5.2 Trainability of the deep narrow neural network 24
5.3 Accumulation of the layer-wise effect 26
6 Conclusion 27
7 Appendix 28
Bibliography 29
초 록 37
감사의 글 38
-
dc.format.extentvi, 38-
dc.language.isoeng-
dc.publisher서울대학교 대학원-
dc.subjectNeural tangent kernel-
dc.subjectGradient flow-
dc.subjectDeep narrow neural network-
dc.subjectArtificial intelligence-
dc.subject.ddc510-
dc.titleNeural Tangent Kernel Analysis of Deep Narrow Neural Networks-
dc.title.alternative좁고 깊은 심층신경망의 뉴럴 탄젠트 커널 분석-
dc.typeThesis-
dc.typeDissertation-
dc.contributor.AlternativeAuthorJongmin Lee-
dc.contributor.department자연과학대학 수리과학부-
dc.description.degree석사-
dc.date.awarded2023-08-
dc.identifier.uciI804:11032-000000179443-
dc.identifier.holdings000000000050▲000000000058▲000000179443▲-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share