Publications

Detailed Information

Understanding Mutual Information in Contrastive Representation Learning : 대조 표현 학습에서 상호 정보의 이해

DC Field Value Language
dc.contributor.advisorWonjong Rhee-
dc.contributor.author이경은-
dc.date.accessioned2023-06-29T02:26:50Z-
dc.date.available2023-06-29T02:26:50Z-
dc.date.issued2023-
dc.identifier.other000000174657-
dc.identifier.urihttps://hdl.handle.net/10371/194081-
dc.identifier.urihttps://dcollection.snu.ac.kr/common/orgView/000000174657ko_KR
dc.description학위논문(박사) -- 서울대학교대학원 : 융합과학기술대학원 융합과학부(디지털정보융합전공), 2023. 2. Wonjong Rhee.-
dc.description.abstractContrastive learning has played a pivotal role in the recent success of unsupervised representation learning. It has been commonly explained with instance discrimination and a mutual information loss, and some of the fundamental explanations are based on mutual information analysis. An analysis based on mutual information, however, can be misleading. First of all, an exact quantification of mutual information over a real-world dataset is challenging. It has not been solved because we cannot access the true joint distribution function of real-world dataset before. Second, previous studies have equated the limitations of contrastive learning with them of mutual information estimation in the absence of the rigorous investigation for a relationship between them. Third, what information is actually being shared by the two views is overlooked. Without carefully examining what information is actually being shared, the interpretation can be completely misleading. In this work, we develop new methods that enable rigorous analysis of mutual information in contrastive learning. We also evaluate the accuracy of variational MI estimators across various data domains, including images and texts. Using the methods, we investigate three existing beliefs and show that they are incorrect. Based on the investigation results, we address two issues in the discussion section. In particular, we question if contrastive learning is indeed an unsupervised representation learning method because the current framework of contrastive learning relies on validation performance for tuning the augmentation design.-
dc.description.tableofcontentsChapter 1. Introduction 1
1.1 Contributions 6
Chapter 2. Background 9
2.1 Contrastive representation learning 9
2.1.1 Previous works to understand contrastive learning 11
2.2 Mutual Information 12
2.3 Variational Mutual Information Estimators 15
2.3.1 Critic function 18
2.3.2 Limitations of the variational MI estimators 19
Chapter 3. Same-class Sampling for Positive Pairing 21
Chapter 4. Understanding the Accuracy of Variational Mutual Information Estimators 27
4.1 Datasets 29
4.1.1 Gaussian dataset 30
4.1.2 Definitions of ds, dr, and Z 30
4.1.3 Details of generating datasets 31
4.2 Experimental setup 33
4.3 Experimental results 34
4.3.1 Critic architecture 34
4.3.2 Critic capacity 38
4.3.3 Choice of the variational MI estimator 39
4.3.4 Number of information sources 39
4.3.5 Representation dimension 40
4.3.6 Nuisance 41
4.3.7 Deep representations 41
4.4 Discussion: How can we make use of MI with practical datasets? 44
4.5 Conclusion 48
Chapter 5. Examining Three Existing Beliefs on Mutual Information in Contrastive Learning 49
5.1 Method 50
5.1.1 Post-training MI estimation 50
5.1.2 CDP dataset 52
5.2 Experimental setups 56
5.2.1 Training 56
5.2.2 Post-training MI estimation 57
5.3 Results 59
5.3.1 A small batch size is a limiting factor for MI estimation but not for contrastive learning. 59
5.3.2 Augmentation-based MI and other metrics are not effective, but MI class is effective. 62
5.3.3 Minimizing task-irrelevant information (InfoMin) is not always necessary. 70
5.4 Discussion 77
5.5 Conclusion 83
Chapter 6. Conclusion 84
6.1 Limitations 86
6.2 Future works 86
Bibliography 88
Appendices 99
-
dc.format.extentxii, 106-
dc.language.isoeng-
dc.publisher서울대학교 대학원-
dc.subjectrepresentation learning-
dc.subjectinformation theory-
dc.subjectmutual information-
dc.subjectvariational bounds of mutual information-
dc.subjectdeep representations-
dc.subjectcontrastive learning-
dc.subject.ddc004-
dc.titleUnderstanding Mutual Information in Contrastive Representation Learning-
dc.title.alternative대조 표현 학습에서 상호 정보의 이해-
dc.typeThesis-
dc.typeDissertation-
dc.contributor.AlternativeAuthorKyungeun Lee-
dc.contributor.department융합과학기술대학원 융합과학부(디지털정보융합전공)-
dc.description.degree박사-
dc.date.awarded2023-02-
dc.identifier.uciI804:11032-000000174657-
dc.identifier.holdings000000000049▲000000000056▲000000174657▲-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share