Publications

Detailed Information

AModifiedLeastAngleRegressionAlgorithmforHierarchicalInteraction : 계층적 교호작용을 고려한 수정된 HLARS 알고리즘 개발

DC Field Value Language
dc.contributor.advisor김용대-
dc.contributor.author김우성-
dc.date.accessioned2017-07-14T00:31:52Z-
dc.date.available2017-07-14T00:31:52Z-
dc.date.issued2016-02-
dc.identifier.other000000133389-
dc.identifier.urihttps://hdl.handle.net/10371/121159-
dc.description학위논문 (박사)-- 서울대학교 대학원 : 통계학과, 2016. 2. 김용대.-
dc.description.abstractVariable selection is important in high dimensional regression. Traditional variable selection methods such as stepwise selection are unstable which means that the set of the selected variables is sensitive according to the change of data sets. As an alternative to those methods, a series of sparse penalized methods are used for estimation and variable selection simultaneously. The full set of LASSO solutions can be calculated by a minor modification of the LARS algorithm.
In many important practical problems, the main effect alone may not be enough to capture the relationship between the response and predictors, and high-order interactions are often of interest to scientific researchers. In considering two-way interaction models with a large number of covariates, we often would like to determine a smaller subset that exhibits strong effects on the response variable have been suggested.
Considering all possible interactions, however, is almost impossible due to computational burden when the number of covariate is large. To resolve this problem, the heredity structure between the main and interaction effect can be considered, algorithms for LASSO with heredity structure. However these algorithms cannot be executed if the number of main effects is large since still computational burden is large. To resolve this issue, we suggest a hierarchical LARS algorithm which can be parallelized easily with MPI.
The proposed hierarchical LARS is a modified version of LARS, but it is more faster than LARS and has comparable prediction accuracy. It can be scaled up since it is possible to be executed in parallel process. MPI is a well known parallel model and we suggest a MPI-version of hierarchical LARS.
-
dc.description.tableofcontents1 Introduction 1
1.1 Overview 1
1.2 Outline of the thesis 6

2 Literature Review : Variable Selection Methods on High Dimensions and MPI 7
2.1 Sparse regularization methods 7
2.2 Sparse regularization methods for two-way interaction linear model 14
2.3 Message passing interface 24
2.3.1 Parallel computational models 25
2.3.2 Advantages of the MPI 29
2.3.3 Basic MPI concepts 31

3 LARS for hierarchical interaction 34
3.1 HLARS algorithm 34
3.2 MPI algorithm for HLARS 41
3.2.1 Modification of HLARS 42

4 Numerical studies 48
4.1 Simulation studies 48
4.1.1 Simulation 1 50
4.1.2 Simulation 2 54
4.1.3 Simulation result for MPI modified HLARS 57
4.2 Real data analysis 59

5 Concluding remarks 62

Bibliography 63

Abstract (in Korean) 67
-
dc.formatapplication/pdf-
dc.format.extent3339885 bytes-
dc.format.mediumapplication/pdf-
dc.language.isoen-
dc.publisher서울대학교 대학원-
dc.subjectHigh dimensional regression-
dc.subjectTwo way interaction model-
dc.subjectLASSO-
dc.subjectLARS-
dc.subjectMassage passing interface-
dc.subject.ddc519-
dc.titleAModifiedLeastAngleRegressionAlgorithmforHierarchicalInteraction-
dc.title.alternative계층적 교호작용을 고려한 수정된 HLARS 알고리즘 개발-
dc.typeThesis-
dc.description.degreeDoctor-
dc.citation.pages68-
dc.contributor.affiliation자연과학대학 통계학과-
dc.date.awarded2016-02-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share