Publications

Detailed Information

Study on Attention-based LSTM Model for Multivariate Time-series Prediction : 다 변수 시계열 예측을 위한 주의 기반 LSTM 모델 연구

DC Field Value Language
dc.contributor.advisorKang, U-
dc.contributor.authorYang Bai-
dc.date.accessioned2019-10-18T15:46:51Z-
dc.date.available2019-10-18T15:46:51Z-
dc.date.issued2019-08-
dc.identifier.other000000157939-
dc.identifier.urihttps://hdl.handle.net/10371/161080-
dc.identifier.urihttp://dcollection.snu.ac.kr/common/orgView/000000157939ko_KR
dc.description학위논문(석사)--서울대학교 대학원 :공과대학 컴퓨터공학부,2019. 8. Kang, U.-
dc.description.abstractGiven previous observations of a multivariate time-series, how can we accurately predict the future value of several steps ahead? With the continuous development of sensor systems and computer systems, time-series prediction techniques are playing more and more important roles in various fields, such as finance, energy, and traffics. Many models have been proposed for time-series prediction tasks, such as Autoregressive model, Vector Autoregressive model, and Recurrent Neural Networks (RNNs). However, these models still have limitations like failure in modeling non-linearity and long-term dependencies in time-series. Among all the proposed approaches, the Temporal Pattern Attention (TPA), which is an attention-based LSTM model, achieves state-of-the-art performance on several real-world multivariate time-series datasets.
In this thesis, we study three factors that effect the prediction performance of TPA model, which are the Recurrent Neural Network RNN layer, the attention mechanism, and the Convolutional Neural Network for temporal patter detection. For recurrent layer, we implement bi-directional LSTMs that can extract information from the input sequence in both forward and backward directions. In addition, we design two attention mechanisms, each of which assigns attention weights in different directions. We study the effect of both attention mechanisms on TPA model. Finally, to validate the Convolutional Neural Network (CNN) for temporal pattern detection, we implement a TPA model without CNN. We test all of these factors using several real-world time-series datasets from different fields. The experimental results indicate the validity of these factors.
-
dc.description.tableofcontentsI. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
II. Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.1 Long Short-term Memory . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.2 Typical Attention Mechanism . . . . . . . . . . . . . . . . . . . . . . 5
2.3 Temporal Pattern Attention Model . . . . . . . . . . . . . . . . . . . . 6
III. Study on Temporal Pattern Attention Model . . . . . . . . . . . . . . 9
3.1 Problem Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3.3 Recurrent Neural Network Layer . . . . . . . . . . . . . . . . . . . . . 10
3.4 Vertical v.s. Horizontal Attention Mechanism . . . . . . . . . . . . . . 12
3.5 Temporal Pattern Attention Model without CNN . . . . . . . . . . . . 14
IV. Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
4.1 Experimental Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
4.2 Performance Comparison (Q1) . . . . . . . . . . . . . . . . . . . . . . 18
4.3 Effects of Bi-directional LSTM (Q2) . . . . . . . . . . . . . . . . . . . 20
4.4 Effects of CNN for Temporal Pattern Detection (Q3) . . . . . . . . . . 22
4.5 Which Attention Direction Is Better (Q4) . . . . . . . . . . . . . . . . 23
V. RelatedWorks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
VI. Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
-
dc.language.isoeng-
dc.publisher서울대학교 대학원-
dc.subjectTime-series-
dc.subjectAttention mechanism-
dc.subjectLSTM-
dc.subjectPrediction-
dc.subject.ddc621.39-
dc.titleStudy on Attention-based LSTM Model for Multivariate Time-series Prediction-
dc.title.alternative다 변수 시계열 예측을 위한 주의 기반 LSTM 모델 연구-
dc.typeThesis-
dc.typeDissertation-
dc.contributor.AlternativeAuthor바이양-
dc.contributor.department공과대학 컴퓨터공학부-
dc.description.degreeMaster-
dc.date.awarded2019-08-
dc.identifier.uciI804:11032-000000157939-
dc.identifier.holdings000000000040▲000000000041▲000000157939▲-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share