Publications

Detailed Information

Stochastic Optimal Energy Management based on Q-Learning for Hybrid Electric Vehicles : Q-Learning 에 기반한 하이브리드 차량의 확률론적 최적 에너지 관리

DC Field Value Language
dc.contributor.advisor차석원-
dc.contributor.author이희윤-
dc.date.accessioned2018-11-12T01:00:13Z-
dc.date.available2018-11-12T01:00:13Z-
dc.date.issued2018-08-
dc.identifier.other000000153142-
dc.identifier.urihttps://hdl.handle.net/10371/143273-
dc.description학위논문 (박사)-- 서울대학교 대학원 : 공과대학 기계항공공학부, 2018. 8. 차석원.-
dc.description.abstractHybrid Electric Vehicles(HEVs) have been widely studied recently with growing concern over sustainability development of the global environment. HEVs use multiple power sources of internal combustion engine and electric battery generally, thus Energy Management Strategy (EMS) determining the split between two power sources needs to be coordinated to maximize the entire vehicle system efficiency. In this study, an EMS based on Stochastic Dynamic Programming(SDP) and reinforcement learning is developed.

SDP is an approaches based on probability, in which optimization problem is defined infinite time horizon, therefore obtained control policy can be used as real-time controller of the vehicle. In order to apply SDP to vehicle control, the characteristics of the driving cycle are expressed as a Transition Probability Matrix (TPM) through the Markov process, in which the vehicle speed and the power demand are discretized. The objective function of the optimization problem in this study is defined as minimizing the expected value of vehicle fuel consumption, deviation of battery state of charge, and frequent engine on / off. As result power split ratio is given as function of vehicle speed, battery SOC, power demand and engine on/off status.

However, SDP has limitation that it is offline policy considering it required TPM, thus in this study, reinforcement learning technique is used to compensate this problem. In the newly proposed control strategy, based on the Q-learning algorithm, the probability information of the driving cycle is updated to the Q value, that is expected cost value of each state variable and control input, and the control rule is updated by calculating the cost function for the all admissible control input at each time step using vehicle model.

To verify control strategy, backward-looking is developed for parallel type HEVs. Simulation results show that fuel economy is improved compared to rule-based strategy and near optimal solution is obtained.
-
dc.description.tableofcontentsCHAPTER 1 INTRODUCTION 1

1.1 Motivation 1

1.2 Background Studies 5

1.3 Contributions 12

1.4 Thesis Outlines 13

CHAPTER 2 VEHICLE MODEL DEVELOPMENT 14

2.1 Target Vehicle Hyundai Sonata Hybrid 14

2.2 Vehicle Modeling 17

CHAPTER 3 STOCHASTIC DYNAMIC PROGAMMING BASED ENERGY MANAGEMENT STRATEGY 24

3.1 Introduction 24

3.2 Deterministic Dynamic Programming 28

3.3 Stochastic Modeling of Driving Cycle Information 33

3.4 Stochastic Dynamic Programming 37

CHAPTER 4 REINFORCEMENT LEARNING BASED ENERGY MANAGEMENT STRATEGY 52

4.1 Introduction 53

4.2 Q-Learning based Energy Management Strategy 56

CHAPTER 5 SIMULATION ANALYSIS 67

5.1 Vehicle Simulation based on Stochastic Dynamic Programming based Energy Management Strategy 67

5.2 Vehicle Simulation using Reinforcement Learning based Energy Management Strategy 75

CHAPTER 6 CONCLUDING REMARKS 92

6.1 Conclusion 92

6.2 Future Work 95

REFERENCE 97

국 문 초 록 108
-
dc.language.isoen-
dc.publisher서울대학교 대학원-
dc.subject.ddc621-
dc.titleStochastic Optimal Energy Management based on Q-Learning for Hybrid Electric Vehicles-
dc.title.alternativeQ-Learning 에 기반한 하이브리드 차량의 확률론적 최적 에너지 관리-
dc.typeThesis-
dc.contributor.AlternativeAuthorHeeyun Lee-
dc.description.degreeDoctor-
dc.contributor.affiliation공과대학 기계항공공학부-
dc.date.awarded2018-08-
Appears in Collections:
Files in This Item:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share