Publications

Detailed Information

DeepPM: Transformer-based Power and Performance Prediction for Energy-Aware Software

Cited 1 time in Web of Science Cited 2 time in Scopus
Authors

Shim, Jun S.; Han, Bogyeong; Kim, Yeseong; Kim, Jihong

Issue Date
2022-03
Publisher
IEEE
Citation
Proceedings -Design, Automation and Test in Europe, DATE, pp.1491-1496
Abstract
Many system-level management and optimization techniques need accurate estimates of power consumption and performance. Earlier research has proposed many high-level/source-level estimation modeling works, particularly for basic blocks. However, most of them still need to execute the target software at least once on a fine-grained simulator or real hardware to extract required features. This paper proposes a performance/power prediction framework, called Deep Power Meter (DeepPM), which estimates them accurately only using the compiled binary. Inspired by the deep learning techniques in natural language processing, we convert the program instructions in the form of vectors and predict the average power and performance of basic blocks based on a transformer model. In addition, unlike existing works based on a Long Short-Term Memory (LSTM) model structure, which only works for basic blocks with a small number of instructions, DeepPM provides highly accurate results for long basic blocks, which takes the majority of the execution time for actual application runs. In our evaluation conducted with SPEC2006 benchmark suite, we show that DeepPM can provide accurate prediction for performance and power consumption with 10.2% and 12.3% error, respectively. DeepPM also outperforms the LSTM-based model by up to 67.2% and 34.9% error for performance and power, respectively.
ISSN
1530-1591
URI
https://hdl.handle.net/10371/186623
DOI
https://doi.org/10.23919/DATE54114.2022.9774589
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share