Publications

Detailed Information

ADC-PIM: Accelerating Convolution on the GPU via In-Memory Approximate Data Comparison

DC Field Value Language
dc.contributor.authorChoi, Jungwoo-
dc.contributor.authorLee, Hyuk-Jae-
dc.contributor.authorRhee, Chae Eun-
dc.date.accessioned2022-10-05T04:10:04Z-
dc.date.available2022-10-05T04:10:04Z-
dc.date.created2022-07-11-
dc.date.issued2022-06-
dc.identifier.citationIEEE Journal on Emerging and Selected Topics in Circuits and Systems, Vol.12 No.2, pp.458-471-
dc.identifier.issn2156-3357-
dc.identifier.urihttps://hdl.handle.net/10371/185300-
dc.description.abstractRecently, convolutional neural networks (CNN) have been widely used in image processing and computer vision. GPUs are often used to accelerate the CNN, but performance is limited by high computational costs and memory usage of the convolution. Prior studies exploited approximate computing to reduce the computational costs. However, they only reduced the amount of the computation, thereby its performance is bottlenecked by the memory bandwidth due to an increased memory intensity. In addition, load imbalance between warps caused by approximation also inhibits the performance improvement. In this paper, we propose a processing-in-memory (PIM) solution that reduces the amount of data movement and computation through the Approximate Data Comparison (ADC-PIM). Instead of determining the value similarity after loading the data to the GPU, the ADC-PIM unit located on 3D-stacked memory compares the similarity and transfers only the selected representative data to the GPU. The GPU performs convolution on the representative data transferred from the ADC-PIM, and reuses the calculated results based on the similarity information. To reduce the increase in memory latency caused by the in-memory data comparison, we propose a two-level PIM architecture that exploits both the DRAM bank and TSV stage. By dividing the comparisons into multiple banks and then merging the results on the TSV stage, the ADC-PIM effectively hides the delay caused by the comparisons. To ease the load balancing on the GPU, the ADC-PIM performs data reorganization by assigning the representative data to addresses that are computed based on the comparison result. Experimental results show that the proposed ADC-PIM provides a 43% speedup and 32% energy saving with less than a 1% accuracy drop.-
dc.language영어-
dc.publisherIEEE Circuits and Systems Society-
dc.titleADC-PIM: Accelerating Convolution on the GPU via In-Memory Approximate Data Comparison-
dc.typeArticle-
dc.identifier.doi10.1109/JETCAS.2022.3167391-
dc.citation.journaltitleIEEE Journal on Emerging and Selected Topics in Circuits and Systems-
dc.identifier.wosid000811585100014-
dc.identifier.scopusid2-s2.0-85128668009-
dc.citation.endpage471-
dc.citation.number2-
dc.citation.startpage458-
dc.citation.volume12-
dc.description.isOpenAccessN-
dc.contributor.affiliatedAuthorLee, Hyuk-Jae-
dc.type.docTypeArticle-
dc.description.journalClass1-
Appears in Collections:
Files in This Item:
There are no files associated with this item.

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share