Publications

Detailed Information

Cooperative Inference of DNNs for Delay-and Memory-Constrained Wireless IoT Systems

Cited 0 time in Web of Science Cited 5 time in Scopus
Authors

Yun, Sangseok; Choi, Wan; Kim, Il-Min

Issue Date
2022-01
Publisher
Institute of Electrical and Electronics Engineers Inc.
Citation
IEEE Internet of Things Journal
Abstract
IEEEThis work studies cooperative inference of deep neural networks (DNNs) in which a memory-constrained end-device performs a delay-constrained inference process with an aid of an edge server. Although several works considered the cooperative inference of DNNs in the literature, it was assumed in those works that the memory footprints at end-devices are unlimited, which is in practice not realistic. To address this issue, in this work, a memory-aware cooperative DNN inference is proposed. Specifically, we propose to adopt knowledge distillation to obtain high-performing lightweight DNNs. To minimize the inference delay, we first analyze the end-to-end delay required for processing the proposed cooperative DNN inference, and then we minimize the delay by jointly optimizing the DNN partitioning point and the intermediate data transmission rate. Also, a dynamic DNN selection scheme is developed by fully exploiting the available memory resource in order to maximize the performance of the inference task in terms of inference accuracy. Experimental results demonstrate that the proposed cooperative DNN inference considerably outperforms the comparable schemes while satisfying both the delay constraint and the memory constraint.
ISSN
2327-4662
URI
https://hdl.handle.net/10371/183909
DOI
https://doi.org/10.1109/JIOT.2022.3152359
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share