Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.authorAhn, Hyunho
dc.contributor.authorLee, Munkyu
dc.contributor.authorSeong, Sihoon
dc.contributor.authorNa, Gap-Joo
dc.contributor.authorChun, In-Geol
dc.contributor.authorVarghese, Blesson
dc.contributor.authorHong, Cheol-Ho
dc.date.accessioned2024-07-10T10:30:13Z
dc.date.available2024-07-10T10:30:13Z
dc.date.issued2024-06-24
dc.identifier302403613
dc.identifier3ff2ab96-02af-4274-add5-9e6c4d8d7acd
dc.identifier.citationAhn , H , Lee , M , Seong , S , Na , G-J , Chun , I-G , Varghese , B & Hong , C-H 2024 , ' ScissionLite: accelerating distributed deep learning with lightweight data compression for IIoT ' , IEEE Transactions on Industrial Informatics , vol. Early Access . https://doi.org/10.1109/TII.2024.3413340en
dc.identifier.issn1551-3203
dc.identifier.urihttps://hdl.handle.net/10023/30140
dc.descriptionFunding: This work was supported in part by the Electronics and Telecommunications Research Institute through the Korean government under Grant 23zs1300 (Research on High Performance Computing Technology to overcome limitations of AI processing) and in part by the Korea Institute for Advancement of Technology (KIAT) through the Korea Government (MOTIE) under Grant P0017011 (HRD Program for Industrial Innovation). Paper no. TII-23-4829.en
dc.description.abstractIndustrial Internet of Things (IIoT) applications can greatly benefit from leveraging edge computing. For instance, applications relying on deep neural network (DNN) models can be sliced and distributed across IIoT devices and the network edge to reduce inference latency. However, low network performance between IIoT devices and the edge often becomes a bottleneck. In this study, we propose ScissionLite, a holistic framework designed to accelerate distributed DNN inference using lightweight data compression. Our compression method features a novel lightweight down/upsampling network tailored for performance-limited IIoT devices, which is inserted at the slicing point of a DNN model to reduce outbound network traffic without causing a significant drop in accuracy. In addition, we have developed a benchmarking tool to accurately identify the optimal slicing point of the DNN for the best inference latency. ScissionLite improves inference latency by up to 15.7× with minimal accuracy degradation.
dc.format.extent11
dc.format.extent1364664
dc.language.isoeng
dc.relation.ispartofIEEE Transactions on Industrial Informaticsen
dc.subjectEdge computingen
dc.subjectIIoTen
dc.subjectDeep neural networksen
dc.subjectModel slicingen
dc.subjectInferenceen
dc.subjectQA75 Electronic computers. Computer scienceen
dc.subject3rd-NDASen
dc.subject.lccQA75en
dc.titleScissionLite: accelerating distributed deep learning with lightweight data compression for IIoTen
dc.typeJournal articleen
dc.contributor.institutionUniversity of St Andrews. School of Computer Scienceen
dc.identifier.doi10.1109/TII.2024.3413340
dc.description.statusPeer revieweden
dc.date.embargoedUntil2024-06-24


This item appears in the following Collection(s)

Show simple item record