Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.authorEccles, Bailey J.
dc.contributor.authorRodgers, Philip
dc.contributor.authorKilpatrick, Peter
dc.contributor.authorSpence, Ivor
dc.contributor.authorVarghese, Blesson
dc.date.accessioned2023-10-30T17:30:03Z
dc.date.available2023-10-30T17:30:03Z
dc.date.issued2024-03
dc.identifier293968823
dc.identifier918adda9-15d0-438a-92ce-8022d06a4498
dc.identifier.citationEccles , B J , Rodgers , P , Kilpatrick , P , Spence , I & Varghese , B 2024 , ' DNNShifter : an efficient DNN pruning system for edge computing ' , Future Generation Computer Systems , vol. 152 , pp. 43-54 . https://doi.org/10.48550/arXiv.2309.06973 , https://doi.org/10.1016/j.future.2023.09.025en
dc.identifier.issn0167-739X
dc.identifier.otherORCID: /0000-0002-5533-7503/work/146006739
dc.identifier.urihttps://hdl.handle.net/10023/28595
dc.descriptionFunding: This research is funded by Rakuten Mobile, Japan .en
dc.description.abstractDeep neural networks (DNNs) underpin many machine learning applications. Production quality DNN models achieve high inference accuracy by training millions of DNN parameters which has a significant resource footprint. This presents a challenge for resources operating at the extreme edge of the network, such as mobile and embedded devices that have limited computational and memory resources. To address this, models are pruned to create lightweight, more suitable variants for these devices. Existing pruning methods are unable to provide similar quality models compared to their unpruned counterparts without significant time costs and overheads or are limited to offline use cases. Our work rapidly derives suitable model variants while maintaining the accuracy of the original model. The model variants can be swapped quickly when system and network conditions change to match workload demand. This paper presents DNNShifter  , an end-to-end DNN training, spatial pruning, and model switching system that addresses the challenges mentioned above. At the heart of DNNShifter  is a novel methodology that prunes sparse models using structured pruning - combining the accuracy-preserving benefits of unstructured pruning with runtime performance improvements of structured pruning. The pruned model variants generated by DNNShifter  are smaller in size and thus faster than dense and sparse model predecessors, making them suitable for inference at the edge while retaining near similar accuracy as of the original dense model. DNNShifter  generates a portfolio of model variants that can be swiftly interchanged depending on operational conditions. DNNShifter  produces pruned model variants up to 93x faster than conventional training methods. Compared to sparse models, the pruned model variants are up to 5.14x smaller and have a 1.67x inference latency speedup, with no compromise to sparse model accuracy. In addition, DNNShifter  has up to 11.9x lower overhead for switching models and up to 3.8x lower memory utilisation than existing approaches. DNNShifter  is available for public use from https://github.com/blessonvar/DNNShifter.
dc.format.extent2659299
dc.language.isoeng
dc.relation.ispartofFuture Generation Computer Systemsen
dc.subjectDeep neural networksen
dc.subjectMachine learningen
dc.subjectInternet of thingsen
dc.subjectEdge computingen
dc.subjectModel compressionen
dc.subjectModel pruningen
dc.subjectQA75 Electronic computers. Computer scienceen
dc.subjectNDASen
dc.subject.lccQA75en
dc.titleDNNShifter : an efficient DNN pruning system for edge computingen
dc.typeJournal articleen
dc.contributor.institutionUniversity of St Andrews. School of Computer Scienceen
dc.identifier.doi10.48550/arXiv.2309.06973
dc.description.statusPeer revieweden


This item appears in the following Collection(s)

Show simple item record