Mid Sweden University

miun.sePublications
Planned maintenance
A system upgrade is planned for 10/12-2024, at 12:00-13:00. During this time DiVA will be unavailable.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Neural Network Compression Through Shunt Connections and Knowledge Distillation for Semantic Segmentation Problems
2021 (English)In: IFIP Advances in Information and Communication Technology, Springer Science and Business Media Deutschland GmbH , 2021, p. 349-361Conference paper, Published paper (Refereed)
Abstract [en]

Employing convolutional neural network models for large scale datasets represents a big challenge. Especially embedded devices with limited resources cannot run most state-of-the-art model architectures in real-time, necessary for many applications. This paper proves the applicability of shunt connections on large scale datasets and narrows this computational gap. Shunt connections is a proposed method for MobileNet compression. We are the first to provide results of shunt connections for the MobileNetV3 model and for segmentation tasks on the Cityscapes dataset, using the DeeplabV3 architecture, on which we achieve compression by 28%, while observing a 3.52 drop in mIoU. The training of shunt-inserted models are optimized through knowledge distillation. The full code used for this work will be available online. © 2021, IFIP International Federation for Information Processing.

Place, publisher, year, edition, pages
Springer Science and Business Media Deutschland GmbH , 2021. p. 349-361
Keywords [en]
Accuracy, CIFAR, Cityscapes, DeepLab, Embedded machine learning, Knowledge distillation, Latency, Machine learning, MobileNet, Optimization, Shunt connections, Convolutional neural networks, Distillation, Large dataset, Network architecture, Semantic Web, Semantics, Embedded device, Large-scale datasets, Network compression, Real time, Semantic segmentation, State of the art, Distilleries
Identifiers
URN: urn:nbn:se:miun:diva-43421DOI: 10.1007/978-3-030-79150-6_28Scopus ID: 2-s2.0-85111872655ISBN: 9783030791490 (print)OAI: oai:DiVA.org:miun-43421DiVA, id: diva2:1604114
Conference
25 June 2021 through 27 June 2021
Available from: 2021-10-18 Created: 2021-10-18 Last updated: 2021-10-18Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Search in DiVA

By author/editor
Jantsch, A.

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 10 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf