Class-specific early exit design methodology for convolutional neural networks.
File(s)Paper___Early_Exit___Applied_Soft_Computing.pdf (786.31 KB)
Accepted version
Author(s)
Bonato, Vanderlei
Bouganis, Christos-Savvas
Type
Journal Article
Abstract
Convolutional Neural Network-based (CNN) inference is a demanding computational task where a long
sequence of operations is applied to an input as dictated by the network topology. Optimisations
by data quantisation, data reuse, network pruning, and dedicated hardware architectures have a
strong impact on reducing both energy consumption and hardware resource requirements, and on
improving inference latency. Implementing new applications from established models available from
both academic and industrial worlds is common nowadays. Further optimisations by preserving model
architecture have been proposed via early exiting approaches, where additional exit points are included
in order to evaluate classifications of samples that produce feature maps with sufficient evidence to
be classified before reaching the final model exit. This paper proposes a methodology for designing
early-exit networks from a given baseline model aiming to improve the average latency for a targeted
subset class constrained by the original accuracy for all classes. Results demonstrate average time
saving in the order of 2.09× to 8.79× for dataset CIFAR10 and 15.00× to 20.71× for CIFAR100 for
baseline models ResNet-21, ResNet-110, Inceptionv3-159, and DenseNet-121.
sequence of operations is applied to an input as dictated by the network topology. Optimisations
by data quantisation, data reuse, network pruning, and dedicated hardware architectures have a
strong impact on reducing both energy consumption and hardware resource requirements, and on
improving inference latency. Implementing new applications from established models available from
both academic and industrial worlds is common nowadays. Further optimisations by preserving model
architecture have been proposed via early exiting approaches, where additional exit points are included
in order to evaluate classifications of samples that produce feature maps with sufficient evidence to
be classified before reaching the final model exit. This paper proposes a methodology for designing
early-exit networks from a given baseline model aiming to improve the average latency for a targeted
subset class constrained by the original accuracy for all classes. Results demonstrate average time
saving in the order of 2.09× to 8.79× for dataset CIFAR10 and 15.00× to 20.71× for CIFAR100 for
baseline models ResNet-21, ResNet-110, Inceptionv3-159, and DenseNet-121.
Date Issued
2021-08
Date Acceptance
2021-03-13
Citation
Applied Soft Computing, 2021, 107, pp.1-12
ISSN
1568-4946
Publisher
Elsevier
Start Page
1
End Page
12
Journal / Book Title
Applied Soft Computing
Volume
107
Copyright Statement
© 2021 Elsevier Ltd. All rights reserved. This manuscript is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Licence http://creativecommons.org/licenses/by-nc-nd/4.0/
Identifier
https://www.sciencedirect.com/science/article/pii/S1568494621002398
Subjects
Artificial Intelligence & Image Processing
0102 Applied Mathematics
0801 Artificial Intelligence and Image Processing
0806 Information Systems
Publication Status
Published
Date Publish Online
2021-04-01