Anticipate, Ensemble and Prune: Improving Convolutional Neural Networks via Aggregated Early Exits

01/28/2023
by   Simone Sarti, et al.
0

Today, artificial neural networks are the state of the art for solving a variety of complex tasks, especially in image classification. Such architectures consist of a sequence of stacked layers with the aim of extracting useful information and having it processed by a classifier to make accurate predictions. However, intermediate information within such models is often left unused. In other cases, such as in edge computing contexts, these architectures are divided into multiple partitions that are made functional by including early exits, i.e. intermediate classifiers, with the goal of reducing the computational and temporal load without extremely compromising the accuracy of the classifications. In this paper, we present Anticipate, Ensemble and Prune (AEP), a new training technique based on weighted ensembles of early exits, which aims at exploiting the information in the structure of networks to maximise their performance. Through a comprehensive set of experiments, we show how the use of this approach can yield average accuracy improvements of up to 15 internal pruning operation also allows reducing the number of parameters by up to 41 latency time to make inference by 16 learn weights that allow early exits to achieve better accuracy values than those obtained from single-output reference models.

READ FULL TEXT
research
06/09/2021

Zero Time Waste: Recycling Predictions in Early Exit Neural Networks

The problem of reducing processing time of large deep learning models is...
research
06/07/2020

BERT Loses Patience: Fast and Robust Inference with Early Exit

In this paper, we propose Patience-based Early Exit, a straightforward y...
research
11/21/2019

MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks

As the development of neural networks, more and more deep neural network...
research
03/03/2020

Distilled Hierarchical Neural Ensembles with Adaptive Inference Cost

Deep neural networks form the basis of state-of-the-art models across a ...
research
07/14/2022

T-RECX: Tiny-Resource Efficient Convolutional Neural Networks with Early-Exit

Deploying Machine learning (ML) on the milliwatt-scale edge devices (tin...
research
06/05/2023

Towards Anytime Classification in Early-Exit Architectures by Enforcing Conditional Monotonicity

Modern predictive models are often deployed to environments in which com...
research
05/28/2021

Early Exiting with Ensemble Internal Classifiers

As a simple technique to accelerate inference of large-scale pre-trained...

Please sign up or login with your details

Forgot password? Click here to reset