Edge-PRUNE: Flexible Distributed Deep Learning Inference

04/27/2022
by   Jani Boutellier, et al.
0

Collaborative deep learning inference between low-resource endpoint devices and edge servers has received significant research interest in the last few years. Such computation partitioning can help reducing endpoint device energy consumption and improve latency, but equally importantly also contributes to privacy-preserving of sensitive data. This paper describes Edge-PRUNE, a flexible but light-weight computation framework for distributing machine learning inference between edge servers and one or more client devices. Compared to previous approaches, Edge-PRUNE is based on a formal dataflow computing model, and is agnostic towards machine learning training frameworks, offering at the same time wide support for leveraging deep learning accelerators such as embedded GPUs. The experimental section of the paper demonstrates the use and performance of Edge-PRUNE by image classification and object tracking applications on two heterogeneous endpoint devices and an edge server, over wireless and physical connections. Endpoint device inference time for SSD-Mobilenet based object tracking, for example, is accelerated 5.8x by collaborative inference.

READ FULL TEXT

page 1

page 5

research
06/16/2022

Fault-Tolerant Collaborative Inference through the Edge-PRUNE Framework

Collaborative inference has received significant research interest in ma...
research
04/23/2023

The Case for Hierarchical Deep Learning Inference at the Network Edge

Resource-constrained Edge Devices (EDs), e.g., IoT sensors and microcont...
research
10/15/2022

The Effects of Partitioning Strategies on Energy Consumption in Distributed CNN Inference at The Edge

Nowadays, many AI applications utilizing resource-constrained edge devic...
research
05/14/2020

Prive-HD: Privacy-Preserved Hyperdimensional Computing

The privacy of data is a major challenge in machine learning as a traine...
research
04/22/2020

OL4EL: Online Learning for Edge-cloud Collaborative Learning on Heterogeneous Edges with Resource Constraints

Distributed machine learning (ML) at network edge is a promising paradig...
research
03/22/2023

Edge Deep Learning Model Protection via Neuron Authorization

With the development of deep learning processors and accelerators, deep ...
research
02/01/2023

Xenos: Dataflow-Centric Optimization to Accelerate Model Inference on Edge Devices

Edge computing has been emerging as a popular scenario for model inferen...

Please sign up or login with your details

Forgot password? Click here to reset