Analyzing the Performance of Smart Industry 4.0 Applications on Cloud Computing Systems

12/11/2020
by   Razin Farhan Hussain, et al.
0

Cloud-based Deep Neural Network (DNN) applications that make latency-sensitive inference are becoming an indispensable part of Industry 4.0. Due to the multi-tenancy and resource heterogeneity, both inherent to the cloud computing environments, the inference time of DNN-based applications are stochastic. Such stochasticity, if not captured, can potentially lead to low Quality of Service (QoS) or even a disaster in critical sectors, such as Oil and Gas industry. To make Industry 4.0 robust, solution architects and researchers need to understand the behavior of DNN-based applications and capture the stochasticity exists in their inference times. Accordingly, in this study, we provide a descriptive analysis of the inference time from two perspectives. First, we perform an application-centric analysis and statistically model the execution time of four categorically different DNN applications on both Amazon and Chameleon clouds. Second, we take a resource-centric approach and analyze a rate-based metric in form of Million Instruction Per Second (MIPS) for heterogeneous machines in the cloud. This non-parametric modeling, achieved via Jackknife and Bootstrap re-sampling methods, provides the confidence interval of MIPS for heterogeneous cloud machines. The findings of this research can be helpful for researchers and cloud solution architects to develop solutions that are robust against the stochastic nature of the inference time of DNN applications in the cloud and can offer a higher QoS to their users and avoid unintended outcomes.

READ FULL TEXT

page 1

page 5

research
05/13/2021

Privacy Inference Attacks and Defenses in Cloud-based Deep Neural Network: A Survey

Deep Neural Network (DNN), one of the most powerful machine learning alg...
research
05/01/2020

Inference Time Optimization Using BranchyNet Partitioning

Deep Neural Network (DNN) applications with edge computing presents a tr...
research
01/01/2023

Federated Fog Computing for Remote Industry 4.0 Applications

Industry 4.0 operates based on IoT devices, sensors, and actuators, tran...
research
07/23/2022

RIBBON: Cost-Effective and QoS-Aware Deep Learning Model Inference using a Diverse Pool of Cloud Computing Instances

Deep learning model inference is a key service in many businesses and sc...
research
05/21/2022

Learning to Dynamically Select Cost Optimal Schedulers in Cloud Computing Environments

The operational cost of a cloud computing platform is one of the most si...
research
01/21/2019

No DNN Left Behind: Improving Inference in the Cloud with Multi-Tenancy

With the rise of machine learning, inference on deep neural networks (DN...
research
05/11/2019

Improving Robustness of Heterogeneous Serverless Computing Systems Via Probabilistic Task Pruning

Cloud-based serverless computing is an increasingly popular computing pa...

Please sign up or login with your details

Forgot password? Click here to reset