Daydream: Accurately Estimating the Efficacy of Optimizations for DNN Training

06/05/2020
by   Hongyu Zhu, et al.
0

Modern deep neural network (DNN) training jobs use complex and heterogeneous software/hardware stacks. The efficacy of software-level optimizations can vary significantly when used in different deployment configurations. It is onerous and error-prone for ML practitioners and system developers to implement each optimization separately, and determine which ones will improve performance in their own configurations. Unfortunately, existing profiling tools do not aim to answer predictive questions such as "How will optimization X affect the performance of my model?". We address this critical limitation, and proposes a new profiling tool, Daydream, to help programmers efficiently explore the efficacy of DNN optimizations. Daydream models DNN execution with a fine-grained dependency graph based on low-level traces collected by CUPTI, and predicts runtime by simulating execution based on the dependency graph. Daydream maps the low-level traces using DNN domain-specific knowledge, and introduces a set of graph-transformation primitives that can easily model a wide variety of optimizations. We show that Daydream is able to model most mainstream DNN optimization techniques, and accurately predict the efficacy of optimizations that will result in significant performance improvements.

READ FULL TEXT
research
10/03/2022

Decompiling x86 Deep Neural Network Executables

Due to their widespread use on heterogeneous hardware devices, deep lear...
research
05/05/2022

dPRO: A Generic Profiling and Optimization System for Expediting Distributed DNN Training

Distributed training using multiple devices (e.g., GPUs) has been widely...
research
06/14/2023

MetaML: Automating Customizable Cross-Stage Design-Flow for Deep Learning Acceleration

This paper introduces a novel optimization framework for deep neural net...
research
06/02/2021

IrEne: Interpretable Energy Prediction for Transformers

Existing software-based energy measurements of NLP models are not accura...
research
03/22/2022

NNReArch: A Tensor Program Scheduling Framework Against Neural Network Architecture Reverse Engineering

Architecture reverse engineering has become an emerging attack against d...
research
02/10/2022

Feasible Low-thrust Trajectory Identification via a Deep Neural Network Classifier

In recent years, deep learning techniques have been introduced into the ...
research
08/23/2023

RamseyRL: A Framework for Intelligent Ramsey Number Counterexample Searching

The Ramsey number is the minimum number of nodes, n = R(s, t), such that...

Please sign up or login with your details

Forgot password? Click here to reset