PRANC: Pseudo RAndom Networks for Compacting deep models

06/16/2022
by   Parsa Nooralinejad, et al.
0

Communication becomes a bottleneck in various distributed Machine Learning settings. Here, we propose a novel training framework that leads to highly efficient communication of models between agents. In short, we train our network to be a linear combination of many pseudo-randomly generated frozen models. For communication, the source agent transmits only the `seed' scalar used to generate the pseudo-random `basis' networks along with the learned linear mixture coefficients. Our method, denoted as PRANC, learns almost $100\times$ fewer parameters than a deep model and still performs well on several datasets and architectures. PRANC enables 1) efficient communication of models between agents, 2) efficient model storage, and 3) accelerated inference by generating layer-wise weights on the fly. We test PRANC on CIFAR-10, CIFAR-100, tinyImageNet, and ImageNet-100 with various architectures like AlexNet, LeNet, ResNet18, ResNet20, and ResNet56 and demonstrate a massive reduction in the number of parameters while providing satisfactory performance on these benchmark datasets. The code is available \href{https://github.com/UCDvision/PRANC}{https://github.com/UCDvision/PRANC}

READ FULL TEXT
research
09/30/2022

Verifiable and Energy Efficient Medical Image Analysis with Quantised Self-attentive Deep Neural Networks

Convolutional Neural Networks have played a significant role in various ...
research
10/13/2022

Parameter-Efficient Masking Networks

A deeper network structure generally handles more complicated non-linear...
research
01/04/2023

Beckman Defense

Optimal transport (OT) based distributional robust optimisation (DRO) ha...
research
05/26/2022

Membership Inference Attack Using Self Influence Functions

Member inference (MI) attacks aim to determine if a specific data sample...
research
03/31/2022

Scaling Up Models and Data with and

Recent neural network-based language models have benefited greatly from ...
research
01/18/2023

Local Learning with Neuron Groups

Traditional deep network training methods optimize a monolithic objectiv...
research
03/13/2022

AugShuffleNet: Improve ShuffleNetV2 via More Information Communication

Based on ShuffleNetV2, we build a more powerful and efficient model fami...

Please sign up or login with your details

Forgot password? Click here to reset