Background: Distributed training is essential for large scale training o...
Power consumption is a major obstacle in the deployment of deep neural
n...
Quantization of the weights and activations is one of the main methods t...
Background: Catastrophic forgetting is the notorious vulnerability of ne...
Neural gradient compression remains a main bottleneck in improving train...
Background: Recently, an extensive amount of research has been focused o...
Background: Recent developments have made it possible to accelerate neur...
Large-batch SGD is important for scaling training of deep neural network...
Unlike traditional approaches that focus on the quantization at the netw...
Quantized Neural Networks (QNNs) are often used to improve network effic...
We suggest a novel approach for the estimation of the posterior distribu...
Over the past few years batch-normalization has been commonly used in de...
Deep convolutional network has been the state-of-the-art approach for a ...
Neural networks are commonly used as models for classification for a wid...
We show that gradient descent on an unregularized logistic regression pr...
Background: Deep learning models are typically trained using stochastic
...
Background: Statistical mechanics results (Dauphin et al. (2014); Chorom...
Convolutional networks have marked their place over the last few years a...
Deep learning has proven itself as a successful set of models for learni...