Differentiable Approximation Bridges For Training Networks Containing Non-Differentiable Functions

05/09/2019
by   Jason Ramapuram, et al.
0

Modern neural network training relies on piece-wise (sub-)differentiable functions in order to use backpropation for efficient calculation of gradients. In this work, we introduce a novel method to allow for non-differentiable functions at intermediary layers of deep neural networks. We do so through the introduction of a differentiable approximation bridge (DAB) neural network which provides smooth approximations to the gradient of the non-differentiable function. We present strong empirical results (performing over 600 experiments) in three different domains: unsupervised (image) representation learning, image classification, and sequence sorting to demonstrate that our proposed method improves state of the art performance. We demonstrate that utilizing non-differentiable functions in unsupervised (image) representation learning improves reconstruction quality and posterior linear separability by 10 also observe an accuracy improvement of 77 25 classification setting with the sort non-linearity. This work enables the usage of functions that were previously not usable in neural networks.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset