Tucker Tensor Layer in Fully Connected Neural Networks

03/14/2019
by   Giuseppe G. Calvi, et al.
0

We introduce the Tucker Tensor Layer (TTL), an alternative to the dense weight-matrices of the fully connected layers of feed-forward neural networks (NNs), to answer the long standing quest to compress NNs and improve their interpretability. This is achieved by treating these weight-matrices as the unfolding of a higher order weight-tensor. This enables us to introduce a framework for exploiting the multi-way nature of the weight-tensor in order to efficiently reduce the number of parameters, by virtue of the compression properties of tensor decompositions. The Tucker Decomposition (TKD) is employed to decompose the weight-tensor into a core tensor and factor matrices. We re-derive back-propagation within this framework, by extending the notion of matrix derivatives to tensors. In this way, the physical interpretability of the TKD is exploited to gain insights into training, through the process of computing gradients with respect to each factor matrix. The proposed framework is validated on synthetic data and on the Fashion-MNIST dataset, emphasizing the relative importance of various data features in training, hence mitigating the "black-box" issue inherent to NNs. Experiments on both MNIST and Fashion-MNIST illustrate the compression properties of the TTL, achieving a 66.63 fold compression whilst maintaining comparable performance to the uncompressed NN.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/22/2015

Tensorizing Neural Networks

Deep neural networks currently demonstrate state-of-the-art performance ...
research
07/21/2015

Compression of Fully-Connected Layer in Neural Network by Kronecker Product

In this paper we propose and study a technique to reduce the number of p...
research
08/19/2021

Residual Tensor Train: a Flexible and Efficient Approach for Learning Multiple Multilinear Correlations

Tensor Train (TT) approach has been successfully applied in the modellin...
research
10/28/2017

Trainable back-propagated functional transfer matrices

Connections between nodes of fully connected neural networks are usually...
research
10/07/2022

Latent Matrices for Tensor Network Decomposition and to Tensor Completion

The prevalent fully-connected tensor network (FCTN) has achieved excelle...
research
05/25/2018

Tensorized Spectrum Preserving Compression for Neural Networks

Modern neural networks can have tens of millions of parameters, and are ...
research
12/31/2017

Learning Relevant Features of Data with Multi-scale Tensor Networks

Inspired by coarse-graining approaches used in physics, we show how simi...

Please sign up or login with your details

Forgot password? Click here to reset