Stable Tensor Neural Networks for Rapid Deep Learning

11/15/2018
by   Elizabeth Newman, et al.
0

We propose a tensor neural network (t-NN) framework that offers an exciting new paradigm for designing neural networks with multidimensional (tensor) data. Our network architecture is based on the t-product (Kilmer and Martin, 2011), an algebraic formulation to multiply tensors via circulant convolution. In this t-product algebra, we interpret tensors as t-linear operators analogous to matrices as linear operators, and hence our framework inherits mimetic matrix properties. To exemplify the elegant, matrix-mimetic algebraic structure of our t-NNs, we expand on recent work (Haber and Ruthotto, 2017) which interprets deep neural networks as discretizations of non-linear differential equations and introduces stable neural networks which promote superior generalization. Motivated by this dynamic framework, we introduce a stable t-NN which facilitates more rapid learning because of its reduced, more powerful parameterization. Through our high-dimensional design, we create a more compact parameter space and extract multidimensional correlations otherwise latent in traditional algorithms. We further generalize our t-NN framework to a family of tensor-tensor products (Kernfeld, Kilmer, and Aeron, 2015) which still induce a matrix-mimetic algebraic structure. Through numerical experiments on the MNIST and CIFAR-10 datasets, we demonstrate the more powerful parameterizations and improved generalizability of stable t-NNs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/24/2020

On some tensor tubal-Krylov subspace methods via the T-product

In the present paper, we introduce new tensor Krylov subspace methods fo...
research
06/17/2018

Tensor-Tensor Product Toolbox

Tensors are higher-order extensions of matrices. In recent work [Kilmer ...
research
12/21/2021

A μ-mode BLAS approach for multidimensional tensor-structured problems

In this manuscript, we present a common tensor framework which can be us...
research
09/16/2022

Deep tensor networks with matrix product operators

We introduce deep tensor networks, which are exponentially wide neural n...
research
10/22/2020

Stability of Algebraic Neural Networks to Small Perturbations

Algebraic neural networks (AlgNNs) are composed of a cascade of layers e...
research
08/23/2017

Classification via Tensor Decompositions of Echo State Networks

This work introduces a tensor-based method to perform supervised classif...
research
03/30/2011

Internal Constraints of the Trifocal Tensor

The fundamental matrix and trifocal tensor are convenient algebraic repr...

Please sign up or login with your details

Forgot password? Click here to reset