Low-rank lottery tickets: finding efficient low-rank neural networks via matrix differential equations

05/26/2022
by   Steffen Schotthöfer, et al.
0

Neural networks have achieved tremendous success in a large variety of applications. However, their memory footprint and computational demand can render them impractical in application settings with limited hardware or energy resources. In this work, we propose a novel algorithm to find efficient low-rank subnetworks. Remarkably, these subnetworks are determined and adapted already during the training phase and the overall time and memory resources required by both training and evaluating them is significantly reduced. The main idea is to restrict the weight matrices to a low-rank manifold and to update the low-rank factors rather than the full matrix during training. To derive training updates that are restricted to the prescribed manifold, we employ techniques from dynamic model order reduction for matrix differential equations. Moreover, our method automatically and dynamically adapts the ranks during training to achieve a desired approximation accuracy. The efficiency of the proposed method is demonstrated through a variety of numerical experiments on fully-connected and convolutional networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/23/2020

A geometry based algorithm for dynamical low-rank approximation

In this paper, we propose a geometry based algorithm for dynamical low-r...
research
09/12/2023

From low-rank retractions to dynamical low-rank approximation and back

In algorithms for solving optimization problems constrained to a smooth ...
research
05/30/2023

Rank-adaptive spectral pruning of convolutional layers during training

The computing cost and memory demand of deep learning pipelines have gro...
research
12/05/2022

A low-rank algorithm for solving Lyapunov operator φ-functions within the matrix-valued exponential integrators

In this work we develop a low-rank algorithm for the computation of low-...
research
01/31/2023

On the Initialisation of Wide Low-Rank Feedforward Neural Networks

The edge-of-chaos dynamics of wide randomly initialized low-rank feedfor...
research
04/12/2023

A parallel rank-adaptive integrator for dynamical low-rank approximation

This work introduces a parallel and rank-adaptive matrix integrator for ...
research
04/07/2015

Efficient SDP Inference for Fully-connected CRFs Based on Low-rank Decomposition

Conditional Random Fields (CRF) have been widely used in a variety of co...

Please sign up or login with your details

Forgot password? Click here to reset