A Recursively Recurrent Neural Network (R2N2) Architecture for Learning Iterative Algorithms

11/22/2022
by   Danimir T. Doncevic, et al.
0

Meta-learning of numerical algorithms for a given task consist of the data-driven identification and adaptation of an algorithmic structure and the associated hyperparameters. To limit the complexity of the meta-learning problem, neural architectures with a certain inductive bias towards favorable algorithmic structures can, and should, be used. We generalize our previously introduced Runge-Kutta neural network to a recursively recurrent neural network (R2N2) superstructure for the design of customized iterative algorithms. In contrast to off-the-shelf deep learning approaches, it features a distinct division into modules for generation of information and for the subsequent assembly of this information towards a solution. Local information in the form of a subspace is generated by subordinate, inner, iterations of recurrent function evaluations starting at the current outer iterate. The update to the next outer iterate is computed as a linear combination of these evaluations, reducing the residual in this space, and constitutes the output of the network. We demonstrate that regular training of the weight parameters inside the proposed superstructure on input/output data of various computational problem classes yields iterations similar to Krylov solvers for linear equation systems, Newton-Krylov solvers for nonlinear equation systems, and Runge-Kutta integrators for ordinary differential equations. Due to its modularity, the superstructure can be readily extended with functionalities needed to represent more general classes of iterative algorithms traditionally based on Taylor series expansions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/04/2021

Personalized Algorithm Generation: A Case Study in Meta-Learning ODE Integrators

We study the meta-learning of numerical algorithms for scientific comput...
research
02/07/2021

Meta-Learning with Neural Tangent Kernels

Model Agnostic Meta-Learning (MAML) has emerged as a standard framework ...
research
09/04/2017

DR-RNN: A deep residual recurrent neural network for model reduction

We introduce a deep residual recurrent neural network (DR-RNN) as an eff...
research
06/18/2023

Meta-Learning for Airflow Simulations with Graph Neural Networks

The field of numerical simulation is of significant importance for the d...
research
01/05/2020

From Learning to Meta-Learning: Reduced Training Overhead and Complexity for Communication Systems

Machine learning methods adapt the parameters of a model, constrained to...
research
06/17/2022

Accelerating numerical methods by gradient-based meta-solving

In science and engineering applications, it is often required to solve s...
research
05/14/2023

A new iterative method for construction of the Kolmogorov-Arnold representation

The Kolmogorov-Arnold representation of a continuous multivariate functi...

Please sign up or login with your details

Forgot password? Click here to reset