OptNet: Differentiable Optimization as a Layer in Neural Networks

03/01/2017
by   Brandon Amos, et al.
0

This paper presents OptNet, a network architecture that integrates optimization problems (here, specifically in the form of quadratic programs) as individual layers in larger end-to-end trainable deep networks. These layers encode constraints and complex dependencies between the hidden states that traditional convolutional and fully-connected layers often cannot capture. In this paper, we explore the foundations for such an architecture: we show how techniques from sensitivity analysis, bilevel optimization, and implicit differentiation can be used to exactly differentiate through these layers and with respect to layer parameters; we develop a highly efficient solver for these layers that exploits fast GPU-based batch solves within a primal-dual interior point method, and which provides backpropagation gradients with virtually no additional cost on top of the solve; and we highlight the application of these approaches in several problems. In one notable example, we show that the method is capable of learning to play mini-Sudoku (4x4) given just input and output games, with no a priori information about the rules of the game; this highlights the ability of our architecture to learn hard constraints better than other neural architectures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/14/2021

Efficient differentiable quadratic programming layers: an ADMM approach

Recent advances in neural-network architecture allow for seamless integr...
research
03/03/2020

Implicitly Defined Layers in Neural Networks

In conventional formulations of multilayer feedforward neural networks, ...
research
08/21/2023

Differentiable Frank-Wolfe Optimization Layer

Differentiable optimization has received a significant amount of attenti...
research
03/21/2021

ProgressiveSpinalNet architecture for FC layers

In deeplearning models the FC (fully connected) layer has biggest import...
research
06/10/2022

Flexible Differentiable Optimization via Model Transformations

We introduce DiffOpt.jl, a Julia library to differentiate through the so...
research
05/07/2018

What game are we playing? End-to-end learning in normal and extensive form games

Although recent work in AI has made great progress in solving large, zer...
research
04/30/2020

Physarum Powered Differentiable Linear Programming Layers and Applications

Consider a learning algorithm, which involves an internal call to an opt...

Please sign up or login with your details

Forgot password? Click here to reset