Conformal Symplectic and Relativistic Optimization

03/11/2019
by   Guilherme França, et al.
0

Although momentum-based optimization methods have had a remarkable impact on machine learning, their heuristic construction has been an obstacle to a deeper understanding. A promising direction to study these accelerated algorithms has been emerging through connections with continuous dynamical systems. Yet, it is unclear whether the main properties of the underlying dynamical system are preserved by the algorithms from which they are derived. Conformal Hamiltonian systems form a special class of dissipative systems, having a distinct symplectic geometry. In this paper, we show that gradient descent with momentum preserves this symplectic structure, while Nesterov's accelerated gradient method does not. More importantly, we propose a generalization of classical momentum based on the special theory of relativity. The resulting conformal symplectic and relativistic algorithm enjoys better stability since it operates on a different space compared to its classical predecessor. Its benefits are discussed and verified in deep learning experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/07/2016

Nesterov's Accelerated Gradient and Momentum as approximations to Regularised Update Descent

We present a unifying framework for adapting the update direction in gra...
research
06/19/2020

How Does Momentum Help Frank Wolfe?

We unveil the connections between Frank Wolfe (FW) type algorithms and t...
research
02/28/2020

Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives

We analyze the convergence rate of various momentum-based optimization a...
research
12/06/2019

Optimization algorithms inspired by the geometry of dissipative systems

Accelerated gradient methods are a powerful optimization tool in machine...
research
07/23/2021

Optimization on manifolds: A symplectic approach

There has been great interest in using tools from dynamical systems and ...
research
02/10/2018

On Symplectic Optimization

Accelerated gradient methods have had significant impact in machine lear...
research
11/27/2017

Accelerated Optimization in the PDE Framework: Formulations for the Active Contour Case

Following the seminal work of Nesterov, accelerated optimization methods...

Please sign up or login with your details

Forgot password? Click here to reset