Stochastic Particle Gradient Descent for Infinite Ensembles

12/14/2017
by   Atsushi Nitanda, et al.
0

The superior performance of ensemble methods with infinite models are well known. Most of these methods are based on optimization problems in infinite-dimensional spaces with some regularization, for instance, boosting methods and convex neural networks use L^1-regularization with the non-negative constraint. However, due to the difficulty of handling L^1-regularization, these problems require early stopping or a rough approximation to solve it inexactly. In this paper, we propose a new ensemble learning method that performs in a space of probability measures, that is, our method can handle the L^1-constraint and the non-negative constraint in a rigorous way. Such an optimization is realized by proposing a general purpose stochastic optimization method for learning probability measures via parameterization using transport maps on base models. As a result of running the method, a transport map to output an infinite ensemble is obtained, which forms a residual-type network. From the perspective of functional gradient methods, we give a convergence rate as fast as that of a stochastic optimization method for finite dimensional nonconvex problems. Moreover, we show an interior optimality property of a local optimality condition used in our analysis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/29/2022

Asymptotic Consistency for Nonconvex Risk-Averse Stochastic Optimization with Infinite Dimensional Decision Spaces

Optimal values and solutions of empirical approximations of stochastic o...
research
05/21/2018

Stochastic Gradient Descent for Stochastic Doubly-Nonconvex Composite Optimization

The stochastic gradient descent has been widely used for solving composi...
research
04/13/2016

Algorithms for stochastic optimization with expectation constraints

This paper considers the problem of minimizing an expectation function o...
research
05/09/2023

Accelerated gradient descent method for functionals of probability measures by new convexity and smoothness based on transport maps

We consider problems of minimizing functionals ℱ of probability measures...
research
08/13/2020

Variance Regularization for Accelerating Stochastic Optimization

While nowadays most gradient-based optimization methods focus on explori...
research
05/24/2011

Ergodic Mirror Descent

We generalize stochastic subgradient descent methods to situations in wh...
research
01/07/2018

Gradient Layer: Enhancing the Convergence of Adversarial Training for Generative Models

We propose a new technique that boosts the convergence of training gener...

Please sign up or login with your details

Forgot password? Click here to reset