A Quasi-Newton Approach to Nonsmooth Convex Optimization Problems in Machine Learning

04/24/2008
by   Jin Yu, et al.
0

We extend the well-known BFGS quasi-Newton method and its memory-limited variant LBFGS to the optimization of nonsmooth convex objectives. This is done in a rigorous fashion by generalizing three components of BFGS to subdifferentials: the local quadratic model, the identification of a descent direction, and the Wolfe line search conditions. We prove that under some technical conditions, the resulting subBFGS algorithm is globally convergent in objective function value. We apply its memory-limited variant (subLBFGS) to L_2-regularized risk minimization with the binary hinge loss. To extend our algorithm to the multiclass and multilabel settings, we develop a new, efficient, exact line search algorithm. We prove its worst-case time complexity bounds, and show that our line search can also be used to extend a recently developed bundle method to the multiclass and multilabel settings. We also apply the direction-finding component of our algorithm to L_1-regularized risk minimization with logistic loss. In all these contexts our methods perform comparable to or better than specialized state-of-the-art solvers on a number of publicly available datasets. An open source implementation of our algorithms is freely available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/11/2019

Regularization of Limited Memory Quasi-Newton Methods for Large-Scale Nonconvex Minimization

This paper deals with the unconstrained optimization of smooth objective...
research
07/31/2023

Line Search for Convex Minimization

Golden-section search and bisection search are the two main principled a...
research
04/26/2017

Stochastic Orthant-Wise Limited-Memory Quasi-Newton Methods

The ℓ_1-regularized sparse model has been popular in machine learning so...
research
02/25/2016

Fast Nonsmooth Regularized Risk Minimization with Continuation

In regularized risk minimization, the associated optimization problem be...
research
02/15/2018

A Progressive Batching L-BFGS Method for Machine Learning

The standard L-BFGS method relies on gradient approximations that are no...
research
09/07/2009

Lower Bounds for BMRM and Faster Rates for Training SVMs

Regularized risk minimization with the binary hinge loss and its variant...
research
07/29/2022

Compact representations of structured BFGS matrices

For general large-scale optimization problems compact representations ex...

Please sign up or login with your details

Forgot password? Click here to reset