Improved Global Guarantees for the Nonconvex Burer–Monteiro Factorization via Rank Overparameterization

07/05/2022
by   Richard Y. Zhang, et al.
0

We consider minimizing a twice-differentiable, L-smooth, and μ-strongly convex objective ϕ over an n× n positive semidefinite matrix M≽0, under the assumption that the minimizer M^⋆ has low rank r^⋆≪ n. Following the Burer–Monteiro approach, we instead minimize the nonconvex objective f(X)=ϕ(XX^T) over a factor matrix X of size n× r. This substantially reduces the number of variables from O(n^2) to as few as O(n) and also enforces positive semidefiniteness for free, but at the cost of giving up the convexity of the original problem. In this paper, we prove that if the search rank r≥ r^⋆ is overparameterized by a constant factor with respect to the true rank r^⋆, namely as in r>1/4(L/μ-1)^2r^⋆, then despite nonconvexity, local optimization is guaranteed to globally converge from any initial point to the global optimum. This significantly improves upon a previous rank overparameterization threshold of r≥ n, which is known to be sharp if ϕ is allowed to be nonsmooth and/or non-strongly convex, but would increase the number of variables back up to O(n^2). Conversely, without rank overparameterization, we prove that such a global guarantee is possible if and only if ϕ is almost perfectly conditioned, with a condition number of L/μ<3. Therefore, we conclude that a small amount of overparameterization can lead to large improvements in theoretical guarantees for the nonconvex Burer–Monteiro factorization.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset