Optimization Guarantees of Unfolded ISTA and ADMM Networks With Smooth Soft-Thresholding

09/12/2023
by   Shaik Basheeruddin Shah, et al.
0

Solving linear inverse problems plays a crucial role in numerous applications. Algorithm unfolding based, model-aware data-driven approaches have gained significant attention for effectively addressing these problems. Learned iterative soft-thresholding algorithm (LISTA) and alternating direction method of multipliers compressive sensing network (ADMM-CSNet) are two widely used such approaches, based on ISTA and ADMM algorithms, respectively. In this work, we study optimization guarantees, i.e., achieving near-zero training loss with the increase in the number of learning epochs, for finite-layer unfolded networks such as LISTA and ADMM-CSNet with smooth soft-thresholding in an over-parameterized (OP) regime. We achieve this by leveraging a modified version of the Polyak-Lojasiewicz, denoted PL^*, condition. Satisfying the PL^* condition within a specific region of the loss landscape ensures the existence of a global minimum and exponential convergence from initialization using gradient descent based methods. Hence, we provide conditions, in terms of the network width and the number of training samples, on these unfolded networks for the PL^* condition to hold. We achieve this by deriving the Hessian spectral norm of these networks. Additionally, we show that the threshold on the number of training samples increases with the increase in the network width. Furthermore, we compare the threshold on training samples of unfolded networks with that of a standard fully-connected feed-forward network (FFNN) with smooth soft-thresholding non-linearity. We prove that unfolded networks have a higher threshold value than FFNN. Consequently, one can expect a better expected error for unfolded networks than FFNN.

READ FULL TEXT
research
04/24/2018

Between hard and soft thresholding: optimal iterative thresholding algorithms

Iterative thresholding algorithms seek to optimize a differentiable obje...
research
03/16/2016

Norm-1 Regularized Consensus-based ADMM for Imaging with a Compressive Antenna

This paper presents a novel norm-one-regularized, consensus-based imagin...
research
07/05/2021

Provable Convergence of Nesterov Accelerated Method for Over-Parameterized Neural Networks

Despite the empirical success of deep learning, it still lacks theoretic...
research
07/22/2011

A Unifying Analysis of Projected Gradient Descent for ℓ_p-constrained Least Squares

In this paper we study the performance of the Projected Gradient Descent...
research
04/14/2023

Convex Dual Theory Analysis of Two-Layer Convolutional Neural Networks with Soft-Thresholding

Soft-thresholding has been widely used in neural networks. Its basic net...
research
10/29/2020

Generalization bounds for deep thresholding networks

We consider compressive sensing in the scenario where the sparsity basis...
research
09/06/2020

An Analysis of Alternating Direction Method of Multipliers for Feed-forward Neural Networks

In this work, we present a hardware compatible neural network training a...

Please sign up or login with your details

Forgot password? Click here to reset