Skip to yearly menu bar Skip to main content


Poster

Learning step sizes for unfolded sparse coding

Pierre Ablin · Thomas Moreau · Mathurin Massias · Alexandre Gramfort

East Exhibition Hall B + C #39

Keywords: [ Algorithms -> Sparsity and Compressed Sensing; Deep Learning; Optimization ] [ Convex Optimization ] [ Algorithms ] [ Sparse Coding and Dimensionality Expansion ]


Abstract:

Sparse coding is typically solved by iterative optimization techniques, such as the Iterative Shrinkage-Thresholding Algorithm (ISTA). Unfolding and learning weights of ISTA using neural networks is a practical way to accelerate estimation. In this paper, we study the selection of adapted step sizes for ISTA. We show that a simple step size strategy can improve the convergence rate of ISTA by leveraging the sparsity of the iterates. However, it is impractical in most large-scale applications. Therefore, we propose a network architecture where only the step sizes of ISTA are learned. We demonstrate that for a large class of unfolded algorithms, if the algorithm converges to the solution of the Lasso, its last layers correspond to ISTA with learned step sizes. Experiments show that our method is competitive with state-of-the-art networks when the solutions are sparse enough.

Live content is unavailable. Log in and register to view live content