arXiv Analytics

Sign in

arXiv:math/0307152 [math.FA]AbstractReferencesReviewsResources

An iterative thresholding algorithm for linear inverse problems with a sparsity constraint

Ingrid Daubechies, Michel Defrise, Christine De Mol

Published 2003-07-10, updated 2003-11-02Version 2

We consider linear inverse problems where the solution is assumed to have a sparse expansion on an arbitrary pre-assigned orthonormal basis. We prove that replacing the usual quadratic regularizing penalties by weighted l^p-penalties on the coefficients of such expansions, with 1 < or = p < or =2, still regularizes the problem. If p < 2, regularized solutions of such l^p-penalized problems will have sparser expansions, with respect to the basis under consideration. To compute the corresponding regularized solutions we propose an iterative algorithm that amounts to a Landweber iteration with thresholding (or nonlinear shrinkage) applied at each iteration step. We prove that this algorithm converges in norm. We also review some potential applications of this method.

Comments: 30 pages, 3 figures; this is version 2 - changes with respect to v1: small correction in proof (but not statement of) lemma 3.15; description of Besov spaces in intro and app A clarified (and corrected); smaller pointsize (making 30 instead of 38 pages)
Categories: math.FA, math.NA
Related articles: Most relevant | Search more
arXiv:1409.7610 [math.FA] (Published 2014-09-26)
Generalized Convergence Rates Results for Linear Inverse Problems in Hilbert Spaces
arXiv:1604.03364 [math.FA] (Published 2016-04-12)
Elastic-net regularization versus $\ell^1$-regularization for linear inverse problems with quasi-sparse solutions
arXiv:1511.02950 [math.FA] (Published 2015-11-10)
Optimal Convergence Rates Results for Linear Inverse Problems in Hilbert Spaces