Sparse and Smooth Signal Estimation: Convexification of L0-Formulations
Abstract
Signal estimation problems with smoothness and sparsity priors can be naturally modeled as quadratic optimization with $\ell_0$-“norm” constraints. Since such problems are non-convex and hard-to-solve, the standard approach is, instead, to tackle their convex surrogates based on $\ell_1$-norm relaxations. In this paper, we propose new iterative (convex) conic quadratic relaxations that exploit not only the $\ell_0$-“norm” terms, but also the fitness and smoothness functions. The iterative convexification approach substantially closes the gap between the $\ell_0$-“norm” and its $\ell_1$ surrogate. These stronger relaxations lead to significantly better estimators than $\ell_1$-norm approaches and also allow one to utilize affine sparsity priors. In addition, the parameters of the model and the resulting estimators are easily interpretable. Experiments with a tailored Lagrangian decomposition method indicate that the proposed iterative convex relaxations yield solutions within 1\% of the exact $\ell_0$-approach, and can tackle instances with up to 100,000 variables under one minute.
Cite
Text
Atamturk et al. "Sparse and Smooth Signal Estimation: Convexification of L0-Formulations." Journal of Machine Learning Research, 2021.Markdown
[Atamturk et al. "Sparse and Smooth Signal Estimation: Convexification of L0-Formulations." Journal of Machine Learning Research, 2021.](https://mlanthology.org/jmlr/2021/atamturk2021jmlr-sparse/)BibTeX
@article{atamturk2021jmlr-sparse,
title = {{Sparse and Smooth Signal Estimation: Convexification of L0-Formulations}},
author = {Atamturk, Alper and Gomez, Andres and Han, Shaoning},
journal = {Journal of Machine Learning Research},
year = {2021},
pages = {1-43},
volume = {22},
url = {https://mlanthology.org/jmlr/2021/atamturk2021jmlr-sparse/}
}