TY - JOUR
T1 - Compressed Sensing with 1D Total Variation: Breaking Sample Complexity Barriers via Non-Uniform Recovery
AU - Genzel, Martin
AU - März, Maximilian
AU - Seidel, Robert
PY - 2022/3
Y1 - 2022/3
N2 - This paper investigates total variation minimization in one spatial dimension for the recovery of gradient-sparse signals from undersampled Gaussian measurements. Recently established bounds for the required sampling rate state that uniform recovery of all $s$-gradient-sparse signals in ${\mathbb{R}}^n$ is only possible with $m \gtrsim \sqrt{s n} \cdot{\operatorname{PolyLog}}(n)$ measurements. Such a condition is especially prohibitive for high-dimensional problems, where $s$ is much smaller than $n$. However, previous empirical findings seem to indicate that this sampling rate does not reflect the typical behavior of total variation minimization. The present work provides a rigorous analysis that breaks the $\sqrt{s n}$-bottleneck for a large class of “natural” signals. The main result shows that non-uniform recovery succeeds with high probability for $m \gtrsim s \cdot{\operatorname{PolyLog}}(n)$ measurements if the jump discontinuities of the signal vector are sufficiently well separated. In particular, this guarantee allows for signals arising from a discretization of piecewise constant functions defined on an interval. The key ingredient of the proof is a novel upper bound for the associated conic Gaussian mean width, which is based on a signal-dependent, non-dyadic Haar wavelet transform. Furthermore, a natural extension to stable and robust recovery is addressed.
AB - This paper investigates total variation minimization in one spatial dimension for the recovery of gradient-sparse signals from undersampled Gaussian measurements. Recently established bounds for the required sampling rate state that uniform recovery of all $s$-gradient-sparse signals in ${\mathbb{R}}^n$ is only possible with $m \gtrsim \sqrt{s n} \cdot{\operatorname{PolyLog}}(n)$ measurements. Such a condition is especially prohibitive for high-dimensional problems, where $s$ is much smaller than $n$. However, previous empirical findings seem to indicate that this sampling rate does not reflect the typical behavior of total variation minimization. The present work provides a rigorous analysis that breaks the $\sqrt{s n}$-bottleneck for a large class of “natural” signals. The main result shows that non-uniform recovery succeeds with high probability for $m \gtrsim s \cdot{\operatorname{PolyLog}}(n)$ measurements if the jump discontinuities of the signal vector are sufficiently well separated. In particular, this guarantee allows for signals arising from a discretization of piecewise constant functions defined on an interval. The key ingredient of the proof is a novel upper bound for the associated conic Gaussian mean width, which is based on a signal-dependent, non-dyadic Haar wavelet transform. Furthermore, a natural extension to stable and robust recovery is addressed.
UR - https://www.mendeley.com/catalogue/975d3e9c-d730-301f-9a52-284521498db1/
U2 - 10.1093/imaiai/iaab001
DO - 10.1093/imaiai/iaab001
M3 - Article
SN - 2049-8764
VL - 11
SP - 203
EP - 250
JO - Information and Inference: A Journal of the IMA
JF - Information and Inference: A Journal of the IMA
IS - 1
ER -