TY - JOUR
T1 - Approximate sparse recovery
T2 - Optimizing time and measurements
AU - Gilbert, Anna C.
AU - Li, Yi
AU - Porat, Ely
AU - Strauss, Martin J.
PY - 2012
Y1 - 2012
N2 - A Euclidean approximate sparse recovery system consists of parameters k,N, an m-by-N measurement matrix, φ, and a decoding algorithm, D. Given a vector, x, the system approximates x by x̂ = D(φx), which must satisfy |̂x - x|2 ≤ C|x - x k|2, where x k denotes the optimal k-term approximation to x. (The output ̂x may have more than k terms.) For each vector x, the system must succeed with probability at least 3/4. Among the goals in designing such systems are minimizing the number m of measurements and the runtime of the decoding algorithm, D. In this paper, we give a system with m = O(k log(N/k)) measurements-matching a lower bound, up to a constant factor-and decoding time k log O(1) N, matching a lower bound up to a polylog(N) factor. We also consider the encode time (i.e., the time to multiply φ by x), the time to update measurements (i.e., the time to multiply φ by a 1-sparse x), and the robustness and stability of the algorithm (resilience to noise before and after the measurements). Our encode and update times are optimal up to log(k) factors. The columns of φ have at most O(log 2(k) log(N/k)) nonzeros, each of which can be found in constant time. Our full result, a fully polynomial randomized approximation scheme, is as follows. If x = x k+ν1, where ν1 and ν2 (below) are arbitrary vectors (regarded as noise), then setting x̂ = D(φx+ν2), and for properly normalized φ, we get |x -x̂| 2 2 ≤ (1+ε) |ν1| 2 2 +ε |ν2| 2 2 using O((k/ε) log(N/k)) measurements and (k/epsi;) log O(1)(N) time for decoding.
AB - A Euclidean approximate sparse recovery system consists of parameters k,N, an m-by-N measurement matrix, φ, and a decoding algorithm, D. Given a vector, x, the system approximates x by x̂ = D(φx), which must satisfy |̂x - x|2 ≤ C|x - x k|2, where x k denotes the optimal k-term approximation to x. (The output ̂x may have more than k terms.) For each vector x, the system must succeed with probability at least 3/4. Among the goals in designing such systems are minimizing the number m of measurements and the runtime of the decoding algorithm, D. In this paper, we give a system with m = O(k log(N/k)) measurements-matching a lower bound, up to a constant factor-and decoding time k log O(1) N, matching a lower bound up to a polylog(N) factor. We also consider the encode time (i.e., the time to multiply φ by x), the time to update measurements (i.e., the time to multiply φ by a 1-sparse x), and the robustness and stability of the algorithm (resilience to noise before and after the measurements). Our encode and update times are optimal up to log(k) factors. The columns of φ have at most O(log 2(k) log(N/k)) nonzeros, each of which can be found in constant time. Our full result, a fully polynomial randomized approximation scheme, is as follows. If x = x k+ν1, where ν1 and ν2 (below) are arbitrary vectors (regarded as noise), then setting x̂ = D(φx+ν2), and for properly normalized φ, we get |x -x̂| 2 2 ≤ (1+ε) |ν1| 2 2 +ε |ν2| 2 2 using O((k/ε) log(N/k)) measurements and (k/epsi;) log O(1)(N) time for decoding.
KW - Approximation
KW - Embedding
KW - Sketching
KW - Sparse approximation
KW - Sublinear algorithms
UR - http://www.scopus.com/inward/record.url?scp=84861624238&partnerID=8YFLogxK
U2 - 10.1137/100816705
DO - 10.1137/100816705
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:84861624238
SN - 0097-5397
VL - 41
SP - 436
EP - 453
JO - SIAM Journal on Computing
JF - SIAM Journal on Computing
IS - 2
ER -