Compressed sensing theory explains why Lasso programs recover structured high-dimensional signals with minimax order-optimal error. Yet, the optimal choice of the program’s governing parameter is often unknown in practice. It is still unclear how variation of the governing parameter impacts recovery error in compressed sensing, which is otherwise provably stable and robust. We establish a novel notion of instability in Lasso programs when the measurement matrix is identity. This is the proximal denoising setup. We prove asymptotic cusp-like behaviour of the risk as a function of the parameter choice, and illustrate the theory with numerical simulations. For example, a 0.1% underestimate of a Lasso parameter can increase the error significantly; and a 50% underestimate can cause the error to increase by a factor of 109. We hope that revealing parameter instability regimes of Lasso programs helps to inform a practitioner’s choice. Finally, we discuss how these results extend to their more general Lasso counterparts.
Event Categories
- IAM Seminar
- IAM Graduate Seminar
- SCAIM Seminar
- Fluids Seminar
- Mathematics of Information and Applications Seminar
- BC Data Science Colloquium
- IAM Distinguished Colloquium
- Probability Seminar
- CS Seminar
- Quantum Information and Computing
- Algorithms Seminars
- Annual Retreat
- PIMS
- Other IAM Events
- Computer Science Distinguished Lectures
- IAM-PIMS Distinguished Colloquium
- IAM Public Lecture
- IAM Distinguished Alumni Lecture
- IAM Seminar / UBC Early-Career Award Lecture
- Mathematical Biology Seminar
- IAM Career Chats