Translator Disclaimer
9 September 2019 Compressed sensing and generative models
Author Affiliations +
The goal of compressed sensing is make use of image structure to estimate an image from a small number of linear measurements. The structure is typically represented by sparsity in a well-chosen basis. We describe how to achieve guarantees similar to standard compressed sensing but without employing sparsity at all -instead, we suppose that vectors lie near the range of a generative model G: Rk → Rn. Our main theorem here is that, if G is L-Lipschitz, then roughly O(k log L) random Gaussian measurements suffice; this is O(kd log n) for typical d-layer neural networks. The above result describes how to use a model to recover a signal from noisy data. But if the data is noisy, how can we learn the generative model in the first place? This paper will describe how to incorporate the measurement process in generative adversarial network (GAN) training. Even if the noisy data does not uniquely identify the non-noisy signal, the distribution of noisy data may still uniquely identify the distribution of non-noisy signals. In presenting the above results we summarize and synthesize the work of [BJPD17] and [BPD18]. We then add some observations on the limitations of the approaches.
© (2019) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Eric Price "Compressed sensing and generative models", Proc. SPIE 11138, Wavelets and Sparsity XVIII, 111380R (9 September 2019);

Back to Top