- This event has passed.
M.Tech.(Research) Thesis Defense of Mr. Vinayak Killedar
November 17, 2021 @ 7:30 PM - 8:30 PM IST
Title of the thesis: Solving Inverse Problems Using a Deep Generative Prior
Supervisor: Prof. Chandra Sekhar Seelamantula (EE)
Examiner: Prof. Sumohana Channappayya (EE, IIT Hyderabad)
Abstract: The objective in an inverse problem is to recover a signal from its measurements, given the knowledge of the measurement operator. In this thesis, we address the problems of compressive sensing (CS) and compressive phase retrieval (CPR) using a generative prior model with sparse latent sampling. These problems are ill-posed and have infinite solutions. Structural assumptions such as smoothness, sparsity and non-negativity are imposed on the solution to obtain a unique solution.
The standard CS and CPR formulations impose a sparsity prior on the signal. Recently, generative modeling approaches have removed the sparsity constraint and shown superior performance over traditional CS and CPR techniques in recovering signals from fewer measurements. Generative model uses a pre-trained network, the generator of a Generative Adversarial Network (GAN) or the decoder of a Variational Autoencoder (VAE) to model the distribution of the signal and impose a Set-Restricted Eigenvalue Condition (S-REC) on the measurement operator. The S-REC property places a condition on the l-2 norm of the difference in signal and measurement domain for signals coming from the set S. Solving CS and CPR using generative models have some limitations. The reconstructed signal is constrained to lie in the range-space of the generator. The reconstruction process is slow because the latent space is optimized through gradient-descent (GD) and requires several restarts. It has been argued that the distribution of natural images is not confined to a single manifold, but a union of submanifolds. To take advantage of this property, we propose a sparsity-driven latent space sampling (SDLSS) framework, where sparsity is imposed in the latent space. The effect is to divide the latent space into subspaces such that the generator models map each subspace into a submanifold. We propose a proximal meta-learning (PML) algorithm to optimize the parameters of the generative model along with the latent code. The PML algorithm reduces the number of gradient steps required during testing and imposes sparsity in the latent space. We derive the sample complexity bounds within the SDLSS framework for the linear CS model, which is a generalization of the result available in the literature. The results demonstrate that, for a higher degree of compression, the SDLSS method is more efficient than the state-of-the-art deep compressive sensing (DCS) method. We consider both linear and learned nonlinear sensing mechanisms, where the nonlinear operator is a learned fully connected neural network or a convolutional neural network and show that the learned nonlinear version is superior to the linear one.
As an application of the nonlinear sensing operator, we consider compressive phase retrieval, wherein the problem is to reconstruct a signal from the magnitude of its compressed linear measurements. We adapt the S-REC imposed on the measurement operator and propose a novel cost function. The SDLSS framework along with PML algorithm is applied to optimize the sparse latent space such that the adapted $\mathcal{S}$-REC loss and data-fitting error are minimized. The reconstruction process is fast and requires few gradient steps during testing compared with the state-of-art deep phase retrieval technique.
Experiments are conducted on standard datasets such as MNIST, Fashion-MNIST, CIFAR-10, and CelebA to validate the efficiency of SDLSS framework for CS and CPR. The results show that, for a given dataset, there exists an effective input latent dimension for the generative model. Performance quantification is carried out by employing three objective metrics: peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM), and reconstruction error (RE) per pixel, which are averaged across the test dataset.
About the speaker: Vinayak Killedar obtained a B.E. (ECE) degree from M. S. Ramaiah Institute of Technology (MSRIT), Bangalore in 2008. During 2008-2010, he worked for Robert Bosch Engineering and Business Solution (RBEI), Coimbatore. He joined the M.Tech.(Signal Processing) program in National Institute of Technology (NIT) Calicut and graduated in 2013. He worked for Continental AG during 2014-2018 in the areas of autonomous driving and radar signal processing. Subsequently, he joined the Spectrum Lab, Department of Electrical Engineering, Indian Institute of Science for M.Tech.(Research) and specialized in Compressed Sensing and Machine Learning. He is presently a Senior Technical Specialist at Ansys, Kempten, Germany.