We introduce a method to infer a variational approximation to the posterior distribution of solutions in computational imaging inverse problems. Machine learning methods applied to computational imaging have proven very successful, but have so far largely focused on retrieving a single optimal solution for a given task. Such retrieval is arguably an incomplete description of the solution space, as in ill-posed inverse problems there may be many similarly likely reconstructions. We minimise an upper bound on the divergence between our approximate distribution and the true intractable posterior, thereby obtaining a probabilistic description of the solution space in imaging inverse problems with empirical prior. We demonstrate the advantage of our technique in quantitative simulations with the CelebA dataset and common image reconstruction tasks. We then apply our method to two of the currently most challenging problems in experimental optics: imaging through highly scattering media and imaging through multi-modal optical fibres. In both settings we report state of the art reconstructions, while providing new capabilities, such as estimation of error-bars and visualisation of multiple likely reconstructions.
oddsidemargin has been altered.
marginparsep has been altered.
topmargin has been altered.
marginparwidth has been altered.
marginparpush has been altered.
paperheight has been altered.
The page layout violates the ICML style. Please do not change the page layout, or include packages like geometry, savetrees, or fullpage, which change it for you. We’re not able to reliably undo arbitrary changes to the style. Please remove the offending package(s), or layout-changing commands and try again.
Variational Inference for Computational Imaging Inverse Problems
Francesco Tonolini 0 Ashley Lyons 0 Piergiorgio Caramazza 0 Daniele Faccio 0 Roderick Murray-Smith 0
We acknowledge funding from Amazon and EPSRC grants EP/M01326X/1, EP/R018634/1.\@xsect
In recent years, machine learning has taken an increasingly central role in solving computational imaging (CI) problems, leading to unprecedented reconstruction quality in a range of settings (Lucas et al., 2018; Kulkarni et al., 2016; Chang et al., 2017) and allowing for new and challenging imaging tasks to be accomplished (Liu et al., 2015a; Isola et al., 2017). However, learning methods have so far been applied to inverse imaging problems to infer a single near-optimal reconstruction, lacking a probabilistic explanation of the solution space. Conversely, as advancements in machine learning allow for increasingly ill-posed imaging inverse problems to be approached, probabilistic descriptions become important; if appreciably different solutions satisfy similarly well the observed measurements and empirical prior, an accurate retrieval should capture their variability.
In this paper, we introduce a variational method for approximating the posterior distribution in ill-posed imaging inverse problems, where the prior is empirically learned from signal examples. Given some model of the forward observation process and a set of signal examples, we minimise stochastically an upper bound on the cross entropy between the true posterior and a flexible parametric distribution. With the new method, we obtain better reconstruction performance than inference using equivalent deterministic neural networks, while at the same time retrieving information which captures the nature of the posterior density, such as accurate pixel-wise error-bars and draws from the distribution of likely solutions.
We first demonstrate the advantage of our method in quantitative simulations, retrieving images of faces from the CelebA dataset (Liu et al., 2015b) under different common degradation conditions. We then perform reconstructions from experimental acquisitions in two particularly ill-posed imaging settings, currently at the focus of optical imaging research: computational diffuse imaging (Badon et al., 2016; Durán et al., 2015; Boccolini et al., 2017; Horisaki et al., 2016) and imaging through multi-modal optical fibres (Plöschner et al., 2015; Turtaev et al., 2018; Moran et al., 2018). In both sets of experiments, we report state of the art reconstructions, while clearly demonstrating the importance of capturing the nature of variation in the solution space through our variational approach.
Computational imaging (CI) broadly identifies a class of methods in which an image of interest is not directly observed, but rather inferred algorithmically from one or more measurements (Bertero & Boccacci, 1998). Many image recovery tasks fall within this definition, such as de-convolution (Starck et al., 2003; Bioucas-Dias et al., 2006), computed tomography (CT) (Ritschl et al., 2011) and structured illumination imaging (Magalhães et al., 2011).
Traditionally, CI tasks are modelled as inverse problems, where a target signal is measured through a forward model , yielding observations . The aim is then to retrieve the signal from the observations (Bertero & Boccacci, 1998; Vogel, 2002). In the following subsections, we review the main advances in solving imaging inverse problems, starting from linear models and user defined regularisation functions, to then focus on the more recent application of learning based methods.
In many CI problems, the forward model can be described as a linear operator and some independent noise , such that (Bertero & Boccacci, 1998; Daubechies et al., 2004). Under such an assumption, the aim is to find a solution image which satisfies the linear observations well, while retaining some properties of interest. The estimate of is then recovered by minimising an objective function of the form
where indicates the Euclidean norm and is a penalty function that enforces some desired property in . In the case of images, it is common to assume that is sparse in some basis, such as frequency or wavelets, leading to -norm penalty functions of the form , where is a unitary operator which maps to a sparse basis (Daubechies et al., 2004; Donoho, 2006; Figueiredo et al., 2007). For such choices of penalty, the objective function of equation 1 is convex. This makes the optimisation problem solvable with a variety of efficient methods (Daubechies et al., 2004; Yang & Zhang, 2011) and provides theoretical guarantees on the recoverability of the solution (Candes et al., 2006).
The aforementioned framework has been widely applied to perform CI. For instance, many image restoration tasks, such as de-blurring, up-sampling and in-painting have been formulated as ill-conditioned linear inverse problems and are solved as described above (Osher et al., 2005). Various more complex sensing models can also be cast as linear operators, leading to the use of constrained optimisation in several CI systems that rely on ill-posed observations, such as sparse CT (Ritschl et al., 2011), single pixel photography (Sun et al., 2013) and imaging of objects hidden from view (Velten et al., 2012).
However, despite its theoretical benefits and wide adoption, sparsity regularisation is often too generic to accurately model the desired assumptions in a given imaging setting (Chang et al., 2017).
Bayesian inference has been applied to approximate the posterior distribution in linear inverse problems such as those described in the previous section, where the prior distribution is defined by some assumed image property, such as sparsity in some basis or image smoothness (Ji et al., 2008; Mohammad-Djafari, 2013). In particular, variational inference has been applied to common inverse problems such as blind deconvolution and image super-resolution (Likas & Galatsanos, 2004; Babacan et al., 2011). In these works, the true intractable posterior of the inverse problem is approximated with a tractable parametric function.
Differently from these methods, we aim to use examples of images as signal prior, rather than defined functional assumptions. This makes the techniques described above not directly applicable to our target case, as the prior can not be point evaluated, but we only have access to a finite number of draws from it in the form of a data set of signal examples.
The increasing availability of image data sets and continuous advancements in learning inference models enabled new possibilities for CI and inverse problems in general. Most learning approaches can arguably be described as inverse mapping models; with enough example pairs available, a neural network can be trained to directly recover a signal from observations (Lucas et al., 2018).
With such approaches, because a direct inverse mapping is learned, there are a number of advantages compared to optimisation with sparse regularisation. First, the model is trained with a set of images the target solution is assumed to belong to, implicitly making the signal assumptions more specific than sparsity in some basis. Second, the observation model is not constrained to be linear, or even differentiable; so long as we have access to a large number of signal-observations pairs we can train the network to perform the inversion. Third, once the model is trained, inference is non-iterative and thus typically much faster.
Several methods have been developed to perform such inference in a variety of CI settings (Egmont-Petersen et al., 2002; McCann et al., 2017). State of the art performance has been demonstrated with specifically designed neural networks models in many common image processing tasks, such as deconvolution and super-resolution (Xu et al., 2014; Ledig et al., 2017), as well as signal recovery from under determined projective measurements (Kulkarni et al., 2016).
Despite their proven ability to recover compelling images in CI inverse problems, empirical inverse mappings often lack model consistency, as recovered solutions are not directly induced to satisfy the observed measurements, but only determined by a flexible approximator trained with arbitrarily many signal-observations example pairs.
A second emerging class of learning based methods in CI is that of maximum a posteriori (MAP) inference with generative models. In these frameworks, similarly to MAP inference with analytical priors, the solution is found by minimising the Euclidean distance to the observations . However, instead of regularising the solution with some penalty function, the signal of interest is assumed to be generated by a pre-trained generative model as , where is the generative model’s latent variable (Bora et al., 2017). As the generative model is trained with an example set that the target signal is assumed to belong to, generated images are expected to retain empirically learned properties (Chang et al., 2017). The solution is then found by performing the following minimisation,
In such a way, the solution is constrained to be within the domain of the generative model, as the recovered is by definition generated from , but at the same time agreement to the measurements is directly induced. Though the optimisation problem is not convex, certain theoretical bounds were derived for the cases of linear observation processes by Bora et al. (2017) and more recently phase-less linear observation processes (Hand et al., 2018).
MAP inference with generative models has been demonstrated in compressed sensing settings (Bora et al., 2017; Mardani et al., 2017) and with several common image processing tasks (Chang et al., 2017). Though it is iterative and requires the definition of a differentiable observation model, it is formally more consistent than inverse mappings; the solution is found by maximising data fidelity under empirically learned signal assumptions.
Notwithstanding their achievements, the rich classes of learning methods for CI described above focus on retrieving a single optimal solution to an inverse problem, lacking a probabilistic description of the solution space. Differently from these, we aim to infer an approximate posterior distribution from a set of measurements.
In our probabilistic modelling of inverse problems, we exploit conditional variational auto-encoders (CVAEs) to fit flexible posterior distributions from observations. CVAEs are latent variable generative models where the prior in the latent space is conditioned on an observation (Sohn et al., 2015). As in standard VAEs, the generative model is trained with an evidence lower bound (ELBO), fitting flexible posteriors to the training data.
CVAEs have been applied to perform a range of inference tasks, such as classification (Sohn et al., 2015), generation conditioned on classes (Nguyen et al., 2017), image-from-text inference (Yan et al., 2016) and missing value imputation (Nazabal et al., 2018). However, within CI, they have been used to infer from simple deterministic observation, such as missing pixels (Nguyen et al., 2017), while in many inverse problems the observation model is probabilistic and with potentially complicated conditional distribution.
Though we exploit CVAEs to build a flexible distribution, we aim to find a variational approximation to the posterior from a defined observation distribution , rather than building a generative model conditioned on deterministic observations. In practice, our method is different in the training of the conditional model; instead of given observations to condition the latent priors we have an observation density determined by our estimate of the observation process.111We use the inverse problems conventional notation for signals and observations , which is inverted compared to the common formulation of CVAEs.
We design a variational method to infer flexible posterior PDFs from observations in inverse problems where the prior is empirically built from image examples, hence capturing the probabilistic nature of the solution space.
We wish to recover the posterior density of signals from measurements , obtained through an observation process . In CI inverse problems, such observation process is often probabilistic, as in most cases sensors readings are subject to noise and the forward model parameters are known with limited precision. Measurements can then be assumed to be drawn from a distribution . The form of the posterior we wish to recover is then
This distribution is intractable to directly estimate, mainly for two reasons. Firstly, in the learning setting we consider here, we do not have access to the prior distribution , but only to a set of signal examples which are assumed to be drawn from it . Secondly, in a general case, we can not define directly a parametric form for , but can only draw from it by stochastically generating observations as , where is drawn from the parametric family of observation processes defined by our estimate. For example, in a convolutional process where we are uncertain about the point spread function (PSF), we can generate an observation by first drawing a PSF and then convolving the signal with it to obtain a draw . The estimate of the forward observation process and its parameters can be completely predefined by some physical model, or partially inferred from signal-observation example pairs, depending on the imaging setting and available data. Furthermore, in many cases a parametric function for is not available, but draws can be generated through numeric simulations of the physical observation process.
Due to the intractability of the true posterior , we aim to find some parametric PDF to approximate it well over the distribution of expected measurements . To this end, we minimise the expectation of the cross entropy under the measurements’ distribution ,
Here, is the set of parameters that define the distribution . The optimisation of Equation 4 is equivalent to fitting to the true posterior over the distribution of measurements we expect to see . We can further simplify this objective function to give
In this form, the objective function can be estimated stochastically, as the training signals are assumed to be sampled from the empirical prior and we can draw measurements from the conditional distribution through an estimate of the observation model, as described in subsection id1.
As images and other natural signals often lie on complicated manifolds and the observation density is arbitrary, the approximate distribution needs to be rather flexible. To this end, we make use of a conditional latent variable model
The latent distribution is an isotropic Gaussian distribution , where the moments and are inferred from a measurement by a neural network. Similarly, is an isotropic Gaussian , where the moments are inferred from both a measurement and latent variable with a second neural network. A schematic representation of the inference model is shown in Fig. 1. This approximate posterior has the same form as the marginal likelihood in CVAEs and is intractable to train directly (Sohn et al., 2015). However, we can make use of a recognition function conditioned on signals and measurements to derive an ELBO for that is instead efficient to estimate and optimise (Kingma & Welling, 2013; Rezende et al., 2014):
The recognition function is also an isotropic Gaussian distribution, where the moments are inferred from measurements and signals with a neural network. We can then define a lower bound for the objective function of Eq. 5 as
The above expression is equivalent to the expectation of a CVAE lower bound under the observation distribution , integrated over the empirical prior . We can estimate this bound stochastically as
where are training examples, assumed to be drawn from the signal prior , are synthetic measurements obtained from example signals through a forward propagation and are drawn from the recognition function .
To train the approximate posterior, we maximise the empirical lower bound of Eq. 9 with respect to the parameters , and , over the training set . Every iteration, to draw from , we generate simulated observations from a physical model of the observation process. With the trained model, we can then draw examples of solutions to the inverse problem given an observation by first drawing a latent variable and subsequently drawing from the conditional . From these, we can compute quantities that describe the aggregate statistics of the solution space, such as pixel-wise means and error bars. A schematic representation of the model is shown in Fig. 1. The method described retains a number of desirable advantages for solving imaging inverse problems:
From an observation, the model infers a parametric distribution of solutions, from which we can draw and obtain uncertainty measures. Such description is much richer than a single best estimate and arguably more informative for further information extraction or decision making, whether this is done by a human or an automated system.
Inference is non-iterative, meaning that once the approximate posterior is trained, drawing from it and estimating quantities of interest is expectedly much faster than MAP methods.
An estimate of the forward model is explicitly incorporated in the training, meaning that the approximate inversion is induced to be consistent with known properties of the physical observation process.
While signal-observation example pairs can be used to refine or discover a forward model, if an accurate estimate is already available, only examples of the target signals are required. In many settings these are easier to obtain than observation examples, or even already available in existing data sets, avoiding the need to carry out lengthy and potentially expensive acquisitions with specific imaging instruments.
To our knowledge, the method we present here is the first to combine all the aforementioned properties. In our evaluation we make use of fully connected layers for all neural networks involved. However, the general approach we propose is compatible with other architectures, such as convolutional layers designed to process the observations in a particular setting and generative networks that model dependencies between pixels in images (Gulrajani et al., 2016).
We carry out a number of experiments to evaluate our proposed method, both in simulation and with physical imaging systems. We first recover faces from the CelebA data set (Liu et al., 2015b) under different common simulated observation conditions, evaluating quantitatively our method. Secondly, we reconstruct images of letters deeply embedded within a scattering medium from experimental time of flight measurements. Lastly, we reconstruct images of hand-written digits and fashion items from the speckle patterns obtained when imaging through a multi-modal optical fiber.
We make use of the CelebA dataset, down-sampled to images. We consider different common image observation conditions, including blurring, down-sampling and missing pixels with additive Gaussian noise, each with different parameters and parameters uncertainties. For each case, we train our variational model using CelebA examples as training set and the corresponding observation model to generate samples . We then infer the approximate posterior from test observations, generated from a separate set of images. An example showing increasingly severe blurring and the resulting inference is given in fig. 2. Additional examples and details of the experiments are given in supplementary section A. As the observed image becomes more degraded, the pixel-wise uncertainty increases and draws from the approximate posterior diversify, reflecting the broader range of likely reconstructions in the solution space.
To quantitatively evaluate the mean reconstruction performance, we compare the peak signal to noise ratios (PSNRs) obtained with our method to those of a standard neural network and a CVAE trained on simulated examples-observation pairs. As shown in Fig. 3, our proposed method proved competitive or better in every considered example. To then evaluate our method in terms of distribution accuracy, we compare the test set ELBO to that of an equivalent CVAE. As shown in table 1, in every considered case, our variational method gave an appreciably higher ELBO, proving the advantage of training through the expectation from a forward model estimate, rather than simply signal-observation pairs.
Imaging through strongly diffusive media remains an outstanding problem in optical CI, with applications in biological and medical imaging and imaging in adverse environmental conditions. Visible or near-infrared light does propagate in turbid media, such as biological tissue or fog, however, its path is strongly affected by scattering, leading to the loss of any direct image information after a short propagation length.
Following the experimental implementation presented by Boccolini et al. (2017), we employ a near-infrared pulsed laser and a single photon sensitive time of flight (ToF) camera with a temporal resolution of to perform transmission diffuse imaging. In these experiments, we place different cut-out shapes of alphabetic letters between two identical thick slabs of diffusive material, with measured absorption and scattering coefficients of and respectively. We then illuminate one surface with the pulsed laser and image the opposite surface with the ToF camera. A schematic representation and a photograph of the set up are shown in fig. 5. Achieving accurate reconstructions with simple objects in this settings, is a first important step towards achieving imaging through biological tissue with near-infrared light and hence non-ionising radiation.
A pulse of light from the laser propagates through the diffusing material, reaches the hidden objects, which partially absorbs it, and then propagates further through the medium to the imaged surface. The ToF camera records a video of the light intensity as a function of time as it exits the medium. We subtract a video recorded with an empty piece of material to that obtained with the object present, thereby obtaining a video of the estimated difference in light intensity caused by the hidden object. such experimental videos are shown in Fig. 4. At this depth, more than times longer than the mean free path, the diffusion effect is so severe that these basic shapes are not distinguishable directly from the videos. Furthermore, the measurements experience low signal-to-noise ratio due to the low light intensity that reaches the imaged surface and the low fill factor of the ToF camera, which is about .
The reconstruction of a hidden object from a ToF video can be cast as a linear inverse problem, where the image is observed through a linear operator to generate an observed video . Detailed descriptions of the observation model and the experiments are given in supplementary section B. This problem has been previously approached by performing MAP inference with -norm and TV-norm penalties (Boccolini et al., 2017). As shown in fig. 4, this technique is capable of retrieving general features of the object embedded in the scattering medium, but sometimes results in severe artefacts that make the images unrecognisable. Furthermore, to obtain the displayed results, it is necessary to carefully tune the penalty coefficients of the MAP objective function for each example, making such retrieval highly dependent on human supervision.
To approach this problem, we train our variational method with a down-sampled subset of the NIST dataset (Johnson, 2010), composed of hand-written numbers and letters, which we assumed to be samples from an empirical prior . Each training iteration, observations are generated through a simulation of the light propagation in the medium. The trained model is then used to infer an approximate posterior of solutions from the experimental ToF videos. Fig. 4 Shows the empirical mean and standard deviation of the recovered posterior, along with examples of drawn samples. Exploiting a more specific empirical prior compared to previous MAP estimate techniques, the proposed variational method retrieves more accurate reconstructions, where the different letters are clearly recognisable. Moreover, this particularly ill-posed inverse problem example highlights the importance of using a variational approach; the solution space given a diffuse video is rather variable and, unlike MAP inference and other single estimate methods, through the approximate posterior we can capture such variability by empirically estimating uncertainty and visualising different drawn samples. We further note that no training experimental acquisition was required, but only simple calibration experiments to measure the coefficients of the material. In this setting this is highly desirable as each example is manually prepared and carefully aligned before an acquisition. In addition, unlike the analytic MAP method, the technique is non-iterative and does not require careful tuning of any parameter to obtain results such as those shown in Fig. 4.
Being able to transmit images through multi-modal optical fibers (MMOFs) is a long standing goal of optical imaging and holds potential in applications ranging from endoscopy to communication. A phase coherent image can be focused into a MMOF and couple with its modes to propagate through the fibre with minimal loss of information. However, the input light field undergoes a linear complex transformation before emerging at the output as a speckle pattern (examples shown in Fig. 6). Such transformation depends on physical factors, such as fibre material, size, arrangement and even temperature. This makes the reconstruction of images from speckle patterns an extremely challenging task and traditionally requires to measure both amplitude and phase of the speckle patterns with a spectrometer (Plöschner et al., 2015).
Recently, machine learning methods have been applied to learn inverse mappings from speckle intensities to images, without requiring phase measurements at the output (Moran et al., 2018; Borhani et al., 2018). We aim to perform this same task, making use of the paired images and speckles data set presented in Moran et al. (2018), but with some important differences in our approach. Instead of learning to infer images from speckle patterns directly, we use the speckles-images pairs in the training set to estimate the coefficients of the complex transmission matrix and subsequently incorporate this learned forward observation model in the training of our approximate posterior. In such a way we are able to impose consistency with the known physical structure of the observation model while exploiting available paired examples to learn an estimate of its unknown coefficients. As in the main experiments presented in Moran et al. (2018), we train with a data set composed of images of hand-written digits, pieces of clothing and random patterns. In our probabilistic framework, this choice of training data reflects a prior which is highly “peaked” around images of numbers and clothes, but presents appreciable density everywhere, as random patterns are included in the set. In Fig. 6 We compare reconstructions obtained with our variational inference method to those recovered with the architecture presented in (Moran et al., 2018).
The average reconstruction recovered with our method is competitive with that learned through the deterministic inverse mapping. However, the variational method is capable of identifying regions of higher uncertainty, as it can be seen in the recovered standard deviations. In the clothes examples, reconstructions are less accurate in regions of fine details, but the recovered standard deviations correctly predicts such regions. In the last example, which does not resemble the training images of numbers and clothes, the reconstruction, though still recognisable, is visibly less accurate. However, this lower performance is predicted by the standard deviation returned with the variational method.
We presented a method to infer a variational approximation to the posterior distribution of solutions in imaging inverse problems. With such method, reconstructions are recovered non-iteratively, retain consistency with a defined physical model of the observation process and provide probabilistic information, such as pixel-wise error bars and different samples from the solution space. Simulated quantitative experiments showed superior performance to deterministic methods and standard CVAEs with equivalent neural architectures, simultaneously providing the aforementioned statistical descriptions. In real imaging experiments, the proposed method performed competitively or better in its average reconstructions compared to previous approaches and enabled new capabilities, such as predicting the accuracy of different image regions and visualising the variation in probable reconstructions through sampling. The method is also widely applicable and extendable to a variety of other systems; There are many settings, ranging from medical imaging to autonomous vehicles sensing, in which observations obey a physical model, reconstructions can greatly benefit from learning and accuracy measures are useful or even critical for decision making.
- Babacan et al. (2011) Babacan, S. D., Molina, R., and Katsaggelos, A. K. Variational Bayesian super resolution. IEEE Transactions on Image Processing, 20(4):984–999, 2011.
- Badon et al. (2016) Badon, A., Li, D., Lerosey, G., Boccara, A. C., Fink, M., and Aubry, A. Smart optical coherence tomography for ultra-deep imaging through highly scattering media. Science advances, 2(11):e1600370, 2016.
- Bertero & Boccacci (1998) Bertero, M. and Boccacci, P. Introduction to inverse problems in imaging. CRC press, 1998.
- Bioucas-Dias et al. (2006) Bioucas-Dias, J. M., Figueiredo, M. A., and Oliveira, J. P. Total variation-based image deconvolution: a majorization-minimization approach. In Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on, volume 2, pp. II–II. IEEE, 2006.
- Boccolini et al. (2017) Boccolini, A., Tonolini, F., Leach, J., Henderson, R., and Faccio, D. Imaging inside highly diffusive media with a space and time-resolving single-photon sensor. In Imaging Systems and Applications, pp. ITu3E–2. Optical Society of America, 2017.
- Bora et al. (2017) Bora, A., Jalal, A., Price, E., and Dimakis, A. G. Compressed sensing using generative models. arXiv preprint arXiv:1703.03208, 2017.
- Borhani et al. (2018) Borhani, N., Kakkava, E., Moser, C., and Psaltis, D. Learning to see through multimode fibers. Optica, 5:960, August 2018.
- Candes et al. (2006) Candes, E. J., Romberg, J. K., and Tao, T. Stable signal recovery from incomplete and inaccurate measurements. Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 59(8):1207–1223, 2006.
- Chang et al. (2017) Chang, J.-H. R., Li, C.-L., Poczos, B., Kumar, B. V., and Sankaranarayanan, A. C. One network to solve them all-solving linear inverse problems using deep projection models. In ICCV, pp. 5889–5898, 2017.
- Daubechies et al. (2004) Daubechies, I., Defrise, M., and De Mol, C. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 57(11):1413–1457, 2004.
- Donoho (2006) Donoho, D. L. Compressed sensing. IEEE Transactions on information theory, 52(4):1289–1306, 2006.
- Durán et al. (2015) Durán, V., Soldevila, F., Irles, E., Clemente, P., Tajahuerce, E., Andrés, P., and Lancis, J. Compressive imaging in scattering media. Optics express, 23(11):14424–14433, 2015.
- Egmont-Petersen et al. (2002) Egmont-Petersen, M., de Ridder, D., and Handels, H. Image processing with neural networks – a review. Pattern recognition, 35(10):2279–2301, 2002.
- Figueiredo et al. (2007) Figueiredo, M. A., Nowak, R. D., and Wright, S. J. Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems. IEEE Journal of selected topics in signal processing, 1(4):586–597, 2007.
- Gulrajani et al. (2016) Gulrajani, I., Kumar, K., Ahmed, F., Taiga, A. A., Visin, F., Vazquez, D., and Courville, A. PixelVAE: A latent variable model for natural images. arXiv preprint arXiv:1611.05013, 2016.
- Hand et al. (2018) Hand, P., Leong, O., and Voroninski, V. Phase retrieval under a generative prior. In Advances in Neural Information Processing Systems, pp. 9154–9164, 2018.
- Horisaki et al. (2016) Horisaki, R., Takagi, R., and Tanida, J. Learning-based imaging through scattering media. Optics express, 24(13):13738–13743, 2016.
- Isola et al. (2017) Isola, P., Zhu, J.-Y., Zhou, T., and Efros, A. A. Image-to-image translation with conditional adversarial networks. arXiv preprint, abs/1611.07004, 2017.
- Ji et al. (2008) Ji, S., Xue, Y., Carin, L., et al. Bayesian compressive sensing. IEEE Transactions on Signal Processing, 56(6):2346, 2008.
- Johnson (2010) Johnson, S. G. NIST Special Database 30. 2010.
- Kingma & Welling (2013) Kingma, D. P. and Welling, M. Auto-encoding variational Bayes. arXiv preprint arXiv, pp. 1312.6114, 2013.
- Kulkarni et al. (2016) Kulkarni, K., Lohit, S., Turaga, P., Kerviche, R., and Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 449–458, 2016.
- Ledig et al. (2017) Ledig, C., Theis, L., Huszár, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A. P., Tejani, A., Totz, J., Wang, Z., et al. Photo-realistic single image super-resolution using a generative adversarial network. In CVPR, volume 2, pp. 4, 2017.
- Likas & Galatsanos (2004) Likas, A. C. and Galatsanos, N. P. A variational approach for Bayesian blind image deconvolution. IEEE Transactions on Signal Processing, 52(8):2222–2233, 2004.
- Liu et al. (2015a) Liu, F., Shen, C., and Lin, G. Deep convolutional neural fields for depth estimation from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5162–5170, 2015a.
- Liu et al. (2015b) Liu, Z., Luo, P., Wang, X., and Tang, X. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV), 2015b.
- Lucas et al. (2018) Lucas, A., Iliadis, M., Molina, R., and Katsaggelos, A. K. Using deep neural networks for inverse problems in imaging: beyond analytical methods. IEEE Signal Processing Magazine, 35(1):20–36, 2018.
- Magalhães et al. (2011) Magalhães, F., Araújo, F. M., Correia, M. V., Abolbashari, M., and Farahi, F. Active illumination single-pixel camera based on compressive sensing. Applied optics, 50(4):405–414, 2011.
- Mardani et al. (2017) Mardani, M., Gong, E., Cheng, J. Y., Vasanawala, S., Zaharchuk, G., Alley, M., Thakur, N., Han, S., Dally, W., Pauly, J. M., et al. Deep generative adversarial networks for compressed sensing automates mri. arXiv preprint arXiv:1706.00051, 2017.
- McCann et al. (2017) McCann, M. T., Jin, K. H., and Unser, M. A review of convolutional neural networks for inverse problems in imaging. arXiv preprint arXiv:1710.04011, 2017.
- Mohammad-Djafari (2013) Mohammad-Djafari, A. Bayesian inference tools for inverse problems. In AIP Conference Proceedings, volume 1553, pp. 163–170. AIP, 2013.
- Moran et al. (2018) Moran, O., Caramazza, P., Faccio, D., and Murray-Smith, R. Deep, complex, invertible networks for inversion of transmission effects in multimode optical fibres. In Advances in Neural Information Processing Systems, pp. 3284–3295, 2018.
- Nazabal et al. (2018) Nazabal, A., Olmos, P. M., Ghahramani, Z., and Valera, I. Handling incomplete heterogeneous data using vaes. arXiv preprint arXiv:1807.03653, 2018.
- Nguyen et al. (2017) Nguyen, A., Clune, J., Bengio, Y., Dosovitskiy, A., and Yosinski, J. Plug & play generative networks: Conditional iterative generation of images in latent space. In CVPR, volume 2, pp. 7, 2017.
- Osher et al. (2005) Osher, S., Burger, M., Goldfarb, D., Xu, J., and Yin, W. An iterative regularization method for total variation-based image restoration. Multiscale Modeling & Simulation, 4(2):460–489, 2005.
- Plöschner et al. (2015) Plöschner, M., Tyc, T., and Čižmár, T. Seeing through chaos in multimode fibres. Nature Photonics, 9(8):529, 2015.
- Rezende et al. (2014) Rezende, D. J., Mohamed, S., and Wierstra, D. Stochastic backpropagation and approximate inference in deep generative models. arXiv preprint arXiv:1401.4082, 2014.
- Ritschl et al. (2011) Ritschl, L., Bergner, F., Fleischmann, C., and Kachelrieß, M. Improved total variation-based CT image reconstruction applied to clinical data. Physics in Medicine & Biology, 56(6):1545, 2011.
- Sohn et al. (2015) Sohn, K., Lee, H., and Yan, X. Learning structured output representation using deep conditional generative models. In Advances in Neural Information Processing Systems, pp. 3483–3491, 2015.
- Starck et al. (2003) Starck, J.-L., Nguyen, M. K., and Murtagh, F. Wavelets and curvelets for image deconvolution: a combined approach. Signal processing, 83(10):2279–2283, 2003.
- Sun et al. (2013) Sun, B., Edgar, M. P., Bowman, R., Vittert, L. E., Welsh, S., Bowman, A., and Padgett, M. 3D computational imaging with single-pixel detectors. Science, 340(6134):844–847, 2013.
- Turtaev et al. (2018) Turtaev, S., Leite, I. T., Altwegg-Boussac, T., Pakan, J. M., Rochefort, N. L., and Čižmár, T. High-fidelity multimode fibre-based endoscopy for deep-brain in vivo imaging. arXiv preprint arXiv:1806.01654, 2018.
- Velten et al. (2012) Velten, A., Willwacher, T., Gupta, O., Veeraraghavan, A., Bawendi, M. G., and Raskar, R. Recovering three-dimensional shape around a corner using ultrafast time-of-flight imaging. Nature communications, 3:745, 2012.
- Vogel (2002) Vogel, C. R. Computational methods for inverse problems, volume 23. SIAM, 2002.
- Xu et al. (2014) Xu, L., Ren, J. S., Liu, C., and Jia, J. Deep convolutional neural network for image deconvolution. In Advances in Neural Information Processing Systems, pp. 1790–1798, 2014.
- Yan et al. (2016) Yan, X., Yang, J., Sohn, K., and Lee, H. Attribute2image: Conditional image generation from visual attributes. In European Conference on Computer Vision, pp. 776–791. Springer, 2016.
- Yang & Zhang (2011) Yang, J. and Zhang, Y. Alternating direction algorithms for -problems in compressive sensing. SIAM journal on scientific computing, 33(1):250–278, 2011.