- Direct Superresolution Technique for Realistic Image Reconstruction
- Computational Depth-Based Deconvolution Technique for Full-Focus
The goal of computational imaging is to extract additional information through advanced image processing algorithms. We have developed computational techniques to improve the image quality over a large depth of field and improve the image resolution over a large field of view. We have also working on superresolution techniques, including structured illumination microscopy and Fourier ptychography.
Traditional superresolution (SR) techniques employ optimizers, priors, and regularizers to deliver stable, appealing restorations even though deviating from the real, ground-truth scene. A non-regularized direct SR technique, illustrated in Fig. 1, has been developed to uniquely solve a set of linear equations representing the multi-shift image reconstruction problem with sufficient measurements to deliver realistic reconstructions without any inherent bias imposed by improper assumptions in the inverse problem.
An adaptive frequency-based filtering scheme (AFFS), shown in Fig. 2, is introduced to gain robustness against poor focus, misestimated shift, boundary variations, and noisy scenarios while maintaining the merit of direct SR to achieve optimal restorations with minimal artifacts.
Simulation results, presented in Fig. 3, demonstrate that more fine features can be resolved with the developed technique as compared with other SR techniques such as bicubic SR and multi-channel blind deconvolution (MCBD) SR. Simulations also show that the retrieved high-resolution features are successfully transferred by the AFFS to produce an artifact-suppressed high-resolution image under realistic imaging conditions such as noisy environment, blurring capturing, boundary tolerance, and shift misestimate.
- B. Salahieh, J. J. Rodriguez, and R. Liang, “Direct superresolution for realistic image reconstruction,” Optics Express (OSA, 2015).
- B. Salahieh, J. J. Rodriguez, R. Liang, “Direct Superresolution Technique for Solving a Miniature Multi-Shift Imaging System,” in Imaging Systems and Applications, JW3A. 5 (OSA, 2015).
Ordinary fixed-focus cameras, such as low-end hand-held cameras, smartphone cameras, and surveillance cameras, have a depth-variant point spread function (PSF) that can be visualized as double cones when focusing at finite distances or as a single cone when focusing at infinity with apex at the on-axis in-focus point located at the hyperfocal distance. Objects become progressively more defocused as their distance to the lens gets smaller than half of the hyperfocal distance. The prior knowledge of the volumetric PSF can be utilized effectively to deblur the captured image at different depth planes.
A computational depth-based deconvolution technique, presented in Fig. 1, has been developed to equip ordinary cameras with an extended depth of field capability. This is achieved by deconvolving the captured scene with pre-calibrated depth-variant point spread function profiles to bring it into focus at different depth planes. Afterward, the focused features from the depth planes are stitched together to form a full-focus image.
To imitate multi-depth (MD) objects, different regions of planar objects traced at various depth planes are stitched together according to a customized depth map, presented in Fig. 2.
The ringing artifacts associated with the deconvolution process are addressed on three levels. First, the boundary artifacts are eliminated by adopting a block-tiling approach. Second, the sharp edges’ ringing artifacts are adaptively controlled by reference maps working as local regularizers through an iterative deconvolution process. Finally, artifacts initiated by different depth surfaces are suppressed by a block-wise deconvolution or depth-based masking approach. The developed algorithm is demonstrated for planar objects and multi-depth objects scenarios.
Note that the depth-transition artifacts can dominate the MD deconvolution results if no special processing is involved (see second column of Fig. 3). The depth-based masking approach can help a lot in reducing such artifacts (see last column in Fig. 3), although it is not fully restoring some of the image details (e.g., compare the vertical bars on the right of restored USAF images between the depth-based masking case and reference MD case).
- B. Salahieh, J. J. Rodriguez, S. Stetson, and R. Liang, “Single-image extended-focus technique using depth-based deconvolution”, submitted for publications (OSA, 2015).
- B. Salahieh, Jeffrey J. Rodriguez, and Rongguang Liang, “Computational Depth-Variant Deconvolution Technique for Full-Focus Imaging, ” Computational Optical Sensing and Imaging, (OSA, 2015).