Demosaicing and Denoising on Simulated Light Field Images

Similar documents
Li, Y., Olsson, R., Sjöström, M. (2018) An analysis of demosaicing for plenoptic capture based on ray optics In: Proceedings of 3DTV Conference 2018

Demosaicing Algorithms

Dictionary Learning based Color Demosaicing for Plenoptic Cameras

Visibility of Uncorrelated Image Noise

Lecture 18: Light field cameras. (plenoptic cameras) Visual Computing Systems CMU , Fall 2013

Simulated Programmable Apertures with Lytro

Modeling the calibration pipeline of the Lytro camera for high quality light-field image reconstruction

A simulation tool for evaluating digital camera image quality

Learning the image processing pipeline

Computational Approaches to Cameras

Introduction. Related Work

Local Linear Approximation for Camera Image Processing Pipelines

Simultaneous Capturing of RGB and Additional Band Images Using Hybrid Color Filter Array

Computational Cameras. Rahul Raguram COMP

Wavefront coding. Refocusing & Light Fields. Wavefront coding. Final projects. Is depth of field a blur? Frédo Durand Bill Freeman MIT - EECS

Lenses, exposure, and (de)focus

Light-Field Database Creation and Depth Estimation

Image Processing by Bilateral Filtering Method

Sensors and Sensing Cameras and Camera Calibration

Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring

Light field sensing. Marc Levoy. Computer Science Department Stanford University

Deconvolution , , Computational Photography Fall 2018, Lecture 12

IMAGE RESTORATION WITH NEURAL NETWORKS. Orazio Gallo Work with Hang Zhao, Iuri Frosio, Jan Kautz

262 JOURNAL OF DISPLAY TECHNOLOGY, VOL. 4, NO. 2, JUNE 2008

Capturing Light. The Light Field. Grayscale Snapshot 12/1/16. P(q, f)

High dynamic range imaging and tonemapping

Color image Demosaicing. CS 663, Ajit Rajwade

A Study of Slanted-Edge MTF Stability and Repeatability

Acquisition Basics. How can we measure material properties? Goal of this Section. Special Purpose Tools. General Purpose Tools

Applications of Flash and No-Flash Image Pairs in Mobile Phone Photography

Image acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

When Does Computational Imaging Improve Performance?

Admin. Lightfields. Overview. Overview 5/13/2008. Idea. Projects due by the end of today. Lecture 13. Lightfield representation of a scene

Lecture 22: Cameras & Lenses III. Computer Graphics and Imaging UC Berkeley CS184/284A, Spring 2017

Dappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing

Gradient-Based Correction of Chromatic Aberration in the Joint Acquisition of Color and Near-Infrared Images

Coded photography , , Computational Photography Fall 2018, Lecture 14

DEPTH FUSED FROM INTENSITY RANGE AND BLUR ESTIMATION FOR LIGHT-FIELD CAMERAS. Yatong Xu, Xin Jin and Qionghai Dai

Adding Realistic Camera Effects to the Computer Graphics Camera Model

ICC Votable Proposal Submission Colorimetric Intent Image State Tag Proposal

Coded Aperture and Coded Exposure Photography

A Framework for Analysis of Computational Imaging Systems

To Do. Advanced Computer Graphics. Outline. Computational Imaging. How do we see the world? Pinhole camera

Unit 1: Image Formation

LENSLESS IMAGING BY COMPRESSIVE SENSING

Design of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems

Why learn about photography in this course?

A Novel Method for Enhancing Satellite & Land Survey Images Using Color Filter Array Interpolation Technique (CFA)

Lytro camera technology: theory, algorithms, performance analysis

lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response

Figure 1 HDR image fusion example

CAMERA BASICS. Stops of light

Optical Performance of Nikon F-Mount Lenses. Landon Carter May 11, Measurement and Instrumentation

What will be on the midterm?

Digital photography , , Computational Photography Fall 2017, Lecture 2

Tomorrow s Digital Photography

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

Color Constancy Using Standard Deviation of Color Channels

Coded photography , , Computational Photography Fall 2017, Lecture 18

Image Formation and Capture

La photographie numérique. Frank NIELSEN Lundi 7 Juin 2010

Smart Interpolation by Anisotropic Diffusion

Deconvolution , , Computational Photography Fall 2017, Lecture 17

Tonemapping and bilateral filtering

Topic 6 - Optics Depth of Field and Circle Of Confusion

DIGITAL CAMERA SENSORS

Privacy Preserving Optics for Miniature Vision Sensors

Disparity Estimation and Image Fusion with Dual Camera Phone Imagery

Introduction to Light Fields

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

VC 11/12 T2 Image Formation

This experiment is under development and thus we appreciate any and all comments as we design an interesting and achievable set of goals.

Lecture 30: Image Sensors (Cont) Computer Graphics and Imaging UC Berkeley CS184/284A

Cameras. CSE 455, Winter 2010 January 25, 2010

The ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do?

Coded Computational Photography!

SURVEILLANCE SYSTEMS WITH AUTOMATIC RESTORATION OF LINEAR MOTION AND OUT-OF-FOCUS BLURRED IMAGES. Received August 2008; accepted October 2008

RGB RESOLUTION CONSIDERATIONS IN A NEW CMOS SENSOR FOR CINE MOTION IMAGING

High Performance Imaging Using Large Camera Arrays

Reikan FoCal Aperture Sharpness Test Report

Basic principles of photography. David Capel 346B IST

Understanding camera trade-offs through a Bayesian analysis of light field projections - A revision Anat Levin, William Freeman, and Fredo Durand

Computer Vision. The Pinhole Camera Model

Improved sensitivity high-definition interline CCD using the KODAK TRUESENSE Color Filter Pattern

Applications of Optics

Double resolution from a set of aliased images

Digital photography , , Computational Photography Fall 2018, Lecture 2

University Of Lübeck ISNM Presented by: Omar A. Hanoun

Color Digital Imaging: Cameras, Scanners and Monitors

Rectifying the Planet USING SPACE TO HELP LIFE ON EARTH

Reikan FoCal Aperture Sharpness Test Report

Generalized Assorted Camera Arrays: Robust Cross-channel Registration and Applications Jason Holloway, Kaushik Mitra, Sanjeev Koppal, Ashok

Image Formation III Chapter 1 (Forsyth&Ponce) Cameras Lenses & Sensors

SYSTEMATIC NOISE CHARACTERIZATION OF A CCD CAMERA: APPLICATION TO A MULTISPECTRAL IMAGING SYSTEM

PHYSICS. Chapter 35 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT

Mirrors, Lenses &Imaging Systems

Wavelengths and Colors. Ankit Mohan MAS.131/531 Fall 2009

Joint Demosaicing and Super-Resolution Imaging from a Set of Unregistered Aliased Images

Image Denoising using Dark Frames

Transcription:

Demosaicing and Denoising on Simulated Light Field Images Trisha Lian Stanford University tlian@stanford.edu Kyle Chiang Stanford University kchiang@stanford.edu Abstract Light field cameras use an array of microlens to capture the 4D radiance of as scene. Standard image processing techniques with light field data do not utilize all four dimensions to demosaic or denoise captured images. In this paper, we formulate demosaicing as an optimization problem and enforce a TV-prior on different dimensions of the light field. We apply our method on simulated light field data created from 3D virtual scenes. Because our data is simulated, we can use ground truth images to evaluate the effectiveness of our method. For certain combinations of dimensions, we achieve better overall PSNR values than the standard demosaicing technique described in Malvar et al. [1]. Despite the improvement in PSNR, we introduce more color artifacts in areas of high frequency in the image. Our method also improves PNSR values for scenes with low illumination levels. 1. Introduction 1.1. Background Unlike standard cameras, light field cameras ( plenoptic cameras) uniquely capture the 4D radiance information of a scene instead of just a 2D intensity image. This is achieved by inserting a microlens array between the camera s main lens and sensor. Each microlens separates incoming rays and allows the sensor to capture both the intensity of a ray as well as the angle from which it arrived (see Figure 1). Each ray can be characterized by its intersection with the microlens plane (s, t) and the main lens (u, v). These four coordinates make up the four dimensions of the light field data: L(u, v, s, t). The 4D data can be post-processed to dynamically change the depth of field and focal plane of the image after it has been acquired. In this paper, we utilize all four dimensions to help improve the demosaicing and denoising steps in the image processing pipeline. Figure 1: A schematic of a light field camera. Each ray can be uniquely characterized by its intersection with the main lens, (u, v) coordinates, and the microlens array, (s, t) coordinates. 1.2. Motivation Standard demosaicing techniques demosaic the Bayer pattern output directly from the camera sensor. For a typical camera, this is the optimal strategy. However, for a light field camera, the microlens array encodes additional information in the sensor image. Demosaicing using traditional techniques ignores this additional information. The objective of our new optimization technique for demosaicing is to try to capture and use all four dimensions when generating the full-color light field. Not much work has been done in utilizing this extra information in light field data. Some researchers [2] have proposed projecting samples of the microlens to the refocus plane before demosaicing. To avoid the random RGB sampling that results from this, the authors resample the radiance according to the parameters of the focal plane in order to achieve even samples for demosaicing. With this method, the authors claim to visually achieve reduced demosaicing artifacts and more detail. Other demosaicing methods use disparity [3] or machine learning [4] to improve color quality. Our method approaches this problem using optimization techniques and uses simulation data to quantify its effectiveness.

Figure 2: A diagram of our camera simulation pipeline. For the light field simulation, we model lenses in PBRT to match a light field camera. The rays are therefore traced through both a main lens and a microlens array. 2. Light Field Simulation In order to test our method against a ground truth image, we use a light field camera simulation currently being developed by one of the authors. This simulation steps through the entire camera pipeline to generate realistic data: from a 3D virtual scene, through the optics of a light field camera, and onto a sensor. To generate the ground truth image, we sample the image with a simulated sensor that has RGB filters at every pixel and no noise parameters. 2.1. Simulation Pipeline Figure 2 summarizes the main steps of the simulation. The simulation starts with a virtual scene created in a 3D modeling program such as Blender or Maya. This scene includes the geometry and material properties of the objects as well as the positions and shapes of lights. Next, a modified version of PBRT [5] is used to trace rays from the sensor, through light field optics (microlens and main lens), and into the scene. PBRT has been modified in this simulation to apply full-spectral rendering. During the raytracing step, the user specifies simulation parameters such as lens types, spectral properties of the light sources, film distance, aperture size, and field of view. The simulation also accounts for realistic lens properties such as chromatic aberration and potential diffraction limited systems. Once all these parameters are specified, the resulting optical image is passed on to ISET (Image Systems Engineering Toolbox) [6]. ISET takes the incoming light information and captures it with a realistic sensor. The user can specify the sensor parameters, such as the Bayer pattern, pixel size, sensor size, exposure time, noise parameters, and illumination levels. The sensor data we obtain from the end of this pipeline is our raw data. 2.2. Simulation Parameters For the data obtained in this paper, we simulated a camera with a 50 mm double gaussian main lens with an aperture setting of f/2.8. The camera had a 500 x 500 microlens array in front of the sensor. The location and size of the array was automatically calculated to cover as many sensor pixels as possible without overlap [7], and therefore has an f-number that matches the main lens. Each microlens covers a 9x9 block of sensor pixels, so we capture 81 different angular views of our scene. The sensor size was 6.7 mm x 6.7 mm with a pixel size of 1.7 x 1.7 um. The resolution of the raw sensor image was 4500 x 4500 pixels and the resolution of the final image was equal to the number of microlenses (500 x 500). The exposure time was set to 1/90 s. Our Bayer pattern had an grbg configuration. See Figure 3 for the transmittance of the three color filters on our simulated sensor. ISET also included shot and electronic noise in the simulated sensor. Figure 3: a) Transmittance plots of the color filters on our simulated sensor. b) Bayer pattern used to obtain our raw data. We render two different scenes. Both are lit with D65 illuminant area lights. One scene contains a chair and a house plant, while the other contains three planar resolution charts at varying distances. The objects in the scene are around 0.5 to 1.5 m away from the camera. 3. Methods 3.1. Baseline - Malvar et al. As a baseline comparison to determine the effectiveness of our new method, we implemented a standard demosaicing algorithm on the raw sensor image. The method we decided to use as a baseline is described in Malvar et al [1]. Because the method performs demosaicing using a linear transformation, it can be implemented using 2D convolutions and computed very efficiently. Furthermore, this method produces very few color artifacts for a typical image. These artifacts only show up in areas of high frequency. 3.2. Optimization Problem For our optimization problem, we wanted to find the most likely 4D light field image that would produce the Bayer filtered image captured by the camera. However,

due to the loss of information when sampling the scene, there are an infinite number of images that could produce the same Bayer filtered image. To choose the most likely image, we note that real world images tend to have sparse gradients and assume an anisotropic TV prior on the 4D image. The optimization problem can then be formulated as follows: 1 min kax bk22 + λkdxk1 x 2 where A is the sampling matrix that generates a Bayer filtered image from the scene, b is the Bayer filtered image captured by the camera, D is the gradient function, and λ is a parameter chosen to weight the TV prior. This approach was inspired by the techniques described in Heide et al. [8]. 3.3. Choice of Gradients For a 2D image, the TV prior would be the sum of the gradient in the X direction and the gradient in the Y direction. However, for our 4D light field data, the TV prior has some ambiguity. There are 2 assumptions made for sparse gradients. The first being that each image captured from slightly different angles should be nearly identical. This would be enforced in the TV prior by setting the gradient function D to the gradient in s and t. The second assumption is that corresponding pixels in images seen through each microlenses should also very similar. To enforce this in the TV prior, D would be set to the gradient in the u and v directions. We chose to investigate these two assumptions both separately and together by looking at 3 cases: sparse gradients in u and v only, sparse gradients in s and t only, and sparse gradients in u, v, s and t. 3.4. ADMM We solved this optimization problem using an iterative ADMM method. To implement this method, we first reformulated the optimization problem in the form: Using the ADMM strategy we then form the augmented Lagrangian ρ 1 kax bk22 +λkzk1 +y T (Dx z)+ kdx zk22 2 2 The iterative ADMM updates can then be derived as follows: 1 AT b + ρdt (z u) v κ v > κ z 0 v κ for v = Dx + u and κ = λ/ρ v + κ v < κ These update rules are repeated until convergence or until the maximum number of iterations is reached. subject to Dx z = 0 x AT A + ρdt D Figure 4: An enlarged portion of the image captured by the sensor. By restructuring the sensor data, we can display the image in this tiled form. Each image corresponds to a single (u,v) index. u u + Dx z 1 min kax bk22 + λkzk1 x 2 Lρ (x, y, z) = 3.5. Image Processing Pipeline It is important to note that we only carry our simulation past the demosaicing portion of the image processing pipeline. We do not perform any gamut mapping, white balancing, or illuminant correction. We chose to do this in order to target the effectiveness of demosaicing and denoising with our method and to not confound our results with processes further down the pipeline. As a result of this purposefully incomplete processing, our images look tinted compared to the original scene.

Figure 4 shows a visualization of the 4D ground truth light field. As described earlier, we produce these ground truth images by capturing the rendered optical image with a full array sensor in ISET. This sensor has color filters for every pixel and has its noise parameters turned off. This image will serve as the reference for all PSNR calculations. 3.6. Gradients in Ground Truth In Figure 5, we calculate and plot the gradients of one of our ground truth image in each of the different light field dimensions. The gradients in these images are mostly dark, which indicates that the gradients are indeed sparse and that the TV assumptions should improve the resulting image.the gradients are more sparse in (u, v), than in (s, t). This is particularly true for the in-focus plane in the center of the image. We would therefore expect our method to perform the best when we assume sparse gradients in the (u, v) dimension. 4. Results (c) For all results, we demosaic our raw data using 1) Malvar et al. s method and 2) our optimization method. For our method, we try three different TV-priors as described above: a) gradients over (u, v), b) gradients over (s, t), and c) gradients over (u, v, s, t). 4.1. Average Illumination For the following results, we set the mean illuminance level to be 12 lux. The maximum illuminance for each image is roughly 70 lux, which is equivalent to standard room lighting. The images shown are taken from the center subaperture (u = 0, v = 0). In other words, it is equivalent to the center tile when you display the data as shown in Figure 4b. By shifting and adding these different tiled images, the user obtains different depth of fields. We calculate PSNR values for both the center sub-aperture image and the mean image (average over all (u, v)). Figure 6 and Figure 7 shows the demosaiced images of our two scenes. The differences (averaged across the color channels) between each image and the ground truth are shown as well. We can see that most errors are centered around the high frequency components of the image, and these errors are higher for our method compared to Malvar et al. Although these errors are difficult to see in the full image, enlarging high frequency sections of the image (Figure 8 and Figure 9) reveals color artifacts for our optimization method. Despite introducing color artifacts, our (u, v) and (u, v, s, t) methods result in higher overall PSNR values than Malvar et al (see Tables 1 and 2). Figure 5: Gradients taken in each of the four light field dimensions. Ground Truth u (c) v s t Center Mean Image Malvar 35.40 db 34.65 db (u,v) 37.26 db 37.56 db (s,t) 34.69 db 32.97 db (u,v,s,t) 36.55 db 35.38 db Table 1: PSNR values for the Chair scene. Center Mean Image Malvar 28.87 db 28.05 db (u,v) 28.07 db 29.01 db (s,t) 28.38 db 26.18 db (u,v,s,t) 28.88 db 28.00 db Table 2: PSNR values for the Resolution Charts scene. 4.2. Changing Illumination Levels Because we assume sparse gradients in the image, our method should perform better under noisier conditions.to

(c) (c) Figure 6: Demosaiced scene of chair [left] along with a visualization of error relative to ground truth [right] Ground truth image. Malvar et al. (c) (u, v). (s, t). (u, v, s, t). Figure 7: Demosaiced image of resolution charts [left] along with a visualization of error relative to ground truth [right] Ground truth image. Malvar et al. (c) (u, v). (s, t). (u, v, s, t).

(c) Figure 8: A enlarged section of arm of the chair. Ground truth image. Malvar et al. (c) (u,v). (s,t). (u,v,s,t). Figure 10: PSNR values for different illumination levels. (c) (c) Figure 9: A enlarged section of resolution chart. Ground truth image. Malvar et al. (c) (u,v). (s,t). (u,v,s,t). test this, we rendered our raw data under different sensor illumination levels in ISET. Lower illumination results in noisier images. Figure 10 and Figure 11 show our results. From the plot, we can see that our (u, v) and (u, v, s, t) methods perform better than Malvar et al. for very low illumination. This is because this baseline technique performs no denoising, while our assumption of sparse gradients automatically smooths out noise. Linear demosaicing (such as Malvar et al. s method) is greatly affected by noise, which is why many image processing pipeline perform denoising before demosaicing. As illumination levels increase, our (u, v) technique con- Figure 11: A comparison of how each technique performs on a noisy image (mean illumination = 1 lux). Ground truth image. Malvar et al. (c) (u, v). (s, t). (u, v, s, t). tinues to outperform Malvar et al. s method in terms of overall image PSNR. (s, t) performs poorly regardless of the illumination. The assumption of sparse gradients in this dimension may not be very strong, which is supported by the number of gradients seen in Figure 5.

5. Conclusion In conclusion, our demosaicing method, which solves an optimization problem, results in an image with better PSNR values than the traditional method when we assume sparse gradients in the (u, v) or (u, v, s, t) dimensions. However, for images with good lighting, we ended up with more color artifacts than demosaicing with traditional methods in areas of high frequencies. We suspect that this is due to the fact that while the TV prior helps create a truer overall image, it is specifically avoiding high frequency signals in the image, resulting in color artifacts in these regions. However, for images with poor illumination or lots of noise, the advantages of running optimization with a TV prior shine, with the best results assuming sparse gradients across u and v. [7] Ng, Ren, et al. Light field photography with a hand-held plenoptic camera. Computer Science Technical Report CSTR 2.11 (2005): 1-11. [8] Heide, Felix, et al. FlexISP: a flexible camera image processing framework. ACM Transactions on Graphics (TOG) 33.6 (2014): 231. 5.1. Future Work While the solution we investigated may not be the optimal demosaicing for a light field camera in all cases, there are several other possible directions to pursue to try to harness the information of the 4D light field to obtain the best demosaiced image. One possible improvement is to look at a cross channel prior that also penalizes the difference in gradients between the color channels. In most images sharp edges result in gradients in all 3 color channels, so enforcing this assumption could result in fewer color artifacts. Another possible route for investigation is to use the true image to train an optimal linear transform similar to the one presented in Malvar et al, extended to 4 dimensions. References [1] Malvar, Henrique S., Li-wei He, and Ross Cutler. Highquality linear interpolation for demosaicing of Bayerpatterned color images. Acoustics, Speech, and Signal Processing, 2004. Proceedings.(ICASSP 04). IEEE International Conference on. Vol. 3. IEEE, 2004. [2] Yu, Zhan, et al. An analysis of color demosaicing in plenoptic cameras. Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on. IEEE, 2012. [3] Seifi, Mozhdeh, et al. Disparity guided demosaicking of light field images. Image Processing (ICIP), 2014 IEEE International Conference on. IEEE, 2014. [4] Huang, Xiang, and Oliver Cossairt. Dictionary learning based color demosaicing for plenoptic cameras. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 2014. [5] Pharr, Matt, and Greg Humphreys. Physically based rendering: From theory to implementation. Morgan Kaufmann, 2004. [6] Farrell, Joyce, et al. A display simulation toolbox for image quality evaluation. Journal of Display Technology 4.2 (2008): 262-270