Computational Approaches to Cameras

Similar documents

Computational Cameras. Rahul Raguram COMP

Coded photography , , Computational Photography Fall 2018, Lecture 14

Coded photography , , Computational Photography Fall 2017, Lecture 18

Dappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing

Wavefront coding. Refocusing & Light Fields. Wavefront coding. Final projects. Is depth of field a blur? Frédo Durand Bill Freeman MIT - EECS

Capturing Light. The Light Field. Grayscale Snapshot 12/1/16. P(q, f)

Deconvolution , , Computational Photography Fall 2018, Lecture 12

Deconvolution , , Computational Photography Fall 2017, Lecture 17

Lenses, exposure, and (de)focus

Lecture 18: Light field cameras. (plenoptic cameras) Visual Computing Systems CMU , Fall 2013

Coded Computational Photography!

The ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do?

Introduction to Light Fields

To Do. Advanced Computer Graphics. Outline. Computational Imaging. How do we see the world? Pinhole camera

Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring

Admin Deblurring & Deconvolution Different types of blur

Coding and Modulation in Cameras

Admin. Lightfields. Overview. Overview 5/13/2008. Idea. Projects due by the end of today. Lecture 13. Lightfield representation of a scene

Recent Advances in Image Deblurring. Seungyong Lee (Collaboration w/ Sunghyun Cho)

Image Enhancement Using Calibrated Lens Simulations

Light field sensing. Marc Levoy. Computer Science Department Stanford University

Understanding camera trade-offs through a Bayesian analysis of light field projections Anat Levin, William T. Freeman, and Fredo Durand

Deblurring. Basics, Problem definition and variants

Building a Real Camera. Slides Credit: Svetlana Lazebnik

Coded Aperture for Projector and Camera for Robust 3D measurement

Understanding camera trade-offs through a Bayesian analysis of light field projections - A revision Anat Levin, William Freeman, and Fredo Durand

Coded Aperture and Coded Exposure Photography

Building a Real Camera

Light field photography and microscopy

Transfer Efficiency and Depth Invariance in Computational Cameras

High dynamic range imaging and tonemapping

Computational Photography and Video. Prof. Marc Pollefeys

Computational Photography: Principles and Practice

Modeling the calibration pipeline of the Lytro camera for high quality light-field image reconstruction

6.098 Digital and Computational Photography Advanced Computational Photography. Bill Freeman Frédo Durand MIT - EECS

Announcement A total of 5 (five) late days are allowed for projects. Office hours

Image Formation III Chapter 1 (Forsyth&Ponce) Cameras Lenses & Sensors

Image Formation and Camera Design

Basic principles of photography. David Capel 346B IST

Cameras. Outline. Pinhole camera. Camera trial #1. Pinhole camera Film camera Digital camera Video camera

Near-Invariant Blur for Depth and 2D Motion via Time-Varying Light Field Analysis

Light-Field Database Creation and Depth Estimation

Sensors and Sensing Cameras and Camera Calibration

Cameras. Shrinking the aperture. Camera trial #1. Pinhole camera. Digital Visual Effects Yung-Yu Chuang. Put a piece of film in front of an object.

Cameras. Digital Visual Effects, Spring 2008 Yung-Yu Chuang 2008/2/26. with slides by Fredo Durand, Brian Curless, Steve Seitz and Alexei Efros

Restoration of Motion Blurred Document Images

Toward Non-stationary Blind Image Deblurring: Models and Techniques

Computational Camera & Photography: Coded Imaging

What will be on the midterm?

Dynamically Reparameterized Light Fields & Fourier Slice Photography. Oliver Barth, 2009 Max Planck Institute Saarbrücken

Aberrations, Camera, Eye

Recent advances in deblurring and image stabilization. Michal Šorel Academy of Sciences of the Czech Republic

Full Resolution Lightfield Rendering

Improved motion invariant imaging with time varying shutter functions

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Chapter 25 Optical Instruments

VC 11/12 T2 Image Formation

Single-shot three-dimensional imaging of dilute atomic clouds

Dr F. Cuzzolin 1. September 29, 2015

Demosaicing and Denoising on Simulated Light Field Images

Gradient-Based Correction of Chromatic Aberration in the Joint Acquisition of Color and Near-Infrared Images

Cameras As Computing Systems

Tonemapping and bilateral filtering

Cameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017

Optical image stabilization (IS)

A Poorly Focused Talk

VC 14/15 TP2 Image Formation

fast blur removal for wearable QR code scanners

University Of Lübeck ISNM Presented by: Omar A. Hanoun

Applications of Optics

ELEC Dr Reji Mathew Electrical Engineering UNSW

multiframe visual-inertial blur estimation and removal for unmodified smartphones

LENSES. INEL 6088 Computer Vision

Unit 1: Image Formation

Wavelengths and Colors. Ankit Mohan MAS.131/531 Fall 2009

6.A44 Computational Photography

Image Deblurring with Blurred/Noisy Image Pairs

SUPER RESOLUTION INTRODUCTION

Point Spread Function Engineering for Scene Recovery. Changyin Zhou

Reflectors vs. Refractors

Image acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor

Computational Photography Introduction

ME 6406 MACHINE VISION. Georgia Institute of Technology

DEPTH FUSED FROM INTENSITY RANGE AND BLUR ESTIMATION FOR LIGHT-FIELD CAMERAS. Yatong Xu, Xin Jin and Qionghai Dai

Image Formation. Light from distant things. Geometrical optics. Pinhole camera. Chapter 36

CPSC 4040/6040 Computer Graphics Images. Joshua Levine

Computational Photography

Chapter 18 Optical Elements

COURSE NAME: PHOTOGRAPHY AND AUDIO VISUAL PRODUCTION (VOCATIONAL) FOR UNDER GRADUATE (FIRST YEAR)

Why learn about photography in this course?

Supplementary Information

The Camera : Computational Photography Alexei Efros, CMU, Fall 2005

Modeling and Synthesis of Aperture Effects in Cameras

Cameras. CSE 455, Winter 2010 January 25, 2010

Time-Lapse Light Field Photography With a 7 DoF Arm

lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response

Computer Vision. The Pinhole Camera Model

Cameras and Sensors. Today. Today. It receives light from all directions. BIL721: Computational Photography! Spring 2015, Lecture 2!

TRUESENSE SPARSE COLOR FILTER PATTERN OVERVIEW SEPTEMBER 30, 2013 APPLICATION NOTE REVISION 1.0

Transcription:

Computational Approaches to Cameras 11/16/17 Magritte, The False Mirror (1935) Computational Photography Derek Hoiem, University of Illinois

Announcements Final project proposal due Monday (see links on webpage)

Conventional cameras Conventional cameras are designed to capture light in a medium that is directly viewable Scene Eye Sensor Lens Light Display

Computational cameras With a computational approach, we can capture light and then figure out what to do with it Eyes Displays Computation Captured Light Light Scene

Questions for today How can we represent all of the information contained in light? What are the fundamental limitations of cameras? What sacrifices have we made in conventional cameras? For what benefits? How else can we design cameras for better focus, deblurring, multiple views, depth, etc.?

Slides from Efros Representing Light: The Plenoptic Function Figure by Leonard McMillan Q: What is the set of all things that we can ever see? A: The Plenoptic Function (Adelson & Bergen) Let s start with a stationary person and try to parameterize everything that he can see

Grayscale snapshot is intensity of light Seen from a single view point At a single time P(q,f) Averaged over the wavelengths of the visible spectrum (can also do P(x,y), but spherical coordinate are nicer)

Color snapshot is intensity of light P(q,f,l) Seen from a single view point At a single time As a function of wavelength

A movie is intensity of light P(q,f,l,t) Seen from a single view point Over time As a function of wavelength

Holographic movie is intensity of light Seen from ANY viewpoint Over time As a function of wavelength P(q,f,l,t,V X,V Y,V Z )

The Plenoptic Function P(q,f,l,t,V X,V Y,V Z ) Can reconstruct every possible view, at every moment, from every position, at every wavelength Contains every photograph, every movie, everything that anyone has ever seen!

Representing light The atomic element of light: a pixel a ray

Fundamental limitations and trade-offs Only so much light in a given area to capture Basic sensor accumulates light at a set of positions from all orientations, over all time We want intensity of light at a given time at one position for a set of orientations Solutions: funnel, constrain, redirect light change the sensor CCD inside camera

Trade-offs of conventional camera Add a pinhole Pixels correspond to small range of orientations at the camera center, instead of all gathered light at one position Much less light hits sensor Add a lens More light hits sensor Limited depth of field Chromatic aberration Add a shutter Capture average intensity at a particular range of times Increase sensor resolution Each pixel represents a smaller range of orientations Less light per pixel Controls: aperture size, focal length, shutter time

How else can we design cameras? What do they sacrifice/gain?

1. Light Field Photography with Plenoptic Camera Conventional Camera Plenoptic Camera Adelson and Wang 1992 Ng et al. Stanford TR, 2005

Light field photography Like replacing the human retina with an insect compound eye Records where light ray hits the lens

Stanford Plenoptic Camera [Ng et al 2005] Contax medium format camera Kodak 16-megapixel sensor Adaptive Optics microlens array 125μ square-sided microlenses 4000 4000 pixels 292 292 lenses = 14 14 pixels per lens

Light field photography: applications

Light field photography: applications Change in viewpoint

Light field photography: applications Change in viewpoint Lateral Along Optical Axis

Digital Refocusing

Light field photography w/ microlenses We gain Ability to refocus or increase depth of field Ability for small viewpoint shifts What do we lose (vs. conventional camera)?

2. Coded apertures

Image and Depth from a Conventional Camera with a Coded Aperture Anat Levin, Rob Fergus, Frédo Durand, William Freeman MIT CSAIL Slides from SIGGRAPH Presentation

Single input image: Output #1: Depth map

Output #1: Depth map Single input image: Output #2: All-focused image

Lens and defocus Lens aperture Image of a point light source Lens Camera sensor Point spread function Focal plane

Lens and defocus Lens aperture Image of a defocused point light source Object Lens Camera sensor Point spread function Focal plane

Lens and defocus Lens aperture Image of a defocused point light source Object Lens Camera sensor Point spread function Focal plane

Lens and defocus Lens aperture Image of a defocused point light source Object Lens Camera sensor Point spread function Focal plane

Lens and defocus Lens aperture Image of a defocused point light source Object Lens Camera sensor Point spread function Focal plane

Depth and defocus Out of focus Depth from defocus: Infer depth by analyzing local scale of defocus blur In focus

Challenges Hard to discriminate a smooth scene from defocus blur? Out of focus Hard to undo defocus blur Input Ringing with conventional deblurring algorithm

Key ideas Exploit prior on natural images - Improve deconvolution - Improve depth discrimination Natural Unnatural Coded aperture (mask inside lens) - make defocus patterns different from natural images and easier to discriminate

Defocus as local convolution Input defocused image y Local sub-window f k Calibrated blur kernels at depth k x Sharp sub-window Depth k=1: y f k x Depth k=2: y f k x Depth k=3: y f k x

Overview Try deconvolving local input windows with different scaled filters:? Larger scale? Correct scale? Smaller scale Somehow: select best scale.

Challenges Hard to deconvolve even when kernel is known Input Ringing with the traditional Richardson-Lucy deconvolution algorithm Hard to identify correct scale:? Larger scale? Correct scale? Smaller scale

Deconvolution is ill posed f x y? =

Deconvolution is ill posed f x y Solution 1:? = Solution 2:? =

Idea 1: Natural images prior What makes images special? Natural Unnatural Image gradient Natural images have sparse gradients put a penalty on gradients

Deconvolution with prior x arg min f x y 2 l i ( x ) i Convolution error Derivatives prior _ 2 +? Equal convolution error Low? _ 2 + High

Recall: Overview Try deconvolving local input windows with different scaled filters: Larger scale? Correct scale? Smaller scale? Somehow: select best scale. Challenge: smaller scale not so different than correct

Idea 2: Coded Aperture Mask (code) in aperture plane - make defocus patterns different from natural images and easier to discriminate Conventional aperture Our coded aperture

Solution: lens with occluder Object Lens Camera sensor Point spread function Focal plane

Solution: lens with occluder Aperture pattern Image of a defocused point light source Object Lens with coded aperture Camera sensor Point spread function Focal plane

Solution: lens with occluder Aperture pattern Image of a defocused point light source Object Lens with coded aperture Camera sensor Point spread function Focal plane

Solution: lens with occluder Aperture pattern Image of a defocused point light source Object Lens with coded aperture Camera sensor Point spread function Focal plane

Solution: lens with occluder Aperture pattern Image of a defocused point light source Object Lens with coded aperture Camera sensor Point spread function Focal plane

Solution: lens with occluder Aperture pattern Image of a defocused point light source Object Lens with coded aperture Camera sensor Point spread function Focal plane

Coded aperture reduces uncertainty in scale identification Conventional Coded Larger scale Correct scale Smaller scale

spectrum spectrum spectrum spectrum Convolution- frequency domain representation spectrum spectrum Sharp Image Filter, 1 st scale 0 Frequency = 1 st observed image 0 Frequency 0 Frequency Sharp Image Filter, 2 nd scale 0 Frequency = 2 nd observed image 0 Frequency Spatial convolution 0 Frequency Output spectrum has zeros frequency multiplication where filter spectrum has zeros

spectrum spectrum spectrum spectrum Coded aperture: Scale estimation and division by zero spectrum Estimated image Filter, correct scale? 0 Frequency 0 Frequency = Observed image 0 Frequency Estimated image Filter, wrong scale? 0 Frequency 0 Frequency = Large magnitude in image to compensate for tiny magnitude in filter spatial ringing

spectrum spectrum spectrum spectrum Division by zero with a conventional aperture? spectrum Estimated image Filter, correct scale? 0 Frequency = Observed image 0 Frequency 0 Frequency Estimated image Filter, wrong scale? 0 Frequency = no spatial ringing 0 Frequency

Filter Design Analytically search for a pattern maximizing discrimination between images at different defocus scales (KL-divergence) Account for image prior and physical constraints More discrimination between scales Score Less discrimination between scales Sampled aperture patterns Conventional aperture

Depth results

Regularizing depth estimation Try deblurring with 10 different aperture scales x arg min f x y 2 l i ( x ) i Convolution error _ Derivatives prior 2 + Keep minimal error scale in each local window + regularization Input Local depth estimation Regularized depth

Regularizing depth estimation Local depth estimation Input Regularized depth

All focused results

Input

All-focused (deconvolved)

Close-up Original image All-focus image

Comparison- conventional aperture result Ringing due to wrong scale estimation

Comparison- coded aperture result

Input

All-focused (deconvolved)

Close-up Original image All-focus image Naïve sharpening

Application: Digital refocusing from a single image

Application: Digital refocusing from a single image

Application: Digital refocusing from a single image

Application: Digital refocusing from a single image

Application: Digital refocusing from a single image

Application: Digital refocusing from a single image

Application: Digital refocusing from a single image

Coded aperture: pros and cons + + + - + -+ Image AND depth at a single shot No loss of image resolution Simple modification to lens Depth is coarse unable to get depth at untextured areas, might need manual corrections. But depth is a pure bonus Lose some light But deconvolution increases depth of field

50mm f/1.8: $79.95 Cardboard: $1 Tape: $1 Depth acquisition: priceless

Some more quick examples

Quickly move camera in a parabola when taking a picture A motion at any speed in the direction of the parabola will give the same blur kernel

Results Static Camera Parabolic Camera

Results Static Camera Parabolic Camera Motion in wrong direction

RGBW Sensors 2007: Kodak Panchromatic Pixels Outperforms Bayer Grid 2X-4X sensitivity (W: no filter loss) May improve dynamic range (W >> RGB sensitivity) http://www.dpreview.com/news/2007/6/14/kodakhighsens

Computational Approaches to Display 3D TV without glasses 20, $2900, available in Japan (2010) You see different images from different angles http://news.cnet.com/8301-13506_3-20018421-17.html Newer version: http://www.pcmag.com/article2/0,2817,2392380,00.asp http://reviews.cnet.com/3dtv-buying-guide/ Toshiba

Recap of questions How can we represent all of the information contained in light? What are the fundamental limitations of cameras? What sacrifices have we made in conventional cameras? For what benefits? How else can we design cameras for better focus, deblurring, multiple views, depth, etc.?

Next class Exam review But first, have a good Thanksgiving break!