Image Formation and Capture. Acknowledgment: some figures by B. Curless, E. Hecht, W.J. Smith, B.K.P. Horn, and A. Theuwissen

Similar documents
Image Formation and Capture

Building a Real Camera

Building a Real Camera. Slides Credit: Svetlana Lazebnik

LENSES. INEL 6088 Computer Vision

Acquisition. Some slides from: Yung-Yu Chuang (DigiVfx) Jan Neumann, Pat Hanrahan, Alexei Efros

Lenses. Overview. Terminology. The pinhole camera. Pinhole camera Lenses Principles of operation Limitations

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

Cameras. CSE 455, Winter 2010 January 25, 2010

Cameras, lenses and sensors

Light. Path of Light. Looking at things. Depth and Distance. Getting light to imager. CS559 Lecture 2 Lights, Cameras, Eyes

VC 11/12 T2 Image Formation

Announcement A total of 5 (five) late days are allowed for projects. Office hours

VC 14/15 TP2 Image Formation

Cameras. Shrinking the aperture. Camera trial #1. Pinhole camera. Digital Visual Effects Yung-Yu Chuang. Put a piece of film in front of an object.

Projection. Readings. Szeliski 2.1. Wednesday, October 23, 13

Unit 1: Image Formation

Projection. Announcements. Müller-Lyer Illusion. Image formation. Readings Nalwa 2.1

Image Formation: Camera Model

Image Formation and Camera Design

CSE 473/573 Computer Vision and Image Processing (CVIP)

The Camera : Computational Photography Alexei Efros, CMU, Fall 2008

Projection. Projection. Image formation. Müller-Lyer Illusion. Readings. Readings. Let s design a camera. Szeliski 2.1. Szeliski 2.

Lenses, exposure, and (de)focus

Getting light to imager. Capturing Images. Depth and Distance. Ideal Imaging. CS559 Lecture 2 Lights, Cameras, Eyes

Basic principles of photography. David Capel 346B IST

Cameras. Outline. Pinhole camera. Camera trial #1. Pinhole camera Film camera Digital camera Video camera

The Camera : Computational Photography Alexei Efros, CMU, Fall 2005

Charged Coupled Device (CCD) S.Vidhya

Image Formation III Chapter 1 (Forsyth&Ponce) Cameras Lenses & Sensors

Chapter 25 Optical Instruments

Image acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor

Megapixels and more. The basics of image processing in digital cameras. Construction of a digital camera

Two strategies for realistic rendering capture real world data synthesize from bottom up

VC 16/17 TP2 Image Formation

Cameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017

CPSC 425: Computer Vision

Cameras. Outline. Pinhole camera. Camera trial #1. Pinhole camera Film camera Digital camera Video camera High dynamic range imaging

Overview. Charge-coupled Devices. MOS capacitor. Charge-coupled devices. Charge-coupled devices:

Cameras. Digital Visual Effects, Spring 2008 Yung-Yu Chuang 2008/2/26. with slides by Fredo Durand, Brian Curless, Steve Seitz and Alexei Efros

CS 443: Imaging and Multimedia Cameras and Lenses

Applications of Optics

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

Optical Components for Laser Applications. Günter Toesko - Laserseminar BLZ im Dezember

Optical Systems: Pinhole Camera Pinhole camera: simple hole in a box: Called Camera Obscura Aristotle discussed, Al-Hazen analyzed in Book of Optics

Chapters 1 & 2. Definitions and applications Conceptual basis of photogrammetric processing

Dr F. Cuzzolin 1. September 29, 2015

Limitations of lenses

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

Waves & Oscillations

Announcements. Image Formation: Outline. The course. How Cameras Produce Images. Earliest Surviving Photograph. Image Formation and Cameras

mm F2.6 6MP IR-Corrected. Sensor size

brief history of photography foveon X3 imager technology description

What will be on the midterm?

CS6670: Computer Vision

Geometric optics and camera, practically

Lecture PowerPoint. Chapter 25 Physics: Principles with Applications, 6 th edition Giancoli

Laboratory experiment aberrations

Astronomical Cameras

Applied Optics. , Physics Department (Room #36-401) , ,

Lecture 4: Geometrical Optics 2. Optical Systems. Images and Pupils. Rays. Wavefronts. Aberrations. Outline

Lecture 7: Camera Models

Last class. This class. CCDs Fancy CCDs. Camera specs scmos

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2015 Version 3

Reflectors vs. Refractors

Chapter 25. Optical Instruments

CSE 527: Introduction to Computer Vision

Digital Photographic Imaging Using MOEMS

Lecture 02 Image Formation 1

Performance Factors. Technical Assistance. Fundamental Optics

How do we see the world?

IMAGE SENSOR SOLUTIONS. KAC-96-1/5" Lens Kit. KODAK KAC-96-1/5" Lens Kit. for use with the KODAK CMOS Image Sensors. November 2004 Revision 2

A Simple Camera Model

COURSE NAME: PHOTOGRAPHY AND AUDIO VISUAL PRODUCTION (VOCATIONAL) FOR UNDER GRADUATE (FIRST YEAR)

AST Lab exercise: aberrations

CS559: Computer Graphics. Lecture 2: Image Formation in Eyes and Cameras Li Zhang Spring 2008

BEAM HALO OBSERVATION BY CORONAGRAPH

Converging and Diverging Surfaces. Lenses. Converging Surface

Optical basics for machine vision systems. Lars Fermum Chief instructor STEMMER IMAGING GmbH

1 Image Formation. 1.1 Optics 1.1

BROADCAST ENGINEERING 5/05 WHITE PAPER TUTORIAL. HEADLINE: HDTV Lens Design: Management of Light Transmission

Cameras and Sensors. Today. Today. It receives light from all directions. BIL721: Computational Photography! Spring 2015, Lecture 2!

Lecture 22: Cameras & Lenses III. Computer Graphics and Imaging UC Berkeley CS184/284A, Spring 2017

EE119 Introduction to Optical Engineering Spring 2003 Final Exam. Name:

The Xiris Glossary of Machine Vision Terminology

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Computer Vision. The Pinhole Camera Model

CS6670: Computer Vision

Photons and solid state detection

General Imaging System

PHY 431 Homework Set #5 Due Nov. 20 at the start of class

Optics: An Introduction

ABC Math Student Copy. N. May ABC Math Student Copy. Physics Week 13(Sem. 2) Name. Light Chapter Summary Cont d 2

INTRODUCTION THIN LENSES. Introduction. given by the paraxial refraction equation derived last lecture: Thin lenses (19.1) = 1. Double-lens systems

Imaging Optics Fundamentals

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Physics 1230 Homework 8 Due Friday June 24, 2016

Cameras, lenses, and sensors

EE-527: MicroFabrication

Design of a digital holographic interferometer for the. ZaP Flow Z-Pinch

Transcription:

Image Formation and Capture Acknowledgment: some figures by B. Curless, E. Hecht, W.J. Smith, B.K.P. Horn, and A. Theuwissen

Image Formation and Capture Real world Optics Sensor Devices Sources of Error

Optics Pinhole camera Lenses Focus, aperture, distortion

Pinhole Camera Camera obscura known since antiquity Image plane Image Object Pinhole Pinhole camera

Pinhole Camera Camera obscura known since antiquity Image plane Image Object Pinhole Pinhole camera First recording in 1826 onto a pewter plate (by Joseph Nicéphore Niepce)

Pinhole Camera Limitations Aperture too big: blurry image Aperture too small: requires long exposure or high intensity Aperture much too small: diffraction through pinhole blurry image Rule of thumb: aperture should be significantly larger than wavelength of light (400-700 nm)

Lenses Focus a bundle of rays from a scene point onto a single point on the imager Result: can make aperture bigger

Ideal Thin Lens Law Relationship between focal distance and focal length of lens: 1/d o + 1/d i = 1/f

Camera Adjustments Focus? Iris? Changes d i Changes aperture Zoom? Changes f and sometimes d i

Zoom Lenses Varifocal

Zoom Lenses Parfocal

Focus and Depth of Field For a given d i, perfect focus at only one d o In practice, OK for some range of depths Circle of confusion smaller than a pixel Better depth of field with smaller apertures Better approximation to pinhole camera Also better depth of field with wide-angle lenses

Field of View Q: What does field of view of camera depend on? Focal length of lens Size of imager Object distance?

Computing Field of View 1/d o + 1/d i = 1/f tan θ /2 = ½ x o / d o x o θ x i x o / d o = x i / d i θ = 2 tan -1 ½ x i (1/f 1/d o ) d o d i Since typically d o >> f, θ 2 tan -1 ½ x i / f θ x i / f

Aperture Controls amount of light Affects depth of field Affects distortion (since thin-lens approximation is better near center of lens)

Aperture Aperture typically given as f-number What is f /4? Aperture diameter is ¼ the focal length One f-stop equals change of f-number by 2 Equals change in aperture area by factor of 2 Equals change in amount of light by factor of 2 Example: f/2 f/2.8 f/4

Sensors Film Vidicon CCD CMOS

Vidicon Best-known in family of photoconductive video cameras Basically television in reverse + + + + Electron Gun Scanning Electron Beam Lens System Photoconductive Plate

MOS Capacitors MOS = Metal Oxide Semiconductor Gate (wire) SiO 2 (insulator) p-type silicon

MOS Capacitors Voltage applied to gate repels positive holes in the semiconductor +10V + + + + + + Depletion region (electron bucket )

MOS Capacitors Photon striking the material creates electron-hole pair Photon +10V + + + + + + +

Charge Transfer Can move charge from one bucket to another by manipulating voltages

CMOS Imagers Recently, can manufacture chips that combine photosensitive elements and processing elements Benefits: Partial readout Signal processing Eliminate some supporting chips low cost

Color 3-chip vs. 1-chip: quality vs. cost Bayer mosaic

Errors in Digital Images What are some sources of error in this image?

Sources of Error Geometric (focus, distortion) Color (1-chip artifacts, chromatic aberration) Radiometric (cosine falloff, vignetting) Bright areas (flare, bloom, clamping) Signal processing (gamma, compression) Noise

Monochromatic Aberrations Real lenses do not follow thin lens approximation because surfaces are spherical (manufacturing constraints) Result: thin-lens approximation only valid iff sin ϕ ϕ

Spherical Aberration Results in blurring of image, focus shifts when aperture is stopped down Can vary with the way lenses are oriented

Distortion Pincushion or barrel radial distortion Varies with placement of aperture

Distortion Varies with placement of aperture

Distortion Varies with placement of aperture

Distortion Varies with placement of aperture

First-Order Radial Distortion Goal: mathematical formula for distortion If small, can be approximated by first-order formula (like Taylor series expansion): r = r (1 + κ r 2 ) r = ideal distance to center of image r = distorted distance to center of image Higher-order models are possible

Chromatic Aberration Due to dispersion in glass (focal length varies with the wavelength of light) Result: color fringes Worst at edges of image Correct by building lens systems with multiple kinds of glass

Correcting for Aberrations High-quality compound lenses use multiple lens elements to cancel out distortion and aberration Often 5-10 elements, more for extreme wide angle

Other Limitations of Lenses Optical vignetting: less power per unit area transferred for light at an oblique angle Transferred power falls off as cos 4 ϕ Result: darkening of edges of image Mechanical vignetting: due to apertures

Other Limitations of Lenses Flare: light reflecting (often multiple times) from glass-air interface Results in ghost images or haziness Worse in multi-lens systems Ameliorated by optical coatings (thin-film interference)

Bloom Overflow of charge in CCD buckets Spills to adjacent buckets Streaks (usually vertical) next to bright areas Some cameras have anti-bloom circuitry

Flare and Bloom Tanaka

Dynamic Range Most common cameras have 8-bit (per color channel) dynamic range Can be nonlinear: more than 255:1 intensity range Too bright: clamp to maximum Too dim: clamp to 0 Specialty cameras with higher dynamic range (usually 10-, 12-, and 16-bit)

High Dynamic Range (HDR) from Ordinary Cameras Take pictures of same scene with different shutter speeds Identify regions clamped to 0 or 255 Average other pixels, scaled by 1 / shutter speed Can extend dynamic range, but limitations of optics and imager (noise, flare, bloom) still apply

Gamma Vidicon tube naturally has signal that varies with light intensity according to a power law: Signal = E γ, γ 1/2.5 CRT (televisions) naturally obey a power law with gamma 2.3 2.5 Result: video signal standard has gamma of 1/2.5 CCDs and CMOS linear, but gamma 2.2 almost always applied

Noise Thermal noise: in all electronics Noise at all frequencies Proportional to temperature Special cooled cameras available for low noise Shot noise: discrete photons / electrons Shows up at extremely low intensities CCDs / CMOS can have high efficiency approaching 1 electron per photon

Noise 1/f noise inversely proportional to frequency Not completely understood shows up in semiconductors Can be dominant source of noise All of the above apply for imager and amplifier

Filtering Noise Most common method simple blur e.g., convolution with Gaussian Adaptive filters to prevent bleed across intensity edges Other filters for specialized situations e.g., despeckling (median filters) for dead pixels