Intorduction to light sources, pinhole cameras, and lenses

Similar documents
Overview. Image formation - 1

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Phys 531 Lecture 9 30 September 2004 Ray Optics II. + 1 s i. = 1 f

ECEN 4606, UNDERGRADUATE OPTICS LAB

Chapter 18 Optical Elements

Unit 1: Image Formation

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

OPTICS I LENSES AND IMAGES

Physics of Light. Light: electromagnetic radiation that can produce a visual sensation.

Chapter 36. Image Formation

PHYSICS. Chapter 35 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT

Why learn about photography in this course?

lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response

Image Formation. Light from distant things. Geometrical optics. Pinhole camera. Chapter 36

THIN LENSES: APPLICATIONS

Fundamental Paraxial Equation for Thin Lenses

E X P E R I M E N T 12

Image Formation by Lenses

Dr F. Cuzzolin 1. September 29, 2015

Assignment: Light, Cameras, and Image Formation

VC 11/12 T2 Image Formation

VC 14/15 TP2 Image Formation

Lecture 1 1 Light Rays, Images, and Shadows

Cameras. CSE 455, Winter 2010 January 25, 2010

Projection. Announcements. Müller-Lyer Illusion. Image formation. Readings Nalwa 2.1

Physics II. Chapter 23. Spring 2018

Light. Path of Light. Looking at things. Depth and Distance. Getting light to imager. CS559 Lecture 2 Lights, Cameras, Eyes

Chapter 36. Image Formation

PHY 431 Homework Set #5 Due Nov. 20 at the start of class

Optics B. Science Olympiad North Regional Tournament at the University of Florida DO NOT WRITE ON THIS BOOKLET. THIS IS AN TEST SET.

Focus on an optical blind spot A closer look at lenses and the basics of CCTV optical performances,

Applications of Optics

General Physics II. Optical Instruments

Chapters 1 & 2. Definitions and applications Conceptual basis of photogrammetric processing

10.2 Images Formed by Lenses SUMMARY. Refraction in Lenses. Section 10.1 Questions

Radiometry I: Illumination. cs348b Matt Pharr

Unit 3 - Foundations of Waves

Chapter 34: Geometric Optics

Lighting Terminologies Introduction

Physics 345 Pre-lab 1

Introduction. Strand F Unit 3: Optics. Learning Objectives. Introduction. At the end of this unit you should be able to;

Εισαγωγική στην Οπτική Απεικόνιση

VC 16/17 TP2 Image Formation

Magnification, stops, mirrors More geometric optics

Digital Camera Technologies for Scientific Bio-Imaging. Part 2: Sampling and Signal

Vision. The eye. Image formation. Eye defects & corrective lenses. Visual acuity. Colour vision. Lecture 3.5

Notes: Light and Optics. Reflection. Refraction. Law of Reflection. Light goes straight 12/13/2012

Big League Cryogenics and Vacuum The LHC at CERN

Geometric Optics. Objective: To study the basics of geometric optics and to observe the function of some simple and compound optical devices.

LENSES. INEL 6088 Computer Vision

19. Ray Optics. S. G. Rajeev. April 2, 2009

O5: Lenses and the refractor telescope

6.014 Recitation 1: Wireless Radio and Optical Links

The eye & corrective lenses

Pinhole Camera. Nuts and Bolts

CPSC 425: Computer Vision

Optics Review (Chapters 11, 12, 13)

Geometrical Optics. Have you ever entered an unfamiliar room in which one wall was covered with a

Laboratory 2: PV Module Current-Voltage Measurements

Geometrical optics Design of imaging systems

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2015 Version 3

Compound Lens Example

Human Retina. Sharp Spot: Fovea Blind Spot: Optic Nerve

CS 428: Fall Introduction to. Image formation Color and perception. Andrew Nealen, Rutgers, /8/2010 1

Algebra Based Physics. Reflection. Slide 1 / 66 Slide 2 / 66. Slide 3 / 66. Slide 4 / 66. Slide 5 / 66. Slide 6 / 66.

Chapter 23. Light Geometric Optics

Lecture Outline Chapter 27. Physics, 4 th Edition James S. Walker. Copyright 2010 Pearson Education, Inc.

Section 3. Imaging With A Thin Lens

2.71 Optics Fall 05 QUIZ 1 Wednesday, Oct. 12, 2005

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Vision Science I Exam 1 23 September ) The plot to the right shows the spectrum of a light source. Which of the following sources is this

GIST OF THE UNIT BASED ON DIFFERENT CONCEPTS IN THE UNIT (BRIEFLY AS POINT WISE). RAY OPTICS

Instructional Resources/Materials: Light vocabulary cards printed (class set) Enough for each student (See card sort below)

Person s Optics Test KEY SSSS

Study guide for Graduate Computer Vision

Chapter 16 Light Waves and Color

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

The Human Eye and a Camera 12.1

CS 443: Imaging and Multimedia Cameras and Lenses

Chapter 25. Optical Instruments

Focal Length of Lenses

Chapter 3 Mirrors. The most common and familiar optical device

CSE 527: Introduction to Computer Vision

Spherical Mirrors. Concave Mirror, Notation. Spherical Aberration. Image Formed by a Concave Mirror. Image Formed by a Concave Mirror 4/11/2014

Real-Time Scanning Goniometric Radiometer for Rapid Characterization of Laser Diodes and VCSELs

PHY385H1F Introductory Optics Term Test 2 November 6, 2012 Duration: 50 minutes. NAME: Student Number:.

Introduction to Visual Perception & the EM Spectrum

Review. Introduction to Visual Perception & the EM Spectrum. Overview (1):

Chapter 3 Solution to Problems

Light has some interesting properties, many of which are used in medicine:

PHY 1160C Homework Chapter 26: Optical Instruments Ch 26: 2, 3, 5, 9, 13, 15, 20, 25, 27

Tangents. The f-stops here. Shedding some light on the f-number. by Marcus R. Hatch and David E. Stoltzmann

GEOMETRICAL OPTICS Practical 1. Part I. BASIC ELEMENTS AND METHODS FOR CHARACTERIZATION OF OPTICAL SYSTEMS

Getting light to imager. Capturing Images. Depth and Distance. Ideal Imaging. CS559 Lecture 2 Lights, Cameras, Eyes

THE TELESCOPE. PART 1: The Eye and Visual Acuity

Image Formation and Capture

Lenses. A lens is any glass, plastic or transparent refractive medium with two opposite faces, and at least one of the faces must be curved.

Breaking Down The Cosine Fourth Power Law

PHYSICS FOR THE IB DIPLOMA CAMBRIDGE UNIVERSITY PRESS

Transcription:

Intorduction to light sources, pinhole cameras, and lenses Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 01003 October 26, 2011 Abstract 1

1 Analyzing the movement of light The goal of this section is to make sure that you understand some basic ideas about the emission of light, its interaction with surfaces, and its propagation through lenses. Some ideas, like the inverse square law, should be easy to remember once you understand the intuition behind them. 1.1 Point sources of light We start with a discussion of point light sources. This is a simplified model of any small or distant light source, such as a distant light bulb, a star, or perhaps the sun, in which we treat it as though it were an infinitesimal point. Of course, we know that no light source is infinitely small, but assuming that it is a point can make certain analyses simpler. Any light source has associated with it a power output, which is the amount of energy consumed per unit of time (for example, joules per second). A common unit of power associated with light sources is the watt. For example, most of us are familiar with using 60-watt or 100-watt light bulbs. While this power rating refers to the amount of power consumed by the bulb, rather than the amount of power produced as light, for this discussion, we will assume that the power consumed and the power output as light are the same. As another example, the wattage of the sun is about 3.846 10 26 watts. Assume that a point light source of W watts emits energy equally in all directions. If a point light source is surrounded by a sphere of radius 1 meter, then all of the light output by the source will land upon the sphere. Since a sphere of radius 1 meter has a surface area of 4π square meters, then the number of watts of power irradiating each square meter is be given by R = W watts 4π m 2. Steradians. One square unit of area of a unit sphere is considered one unit of solid angle, or one steradian. One steradian represents a cone of directions large enough to intersect one unit of area of a unit sphere, when the tip of the cone is placed at the center of the sphere. Since there are 4π units of area on the surface of a unit sphere, there are 4π steradians of solid angle in a sphere. Compare this to the number of radians in a circle (2π), which is equal to the circumference of a unit circle. The amount of light falling on a surface is a function not only of the power of the light, but also of our distance from it. However, the amount of light emitted into a unit solid angle by a light source is not related to one s distance from a light source. Hence, the emitting power of point light sources, or their radiance, is often measured in watts per unit solid angle or watts per steradian. Alternatively, it may be simply given in watts, which can be thought of as watts per 4π steradians if desired. Now imagine that instead of surrounding the light source by a sphere of radius one meter, we have surrounded it with a sphere of radius 3 meters. The 2

amount of light radiated by the source has not changed. However, now the amount of light falling on each square meter of the surrounding sphere is R = W watts 4π 3 2 m 2 = W watts 36π m 2. Thus, increasing the distance of a surface from the light source by a factor of 3 reduces the power radiated on each unit of that surface by a factor of 9. This is a simple consequence of the fact that the surface area of a sphere is proportional to the square of its radius. The falloff in the power of light per unit of area with the square of distance is an example of an inverse square law. Example 1. Suppose you have just mounted 10 square meters of solar panels on your house, and that they are oriented so that they are perfectly perpendicular to the direction of the sun. (Assume the sun is a point light source.) Discounting the absorption and scattering of the atmosphere, compute the number of watts of power being radiated upon your solar panels, assuming the sun is exactly 150 million kilometers from your solar panels. 2 Extended light sources Most real light sources are better modeled as extended light sources rather than point light sources. That is, the light source has some extent, whether it is small or large. Common examples of extended light sources are the entire sky on an overcast day, the light coming in through an office window (again, on a cloudy day), or a large fluorescent bulb. In these cases, the power of the light emanating from a single infinitesimal point may be considered negligible or even 0. Rather, it is the power emanating over a finite area that is of interest. Hence, for extended light sources, we measure power not just as watts per steradian, but rather, as watts per steradian per unit area, where the area refers to the area of the emitter, not to the surface on which light is falling. Example 2. Imagine that you have a light meter set up about 10 meters from a fluorescent light bulb, and that the recording surface of the light meter is perpendicular to the line connecting the center of the light meter to the center of the bulb. Imagine that your light meter records one milliwatt of light emanating from the bulb when half of the bulb (as viewed from the light meter) is covered with opaque electrical tape. If the tape is removed, we expect the light meter to record two milliwatts, since the surface area of the emitter has doubled. This discussion is particularly germaine to computer vision, since each element of the sensor in a video camera (either a CCD or a CMOS device), or each rod or cone cell in the retina is effectively measuring the total light energy emanating from a very small patch of a scene in the cone of directions toward the sensor. To understand the light reaching a particular sensor element, we must think of the surface as an emitter of light, and integrate the output of that surface (in the direction of the viewer) over the area of the surface which reaches the sensor element. If a video camera is pointed directly as a light source, such as a fluorescent bulb, then we are simply measuring the direct output of the 3

light source. However, if the video camera is aimed at a passive non-emitting surface, like a table, then we are measuring the reflected light or scattered light from the surface. However, this reflected light can still be conceptualized as a light source. 2.1 A note about foreshortening Foreshortening is the phonomenon in which a flat surface appears smaller because it is being viewed at an angle, rather than straight on. For example, a white sheet of paper which subtends (or takes up ) 0.2 steradians of solid angle when viewed perpendicularly (or 0 degrees) relative to the viewer would subtend only 0.1 steradians when viewed at an angle of 60 degrees. This computation, A = A cos(θ), where θ is the angle at which the surface is being viewed (relative to perpendicular), A is the solid angle subtended at 0 degrees, and A is the solid angle subtended at an angle of θ, is a consequence of basic trigonometry. To simplify the examples in this document, we have assumed that both the extended light sources and the light detectors are perpendicular to the line connecting them. When this is true, there is no foreshortening, and hence we can ignore this additional mathematical step. However, in real applications, it is important to consider the significant effect that foreshortening can have in many applications. As a real-world example, a solar panel that is always pointed directly at the sun will collect the maximum possible power from the sun. A solar panel that is oriented at 45 degrees relative to the direction of the sun will collect only 2 2 71% of this power. 3 Lenses and cameras Recall the theoretical pinhole camera model discussed in class. The pinhole camera has an infinitely small hole which lets light through, forming an inverted image of the environment on the image plane at the back of the camera. Let s calculate how much light, in watts, from an extended source falls onto a particular element E of a sensor mounted at the image plane of our pinhole camera. Let s assume our camera is pointed at a flat fluorescent lamp which is 10mm wide and 1000mm long, and so has an emitting area of 10, 000mm 2 as viewed from our camera. Furthermore, let s assume that the fluorescent lamp outputs 0.001 watts per steradian per mm 2. Without going into the details of exactly how large the imaging elements of our sensor are, or how far away the lamp is, let s assume that an area A of the lamp which is 10mm by 10mm, or 100mm 2, is imaged by the single element E of our sensor array. That is, for each point in the small area A of the fluorescent lamp, each ray traced from A, through the camera s pinhole, lands on sensor element E. 4

We can now see that the total number of watts W E landing on sensor element E is given by: W E watts = watts 0.0001 steradian mm 2 100mm2 S steradians (1) = watts 0.01 S steradians. steradian (2) Thus, all we need to know to complete the calculation is the size S in steradians of the solid angle of light which makes it through the pinhole in the camera. But since the pinhole is infinitely small, the answer to this question is S = 0 steradians, and hence W E watts = 0 watts. Here, we have demonstrated a well-known result, which is that ideal pinhole cameras, whose apertures are infinitely small, cannot capture any finite amount of light. There are two options for solving this problem. One is to give our pinhole a finite extent, so that the S in the above equation is non-zero. This is a simple solution, but will introduce blurring, or de-focus, into our images. Alternatively, we can construct a camera with a lens, which can capture light from a large solid angle of directions from a source, and focus that light to a single point again. Now let s redo our calculation of W E assuming we have replaced our pinhole camera with a camera using a certain lens. Let the lens have ideal optics, meaning that it does not absorb or scatter light, and that it focuses all of the light coming into it onto the image plane. Suppose the lens has a radius of 10mm and is 10m, or 10, 000mm, from the fluorescent light source. The cone of directions subtended (or taken up ) by the lens, as measured in steradians, from a point on the light source can be obtained by setting up a proportionality expression in the following way: or area of lens area of 10m radius sphere steradians S subtending lens =, steradians in a sphere π100mm 2 π10 8 mm 2 = S 4π. Hence, the number of steradians S subtended by the lens, is about 4π 10 6 1.26 10 5. Substituting this into Eq.2 gives 1.26 10 7 watts landing on the sensor element E. 5