Breaking Down The Cosine Fourth Power Law

Similar documents
ABSTRACT 1. INTRODUCTION

Lecture Notes 10 Image Sensor Optics. Imaging optics. Pixel optics. Microlens

Chapter 3. Introduction to Zemax. 3.1 Introduction. 3.2 Zemax

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

Speed and Image Brightness uniformity of telecentric lenses

Method for the characterization of Fresnel lens flux transfer performance

Design Description Document

GEOMETRICAL OPTICS AND OPTICAL DESIGN

Compact camera module testing equipment with a conversion lens

ECEN 4606, UNDERGRADUATE OPTICS LAB

Opto Engineering S.r.l.

Imaging with microlenslet arrays

Opti 415/515. Introduction to Optical Systems. Copyright 2009, William P. Kuhn

Lens Design I. Lecture 3: Properties of optical systems II Herbert Gross. Summer term

Lens Design I. Lecture 3: Properties of optical systems II Herbert Gross. Summer term

Image Formation by Lenses

Lens Design I. Lecture 5: Advanced handling I Herbert Gross. Summer term

OPTICAL IMAGING AND ABERRATIONS

OPTICAL SYSTEMS OBJECTIVES

VISUAL PHYSICS ONLINE DEPTH STUDY: ELECTRON MICROSCOPES

INTRODUCTION TO ABERRATIONS IN OPTICAL IMAGING SYSTEMS

Section 3. Imaging With A Thin Lens

Tangents. The f-stops here. Shedding some light on the f-number. by Marcus R. Hatch and David E. Stoltzmann

IMAGE SENSOR SOLUTIONS. KAC-96-1/5" Lens Kit. KODAK KAC-96-1/5" Lens Kit. for use with the KODAK CMOS Image Sensors. November 2004 Revision 2

Section 11. Vignetting

Thin Lenses * OpenStax

This document explains the reasons behind this phenomenon and describes how to overcome it.

Phys 531 Lecture 9 30 September 2004 Ray Optics II. + 1 s i. = 1 f

Some of the important topics needed to be addressed in a successful lens design project (R.R. Shannon: The Art and Science of Optical Design)

Photometry for Traffic Engineers...

Bias errors in PIV: the pixel locking effect revisited.

Lens Design I. Lecture 5: Advanced handling I Herbert Gross. Summer term

INTRODUCTION THIN LENSES. Introduction. given by the paraxial refraction equation derived last lecture: Thin lenses (19.1) = 1. Double-lens systems

Photometry for Traffic Engineers...

Lecture 8. Lecture 8. r 1

Applications of Optics

EC-433 Digital Image Processing

CHAPTER 33 ABERRATION CURVES IN LENS DESIGN

Actually, you only need to design one monocular of the binocular.

Big League Cryogenics and Vacuum The LHC at CERN

Intorduction to light sources, pinhole cameras, and lenses

NIRCam optical calibration sources

X-ray generation by femtosecond laser pulses and its application to soft X-ray imaging microscope

Better Imaging with a Schmidt-Czerny-Turner Spectrograph

Cameras. CSE 455, Winter 2010 January 25, 2010

Chapter 29/30. Wave Fronts and Rays. Refraction of Sound. Dispersion in a Prism. Index of Refraction. Refraction and Lenses

Adding Realistic Camera Effects to the Computer Graphics Camera Model

PROCEEDINGS OF SPIE. Measurement of low-order aberrations with an autostigmatic microscope

ptical Short Course International

ISS-30-VA. Product tags: Integrating Sphere Source. Gigahertz-Optik GmbH 1/5

Patents of eye tracking system- a survey

NFMS THEORY LIGHT AND COLOR MEASUREMENTS AND THE CCD-BASED GONIOPHOTOMETER. Presented by: January, 2015 S E E T H E D I F F E R E N C E

Aperture and Digi scoping. Thoughts on the value of the aperture of a scope digital camera combination.

Optical Design with Zemax

Vision. The eye. Image formation. Eye defects & corrective lenses. Visual acuity. Colour vision. Lecture 3.5

Imaging Optics Fundamentals

PHYSICS. Chapter 35 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT

Yokohama City University lecture INTRODUCTION TO HUMAN VISION Presentation notes 7/10/14

Vision 1. Physical Properties of Light. Overview of Topics. Light, Optics, & The Eye Chaudhuri, Chapter 8

CHAPTER 9 POSITION SENSITIVE PHOTOMULTIPLIER TUBES

( ) Deriving the Lens Transmittance Function. Thin lens transmission is given by a phase with unit magnitude.

Optical Design of. Microscopes. George H. Seward. Tutorial Texts in Optical Engineering Volume TT88. SPIE PRESS Bellingham, Washington USA

Magnification, stops, mirrors More geometric optics

Design and assessment of microlenslet-array relay optics

LENSES. INEL 6088 Computer Vision

Optoliner NV. Calibration Standard for Sighting & Imaging Devices West San Bernardino Road West Covina, California 91790

Reflection! Reflection and Virtual Image!

Unit 1: Image Formation

New foveated wide angle lens with high resolving power and without brightness loss in the periphery

Scanning Long-wave Optical Test System a new ground optical surface slope test system

Extended depth-of-field in Integral Imaging by depth-dependent deconvolution

CPSC 425: Computer Vision

Optical Design with Zemax

DOING PHYSICS WITH MATLAB COMPUTATIONAL OPTICS. GUI Simulation Diffraction: Focused Beams and Resolution for a lens system

Solution of Exercises Lecture Optical design with Zemax Part 6

Photographic zoom fisheye lens design for DSLR cameras

Be aware that there is no universal notation for the various quantities.

Lens Principal and Nodal Points

Geometric optics & aberrations

Exam Preparation Guide Geometrical optics (TN3313)

Introduction. Geometrical Optics. Milton Katz State University of New York. VfeWorld Scientific New Jersey London Sine Singapore Hong Kong

Science Focus 8. Light and Optical Systems. Pop Quiz Master (5 questions) for each Topic A C B D C C B C C A D B C A B B C C A C A C D B A C B B C D

Warren J. Smith Chief Scientist, Consultant Rockwell Collins Optronics Carlsbad, California

Optical Components - Scanning Lenses

System/Prescription Data

doi: /

Optical Design with Zemax for PhD - Basics

Ron Liu OPTI521-Introductory Optomechanical Engineering December 7, 2009

Properties of optical instruments. Projection optical systems

Optics Day 3 Kohler Illumination (Philbert Tsai July 2004) Goal : To build an bright-field microscope with a Kohler illumination pathway

Testing Aspherics Using Two-Wavelength Holography

Testing Aspheric Lenses: New Approaches

Unit 8: Light and Optics

Chapter Ray and Wave Optics

Chapter 23 Study Questions Name: Class:

NANO 703-Notes. Chapter 9-The Instrument

CSE 527: Introduction to Computer Vision

Finite conjugate spherical aberration compensation in high numerical-aperture optical disc readout

Chapter 36: diffraction

Transcription:

Breaking Down The Cosine Fourth Power Law By Ronian Siew, inopticalsolutions.com Why are the corners of the field of view in the image captured by a camera lens usually darker than the center? For one thing, camera lenses by design often introduce vignetting into the image, which is the deliberate clipping of rays at the corners of the field of view in order to cut away excessive lens aberrations. But, it is also known that corner areas in an image can get dark even without vignetting, due in part to the so-called cosine fourth power law. 1 According to this law, when a lens projects the image of a uniform source onto a screen, in the absence of vignetting, the illumination flux density (i.e., the optical power per unit area) across the screen from the center to the edge varies according to the fourth power of the cosine of the angle between the optic axis and the oblique ray striking the screen. Actually, optical designers know this law does not apply generally to all lens conditions. 2 10 Fundamental principles of optical radiative flux transfer in lens systems allow one to tune the illumination distribution across the image by varying lens design characteristics. In this article, we take a tour into the fascinating physics governing the illumination of images in lens systems. Relative Illumination In Lens Systems In lens design, one characterizes the illumination distribution across the screen where the image resides in terms of a quantity known as the lens relative illumination the ratio of the irradiance (i.e., the power per unit area) at any off-axis position of the image to the irradiance at the center of the image. In the absence of deliberate vignetting, the relative illumination produced by a lens across the screen depends on a number of factors, including the nature of rays scattering or emitting from the source, the shape and size of the lens pupils, the distortion in the image, the oblique ray angles behind (and in front of) the lens, and a little-known quantity called differential distortion. Let s look at a simple example. Fig. 1 shows what s called a rear landscape lens a single lens element placed behind the iris (known in lens design terminology as the aperture stop ). The object is a uniform, circular disk of diameter 200 mm, centered on the optic axis, and it emits incoherent light at a wavelength of 550 nm. Additionally, this object is assumed to be a Lambertian source, which is the condition that most optical design programs assume when computing the image s relative illumination. A Lambertian source emits (or scatters) rays in such a way that its radiance (i.e., the flux per unit source area within a unit solid angle) in any direction is constant. Perhaps an approximate example of such a source is the surface of white print paper, such as that used in laser printers and

photocopiers. The prescription for this lens system is given in Table 1, so if you have access to an optical design program, you can enter this prescription and examine this lens in the manner that I m about to describe. Fig. 1 A rear landscape lens example. Table 1 Prescription for the lens shown in Fig. 1 (dimension units in mm) Fig. 2a shows a plot of this lens relative illumination, which has been computed using an optical design program. 11 Note that the relative illumination is conventionally plotted as a function of the image s corresponding object height. Since the object s full diameter is 200 mm, its height above the optic axis is 100 mm, which is the magnitude of the maximum along the horizontal axis in Fig. 2a. From this plot, it is quite evident that the center of the image is actually darker than the edge! Fig. 2b shows a false-color plot of the surface irradiance at the image. Note that the irradiance at the farthest edge of the image is slightly dimmed due to aberrations at the image s edge (note the large blur spot at the farthest y-position above the optic axis in Fig. 1).

Fig. 2 (a) Relative Illumination of the lens in Fig. 1. (b) False color representation of the relative illumination across the image. Now, let s examine what happens to the relative illumination for this lens as a function of the position of the iris, at three different locations relative to the lens front vertex (Figs. 3 5). In Fig. 3a, the iris is maintained at the same location as in Fig. 1, yielding the relative illumination shown in Fig. 3b (which is the same as in Fig. 2a). Fig. 3 (a) Iris at 38 mm from front vertex. (b) Resulting relative illumination. In Fig. 4a, the iris is at 29 mm from the lens s front vertex (note that the total distance from the object to the front vertex of the lens has been maintained at 290 mm), yielding the relative illumination shown in Fig. 4b.

Fig. 4 (a) Iris at 29 mm from front vertex. (b) Resulting relative illumination. In Fig. 5a, the iris is even closer to the lens, and is at 10 mm from the lens s front vertex (whilst again maintaining the object distance to the front vertex of the lens at 290 mm), resulting in the relative illumination plot shown in Fig. 5b. Clearly, through simple manipulation of the iris position, one can tune the relative illumination for this lens. Fig. 5 (a) Iris at 10 mm from front vertex. (b) Resulting relative illumination. The Science Behind Relative Illumination Let s try to understand the results in Figs. 2 5 based on the science behind relative illumination. The general principle is that, as a consequence of the presence of an iris in a lens system, there results what s called the entrance and exit pupils of the lens system. The entrance pupil is what the iris looks like from the perspective of the object, while the exit pupil is what the iris looks like from the perspective of the image. It is the geometry in which rays pass through the entrance and exit pupils that gives rise to the form of the relative illumination observed. The technical term for this ray geometry is the projected solid angle.

From the perspective of the image, as one looks at the exit pupil from the optic axis and up (i.e., in the direction of y shown in Fig. 1), lens elements situated between the iris and the image can change the apparent shape and size of the exit pupil, which can yield either lesser or greater flux passing through the exit pupil above the optic axis. Thus, whenever the flux is greater (as in Figs. 2 3), the relative illumination increases from the center to the edge, and when the flux is less, the relative illumination decreases. But, when the shape and size of the exit pupil match the effect of the obliquity of the rays striking the image, the relative illumination can be made constant (Fig. 4). Optical design programs usually apply numerical algorithms to compute relative illumination based on ray geometries at the exit pupil. 12 Due to the law of conservation of energy (i.e., optical flux passing through the entrance pupil is equal to the flux passing out through the exit pupil), relative illumination can also be computed based on ray geometries at the entrance pupil. In fact, it can be shown 10, 13 that, under the condition of having a small iris diameter (i.e., a large f/number or low numerical aperture), by applying the ray geometry shown in Fig. 6, the relative illumination ( ) of a lens system may be approximated by the following equation: ( ) ( )/ (0) cos (1+ ) 1+ + ( / ) Eq.(1) Fig. 6 Ray geometry for Eq. 1. In Eq. (1), the quantity ( )/ (0) is the ratio of the entrance pupil area at an off-axis point to its area at the optic axis from the perspective of the object, and D is the image distortion (e.g., if the image distortion were 10 percent, then D = 0.1). The quantity / has been called differential distortion 10, 13 and it is the instantaneous rate of change of image distortion with respect to the object s height above the optic axis. Due to differential distortion, the relative illumination

at any position in the image can be affected by the variation of the image s distortion, even if the image distortion itself were zero at that position. Let s apply Eq. (1) to the conditions shown in Figs. 3 5 and compare the resulting computations with the relative illumination plots provided in those figures. In all three figures, there are no lens elements between the iris and the object, so the ratio ( )/ (0) is equal to 1 (i.e., the iris in this case is the entrance pupil). Applying the lens system conditions in Fig. 3 to an optical design program will yield =21.644, = 0.145, and / = 0.002907 mm -1. Inserting these values in Eq. (1), we have ( ) 1.55 at the full height of the object. According to Fig. 3b, the relative illumination from the plot reads 1/0.75 1.33, so our computation is somewhat close to this value (note that the relative illumination in the plots is always normalized to the peak value). Applying the conditions from Fig. 4 into the ray tracing program I am using, I obtain =20.964, = 0.066, and / = 0.001249 mm -1. Inserting these values into Eq. (1) yields ( ) 1.01, which seems close to the maximum value for relative illumination shown in Fig. 4b. Finally, the condition in Fig. 5 results in =19.654, = 0.033, and / = 0.0006312 mm -1. Inserting these values into Eq. (1) yields ( ) 0.9, which is close to the relative illumination at the full object height shown in Fig. 5b. Thus, the equation works pretty well, proving that relative illumination may be viewed as a function of the ray geometry at either the entrance pupil or the exit pupil. Moreover, it is quite evident that differential distortion / plays a significant role in the relative illumination of a lens system. Relative Illumination In Projection And Machine Vision The relative illumination in projection and machine vision systems generally is a more complex function of lens and system variables. When designing the illumination subsystem in, say, a projector, one often applies optical components, such as condensers, light pipes, and lens arrays (also known as fly s eye lenses ). A condenser serves to collect as much flux as possible from a source, while the light pipes and lens arrays serve to homogenize the illumination distribution across the projection target plane. In some other cases of illumination design, the surfaces of optical components may take on any shape, and such freeform lenses and mirrors can produce virtually any desired irradiance distribution across a screen. The result is that relative illumination for such systems cannot be described by Eq. (1), nor is it a simple function of the exit pupil s ray geometry. Rather, the irradiance distribution across a screen is best described using the methods of illumination 14 17 engineering and nonimaging optics.

In machine vision systems, lenses form images onto electronic array sensors, such as CMOS image sensors, whose pixels often have microlenses assembled in front of them. Consequently, light rays in such systems can experience clipping by the microlenses at the edge of the field of view, unless the imaging lens oblique ray angles have been designed to match specific geometries of the microlenses. 18 Additionally, in machine vision systems, one generally mounts light sources next to imaging lenses to illuminate the region of interest for the vision system. A ring of LEDs, for example, may be mounted surrounding the camera lens, and the LED arrangement can be made to balance the inherent relative illumination distribution given by the lens design. Therefore, the final irradiance distribution at the image can be tuned by a virtually unlimited arrangement of illumination geometries. Where Does The Cosine Fourth Power Law Apply? There is but one condition in which the cosine fourth power law applies. In the subject of optical radiometry, one learns that the irradiance across a screen from a small Lambertian source does, indeed, follow a fourth-power cosine rule (Fig. 7). By small it is meant that the surface area of the source is significantly less than the source s distance to the screen. Under this condition, the irradiance ( ) at the screen is well-described by the following equation: ( ) cos Eq.(2) Fig. 7 Small Lambertian source illuminating a distant screen. In Eq. (2), is the source s radiance, and is the source s surface area. Since it is known that the quantity in Eq. (2) is invariant in a lossless optical system, it has been said that the exit pupil of an imaging lens appears as a Lambertian source from the perspective of the image. 19 This is quite accurate, as long as we think of the exit pupil as a Lambertian source of varying shape and size as

it is viewed from across the image plane. Unfortunately, this has, at times, also led to the misconception that the relative illumination across the image plane would be described by Eq. (2). But, as we have seen in this article, this simple cosine rule does not generally apply to lens systems. Rather, the relative illumination of a lens system is a function of the source s emission characteristics, the shape and size of the entrance and exit pupils, the obliquity of rays relative to the optic axis, the distortion in the image, and the differential distortion across the image. About The Author Ronian Siew is a consultant in optical engineering. His optical designs have included objectives, eyepieces, relays, fluorescence detection systems for DNA analysis, and more. He has authored two books and a number of scientific peer-reviewed and conference papers. Ronian holds a MS in Optics and a degree in Physics from the University of Rochester. He may be reached at ronian@inopticalsolutions.com. References: 1. R. Kingslake, Optics in Photography, (SPIE Press, 1992), p. 121. 2. See, for example, Ref. 1, p. 122.

3. P. Foote, Illumination from a radiating disk, in Bulletin of the Bureau of Standards 12, pp. 583 586 (1915). 4. M. Reiss, The Cos 4 Law of Illumination, J. Opt. Soc. Am. 35(4), pp. 283 288 (1945). 5. G. Slussareff, A Reply to Max Reiss, J. Opt. Soc. Am. 36(12), p. 707 (1946). 6. I. C. Gardner, Validity of the Cosine-Fourth Power Law of Illumination, J. Res. Net. Bur. Stand. 39, pp. 213 219 (1947). 7. M. Reiss, Notes on the Cos 4 Law of Illumination, J. Opt. Soc. Am. 38(11), pp. 980 986 (1948). 8. F. Wachendorf, The Condition of Equal Irradiance and the Distribution of Light in Images Formed by Optical Systems Without Artificial Vignetting, J. Opt. Soc. Am. 43(12), pp. 1205 1208 (1953). 9. D. Reshidko and J. Sasian, Role of aberrations in the relative illumination of a lens system, Opt. Eng. 55(11), 115105 (2016). 10. R. Siew, Relative illumination and image distortion, Opt. Eng. 56(4), 049701 (2017). 11. While I have used the Zemax OpticStudio optical design program for simulation, any other optical design program may be used, and you should choose the software that best suits your needs. 12. M. P. Rimmer, Relative Illumination Calculations, in Optical System Design, Analysis, Production for Advanced Technology Systems, R. E. Fischer and P. J. Rogers (editors), Proc. SPIE 0655 (1986). 13. R. Siew, Perspectives on Modern Optics and Imaging: With Practical Examples Using Zemax OpticStudio, (Independently Published, 2017), pp. 116 122. 14. A. V. Arecchi, T. Messadi, and R. J. Koshel, Field Guide to Illumination, (SPIE, 2007). 15. R. J. Koshel, Illumination Engineering: Design with Nonimaging Optics, (Wiley, 2013). 16. J. Chaves, Introduction to Nonimaging Optics, (CRC Press, 2016). 17. R. Winston, J. C. Minano, and P. Benitez, with contributions by N. Shatz and J. C. Bortz, Nonimaging Optics, (Elsevier, 2005). 18. P. P. Clark, Lens Design and Advanced Function for Mobile Cameras, in Smart Mini- Cameras, T. V. Galstian, ed., (CRC Press, 2014), p. 12. 19. R. W. Boyd, Radiometry and the Detection of Optical Radiation, (Wiley, 1983), pp.84 85.