CORRECTED VISION. Here be underscores THE ROLE OF CAMERA AND LENS PARAMETERS IN REAL-WORLD MEASUREMENT

Similar documents
TECHSPEC COMPACT FIXED FOCAL LENGTH LENS

Basler Accessories. Technical Specification BASLER LENS C M. Order Number

Optical design of a high resolution vision lens

Basler Accessories. Technical Specification BASLER LENS C M. Order Number

Imaging Optics Fundamentals

Using Optics to Optimize Your Machine Vision Application

Exercise questions for Machine vision

NOVA S12. Compact and versatile high performance camera system. 1-Megapixel CMOS Image Sensor: 1024 x 1024 pixels at 12,800fps

Cameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017

Eyedentify MMR SDK. Technical sheet. Version Eyedea Recognition, s.r.o.

Laboratory experiment aberrations

LaserBeam ProfilingSolutions. IRLaserBeam Profiler

Image Processing & Projective geometry

AR 2 kanoid: Augmented Reality ARkanoid

A 3D Multi-Aperture Image Sensor Architecture

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

Sensors and Sensing Cameras and Camera Calibration

Advanced 3D Optical Profiler using Grasshopper3 USB3 Vision camera

ECEN 4606, UNDERGRADUATE OPTICS LAB

Coherent Laser Measurement and Control Beam Diagnostics

Computer Vision. The Pinhole Camera Model

IMAGE SENSOR SOLUTIONS. KAC-96-1/5" Lens Kit. KODAK KAC-96-1/5" Lens Kit. for use with the KODAK CMOS Image Sensors. November 2004 Revision 2

Chapter 25 Optical Instruments

mm F2.6 6MP IR-Corrected. Sensor size

Optical Components for Laser Applications. Günter Toesko - Laserseminar BLZ im Dezember

Optical and mechanical parameters. 100 mm N. of elements 20.5 mm Dimensions 11.7 degrees Weight F/N = 4 (fixed) N.A.

Macro Varon 4.5/85. Key features. Applications. Web and surface inspections

Using molded chalcogenide glass technology to reduce cost in a compact wide-angle thermal imaging lens

Opto Engineering S.r.l.

A shooting direction control camera based on computational imaging without mechanical motion

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Chapters 1 & 2. Definitions and applications Conceptual basis of photogrammetric processing

panda family ultra compact scmos cameras

Optical basics for machine vision systems. Lars Fermum Chief instructor STEMMER IMAGING GmbH

Xenon-Diamond 2.9/106 With beam splitter

Data Sheet SMX-160 Series USB2.0 Cameras

Compatible with Windows 8/7/XP, and Linux; Universal programming interfaces for easy custom programming.

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

Vision. The eye. Image formation. Eye defects & corrective lenses. Visual acuity. Colour vision. Lecture 3.5

OPTICAL SYSTEMS OBJECTIVES

Testing Aspheric Lenses: New Approaches

Telecentric lenses.

LEICA VARIO-ELMARIT-R mm f/2,8-4,5 ASPH. 1

23270: AUGMENTED REALITY FOR NAVIGATION AND INFORMATIONAL ADAS. Sergii Bykov Technical Lead Machine Learning 12 Oct 2017

ME 297 L4-2 Optical design flow Analysis

IHV means Independent Hardware Vendor. Example is Qualcomm Technologies Inc. that makes Snapdragon processors. OEM means Original Equipment

Novel Hemispheric Image Formation: Concepts & Applications

Method for out-of-focus camera calibration

Overview: Integration of Optical Systems Survey on current optical system design Case demo of optical system design

The principles of CCTV design in VideoCAD

A LATERAL SENSOR FOR THE ALIGNMENT OF TWO FORMATION-FLYING SATELLITES

pco.edge 4.2 LT 0.8 electrons 2048 x 2048 pixel 40 fps up to :1 up to 82 % pco. low noise high resolution high speed high dynamic range

Compact camera module testing equipment with a conversion lens

SUPRA Optix 3D Optical Profiler

New Paradigm in Testing Heads & Media for HDD. Dr. Lutz Henckels September 2010

IR Laser Illuminators

Astigmatism Particle Tracking Velocimetry for Macroscopic Flows

High Performance Computing for Engineers

EF-45 Iris Recognition System

HMD based VR Service Framework. July Web3D Consortium Kwan-Hee Yoo Chungbuk National University

NanEye GS NanEye GS Stereo. Camera System

e2v Launches New Onyx 1.3M for Premium Performance in Low Light Conditions

A collection of hyperspectral images for imaging systems research Torbjørn Skauli a,b, Joyce Farrell *a

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

New foveated wide angle lens with high resolving power and without brightness loss in the periphery

arxiv: v1 [physics.optics] 2 Nov 2012

User Manual. TE Q1 Development Kit TE V1 Development Kit. Intelligent Image & Information System 1 of 23

Ideal for display mura (nonuniformity) evaluation and inspection on smartphones and tablet PCs.

Fiber Optic Communications

LENSES. INEL 6088 Computer Vision

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

Perspective platforms for BOINC distributed computing network

CSE 473/573 Computer Vision and Image Processing (CVIP)

IPD3. Imaging Photon Detector APPLICATIONS KEY ATTRIBUTES

www. riseeyetracker.com TWO MOONS SOFTWARE LTD RISEBETA EYE-TRACKER INSTRUCTION GUIDE V 1.01

Mobile Virtual Reality what is that and how it works? Alexey Rybakov, Senior Engineer, Technical Evangelist at DataArt

Lecture 4: Geometrical Optics 2. Optical Systems. Images and Pupils. Rays. Wavefronts. Aberrations. Outline

Lecture PowerPoint. Chapter 25 Physics: Principles with Applications, 6 th edition Giancoli

PHYSICS FOR THE IB DIPLOMA CAMBRIDGE UNIVERSITY PRESS

Sharpness, Resolution and Interpolation

Chapter 4: Fourier Optics

CUDA 를활용한실시간 IMAGE PROCESSING SYSTEM 구현. Chang Hee Lee

Norsk Elektro Optikk AS (NEO) HySpex Airborne Sensors System Overview

General Imaging System

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Reikan FoCal Aperture Sharpness Test Report

Acquisition. Some slides from: Yung-Yu Chuang (DigiVfx) Jan Neumann, Pat Hanrahan, Alexei Efros

Visione per il veicolo Paolo Medici 2017/ Visual Perception

Single Slit Diffraction

Camera Selection Criteria. Richard Crisp May 25, 2011

MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

Section 8. Objectives

DESIGNING AND IMPLEMENTING AN ADAPTIVE OPTICS SYSTEM FOR THE UH HOKU KE`A OBSERVATORY ABSTRACT

Development of a Low-order Adaptive Optics System at Udaipur Solar Observatory

How to Choose a Machine Vision Camera for Your Application.

ULS24 Frequently Asked Questions

Xenon-Zirconia 3.3/92

PandroidWiz and Presets

Reikan FoCal Aperture Sharpness Test Report

Transcription:

Here be underscores CORRECTED VISION THE ROLE OF CAMERA AND LENS PARAMETERS IN REAL-WORLD MEASUREMENT JOSEPH HOWSE, NUMMIST MEDIA CIG-GANS WORKSHOP: 3-D COLLECTION, ANALYSIS AND VISUALIZATION LAWRENCETOWN, NS: JANUARY 31, 2018

MOTIVATING QUESTIONS 1. Are my lens and camera good enough to measure what I want to measure? 2. In software, how can I model the perspective and distortion? 3. Is my computer fast enough to process my camera feed in real time? This presentation approaches these questions: Quantitatively Predictively. Recognize feasibility problems at the start of the project, not the end.

MOTIVATING QUESTIONS 1. Are my lens and camera good enough to measure what I want to measure? 2. In software, how can I model the perspective and distortion? 3. Is my computer fast enough to process my camera feed in real time? This presentation approaches these questions: Quantitatively Predictively. Recognize feasibility problems at the start of the project, not the end. Smartphones aren t always the right tool for the job!

OUTLINE High-performance imaging Spatial resolution Temporal resolution Camera calibration The camera matrix Distortion coefficients Computational performance Budgeting operations per pixel per frame Comparing ultra-compact computers

HIGH-PERFORMANCE IMAGING SPATIAL RESOLUTION TEMPORAL RESOLUTION

SPATIAL RESOLUTION Measurable in line pairs per millimeter (lp/mm) Max density of dark-and-light lines that given combo of lens and sensor can resolve Refers to mm on sensor surface, not subject surface combo of lp/mm and magnification determines smallest resolvable detail as measured on subject surface Magnification is easy to change by re-focusing lens

SPATIAL RESOLUTION Example: A microfilm system using the Zeiss S-Orthoplanar 50mm f/4 lens resolves 360 lp/mm At 1:5 magnification (one-fifth life size), smallest resolvable detail on subject surface is: 1 mm / 360 / (1/5) = 13.9 μm At 1:30 magnification, it is: 1 mm / 360 / (1/30) = 83.3 μm

SPATIAL RESOLUTION Limiting factor may be: Diffraction Smaller aperture (larger f-number) is more limiting Longer wavelength of light is more limiting Pixel pitch Distance between photosites; larger is more limiting Lens imperfections Sensor imperfections

SPATIAL RESOLUTION R diffraction =, -., where: R diffraction is the diffraction-limited resolution in lp/mm N is the f-number λ is wavelength of light in mm Human eye s sensitivity peaks at λ = 0.000555 (yellow-green) R pitch =, 2, where: R pitch is the pixel-pitch-limited resolution in lp/mm p is the pixel pitch in mm

SPATIAL RESOLUTION Example: The Nokia Lumia 1020 smartphone has a lens with a maximum aperture of f/2.2 and a sensor with a size of 8.8mm 6.6mm and pixel resolution of 7712 5360 R diffraction = R pitch =, = 819 lp/mm 4.4 6.666777, >.>/??,4 = 876 lp/mm Conclusion: The high pixel density is an irrational design choice. The resolution is limited theoretically by diffraction and realistically by lens imperfections.

TEMPORAL RESOLUTION Things move fast! Waves on the ocean surface Average around 10km/h near shore Cars on the road Conveyor belts in an assembly line Our eyes and eyelids Normal blink lasts 100ms to 400ms

TEMPORAL RESOLUTION Faster motion causes problems: The subject appears in fewer frames (before it goes away) Fewer samples to give to detection algorithm Smaller likelihood of detection When the subject does appear, it is blurrier Effectively, less spatial resolution

TEMPORAL RESOLUTION Example: Suppose a blink detector s true positive rate is 10% (and its false positive rate is negligible). Each of the subject s blinks lasts 300ms on average. A camera running at 60 FPS captures 18 frames during the average blink. Trying 18 times, the blink detector is 1-(0.9 18 )=85% likely to detect the blink at least once. A camera running at 120 FPS captures 36 frames during the average blink. Trying 36 times, the blink detector is 1-(0.9 36 )=98% likely to detect the blink at least once.

CAMERA CALIBRATION THE CAMERA MATRIX DISTORTION COEFFICIENTS

THE CAMERA MATRIX The Ideal Camera Matrix f 0 c x = w/2 0 f c y = h/2 0 0 1 f is focal length (c x, c y ) is center or principal point of image within image plane (w, h) are width and height of image plane (θ, φ) are horizontal and vertical field of view (FOV) f = w 4 + h 4 4 2 tan θ + tan φ 2 2 4

THE CAMERA MATRIX The Ideal Camera Matrix f 0 c x = w/2 0 f c y = h/2 0 0 1 f is focal length (c x, c y ) is center or principal point of image within image plane (w, h) are width and height of image plane (θ, φ) are horizontal and vertical field of view (FOV) f = w 4 + h 4 4 4 2 tan θ + tan φ 2 2 Units must be consistent, e.g.: ü f and (cx, cy) are all in mm ü Or, f and (cx, cy) are all in pixels Spec sheets may give lens s f in mm and image sensor s (w, h) in mm Or, APIs may give (θ, φ) in degrees or radians and image s (w, h) in pixels Camera API in Android SDK

THE CAMERA MATRIX The Ideal Camera Matrix f 0 c x = w/2 0 f c y = h/2 0 0 1 f is focal length (c x, c y ) is center or principal point of image within image plane (w, h) are width and height of image plane (θ, φ) are horizontal and vertical field of view (FOV) f = w 4 + h 4 4 4 2 tan θ + tan φ 2 2 f is useful in calculating size or distance For ideal lens and camera, K LMNOP Q s image is object s size in image = K RPNS T, where: e.g. in pixels, or in mm on sensor surface S real is object s real size d is distance between camera and object

DISTORTION COEFFICIENTS The Ideal Distortion Coefficients k 1 = 0 k 2 = 0 p 1 = 0 p 2 = 0 k 3 = 0 k n is the nth radial distortion coefficient k 1 < 0 usually implies barrel distortion k 1 > 0 usually implies pincushion distortion Changing sign across k n series may imply moustache distortion Barrel Distortion Pincushion Distortion Moustache Distortion p n is the nth tangential distortion coefficient Sign depends on direction of lens s tilt relative to image plane

DISTORTION COEFFICIENTS The Ideal Distortion Coefficients k 1 = 0 k 2 = 0 p 1 = 0 p 2 = 0 k 3 = 0 k n is the nth radial distortion coefficient k 1 < 0 usually implies barrel distortion k 1 > 0 usually implies pincushion distortion Changing sign across k n series may imply moustache distortion p n is the nth tangential distortion coefficient Sign depends on direction of lens s tilt relative to image plane Rarely, lens manufacturer may specify distortion coefficients in spec sheets or code samples Or, third-party libraries may provide distortion coefficients for various lenses: lensfun: http://lensfun.sourceforge.net Python wrapper, lensfunpy: https://github.com/letmaik/lensfunpy Interoperable with OpenCV and SciPy Or, we may have to use calibration process Chessboard calibration in OpenCV

COMPUTATIONAL PERFORMANCE BUDGETING OPERATIONS PER PIXEL PER FRAME COMPARING ULTRA-COMPACT COMPUTERS

BUDGETING OPERATIONS PER PIXEL PER FRAME Peak performance is often specified in FLOPS: floating point operations per second 1 GFLOPS = 1 billion FLOPS Beware, not all FLOPS are equal! Precision may be half (16-bit), single (32-bit), or double (64-bit) Different architectures have different operations Number of FLOPS in higher-level functions, e.g. in OpenCL, varies depending on drivers

BUDGETING OPERATIONS PER PIXEL PER FRAME For a given camera and computer, b = 2 V W X, where: b is the budget in floating point operations per pixel per frame p is the computer s peak performance in FLOPS v is the camera s frequency, i.e. the FPS, i.e. the frame rate in Hz (w, h) are the width and height of the image in pixels

BUDGETING OPERATIONS PER PIXEL PER FRAME Example: Suppose we capture frames from a Point Grey GS3-U3-23S6C-C camera, with 1920x1200 pixels @ 163 FPS. For an Intel Iris Pro Graphics 580 GPU, capable of 1,152 GFLOPS: b =,.,74,6YZ = 3067 floating-point operations per pixel per frame,[\,]46,466 For an AMD HD 8210E GPU, capable of 85 GFLOPS: b = >.7,6 Y`,[\,]46,466 = 226 floating-point operations per pixel per frame

COMPARING ULTRA-COMPACT COMPUTERS: X86 System Intel NUC Kit NUC6i7KYK Skull Canyon Camera Interfaces USB 3.0 + USB 3.1, Thunderbolt, Ethernet Gizmo 2 USB 2.0 + USB 3.0, Ethernet CPU GPU Peak GPU Performance (Float32) Quad-core i7-6770hq Dual-core Jaguar GX- 210HA Iris Pro Graphics 580, 72 execution units, OpenCL 2.0, 128 MB edram HD 8210E, 128 stream processors, OpenCL 1.2 1,152 GFLOPS @ 1,000 MHz 85 GFLOPS @ 300 MHz Peak Power Use* Price 85 W US$595 9 W US$199 * Excluding peripherals

COMPARING ULTRA-COMPACT COMPUTERS: X86 System Intel NUC Kit NUC6i7KYK Skull Canyon Camera Interfaces USB 3.0 + USB 3.1, Thunderbolt, Ethernet Gizmo 2 USB 2.0 + USB 3.0, Ethernet CPU GPU Peak GPU Performance (Float32) Quad-core i7-6770hq Dual-core Jaguar GX- 210HA Iris Pro Graphics 580, 72 execution units, OpenCL 2.0, 128 MB edram HD 8210E, 128 stream processors, OpenCL 1.2 1,152 GFLOPS @ 1,000 MHz 85 GFLOPS @ 300 MHz Peak Power Use* Price 85 W US$595 9 W US$199 Real-world performance difference in one of my OpenCV applications is 5x but not 13x as GFLOPS specs suggest * Excluding peripherals

COMPARING ULTRA-COMPACT COMPUTERS: ARM System NVIDIA Jetson TX2 Camera Interfaces USB 2.0 + USB 3.0, CSI, Ethernet Odroid-XU4 USB 2.0 + USB 3.0, Ethernet CPU GPU Peak GPU Performance (Float32) Dual-core Denver2 + quad-core Cortex-A57 Quad-core Cortex-A15 + quad-core Cortex-A7 NVIDIA Pascal, 256 CUDA cores Mali-T624, 6 cores, OpenCL 1.2 750 GFLOPS @ 1,465 MHz 142 GFLOPS @ 695 MHz Peak Power Use* Price 15 W US$599 16 W US$59 * Excluding peripherals

CONCLUSIONS Feasibility assessments should include: Spatial resolution: lp/mm, diffraction, pixel pitch, magnification level Temporal resolution: speed of subject, need for redundancy A detector s miss rate decreases exponentially with temporal resolution Camera matrix: availability of data on either focal length or FOV Distortion coefficients: availability of either reference data or run-time calibration results Computational performance: GFLOPS, operations per pixel per frame A good lens needs a good camera A good camera needs a good processor and good software optimizations

Here be underscores QUESTIONS? JOSEPH HOWSE, NUMMISTMEDIA HTTP://NUMMIST.COM JOSEPHHOWSE@NUMMIST.COM