Computer Vision. Thursday, August 30

Similar documents
Unit 1: Image Formation

Computer Vision Lecture 1

CPSC 425: Computer Vision

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2015 Version 3

Today. CS 395T Visual Recognition. Course content. Administration. Expectations. Paper reviews

Image Formation III Chapter 1 (Forsyth&Ponce) Cameras Lenses & Sensors

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2014 Version 1

Cameras. CSE 455, Winter 2010 January 25, 2010

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

Lecture 02 Image Formation 1

How do we see the world?

Computer Vision. The Pinhole Camera Model

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

VC 11/12 T2 Image Formation

VC 14/15 TP2 Image Formation

Today I t n d ro ucti tion to computer vision Course overview Course requirements

ME 6406 MACHINE VISION. Georgia Institute of Technology

CSE 527: Introduction to Computer Vision

VC 16/17 TP2 Image Formation

Projection. Announcements. Müller-Lyer Illusion. Image formation. Readings Nalwa 2.1

Digital image processing vs. computer vision Higher-level anchoring

CSE 473/573 Computer Vision and Image Processing (CVIP)

Projection. Readings. Szeliski 2.1. Wednesday, October 23, 13

Dr F. Cuzzolin 1. September 29, 2015

Lecture 2 Camera Models

Image Processing & Projective geometry

Two strategies for realistic rendering capture real world data synthesize from bottom up

Projection. Projection. Image formation. Müller-Lyer Illusion. Readings. Readings. Let s design a camera. Szeliski 2.1. Szeliski 2.

The Camera : Computational Photography Alexei Efros, CMU, Fall 2005

Announcements. Image Formation: Outline. The course. How Cameras Produce Images. Earliest Surviving Photograph. Image Formation and Cameras

6.098 Digital and Computational Photography Advanced Computational Photography. Bill Freeman Frédo Durand MIT - EECS

Image formation - Cameras. Grading & Project. About the course. Tentative Schedule. Course Content. Students introduction

Spring 2018 CS543 / ECE549 Computer Vision. Course webpage URL:

Digital Image Processing COSC 6380/4393

Image Formation: Camera Model

Image Formation. World Optics Sensor Signal. Computer Vision. Introduction to. Light (Energy) Source. Surface Imaging Plane. Pinhole Lens.

MEM455/800 Robotics II/Advance Robotics Winter 2009

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

CS6670: Computer Vision

The Camera : Computational Photography Alexei Efros, CMU, Fall 2008

Digital Image Processing COSC 6380/4393

To Do. Advanced Computer Graphics. Outline. Computational Imaging. How do we see the world? Pinhole camera

Range Sensing strategies

Computer Vision Slides curtesy of Professor Gregory Dudek

FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM

Computational Photography and Video. Prof. Marc Pollefeys

Practical Image and Video Processing Using MATLAB

LENSES. INEL 6088 Computer Vision

DIGITAL IMAGE PROCESSING

COMP 776: Computer Vision

CSCE 763: Digital Image Processing

CS 376b Computer Vision

Computer Vision Introduction or

Course Syllabus OSE 3200 Geometric Optics

ECEN 4606, UNDERGRADUATE OPTICS LAB

Single-view Metrology and Cameras

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

MIT CSAIL Advances in Computer Vision Fall Problem Set 6: Anaglyph Camera Obscura

Building a Real Camera. Slides Credit: Svetlana Lazebnik

Intorduction to light sources, pinhole cameras, and lenses

ELE 882: Introduction to Digital Image Processing (DIP)

A Geometric Correction Method of Plane Image Based on OpenCV

Background. Computer Vision & Digital Image Processing. Improved Bartlane transmitted image. Example Bartlane transmitted image

Digital Image Processing

GEOMETRICAL OPTICS AND OPTICAL DESIGN

Introduction. BIL719 Computer Vision Pinar Duygulu Hacettepe University

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu

Image Formation. Light from distant things. Geometrical optics. Pinhole camera. Chapter 36

Cameras, lenses and sensors

Princeton University COS429 Computer Vision Problem Set 1: Building a Camera

Course Syllabus OSE 3200 Geometric Optics

Cameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017

Single Camera Catadioptric Stereo System

Exercise questions for Machine vision

Lecture 2 Camera Models

College of Optics & Photonics

Chapter 23. Light Geometric Optics

Chapter 36. Image Formation

Image Formation and Capture

Cameras, lenses, and sensors

3D Viewing. Introduction to Computer Graphics Torsten Möller / Manfred Klaffenböck. Machiraju/Zhang/Möller

CIS581: Computer Vision and Computational Photography Homework: Cameras and Convolution Due: Sept. 14, 2017 at 3:00 pm

Lecture 8 Camera Models

Lecture 1 Introduction to Computer Vision. Lin ZHANG, PhD School of Software Engineering, Tongji University Spring 2014

EC-433 Digital Image Processing

CPSC 4040/6040 Computer Graphics Images. Joshua Levine

Perspective. Cornell CS4620/5620 Fall 2012 Lecture Kavita Bala 1 (with previous instructors James/Marschner)

Building a Real Camera

Opto Engineering S.r.l.

Lecture 1 Introduction to Computer Vision. Lin ZHANG, PhD School of Software Engineering, Tongji University Spring 2015

Focus on an optical blind spot A closer look at lenses and the basics of CCTV optical performances,

Image Processing - Intro. Tamás Szirányi

LENSLESS IMAGING BY COMPRESSIVE SENSING

Outline. Image formation: the pinhole camera model Images as functions Digital images Color, light and shading. Reading: textbook: 2.1, 2.2, 2.

Introduction. Ioannis Rekleitis

Virtual and Digital Cameras

CS6670: Computer Vision

Spherical Mirrors. Concave Mirror, Notation. Spherical Aberration. Image Formed by a Concave Mirror. Image Formed by a Concave Mirror 4/11/2014

Physics 4L Spring 2010 Problem set 1 Due Tuesday 26 January in class

Transcription:

Computer Vision Thursday, August 30 1

Today Course overview Requirements, logistics Image formation 2

Introductions Instructor: Prof. Kristen Grauman grauman @ cs TAY 4.118, Thurs 2-4 pm TA: Sudheendra Vijayanarasimhan svnaras @ cs ENS 31 NQ, Mon/Wed 1-2 pm Class page: Check for updates to schedule, assignments, etc. http://www.cs.utexas.edu/~grauman/courses/378/main.htm 3

Introductions 4

Computer vision Automatic understanding of images and video Computing properties of the 3D world from visual data Algorithms and representations to allow a machine to recognize objects, people, scenes, and activities. 5

Why vision? As image sources multiply, so do applications Relieve humans of boring, easy tasks Enhance human abilities Advance human-computer interaction, visualization Perception for robotics / autonomous agents Possible insights into human vision 6

Some applications Factory inspection (Cognex) Monitoring for safety (Poseidon) Surveillance Visualization and tracking License plate reading Visualization 7

Some applications Autonomous robots Navigation, driver safety Assistive technology Visual effects (the Matrix) Medical imaging 8

Some applications Multi-modal interfaces Situated search Image and video databases - CBIR Tracking, activity recognition 9

Why is vision difficult? Ill-posed problem: real world much more complex than what we can measure in images 3D 2D Impossible to literally invert image formation process 10

Challenges: robustness Illumination Object pose Clutter Occlusions Intra-class appearance Viewpoint 11

Challenges: context and human experience Context cues Function Dynamics 12

Challenges: complexity Thousands to millions of pixels in an image 3,000-30,000 human recognizable object categories 30+ degrees of freedom in the pose of articulated objects (humans) Billions of images indexed by Google Image Search 18 billion+ prints produced from digital camera images in 2004 295.5 million camera phones sold in 2005 About half of the cerebral cortex in primates is devoted to processing visual information [Felleman and van Essen 1991] 13

Why is vision difficult? Ill-posed problem: real world much more complex than what we can measure in images 3D 2D Not possible to invert image formation process Generally requires assumptions, constraints; exploitation of domainspecific knowledge 14

Related disciplines Geometry, physics Image processing Artificial intelligence Computer vision Algorithms Pattern recognition Cognitive science 15

Vision and graphics Images Vision Model Graphics Inverse problems: analysis and synthesis. 16

Research problems vs. application areas Feature detection Contour representation Segmentation Stereo vision Shape modeling Color vision Motion analysis Invariants Uncalibrated, selfcalibrating systems Object detection Object recognition Industrial inspection and quality control Reverse engineering Surveillance and security Face, gesture recognition Road monitoring Autonomous vehicles Military applications Medical image analysis Image databases Virtual reality List from [Trucco & Verri 1998] 17

Goals of this course Introduction to primary topics Hands-on experience with algorithms Views of vision as a research area 18

Topics overview Image formation, cameras Color Features Grouping Multiple views Recognition and learning Motion and tracking 19

We will not cover (extensively) Image processing Human visual system Particular machine vision systems or applications 20

Image formation Inverse process of vision: how does light in 3d world project to form 2d images? 21

Features and filters Transforming and describing images; textures and colors 22

Grouping Clustering, segmentation, fitting; what parts belong together? [fig from Shi et al] 23

Multiple views Lowe Hartley and Zisserman Multi-view geometry and matching, stereo Tomasi and Kanade 24

Recognition and learning Shape matching, recognizing objects and categories, learning techniques 25

Motion and tracking Tracking objects, video analysis, low level motion Tomas Izo 26

27

Requirements Biweekly (approx) problem sets Concept questions Implementation problems Two exams, midterm and final Current events (optional) In addition, for graduate students: Research paper summary and review Implementation extension 28

Grading policy Final grade breakdown: Problem sets (50%) Midterm quiz (15%) Final exam (20%) Class participation (15%) 29

Due dates Assignments due before class starts on due date Lose half of possible remaining credit each day late Three free late days, total 30

Collaboration policy You are welcome to discuss problem sets, but all responses and code must be written individually. Students submitting solutions found to be identical or substantially similar (due to inappropriate collaboration) risk failing the course. 31

Current events (optional) Any vision-related piece of news; may revolve around policy, editorial, technology, new product, Brief overview to the class Must be current No ads Email relevant links or information to TA 32

Paper review guidelines Thorough summary in your own words Main contribution Strengths? Weaknesses? How convincing are the experiments? Suggestions to improve them? Extensions? 4 pages max May require reading additional references 33

Miscellaneous Check class website Make sure you get on class mailing list No laptops in class please Feedback welcome and useful 34

35

Image formation How are objects in the world captured in an image? 36

Physical parameters of image formation Photometric Type, direction, intensity of light reaching sensor Surfaces reflectance properties Optical Sensor s lens type focal length, field of view, aperture Geometric Type of projection Camera pose Perspective distortions 37

Radiometry Images formed depend on amount of light from light sources and surface reflectance properties (See F&P Ch 4) 38

Light source direction Image credit: Don Deering 39

Surface reflectance properties Specular [fig from Fleming, Torralba, & Adelson, 2004] Lambertian 40

Perspective projection Pinhole camera: simple model to approximate imaging process Forsyth and Ponce If we treat pinhole as a point, only one ray from any given point can enter the camera 41

Camera obscura In Latin, means dark room "Reinerus Gemma-Frisius, observed an eclipse of the sun at Louvain on January 24, 1544, and later he used this illustration of the event in his book De Radio Astronomica et Geometrica, 1545. It is thought to be the first published illustration of a camera obscura..." Hammond, John H., The Camera Obscura, A Chronicle http://www.acmi.net.au/aic/camera_obscura.html 42

Camera obscura Jetty at Margate England, 1898. An attraction in the late 19 th century Around 1870s http://brightbytes.com/cosite/collection2.html Adapted from R. Duraiswami 43

Perspective effects Far away objects appear smaller Forsyth and Ponce 44

Perspective effects Parallel lines in the scene intersect in the image Forsyth and Ponce 45

Perspective projection equations 3d world mapped to 2d projection Image plane Focal length Camera frame Optical axis Board Forsyth and Ponce 46

Perspective projection equations Image plane Focal length Camera frame Optical axis Non-linear Scene point Image coordinates Forsyth and Ponce 47

Projection properties Many-to-one: any points along same ray map to same point in image Points points Lines lines (collinearity preserved) Distances and angles are not preserved Degenerate cases: Line through focal point projects to a point. Plane through focal point projects to line Plane perpendicular to image plane projects to part of the image. 48

Perspective and art Use of correct perspective projection indicated in 1 st century B.C. frescoes Skill resurfaces in Renaissance: artists develop systematic methods to determine perspective projection (around 1480-1515) Raphael Durer, 1525 49

Weak perspective Approximation: treat magnification as constant Assumes scene depth << average distance to camera Makes perspective equations linear Image plane World points: 50

Orthographic projection Given camera at constant distance from scene World points projected along rays parallel to optical access Limit of perspective projection as 51

Planar pinhole perspective Orthographic projection From M. Pollefeys 52

Which projection model? Weak perspective: Accurate for small, distant objects; recognition Linear projection equations - simplifies math Pinhole perspective: More accurate but more complex Structure from motion 53

54

Pinhole size / aperture Larger Brighter, blurrier Dimmer, more focus Smaller Dimmer, blur from defraction 55

Pinhole vs. lens 56

Cameras with lenses Gather more light, while keeping focus; make pinhole perspective projection practical Left focus Thin lens Right focus Rays entering parallel on one side go through focus on other, and vice versa. In ideal case all rays from P imaged at P. Lens diameter d Focal length f Field of view (portion of 3d space seen by camera) depends on d and f. 57

Field of view As f gets smaller, image becomes more wide angle (more world points project onto the finite image plane) As f gets larger, image becomes more telescopic (smaller part of the world projects onto the finite image plane) from R. Duraiswami 58

Focus and depth of field Depth of field: distance between image planes where blur is tolerable Thin lens: scene points at distinct depths come in focus at different image planes. (Real camera lens systems have greater depth of field.) circles of confusion Shapiro and Stockman 59

Focus and depth of field Image credit: cambridgeincolour.com 60

Depth from focus Images from same point of view, different camera parameters 3d shape / depth estimates [figs from H. Jin and P. Favaro, 2002] 61

Camera parameters How do points in real world relate to positions in the image? Perspective equations so far in terms of camera s reference frame 62

Camera parameters Need to estimate camera s intrinsic and extrinsic parameters to calibrate geometry. World frame Extrinsic: Camera frame World frame Camera frame Intrinsic: Image coordinates relative to camera Pixel coordinates 63

Camera calibration Extrinsic params: rotation matrix and translation vector Intrinsic params: focal length, pixel sizes (mm), image center point, radial distortion parameters Knowing the relationship between real world and image coordinates useful for estimating 3d shape More on this later 64

Articulated tracking Demirdjian et al. 65

3d skeleton extraction Brostow et al, 2004 66

Human eye Pupil/Iris control amount of light passing through lens Retina - contains sensor cells, where image is formed Fovea highest concentration of cones Shapiro and Stockman 67

Sensors Often CCD camera: charge coupled device Record amount of light reaching grid photosensors, which convert light energy into voltage Read digital output row-by-row CCD array camera optics frame grabber computer 68

Think of images as matrices taken from CCD array. Digital images 69

Digital images Intensity : [0,255] j=1 width 520 i=0 500 height im[176][201] has value 164 im[194][203] has value 37 70

Color images, RGB color space R G B 71

Resolution sensor: size of real world scene element a that images to a single pixel image: number of pixels Influences what analysis is feasible, affects best representation choice. [Mori et al] 72

Resolution though not necessarily for the human visual system with familiar faces [Sinha et al] 73

Stereo cameras MRI scans Xray LIDAR devices Other sensors [Jim Gasperini] geospatial-online.com 74

Summary Image formation affected by geometry, photometry, and optics. Projection equations express how world points mapped to 2d image. Lenses make pinhole model practical. Imaged points related to real world coordinates via calibrated cameras. 75

Next Problem set 0 due Sept 6 Matlab warmup Image formation questions Read F&P Chapter 1 Reading for next lecture: F&P Chapter 6 76