# Image and Multidimensional Signal Processing

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 Image and Multidimensional Signal Processing Professor William Hoff Dept of Electrical Engineering &Computer Science

2 Digital Image Fundamentals 2

3 Digital Image Fundamentals Topics Image acquisition and formation Resolution and size of objects Spatial sampling and quantization Matlab exercises Objectives Know how to estimate properties of image acquisition and formation, such as resolution and the size of objects that can be discerned Be familiar with the effects of spatial sampling and quantization on the quality of an image and its storage size 3

4 Notes: Eyeball is about 20 mm in diameter Retina contains both rods and cones Fovea is about 1.5 mm in width, contains about 337,000 cones Focal length about 17 mm 4

5 Rods sensitive to low light (scotopic vision) Cones detect color, work in bright light (photopic vision) 5

6 Image Formation h Treat as pinhole camera good approximation Similar triangles H f h What is the size of the tree on the retina? What if the tree were twice as far away? 15/100 = h/17 or h=2.55 mm 6

7 Problem 2.1 Assume the retina Contains 337,000 sensor elements arranged in a square array, 1.5x1.5 mm And the space between cones is equal to width of cone Estimate the diameter of the smallest printed dot that the eye can discern if the page on which the dot is printed is 0.2 m away from the eyes Assume the visual system ceases to detect the dot when the image of the dot on the fovea becomes smaller than the diameter of one receptor (cone) in that area of the retina 7

8 8

9 Light and the Electromagnetic Spectrum Wavelength (l) and frequency (n) are related by l c To resolve an object, we must use a wavelength equal to the size of the object or smaller 9

10 10

11 CCD (Charge coupled device) A 2D array of photosensitive transistors Charge accumulates during exposure Charges are transferred out to shift registers, read out sequentially Output is an analog signal (eg. RS-170, NTSC, PAL format) Can also be digitized and output in digital (via Firewire, USB, etc) 11

12 Typical CCD cameras Example: Panasonic GP-MF130 Sensor is 6x5 mm, 768x494 pixels Typical lens ~ 6mm focal length Field of view (see next slide)? 12

13 f = focal length q = field of view h = image plane size Field of View q/2 H/2 Center of projection f Image plane h/2 D Object plane being imaged tan(q/2) = (h/2) / f = (H/2) / D Horizontal field of view: tan(q/2) = (6mm/2) /(6mm) = 0.5 q = 2 arctan(0.5) = 53 degrees f, h can be in pixels or mm Horizontal, vertical fov could be different 13

14 Image Acquisition Using Sensor Strips Images can also be formed from (a) A linear array of sensors that is swept (b) A circular array, where the object is moved 14

15 Example Satellite Camera NOAA weather satellite KLM series for polar (low Earth) orbits 15

16 Image Formation 1 ), ( 0 ), ( 0 ), ( ), ( ), ( y x r y x i y x r y x i y x f Components Illumination i(x,y) Reflection r(x,y) 16

17 Sampling and Quantization N bits per pixel allows 2^N values 17

18 Examples An image has 8 bits per pixel If unsigned, range of values is? If signed (ie., two s complement), range is? Number of bytes in a 3872 x 2592 pixel image? 18

19 19

20 Image Representation Image can be represented by an xy plane Sampling partitions the plane into a grid of pixels, whose indices are integers We can also think of it as an MxN matrix of numbers We can index an element either by (row,col) or (x,y) (1,1) row (or y im ) col (or x im ) Image We will use the convention Top left pixel is (1,1) x index (or column) increases to the right y index (or row) increases down 20

21 Summary In digital cameras, the image is projected onto a 2D array of sensor elements. The image is (spatially) sampled by the sensor elements, and the intensities are quantized into discrete values. We can use the pinhole camera model to estimate the field of view and the size of objects projected onto the image. 21

22 Questions About how small an object can you see, using (a) visible light, (b) X-rays, or (c) microwaves? How does the human eye compare with a digital camera (e.g., in terms of resolution and sensitivity)? How could you measure the focal length of a camera? What units can it be expressed in? 22

### Digital Image Processing

Digital Image Processing Digital Imaging Fundamentals Christophoros Nikou cnikou@cs.uoi.gr Images taken from: R. Gonzalez and R. Woods. Digital Image Processing, Prentice Hall, 2008. Digital Image Processing

### Digital Image Fundamentals. Digital Image Processing. Human Visual System. Contents. Structure Of The Human Eye (cont.) Structure Of The Human Eye

Digital Image Processing 2 Digital Image Fundamentals Digital Imaging Fundamentals Christophoros Nikou cnikou@cs.uoi.gr Those who wish to succeed must ask the right preliminary questions Aristotle Images

### Digital Image Processing

Digital Image Processing Digital Imaging Fundamentals Christophoros Nikou cnikou@cs.uoi.gr Images taken from: R. Gonzalez and R. Woods. Digital Image Processing, Prentice Hall, 2008. Digital Image Processing

### Human Visual System. Digital Image Processing. Digital Image Fundamentals. Structure Of The Human Eye. Blind-Spot Experiment.

Digital Image Processing Digital Imaging Fundamentals Christophoros Nikou cnikou@cs.uoi.gr 4 Human Visual System The best vision model we have! Knowledge of how images form in the eye can help us with

### Chapter 2: Digital Image Fundamentals. Digital image processing is based on. Mathematical and probabilistic models Human intuition and analysis

Chapter 2: Digital Image Fundamentals Digital image processing is based on Mathematical and probabilistic models Human intuition and analysis 2.1 Visual Perception How images are formed in the eye? Eye

### Digital Image Processing

Digital Image Processing Lecture # 3 Digital Image Fundamentals ALI JAVED Lecturer SOFTWARE ENGINEERING DEPARTMENT U.E.T TAXILA Email:: ali.javed@uettaxila.edu.pk Office Room #:: 7 Presentation Outline

### Visual perception basics. Image aquisition system. IE PŁ P. Strumiłło

Visual perception basics Image aquisition system Light perception by humans Humans perceive approx. 90% of information about the environment by means of visual system. Efficiency of the human visual system

### Image Processing (EA C443)

Image Processing (EA C443) OBJECTIVES: To study components of the Image (Digital Image) To Know how the image quality can be improved How efficiently the image data can be stored and transmitted How the

### Graphics and Image Processing Basics

EST 323 / CSE 524: CG-HCI Graphics and Image Processing Basics Klaus Mueller Computer Science Department Stony Brook University Julian Beever Optical Illusion: Sidewalk Art Julian Beever Optical Illusion:

### Vision. The eye. Image formation. Eye defects & corrective lenses. Visual acuity. Colour vision. Lecture 3.5

Lecture 3.5 Vision The eye Image formation Eye defects & corrective lenses Visual acuity Colour vision Vision http://www.wired.com/wiredscience/2009/04/schizoillusion/ Perception of light--- eye-brain

### CSCE 763: Digital Image Processing

CSCE 763: Digital Image Processing Spring 2018 Yan Tong Department of Computer Science and Engineering University of South Carolina Today s Agenda Welcome Tentative Syllabus Topics covered in the course

### ME 6406 MACHINE VISION. Georgia Institute of Technology

ME 6406 MACHINE VISION Georgia Institute of Technology Class Information Instructor Professor Kok-Meng Lee MARC 474 Office hours: Tues/Thurs 1:00-2:00 pm kokmeng.lee@me.gatech.edu (404)-894-7402 Class

### COURSE ECE-411 IMAGE PROCESSING. Er. DEEPAK SHARMA Asstt. Prof., ECE department. MMEC, MM University, Mullana.

COURSE ECE-411 IMAGE PROCESSING Er. DEEPAK SHARMA Asstt. Prof., ECE department. MMEC, MM University, Mullana. Why Image Processing? For Human Perception To make images more beautiful or understandable

### Capturing Light in man and machine

Capturing Light in man and machine CS194: Image Manipulation & Computational Photography Alexei Efros, UC Berkeley, Fall 2014 Etymology PHOTOGRAPHY light drawing / writing Image Formation Digital Camera

### Assignment: Light, Cameras, and Image Formation

Assignment: Light, Cameras, and Image Formation Erik G. Learned-Miller February 11, 2014 1 Problem 1. Linearity. (10 points) Alice has a chandelier with 5 light bulbs sockets. Currently, she has 5 100-watt

### Digital Image Processing COSC 6380/4393

Digital Image Processing COSC 6380/4393 Lecture 2 Aug 24 th, 2017 Slides from Dr. Shishir K Shah, Rajesh Rao and Frank (Qingzhong) Liu 1 Instructor TA Digital Image Processing COSC 6380/4393 Pranav Mantini

### The Science Seeing of process Digital Media. The Science of Digital Media Introduction

The Human Science eye of and Digital Displays Media Human Visual System Eye Perception of colour types terminology Human Visual System Eye Brains Camera and HVS HVS and displays Introduction 2 The Science

### IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

IMAGE FORMATION Light source properties Sensor characteristics Surface Exposure shape Optics Surface reflectance properties ANALOG IMAGES An image can be understood as a 2D light intensity function f(x,y)

### 2013 LMIC Imaging Workshop. Sidney L. Shaw Technical Director. - Light and the Image - Detectors - Signal and Noise

2013 LMIC Imaging Workshop Sidney L. Shaw Technical Director - Light and the Image - Detectors - Signal and Noise The Anatomy of a Digital Image Representative Intensities Specimen: (molecular distribution)

### Image Formation: Camera Model

Image Formation: Camera Model Ruigang Yang COMP 684 Fall 2005, CS684-IBMR Outline Camera Models Pinhole Perspective Projection Affine Projection Camera with Lenses Digital Image Formation The Human Eye

### Basic principles of photography. David Capel 346B IST

Basic principles of photography David Capel 346B IST Latin Camera Obscura = Dark Room Light passing through a small hole produces an inverted image on the opposite wall Safely observing the solar eclipse

### Sensors and Sensing Cameras and Camera Calibration

Sensors and Sensing Cameras and Camera Calibration Todor Stoyanov Mobile Robotics and Olfaction Lab Center for Applied Autonomous Sensor Systems Örebro University, Sweden todor.stoyanov@oru.se 20.11.2014

### 2. Pixels and Colors. Introduction to Pixels. Chapter 2. Investigation Pixels and Digital Images

2. Pixels and Colors Introduction to Pixels The term pixel is a truncation of the phrase picture element which is exactly what a pixel is. A pixel is the smallest block of color in a digital picture. The

### Geometry of Aerial Photographs

Geometry of Aerial Photographs Aerial Cameras Aerial cameras must be (details in lectures): Geometrically stable Have fast and efficient shutters Have high geometric and optical quality lenses They can

### Capturing light and color

Capturing light and color Friday, 10/02/2017 Antonis Argyros e-mail: argyros@csd.uoc.gr Szeliski 2.2, 2.3, 3.1 1 Recap from last lecture Pinhole camera model Perspective projection Focal length and depth/field

### Physics 2020 Lab 9 Wave Interference

Physics 2020 Lab 9 Wave Interference Name Section Tues Wed Thu 8am 10am 12pm 2pm 4pm Introduction Consider the four pictures shown below, showing pure yellow lights shining toward a screen. In pictures

### Introduction & Colour

Introduction & Colour Eric C. McCreath School of Computer Science The Australian National University ACT 0200 Australia ericm@cs.anu.edu.au Overview 2 Computer Graphics Uses (Chapter 1) Basic Hardware

### Background. Computer Vision & Digital Image Processing. Improved Bartlane transmitted image. Example Bartlane transmitted image

Background Computer Vision & Digital Image Processing Introduction to Digital Image Processing Interest comes from two primary backgrounds Improvement of pictorial information for human perception How

### Image Acquisition and Representation. Camera. CCD Camera. Image Acquisition Hardware

Image Acquisition and Representation Camera Slide 1 how digital images are produced how digital images are represented Slide 3 First photograph was due to Niepce of France in 1827. Basic abstraction is

### Colour (1) Graphics 2

Colour (1) raphics 2 06-02408 Level 3 10 credits in Semester 2 Professor Aleš Leonardis Slides by Professor Ela Claridge Colours and their origin - spectral characteristics - human visual perception Colour

### 3/5/17. Detector Basics. Quantum Efficiency (QE) and Spectral Response. Quantum Efficiency (QE) and Spectral Response

3/5/17 Detector Basics The purpose of any detector is to record the light collected by the telescope. All detectors transform the incident radiation into a some other form to create a permanent record,

### IMAGE SENSOR SOLUTIONS. KAC-96-1/5" Lens Kit. KODAK KAC-96-1/5" Lens Kit. for use with the KODAK CMOS Image Sensors. November 2004 Revision 2

KODAK for use with the KODAK CMOS Image Sensors November 2004 Revision 2 1.1 Introduction Choosing the right lens is a critical aspect of designing an imaging system. Typically the trade off between image

### A CAMERA IS A LIGHT TIGHT BOX

HOW CAMERAS WORK A CAMERA IS A LIGHT TIGHT BOX Pinhole Principle All contemporary cameras have the same basic features A light-tight box to hold the camera parts and recording material A viewing system

### Image and Video Processing

Image and Video Processing () Image Representation Dr. Miles Hansard miles.hansard@qmul.ac.uk Segmentation 2 Today s agenda Digital image representation Sampling Quantization Sub-sampling Pixel interpolation

### Sharpness, Resolution and Interpolation

Sharpness, Resolution and Interpolation Introduction There are a lot of misconceptions about resolution, camera pixel count, interpolation and their effect on astronomical images. Some of the confusion

### CSE1710. Big Picture. Reminder

CSE1710 Click to edit Master Week text 10, styles Lecture 19 Second level Third level Fourth level Fifth level Fall 2013 Thursday, Nov 14, 2013 1 Big Picture For the next three class meetings, we will

### Important Missions. weather forecasting and monitoring communication navigation military earth resource observation LANDSAT SEASAT SPOT IRS

Fundamentals of Remote Sensing Pranjit Kr. Sarma, Ph.D. Assistant Professor Department of Geography Mangaldai College Email: prangis@gmail.com Ph. No +91 94357 04398 Remote Sensing Remote sensing is defined

### Determining MTF with a Slant Edge Target ABSTRACT AND INTRODUCTION

Determining MTF with a Slant Edge Target Douglas A. Kerr Issue 2 October 13, 2010 ABSTRACT AND INTRODUCTION The modulation transfer function (MTF) of a photographic lens tells us how effectively the lens

### MATLAB Image Processing Toolbox

MATLAB Image Processing Toolbox Copyright: Mathworks 1998. The following is taken from the Matlab Image Processing Toolbox users guide. A complete online manual is availabe in the PDF form (about 5MB).

### Int n r t o r d o u d c u ti t on o n to t o Remote Sensing

Introduction to Remote Sensing Definition of Remote Sensing Remote sensing refers to the activities of recording/observing/perceiving(sensing)objects or events at far away (remote) places. In remote sensing,

### Telescopes and their configurations. Quick review at the GO level

Telescopes and their configurations Quick review at the GO level Refraction & Reflection Light travels slower in denser material Speed depends on wavelength Image Formation real Focal Length (f) : Distance

### Visual Perception of Images

Visual Perception of Images A processed image is usually intended to be viewed by a human observer. An understanding of how humans perceive visual stimuli the human visual system (HVS) is crucial to the

### INTRODUCTION THIN LENSES. Introduction. given by the paraxial refraction equation derived last lecture: Thin lenses (19.1) = 1. Double-lens systems

Chapter 9 OPTICAL INSTRUMENTS Introduction Thin lenses Double-lens systems Aberrations Camera Human eye Compound microscope Summary INTRODUCTION Knowledge of geometrical optics, diffraction and interference,

### Mahdi Amiri. March Sharif University of Technology

Course Presentation Multimedia Systems Color Space Mahdi Amiri March 2014 Sharif University of Technology The wavelength λ of a sinusoidal waveform traveling at constant speed ν is given by Physics of

### Refraction, Lenses, and Prisms

CHAPTER 16 14 SECTION Sound and Light Refraction, Lenses, and Prisms KEY IDEAS As you read this section, keep these questions in mind: What happens to light when it passes from one medium to another? How

### Human Retina. Sharp Spot: Fovea Blind Spot: Optic Nerve

I am Watching YOU!! Human Retina Sharp Spot: Fovea Blind Spot: Optic Nerve Human Vision Optical Antennae: Rods & Cones Rods: Intensity Cones: Color Energy of Light 6 10 ev 10 ev 4 1 2eV 40eV KeV MeV Energy

### How does prism technology help to achieve superior color image quality?

WHITE PAPER How does prism technology help to achieve superior color image quality? Achieving superior image quality requires real and full color depth for every channel, improved color contrast and color

### ROBOT VISION. Dr.M.Madhavi, MED, MVSREC

ROBOT VISION Dr.M.Madhavi, MED, MVSREC Robotic vision may be defined as the process of acquiring and extracting information from images of 3-D world. Robotic vision is primarily targeted at manipulation

### INTRODUCTION TO CCD IMAGING

ASTR 1030 Astronomy Lab 85 Intro to CCD Imaging INTRODUCTION TO CCD IMAGING SYNOPSIS: In this lab we will learn about some of the advantages of CCD cameras for use in astronomy and how to process an image.

### Chapter 20 Human Vision

Chapter 20 GOALS When you have mastered the contents of this chapter, you will be able to achieve the following goals: Characterize the physical parameters that are significant in human vision. Visual

### Photography (cont d)

Lecture 13 Ch. 4 Photography continued Ch. 5 The Eye Feb. 23, 2010 Exams will be back on Feb. 25 Homework 5 is due Feb. 25 Read all of Ch. 5. on The Eye. 1 Photography (cont d) Polarizing and haze filters

### Visual System I Eye and Retina

Visual System I Eye and Retina Reading: BCP Chapter 9 www.webvision.edu The Visual System The visual system is the part of the NS which enables organisms to process visual details, as well as to perform

### Optics B. Science Olympiad North Regional Tournament at the University of Florida DO NOT WRITE ON THIS BOOKLET. THIS IS AN TEST SET.

Optics B Science Olympiad North Regional Tournament at the University of Florida 1 DO NOT WRITE ON THIS BOOKLET. THIS IS AN TEST SET. Part I: General Body Knowledge Questions 2 1) (3 PTS) For much of the

### Education in Microscopy and Digital Imaging

Contact Us Carl Zeiss Education in Microscopy and Digital Imaging ZEISS Home Products Solutions Support Online Shop ZEISS International ZEISS Campus Home Interactive Tutorials Basic Microscopy Spectral

### Image acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor

Image acquisition Digital images are acquired by direct digital acquisition (digital still/video cameras), or scanning material acquired as analog signals (slides, photographs, etc.). In both cases, the

### FRAUNHOFER AND FRESNEL DIFFRACTION IN ONE DIMENSION

FRAUNHOFER AND FRESNEL DIFFRACTION IN ONE DIMENSION Revised November 15, 2017 INTRODUCTION The simplest and most commonly described examples of diffraction and interference from two-dimensional apertures

### CMVision and Color Segmentation. CSE398/498 Robocup 19 Jan 05

CMVision and Color Segmentation CSE398/498 Robocup 19 Jan 05 Announcements Please send me your time availability for working in the lab during the M-F, 8AM-8PM time period Why Color Segmentation? Computationally

### Measurement of Visual Resolution of Display Screens

Measurement of Visual Resolution of Display Screens Michael E. Becker Display-Messtechnik&Systeme D-72108 Rottenburg am Neckar - Germany Abstract This paper explains and illustrates the meaning of luminance

### E X P E R I M E N T 12

E X P E R I M E N T 12 Mirrors and Lenses Produced by the Physics Staff at Collin College Copyright Collin College Physics Department. All Rights Reserved. University Physics II, Exp 12: Mirrors and Lenses

### NFMS THEORY LIGHT AND COLOR MEASUREMENTS AND THE CCD-BASED GONIOPHOTOMETER. Presented by: January, 2015 S E E T H E D I F F E R E N C E

NFMS THEORY LIGHT AND COLOR MEASUREMENTS AND THE CCD-BASED GONIOPHOTOMETER Presented by: January, 2015 1 NFMS THEORY AND OVERVIEW Contents Light and Color Theory Light, Spectral Power Distributions, and

### Εισαγωγική στην Οπτική Απεικόνιση

Εισαγωγική στην Οπτική Απεικόνιση Δημήτριος Τζεράνης, Ph.D. Εμβιομηχανική και Βιοϊατρική Τεχνολογία Τμήμα Μηχανολόγων Μηχανικών Ε.Μ.Π. Χειμερινό Εξάμηνο 2015 Light: A type of EM Radiation EM radiation:

### CS101 Lecture 12: Digital Images. What You ll Learn Today

CS101 Lecture 12: Digital Images Sampling and Quantizing Using bits to Represent Colors and Images Aaron Stevens (azs@bu.edu) 20 February 2013 What You ll Learn Today What is digital information? How to

### 6.098/6.882 Computational Photography 1. Problem Set 1. Assigned: Feb 9, 2006 Due: Feb 23, 2006

6.098/6.882 Computational Photography 1 Problem Set 1 Assigned: Feb 9, 2006 Due: Feb 23, 2006 Note The problems marked with 6.882 only are for the students who register for 6.882. (Of course, students

### 9/12/2011. Training Course Remote Sensing Basic Theory & Image Processing Methods September 2011

Training Course Remote Sensing Basic Theory & Image Processing Methods 19 23 September 2011 Remote Sensing Platforms Michiel Damen (September 2011) damen@itc.nl 1 Overview Platforms & missions aerial surveys

### Geo/SAT 2 INTRODUCTION TO REMOTE SENSING

Geo/SAT 2 INTRODUCTION TO REMOTE SENSING Paul R. Baumann, Professor Emeritus State University of New York College at Oneonta Oneonta, New York 13820 USA COPYRIGHT 2008 Paul R. Baumann Introduction Remote

### Lab Report 3: Speckle Interferometry LIN PEI-YING, BAIG JOVERIA

Lab Report 3: Speckle Interferometry LIN PEI-YING, BAIG JOVERIA Abstract: Speckle interferometry (SI) has become a complete technique over the past couple of years and is widely used in many branches of

### Leica TCS SP8 Quick Start Guide

Leica TCS SP8 Quick Start Guide Leica TCS SP8 System Overview Start-Up Procedure 1. Turn on the CTR Control Box, EL6000 fluorescent light source for the microscope stand. 2. Turn on the Scanner Power

### Chapter 23 Study Questions Name: Class:

Chapter 23 Study Questions Name: Class: Multiple Choice Identify the letter of the choice that best completes the statement or answers the question. 1. When you look at yourself in a plane mirror, you

### Why is blue tinted backlight better?

Why is blue tinted backlight better? L. Paget a,*, A. Scott b, R. Bräuer a, W. Kupper a, G. Scott b a Siemens Display Technologies, Marketing and Sales, Karlsruhe, Germany b Siemens Display Technologies,

### Work environment. Retina anatomy. A human eyeball is like a simple camera! The way of vision signal. Directional sensitivity. Lighting.

Eye anatomy Work environment Lighting 1 2 A human eyeball is like a simple camera! Sclera: outer walls, hard like a light-tight box. Cornea and crystalline lens (eyelens): the two lens system. Retina:

### DESIGNING AND IMPLEMENTING AN ADAPTIVE OPTICS SYSTEM FOR THE UH HOKU KE`A OBSERVATORY ABSTRACT

DESIGNING AND IMPLEMENTING AN ADAPTIVE OPTICS SYSTEM FOR THE UH HOKU KE`A OBSERVATORY University of Hawai`i at Hilo Alex Hedglen ABSTRACT The presented project is to implement a small adaptive optics system

### Consumer digital CCD cameras

CAMERAS Consumer digital CCD cameras Leica RC-30 Aerial Cameras Zeiss RMK Zeiss RMK in aircraft Vexcel UltraCam Digital (note multiple apertures Lenses for Leica RC-30. Many elements needed to minimize

### Advanced Camera and Image Sensor Technology. Steve Kinney Imaging Professional Camera Link Chairman

Advanced Camera and Image Sensor Technology Steve Kinney Imaging Professional Camera Link Chairman Content Physical model of a camera Definition of various parameters for EMVA1288 EMVA1288 and image quality

### Princeton University COS429 Computer Vision Problem Set 1: Building a Camera

Princeton University COS429 Computer Vision Problem Set 1: Building a Camera What to submit: You need to submit two files: one PDF file for the report that contains your name, Princeton NetID, all the

### Digital Image Fundamentals and Image Enhancement in the Spatial Domain

Digital Image Fundamentals and Image Enhancement in the Spatial Domain Mohamed N. Ahmed, Ph.D. Introduction An image may be defined as 2D function f(x,y), where x and y are spatial coordinates. The amplitude

### CS 443: Imaging and Multimedia Cameras and Lenses

CS 443: Imaging and Multimedia Cameras and Lenses Spring 2008 Ahmed Elgammal Dept of Computer Science Rutgers University Outlines Cameras and lenses! 1 They are formed by the projection of 3D objects.

### Introduction to Remote Sensing Fundamentals of Satellite Remote Sensing. Mads Olander Rasmussen

Introduction to Remote Sensing Fundamentals of Satellite Remote Sensing Mads Olander Rasmussen (mora@dhi-gras.com) 01. Introduction to Remote Sensing DHI What is remote sensing? the art, science, and technology

### Chapter 1 Overview of imaging GIS

Chapter 1 Overview of imaging GIS Imaging GIS, a term used in the medical imaging community (Wang 2012), is adopted here to describe a geographic information system (GIS) that displays, enhances, and facilitates

### Compact Dual Field-of-View Telescope for Small Satellite Payloads

Compact Dual Field-of-View Telescope for Small Satellite Payloads James C. Peterson Space Dynamics Laboratory 1695 North Research Park Way, North Logan, UT 84341; 435-797-4624 Jim.Peterson@sdl.usu.edu

### NON-PHOTOGRAPHIC SYSTEMS: Multispectral Scanners Medium and coarse resolution sensor comparisons: Landsat, SPOT, AVHRR and MODIS

NON-PHOTOGRAPHIC SYSTEMS: Multispectral Scanners Medium and coarse resolution sensor comparisons: Landsat, SPOT, AVHRR and MODIS CLASSIFICATION OF NONPHOTOGRAPHIC REMOTE SENSORS PASSIVE ACTIVE DIGITAL

### AgilEye Manual Version 2.0 February 28, 2007

AgilEye Manual Version 2.0 February 28, 2007 1717 Louisiana NE Suite 202 Albuquerque, NM 87110 (505) 268-4742 support@agiloptics.com 2 (505) 268-4742 v. 2.0 February 07, 2007 3 Introduction AgilEye Wavefront

### Instructional Resources/Materials: Light vocabulary cards printed (class set) Enough for each student (See card sort below)

Grade Level/Course: Grade 7 Life Science Lesson/Unit Plan Name: Light Card Sort Rationale/Lesson Abstract: Light vocabulary building, students identify and share vocabulary meaning. Timeframe: 10 to 20

### AS and A LEVEL. Delivery Guide H156/H556 PHYSICS B (ADVANCING PHYSICS) Theme: Imaging and signalling. April 2015

AS and A LEVEL Delivery Guide H156/H556 PHYSICS B (ADVANCING PHYSICS) Theme: Imaging and signalling April 2015 We will inform centres about any changes to the specification. We will also publish changes

### The Optics of Mirrors

Use with Text Pages 558 563 The Optics of Mirrors Use the terms in the list below to fill in the blanks in the paragraphs about mirrors. reversed smooth eyes concave focal smaller reflect behind ray convex

### Vision and Visibility. Human Eye. Eye Components. Cones and Rods. Typical Vision Impairments. CVEN 457 & 696 Lecture #3 Gene Hawkins

Vision and Visibility CVEN 457 & 696 Lecture #3 Gene Hawkins From Allen Chapter 2 Human Eye Eye Components Cornea & lens focuses the image Lens loses ability to focus on close objects with age (presbyopia)

### Parallel Mode Confocal System for Wafer Bump Inspection

Parallel Mode Confocal System for Wafer Bump Inspection ECEN5616 Class Project 1 Gao Wenliang wen-liang_gao@agilent.com 1. Introduction In this paper, A parallel-mode High-speed Line-scanning confocal

### Making NDVI Images using the Sony F717 Nightshot Digital Camera and IR Filters and Software Created for Interpreting Digital Images.

Making NDVI Images using the Sony F717 Nightshot Digital Camera and IR Filters and Software Created for Interpreting Digital Images Draft 1 John Pickle Museum of Science October 14, 2004 Digital Cameras

### DOING PHYSICS WITH MATLAB COMPUTATIONAL OPTICS. GUI Simulation Diffraction: Focused Beams and Resolution for a lens system

DOING PHYSICS WITH MATLAB COMPUTATIONAL OPTICS GUI Simulation Diffraction: Focused Beams and Resolution for a lens system Ian Cooper School of Physics University of Sydney ian.cooper@sydney.edu.au DOWNLOAD

### APPLICATION OF COMPUTER VISION FOR DETERMINATION OF SYMMETRICAL OBJECT POSITION IN THREE DIMENSIONAL SPACE

APPLICATION OF COMPUTER VISION FOR DETERMINATION OF SYMMETRICAL OBJECT POSITION IN THREE DIMENSIONAL SPACE Najirah Umar 1 1 Jurusan Teknik Informatika, STMIK Handayani Makassar Email : najirah_stmikh@yahoo.com

### Experiment 1: Fraunhofer Diffraction of Light by a Single Slit

Experiment 1: Fraunhofer Diffraction of Light by a Single Slit Purpose 1. To understand the theory of Fraunhofer diffraction of light at a single slit and at a circular aperture; 2. To learn how to measure

### 8.2 IMAGE PROCESSING VERSUS IMAGE ANALYSIS Image processing: The collection of routines and

8.1 INTRODUCTION In this chapter, we will study and discuss some fundamental techniques for image processing and image analysis, with a few examples of routines developed for certain purposes. 8.2 IMAGE

### Intorduction to light sources, pinhole cameras, and lenses

Intorduction to light sources, pinhole cameras, and lenses Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 01003 October 26, 2011 Abstract 1 1 Analyzing

### Metis H311 / H322. Highly Advanced, Full Featured 2-Color High-Speed Pyrometers

Metis H311 / H322 Highly Advanced, Full Featured 2-Color High-Speed Pyrometers High-Speed 2-color pyrometer for non-contact temperature measurement in short wavelength spectral range, primarily for measurements

### Images and Graphics. 4. Images and Graphics - Copyright Denis Hamelin - Ryerson University

Images and Graphics Images and Graphics Graphics and images are non-textual information that can be displayed and printed. Graphics (vector graphics) are an assemblage of lines, curves or circles with

### Image Processing COS 426

Image Processing COS 426 What is a Digital Image? A digital image is a discrete array of samples representing a continuous 2D function Continuous function Discrete samples Limitations on Digital Images

### Color Cameras: Three kinds of pixels

Color Cameras: Three kinds of pixels 3 Chip Camera Introduction to Computer Vision CSE 252a Lecture 9 Lens Dichroic prism Optically split incoming light onto three sensors, each responding to different