Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera

Size: px
Start display at page:

Download "Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera"

Transcription

1 Facoltà di Ingegneria Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera Mariolino De Cecco, Nicolo Biasi, Ilya Afanasyev Trento, /20

2 Content 1. Kinect installation. 2. Kinect calibration. 3. Kinect outputs. 4. MATLAB Preprocessing Kinect data. 5. MATLAB Inputs. 6. SQ Curvature Software. 7. Tests. 8. Results. 9. Links. 2/20

3 Kinect installation 1 st attempt 1. OpenNI backend: 1. Nicolas Burrus software [1]: Kinect RGBDemo v > supports OpenNI/Nite backends [2] and has the experimental infrared support with OpenNI (still buggy) -> I tried to check this work for OpenNI backend under Windows 64bit. 2. Install Sensor Kinect drivers under Windows 32 bit [3] and OpenNI/Nite modules (Win32) [2] according to the instruction [4] and verified that they work properly. 3. Use RGB-D Capture [1] to grab Kinect RGB and IR images with intensity.raw and depth.raw files. -> the software doesn t grab IR images!! -> only RGB image, intensity.raw, depth.raw and calibration file (calibration.yml without distortion parameters)!! 4. calibrate-openni-intrinsics --pattern-size grab1 calibration.yml -> gives openni_calibration.yml with Intrinsic matrix and distortion coefficients for Kinect RGB camera!! 5. calibrate-openni-depth.exe --pattern-size grab1 gives partly processed figures with a message about errors. 3/20

4 The comparison of Kinect calibration results for - Libfreenect backend (under OS Linux / Ubuntu bit) - OpenNI backend (under OS Windows 32bit) Libfreenect backend OpenNI backend 23/11/2011 The same figures of Intrinsic Matrixes for Kinect RGB and Depth Cameras The Depth Camera Calibration for OpenNI backend is NOT available!!! -

5 The comparison of Kinect calibration results for - Libfreenect backend (under OS Linux / Ubuntu bit) - OpenNI backend (under OS Windows 32bit) Libfreenect backend OpenNI backend 23/11/2011 There is NO extrinsic mapping between Kinect Depth and RGB cameras for OpenNI backend The reprojection of 3D points from Kinect Depth Camera to RGB image for OpenNI is NOT possible!! -

6 Kinect installation 2 nd attempt 1. Libfreenect backend (Windows): 1. Nicolas Burrus software [1]: Kinect RGBDemo v > supports Libfreenect backend [2] and has RGB and infrared support -> I tried to check the work under Windows 64bit. 2. Install Xbox NUI Motor drivers under Windows 32 bit [3] and OpenNI/Nite modules (Win32) [2] according to the instruction [4] and verified that they work properly. 3. Use RGB-D Capture [1] to grab Kinect RGB and IR images with intensity.raw and depth.raw files. -> the software doesn t grab IR images!! -> only RGB image, intensity.raw, depth.raw and calibration file (calibration.yml without distortion paraeters)!! 4. calibrate-openni-intrinsics --pattern-size grab1 calibration.yml -> gives openni_calibration.yml with Intrinsic matrix and distortion coefficients for Kinect RGB camera!! 5. calibrate-openni-depth.exe --pattern-size grab1 gives partly processed figures with a message about errors. 6/20

7 Kinect installation New Idea 1. OpenNI backend: 1. Nicolas Burrus software [1]: Kinect RGBDemo v > supports OpenNI/Nite backends [2] and has the experimental infrared support with OpenNI (still buggy) -> I tried to check this work for OpenNI backend under Windows 64bit. 2. Install Sensor Kinect drivers under Windows 32 bit [3] and OpenNI/Nite modules (Win32) [2] according to the instruction [4] and verified that they work properly. 3. Use RGB-D Capture [1] to grab Kinect RGB and IR images with intensity.raw and depth.raw files. -> the software doesn t grab IR images!! -> only RGB image, intensity.raw, depth.raw and calibration file (calibration.yml without distortion paraeters)!! 4. calibrate-openni-intrinsics --pattern-size grab1 calibration.yml -> gives openni_calibration.yml with Intrinsic matrix and distortion coefficients for Kinect RGB camera!! 5. calibrate-openni-depth.exe --pattern-size grab1 gives partly processed figures with a message about errors. 7/20

8 5. MATLAB Inputs. Cube_Curvatura_ \ main_reproject.m -> color.png (480x640 pixels) -> calibdata.mat: - matrixes of intrinsic parameters for IR camera (K_ir) and RGB camera (K_rgb) - extrinsic mapping between ir (depth) and rgb Kinect camera (R,T) - distortion coefficients (kc_ir, kc_rgb) -> Points.mat (N x 6), where N number of points with (x,y,z,r,g,b) info 8/20

9 6. MATLAB Software. 6.1 Elimination of the ground. RANSAC search of the ground by using the plane equation. 9/20

10 6.2 3D Points Reprojection. Transformation from IR (Depth) to RGB camera reference system: P_rgb = R * P_ir + T, where R,T - extrinsic mapping between ir and rgb Kinect cameras; P_rgb, P_ir points in RGB and IR camera reference systems. Cube Reprojection without considering distortion (up figures). The distance from Kinect to the cube z = 0.5 m Cube Reprojection with considering distortion (bottom figures).??? Bad calibration? 10/2 0

11 6.2 3D Points Reprojection. Cube and Cylinder Reprojection without considering distortion (up figures). The distance from Kinect to the objects z = 0.6 m??? Bad calibration?! 23/11/2011 Cube and Cylinder Reprojection with considering distortion (bottom figures). 11/2 0

12 6.2 3D Points Reprojection. Transformation from IR (Depth) to RGB camera reference system: P_rgb = R * P_ir + T, where R,T - extrinsic mapping between ir and rgb Kinect cameras; P_rgb, P_ir points in RGB and IR camera reference systems. Cube Reprojection with calibration parameters from Nicolo (upper figures). The distance from Kinect to the cube z = 0.5 m Bad calibration?! Cube Reprojection with calibration parameters from Alberto (bottom figures). 12/2 0

13 6.2 3D Points Reprojection. Camera Calibration and 3d Reconstruction [9] for so-called pinhole camera model. The scene view is formed by projecting 3D points into the image plane using a perspective transformation. H where A R(T) A R(T) A = K_rgb R(T) = [R, T] - are the coordinates of a 3D point in the world coordinate space. - are the coordinates of the projection point in pixels. - matrix of intrinsic parameters (does not depend on the scene viewed). - a principal point (that is usually at the image center) - are the focal lengths expressed in pixel-related units. - the matrix of extrinsic parameters. Cube Reprojection with calibration parameters from Alberto (bottom figures). 13/2 0

14 6.3 RANSAC fitting SQ to 3D data points. With using Levenberg-Marquardt algorithm of distance minimization from SQ to 3D points. Red points outliers. Blue points inliers. Green points SQ model. 14/2 0

15 6.3 Object structure creation and reprojection on image. Reprojection of the lines between the 3D points of cube vertexes to CCD of Kinect RGB camera (figure a) and to the image with reprojected 3D points (figure b). The information about vertex position was obtained in the previous stage of RANSAC cube pose estimation. Red lines the cube frameworks Figure a Figure b 15/2 0

16 9. References 1. Nicolas Burrus. Kinect. RGBDemo, calibrate and visualize Kinect output. // OpenNI Modules. // 3. SensorKinect drivers. // github social coding How-to: Successfully install Kinect on Windows (OpenNI and NITE). // Vangos Pterneas blog: 5. Install OpenKinect for Windows 7 and XP. // 6. Camera Calibration and 3d Reconstruction. // OpenCV (Open Source Computer Vision) v2.1 documentation: ation_and_3d_reconstruction.html 16/20

Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera

Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera Facoltà di Ingegneria Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera Mariolino De Cecco, Nicolo Biasi, Ilya Afanasyev Trento, 2011 1/20 Content

More information

USING ROBOCOMP AND KINECT IN AUGMENTED REALITY APPLICATIONS. Leandro P. Serrano July 2011, Coimbra

USING ROBOCOMP AND KINECT IN AUGMENTED REALITY APPLICATIONS. Leandro P. Serrano July 2011, Coimbra USING ROBOCOMP AND KINECT IN AUGMENTED REALITY APPLICATIONS Leandro P. Serrano July 2011, Coimbra Augmented Reality What is the Augmented Reality? Advantages Tools Problems Proposed solution Future work

More information

Computer Vision Slides curtesy of Professor Gregory Dudek

Computer Vision Slides curtesy of Professor Gregory Dudek Computer Vision Slides curtesy of Professor Gregory Dudek Ioannis Rekleitis Why vision? Passive (emits nothing). Discreet. Energy efficient. Intuitive. Powerful (works well for us, right?) Long and short

More information

Sensors and Sensing Cameras and Camera Calibration

Sensors and Sensing Cameras and Camera Calibration Sensors and Sensing Cameras and Camera Calibration Todor Stoyanov Mobile Robotics and Olfaction Lab Center for Applied Autonomous Sensor Systems Örebro University, Sweden todor.stoyanov@oru.se 20.11.2014

More information

Computer Vision. The Pinhole Camera Model

Computer Vision. The Pinhole Camera Model Computer Vision The Pinhole Camera Model Filippo Bergamasco (filippo.bergamasco@unive.it) http://www.dais.unive.it/~bergamasco DAIS, Ca Foscari University of Venice Academic year 2017/2018 Imaging device

More information

Image Processing & Projective geometry

Image Processing & Projective geometry Image Processing & Projective geometry Arunkumar Byravan Partial slides borrowed from Jianbo Shi & Steve Seitz Color spaces RGB Red, Green, Blue HSV Hue, Saturation, Value Why HSV? HSV separates luma,

More information

Road Boundary Estimation in Construction Sites Michael Darms, Matthias Komar, Dirk Waldbauer, Stefan Lüke

Road Boundary Estimation in Construction Sites Michael Darms, Matthias Komar, Dirk Waldbauer, Stefan Lüke Road Boundary Estimation in Construction Sites Michael Darms, Matthias Komar, Dirk Waldbauer, Stefan Lüke Lanes in Construction Sites Roadway is often bounded by elevated objects (e.g. guidance walls)

More information

FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM

FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM Takafumi Taketomi Nara Institute of Science and Technology, Japan Janne Heikkilä University of Oulu, Finland ABSTRACT In this paper, we propose a method

More information

ME 6406 MACHINE VISION. Georgia Institute of Technology

ME 6406 MACHINE VISION. Georgia Institute of Technology ME 6406 MACHINE VISION Georgia Institute of Technology Class Information Instructor Professor Kok-Meng Lee MARC 474 Office hours: Tues/Thurs 1:00-2:00 pm kokmeng.lee@me.gatech.edu (404)-894-7402 Class

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Sensors and Image Formation Imaging sensors and models of image formation Coordinate systems Digital

More information

AR 2 kanoid: Augmented Reality ARkanoid

AR 2 kanoid: Augmented Reality ARkanoid AR 2 kanoid: Augmented Reality ARkanoid B. Smith and R. Gosine C-CORE and Memorial University of Newfoundland Abstract AR 2 kanoid, Augmented Reality ARkanoid, is an augmented reality version of the popular

More information

Coded Aperture for Projector and Camera for Robust 3D measurement

Coded Aperture for Projector and Camera for Robust 3D measurement Coded Aperture for Projector and Camera for Robust 3D measurement Yuuki Horita Yuuki Matugano Hiroki Morinaga Hiroshi Kawasaki Satoshi Ono Makoto Kimura Yasuo Takane Abstract General active 3D measurement

More information

BEAMFORMING WITH KINECT V2

BEAMFORMING WITH KINECT V2 BEAMFORMING WITH KINECT V2 Stefan Gombots, Felix Egner, Manfred Kaltenbacher Institute of Mechanics and Mechatronics, Vienna University of Technology Getreidemarkt 9, 1060 Wien, AUT e mail: stefan.gombots@tuwien.ac.at

More information

Various Calibration Functions for Webcams and AIBO under Linux

Various Calibration Functions for Webcams and AIBO under Linux SISY 2006 4 th Serbian-Hungarian Joint Symposium on Intelligent Systems Various Calibration Functions for Webcams and AIBO under Linux Csaba Kertész, Zoltán Vámossy Faculty of Science, University of Szeged,

More information

Image stitching. Image stitching. Video summarization. Applications of image stitching. Stitching = alignment + blending. geometrical registration

Image stitching. Image stitching. Video summarization. Applications of image stitching. Stitching = alignment + blending. geometrical registration Image stitching Stitching = alignment + blending Image stitching geometrical registration photometric registration Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2005/3/22 with slides by Richard Szeliski,

More information

Geometry-Based Populated Chessboard Recognition

Geometry-Based Populated Chessboard Recognition Geometry-Based Populated Chessboard Recognition whoff@mines.edu Colorado School of Mines Golden, Colorado, USA William Hoff bill.hoff@daqri.com DAQRI Vienna, Austria My co-authors: Youye Xie, Gongguo Tang

More information

Princeton University COS429 Computer Vision Problem Set 1: Building a Camera

Princeton University COS429 Computer Vision Problem Set 1: Building a Camera Princeton University COS429 Computer Vision Problem Set 1: Building a Camera What to submit: You need to submit two files: one PDF file for the report that contains your name, Princeton NetID, all the

More information

3DUNDERWORLD-SLS v.3.0

3DUNDERWORLD-SLS v.3.0 3DUNDERWORLD-SLS v.3.0 Rapid Scanning and Automatic 3D Reconstruction of Underwater Sites FP7-PEOPLE-2010-RG - 268256 3DUNDERWORLD Software Developer(s): Kyriakos Herakleous Researcher(s): Kyriakos Herakleous,

More information

Lane Detection in Automotive

Lane Detection in Automotive Lane Detection in Automotive Contents Introduction... 2 Image Processing... 2 Reading an image... 3 RGB to Gray... 3 Mean and Gaussian filtering... 5 Defining our Region of Interest... 6 BirdsEyeView Transformation...

More information

Dual-fisheye Lens Stitching for 360-degree Imaging & Video. Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington

Dual-fisheye Lens Stitching for 360-degree Imaging & Video. Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington Dual-fisheye Lens Stitching for 360-degree Imaging & Video Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington Introduction 360-degree imaging: the process of taking multiple photographs and

More information

Active Stereo Vision. COMP 4102A Winter 2014 Gerhard Roth Version 1

Active Stereo Vision. COMP 4102A Winter 2014 Gerhard Roth Version 1 Active Stereo Vision COMP 4102A Winter 2014 Gerhard Roth Version 1 Why active sensors? Project our own texture using light (usually laser) This simplifies correspondence problem (much easier) Pluses Can

More information

MIT CSAIL Advances in Computer Vision Fall Problem Set 6: Anaglyph Camera Obscura

MIT CSAIL Advances in Computer Vision Fall Problem Set 6: Anaglyph Camera Obscura MIT CSAIL 6.869 Advances in Computer Vision Fall 2013 Problem Set 6: Anaglyph Camera Obscura Posted: Tuesday, October 8, 2013 Due: Thursday, October 17, 2013 You should submit a hard copy of your work

More information

Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image

Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image Takahiro Hasegawa, Ryoji Tomizawa, Yuji Yamauchi, Takayoshi Yamashita and Hironobu Fujiyoshi Chubu University, 1200, Matsumoto-cho,

More information

CSE 527: Introduction to Computer Vision

CSE 527: Introduction to Computer Vision CSE 527: Introduction to Computer Vision Week 2 - Class 2: Vision, Physics, Cameras September 7th, 2017 Today Physics Human Vision Eye Brain Perspective Projection Camera Models Image Formation Digital

More information

Lecture 02 Image Formation 1

Lecture 02 Image Formation 1 Institute of Informatics Institute of Neuroinformatics Lecture 02 Image Formation 1 Davide Scaramuzza http://rpg.ifi.uzh.ch 1 Lab Exercise 1 - Today afternoon Room ETH HG E 1.1 from 13:15 to 15:00 Work

More information

Assignment: Cameras and Light

Assignment: Cameras and Light Assignment: Cameras and Light Erik G. Learned-Miller April 5, 2011 1 For this assignment, I do not want you to do ANY collaboration whatsoever. It is important that you work through this assignment on

More information

Unit 1: Image Formation

Unit 1: Image Formation Unit 1: Image Formation 1. Geometry 2. Optics 3. Photometry 4. Sensor Readings Szeliski 2.1-2.3 & 6.3.5 1 Physical parameters of image formation Geometric Type of projection Camera pose Optical Sensor

More information

A Publicly Available RGB-D Data Set of Muslim Prayer Postures Recorded Using Microsoft Kinect for Windows

A Publicly Available RGB-D Data Set of Muslim Prayer Postures Recorded Using Microsoft Kinect for Windows J Basic Appl Sci Res, 4(7)115-125, 2014 2014, TextRoad Publication ISSN 2090-4304 Journal of Basic and Applied Scientific Research wwwtextroadcom A Publicly Available RGB-D Data Set of Muslim Prayer Postures

More information

Acquisition Basics. How can we measure material properties? Goal of this Section. Special Purpose Tools. General Purpose Tools

Acquisition Basics. How can we measure material properties? Goal of this Section. Special Purpose Tools. General Purpose Tools Course 10 Realistic Materials in Computer Graphics Acquisition Basics MPI Informatik (moving to the University of Washington Goal of this Section practical, hands-on description of acquisition basics general

More information

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image Camera & Color Overview Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image Book: Hartley 6.1, Szeliski 2.1.5, 2.2, 2.3 The trip

More information

Image Mosaicing. Jinxiang Chai. Source: faculty.cs.tamu.edu/jchai/cpsc641_spring10/lectures/lecture8.ppt

Image Mosaicing. Jinxiang Chai. Source: faculty.cs.tamu.edu/jchai/cpsc641_spring10/lectures/lecture8.ppt CSCE 641 Computer Graphics: Image Mosaicing Jinxiang Chai Source: faculty.cs.tamu.edu/jchai/cpsc641_spring10/lectures/lecture8.ppt Outline Image registration - How to break assumptions? 3D-2D registration

More information

Toward an Augmented Reality System for Violin Learning Support

Toward an Augmented Reality System for Violin Learning Support Toward an Augmented Reality System for Violin Learning Support Hiroyuki Shiino, François de Sorbier, and Hideo Saito Graduate School of Science and Technology, Keio University, Yokohama, Japan {shiino,fdesorbi,saito}@hvrl.ics.keio.ac.jp

More information

Projection. Announcements. Müller-Lyer Illusion. Image formation. Readings Nalwa 2.1

Projection. Announcements. Müller-Lyer Illusion. Image formation. Readings Nalwa 2.1 Announcements Mailing list (you should have received messages) Project 1 additional test sequences online Projection Readings Nalwa 2.1 Müller-Lyer Illusion Image formation object film by Pravin Bhat http://www.michaelbach.de/ot/sze_muelue/index.html

More information

ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES

ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES Petteri PÖNTINEN Helsinki University of Technology, Institute of Photogrammetry and Remote Sensing, Finland petteri.pontinen@hut.fi KEY WORDS: Cocentricity,

More information

Midterm Examination CS 534: Computational Photography

Midterm Examination CS 534: Computational Photography Midterm Examination CS 534: Computational Photography November 3, 2015 NAME: SOLUTIONS Problem Score Max Score 1 8 2 8 3 9 4 4 5 3 6 4 7 6 8 13 9 7 10 4 11 7 12 10 13 9 14 8 Total 100 1 1. [8] What are

More information

Impact of Thermal and Environmental Conditions on the Kinect Sensor

Impact of Thermal and Environmental Conditions on the Kinect Sensor Impact of Thermal and Environmental Conditions on the Kinect Sensor David Fiedler and Heinrich Müller Department of Computer Science VII, Technische Universität Dortmund, Otto-Hahn-Straße 16, 44227 Dortmund,

More information

Catadioptric Stereo For Robot Localization

Catadioptric Stereo For Robot Localization Catadioptric Stereo For Robot Localization Adam Bickett CSE 252C Project University of California, San Diego Abstract Stereo rigs are indispensable in real world 3D localization and reconstruction, yet

More information

PLazeR. a planar laser rangefinder. Robert Ying (ry2242) Derek Xingzhou He (xh2187) Peiqian Li (pl2521) Minh Trang Nguyen (mnn2108)

PLazeR. a planar laser rangefinder. Robert Ying (ry2242) Derek Xingzhou He (xh2187) Peiqian Li (pl2521) Minh Trang Nguyen (mnn2108) PLazeR a planar laser rangefinder Robert Ying (ry2242) Derek Xingzhou He (xh2187) Peiqian Li (pl2521) Minh Trang Nguyen (mnn2108) Overview & Motivation Detecting the distance between a sensor and objects

More information

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics IMAGE FORMATION Light source properties Sensor characteristics Surface Exposure shape Optics Surface reflectance properties ANALOG IMAGES An image can be understood as a 2D light intensity function f(x,y)

More information

Checkerboard Tracker for Camera Calibration. Andrew DeKelaita EE368

Checkerboard Tracker for Camera Calibration. Andrew DeKelaita EE368 Checkerboard Tracker for Camera Calibration Abstract Andrew DeKelaita EE368 The checkerboard extraction process is an important pre-preprocessing step in camera calibration. This project attempts to implement

More information

ROAD TO THE BEST ALPR IMAGES

ROAD TO THE BEST ALPR IMAGES ROAD TO THE BEST ALPR IMAGES INTRODUCTION Since automatic license plate recognition (ALPR) or automatic number plate recognition (ANPR) relies on optical character recognition (OCR) of images, it makes

More information

A Personal Surround Environment: Projective Display with Correction for Display Surface Geometry and Extreme Lens Distortion

A Personal Surround Environment: Projective Display with Correction for Display Surface Geometry and Extreme Lens Distortion A Personal Surround Environment: Projective Display with Correction for Display Surface Geometry and Extreme Lens Distortion Tyler Johnson, Florian Gyarfas, Rick Skarbez, Herman Towles and Henry Fuchs

More information

Blind navigation with a wearable range camera and vibrotactile helmet

Blind navigation with a wearable range camera and vibrotactile helmet Blind navigation with a wearable range camera and vibrotactile helmet (author s name removed for double-blind review) X university 1@2.com (author s name removed for double-blind review) X university 1@2.com

More information

Rectifying the Planet USING SPACE TO HELP LIFE ON EARTH

Rectifying the Planet USING SPACE TO HELP LIFE ON EARTH Rectifying the Planet USING SPACE TO HELP LIFE ON EARTH About Me Computer Science (BS) Ecology (PhD, almost ) I write programs that process satellite data Scientific Computing! Land Cover Classification

More information

Goal of this Section. Capturing Reflectance From Theory to Practice. Acquisition Basics. How can we measure material properties? Special Purpose Tools

Goal of this Section. Capturing Reflectance From Theory to Practice. Acquisition Basics. How can we measure material properties? Special Purpose Tools Capturing Reflectance From Theory to Practice Acquisition Basics GRIS, TU Darmstadt (formerly University of Washington, Seattle Goal of this Section practical, hands-on description of acquisition basics

More information

Digital Photographic Imaging Using MOEMS

Digital Photographic Imaging Using MOEMS Digital Photographic Imaging Using MOEMS Vasileios T. Nasis a, R. Andrew Hicks b and Timothy P. Kurzweg a a Department of Electrical and Computer Engineering, Drexel University, Philadelphia, USA b Department

More information

Computational Rephotography

Computational Rephotography Computational Rephotography SOONMIN BAE MIT Computer Science and Artificial Intelligence Laboratory ASEEM AGARWALA Abobe Systems, Inc. and FRÉDO DURAND MIT Computer Science and Artificial Intelligence

More information

Sensory Fusion for Image

Sensory Fusion for Image , pp.34-38 http://dx.doi.org/10.14257/astl.2014.45.07 Sensory Fusion for Image Sungjun Park, Wansik Yun, and Gwanggil Jeon 1 Department of Embedded Systems Engineering, Incheon National University, 119

More information

Computer Vision. Thursday, August 30

Computer Vision. Thursday, August 30 Computer Vision Thursday, August 30 1 Today Course overview Requirements, logistics Image formation 2 Introductions Instructor: Prof. Kristen Grauman grauman @ cs TAY 4.118, Thurs 2-4 pm TA: Sudheendra

More information

Computational Re-Photography Soonmin Bae, Aseem Agarwala, and Fredo Durand

Computational Re-Photography Soonmin Bae, Aseem Agarwala, and Fredo Durand Computer Science and Artificial Intelligence Laboratory Technical Report MIT-CSAIL-TR-2010-016 CBCL-287 April 7, 2010 Computational Re-Photography Soonmin Bae, Aseem Agarwala, and Fredo Durand massachusetts

More information

Method for out-of-focus camera calibration

Method for out-of-focus camera calibration 2346 Vol. 55, No. 9 / March 20 2016 / Applied Optics Research Article Method for out-of-focus camera calibration TYLER BELL, 1 JING XU, 2 AND SONG ZHANG 1, * 1 School of Mechanical Engineering, Purdue

More information

A Comparison Between Camera Calibration Software Toolboxes

A Comparison Between Camera Calibration Software Toolboxes 2016 International Conference on Computational Science and Computational Intelligence A Comparison Between Camera Calibration Software Toolboxes James Rothenflue, Nancy Gordillo-Herrejon, Ramazan S. Aygün

More information

Detecting Greenery in Near Infrared Images of Ground-level Scenes

Detecting Greenery in Near Infrared Images of Ground-level Scenes Detecting Greenery in Near Infrared Images of Ground-level Scenes Piotr Łabędź Agnieszka Ozimek Institute of Computer Science Cracow University of Technology Digital Landscape Architecture, Dessau Bernburg

More information

Computer and Machine Vision

Computer and Machine Vision Computer and Machine Vision Lecture Week 7 Part-2 (Exam #1 Review) February 26, 2014 Sam Siewert Outline of Week 7 Basic Convolution Transform Speed-Up Concepts for Computer Vision Hough Linear Transform

More information

Visual Servoing. Charlie Kemp. 4632B/8803 Mobile Manipulation Lecture 8

Visual Servoing. Charlie Kemp. 4632B/8803 Mobile Manipulation Lecture 8 Visual Servoing Charlie Kemp 4632B/8803 Mobile Manipulation Lecture 8 From: http://www.hsi.gatech.edu/visitors/maps/ 4 th floor 4100Q M Building 167 First office on HSI side From: http://www.hsi.gatech.edu/visitors/maps/

More information

Face Detection System on Ada boost Algorithm Using Haar Classifiers

Face Detection System on Ada boost Algorithm Using Haar Classifiers Vol.2, Issue.6, Nov-Dec. 2012 pp-3996-4000 ISSN: 2249-6645 Face Detection System on Ada boost Algorithm Using Haar Classifiers M. Gopi Krishna, A. Srinivasulu, Prof (Dr.) T.K.Basak 1, 2 Department of Electronics

More information

Leica RCD30 Calibration Certificate

Leica RCD30 Calibration Certificate Leica RCD30 Calibration Certificate Camera Head Serial Number Lens Serial Number This certificate is valid for CH62 62001 NAG-D 3.5/50 50002 Inspector Calibration certificate issued on 23 June 2011 Udo

More information

Panoramic Mosaicing with a 180 Field of View Lens

Panoramic Mosaicing with a 180 Field of View Lens CENTER FOR MACHINE PERCEPTION CZECH TECHNICAL UNIVERSITY Panoramic Mosaicing with a 18 Field of View Lens Hynek Bakstein and Tomáš Pajdla {bakstein, pajdla}@cmp.felk.cvut.cz REPRINT Hynek Bakstein and

More information

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2014 Version 1

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2014 Version 1 Image Formation Dr. Gerhard Roth COMP 4102A Winter 2014 Version 1 Image Formation Two type of images Intensity image encodes light intensities (passive sensor) Range (depth) image encodes shape and distance

More information

A Case Study of Security and Privacy Threats from Augmented Reality (AR)

A Case Study of Security and Privacy Threats from Augmented Reality (AR) A Case Study of Security and Privacy Threats from Augmented Reality (AR) Song Chen, Zupei Li, Fabrizio DAngelo, Chao Gao, Xinwen Fu Binghamton University, NY, USA; Email: schen175@binghamton.edu of Computer

More information

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2 CSE 165: 3D User Interaction Lecture #7: Input Devices Part 2 2 Announcements Homework Assignment #2 Due tomorrow at 2pm Sony Move check out Homework discussion Monday at 6pm Input Devices CSE 165 -Winter

More information

LIST 04 Submission Date: 04/05/2017; Cut-off: 14/05/2017. Part 1 Theory. Figure 1: horizontal profile of the R, G and B components.

LIST 04 Submission Date: 04/05/2017; Cut-off: 14/05/2017. Part 1 Theory. Figure 1: horizontal profile of the R, G and B components. Universidade de Brasília (UnB) Faculdade de Tecnologia (FT) Departamento de Engenharia Elétrica (ENE) Course: Image Processing Prof. Mylène C.Q. de Farias Semester: 2017.1 LIST 04 Submission Date: 04/05/2017;

More information

Cameras. CSE 455, Winter 2010 January 25, 2010

Cameras. CSE 455, Winter 2010 January 25, 2010 Cameras CSE 455, Winter 2010 January 25, 2010 Announcements New Lecturer! Neel Joshi, Ph.D. Post-Doctoral Researcher Microsoft Research neel@cs Project 1b (seam carving) was due on Friday the 22 nd Project

More information

Sébastien Equis, and Pierre Jacquot EPFL, Nanophotonics and Metrology Laboratory, Lausanne, Switzerland

Sébastien Equis, and Pierre Jacquot EPFL, Nanophotonics and Metrology Laboratory, Lausanne, Switzerland Sébastien Equis, and Pierre Jacquot EPFL, Nanophotonics and Metrology Laboratory, Lausanne, Switzerland sebastien.equis@epfl.ch Outline Conventional fringe projection and phase extraction Motivation of

More information

Finger rotation detection using a Color Pattern Mask

Finger rotation detection using a Color Pattern Mask Finger rotation detection using a Color Pattern Mask V. Shishir Reddy 1, V. Raghuveer 2, R. Hithesh 3, J. Vamsi Krishna 4,, R. Pratesh Kumar Reddy 5, K. Chandra lohit 6 1,2,3,4,5,6 Electronics and Communication,

More information

Lane Detection in Automotive

Lane Detection in Automotive Lane Detection in Automotive Contents Introduction... 2 Image Processing... 2 Reading an image... 3 RGB to Gray... 3 Mean and Gaussian filtering... 6 Defining our Region of Interest... 10 BirdsEyeView

More information

Dr F. Cuzzolin 1. September 29, 2015

Dr F. Cuzzolin 1. September 29, 2015 P00407 Principles of Computer Vision 1 1 Department of Computing and Communication Technologies Oxford Brookes University, UK September 29, 2015 September 29, 2015 1 / 73 Outline of the Lecture 1 2 Basics

More information

Estimation of spectral response of a consumer grade digital still camera and its application for temperature measurement

Estimation of spectral response of a consumer grade digital still camera and its application for temperature measurement Indian Journal of Pure & Applied Physics Vol. 47, October 2009, pp. 703-707 Estimation of spectral response of a consumer grade digital still camera and its application for temperature measurement Anagha

More information

According to the proposed AWB methods as described in Chapter 3, the following

According to the proposed AWB methods as described in Chapter 3, the following Chapter 4 Experiment 4.1 Introduction According to the proposed AWB methods as described in Chapter 3, the following experiments were designed to evaluate the feasibility and robustness of the algorithms.

More information

HDR videos acquisition

HDR videos acquisition HDR videos acquisition dr. Francesco Banterle francesco.banterle@isti.cnr.it How to capture? Videos are challenging: We need to capture multiple frames at different exposure times and everything moves

More information

Projection. Readings. Szeliski 2.1. Wednesday, October 23, 13

Projection. Readings. Szeliski 2.1. Wednesday, October 23, 13 Projection Readings Szeliski 2.1 Projection Readings Szeliski 2.1 Müller-Lyer Illusion by Pravin Bhat Müller-Lyer Illusion by Pravin Bhat http://www.michaelbach.de/ot/sze_muelue/index.html Müller-Lyer

More information

MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS

MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS INFOTEH-JAHORINA Vol. 10, Ref. E-VI-11, p. 892-896, March 2011. MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS Jelena Cvetković, Aleksej Makarov, Sasa Vujić, Vlatacom d.o.o. Beograd Abstract -

More information

Panoramic Vision System for an Intelligent Vehicle using. a Laser Sensor and Cameras

Panoramic Vision System for an Intelligent Vehicle using. a Laser Sensor and Cameras Panoramic Vision System for an Intelligent Vehicle using a Laser Sensor and Cameras Min Woo Park PH.D Student, Graduate School of Electrical Engineering and Computer Science, Kyungpook National University,

More information

CS415 Human Computer Interaction

CS415 Human Computer Interaction CS415 Human Computer Interaction Lecture 10 Advanced HCI Universal Design & Intro to Cognitive Models October 30, 2016 Sam Siewert Summary of Thoughts on ITS Collective Wisdom of Our Classes (2015, 2016)

More information

Opto Engineering S.r.l.

Opto Engineering S.r.l. TUTORIAL #1 Telecentric Lenses: basic information and working principles On line dimensional control is one of the most challenging and difficult applications of vision systems. On the other hand, besides

More information

GESTURE BASED HUMAN MULTI-ROBOT INTERACTION. Gerard Canal, Cecilio Angulo, and Sergio Escalera

GESTURE BASED HUMAN MULTI-ROBOT INTERACTION. Gerard Canal, Cecilio Angulo, and Sergio Escalera GESTURE BASED HUMAN MULTI-ROBOT INTERACTION Gerard Canal, Cecilio Angulo, and Sergio Escalera Gesture based Human Multi-Robot Interaction Gerard Canal Camprodon 2/27 Introduction Nowadays robots are able

More information

Recognizing Panoramas

Recognizing Panoramas Recognizing Panoramas Kevin Luo Stanford University 450 Serra Mall, Stanford, CA 94305 kluo8128@stanford.edu Abstract This project concerns the topic of panorama stitching. Given a set of overlapping photos,

More information

DETERMINING CALIBRATION PARAMETERS FOR A HARTMANN- SHACK WAVEFRONT SENSOR

DETERMINING CALIBRATION PARAMETERS FOR A HARTMANN- SHACK WAVEFRONT SENSOR DETERMINING CALIBRATION PARAMETERS FOR A HARTMANN- SHACK WAVEFRONT SENSOR Felipe Tayer Amaral¹, Luciana P. Salles 2 and Davies William de Lima Monteiro 3,2 Graduate Program in Electrical Engineering -

More information

CALIBRATION REPORT SUMMARY

CALIBRATION REPORT SUMMARY CALIBRATION REPORT SUMMARY Material Description Assembly 2PADI080 1027 0997 A / 09A Camera Module 1 CCD KODAK KAF 39Mp Full Frame Color Image Sensor SN CQ011027 Lens Schneider Apo Digitar 4.0/80 N SN 15006871

More information

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2015 Version 3

Image Formation. Dr. Gerhard Roth. COMP 4102A Winter 2015 Version 3 Image Formation Dr. Gerhard Roth COMP 4102A Winter 2015 Version 3 1 Image Formation Two type of images Intensity image encodes light intensities (passive sensor) Range (depth) image encodes shape and distance

More information

Automatic source camera identification using the intrinsic lens radial distortion

Automatic source camera identification using the intrinsic lens radial distortion Automatic source camera identification using the intrinsic lens radial distortion Kai San Choi, Edmund Y. Lam, and Kenneth K. Y. Wong Department of Electrical and Electronic Engineering, University of

More information

Digital deformation model for fisheye image rectification

Digital deformation model for fisheye image rectification Digital deformation model for fisheye image rectification Wenguang Hou, 1 Mingyue Ding, 1 Nannan Qin, 2 and Xudong Lai 2, 1 Department of Bio-medical Engineering, Image Processing and Intelligence Control

More information

Panoramas. CS 178, Spring Marc Levoy Computer Science Department Stanford University

Panoramas. CS 178, Spring Marc Levoy Computer Science Department Stanford University Panoramas CS 178, Spring 2010 Marc Levoy Computer Science Department Stanford University What is a panorama?! a wider-angle image than a normal camera can capture! any image stitched from overlapping photographs!

More information

Light Field based 360º Panoramas

Light Field based 360º Panoramas 1 Light Field based 360º Panoramas André Alexandre Rodrigues Oliveira Abstract This paper describes in detail the developed light field based 360º panorama creation solution, named as multiperspective

More information

A 3D Profile Parallel Detecting System Based on Differential Confocal Microscopy. Y.H. Wang, X.F. Yu and Y.T. Fei

A 3D Profile Parallel Detecting System Based on Differential Confocal Microscopy. Y.H. Wang, X.F. Yu and Y.T. Fei Key Engineering Materials Online: 005-10-15 ISSN: 166-9795, Vols. 95-96, pp 501-506 doi:10.408/www.scientific.net/kem.95-96.501 005 Trans Tech Publications, Switzerland A 3D Profile Parallel Detecting

More information

NUST FALCONS. Team Description for RoboCup Small Size League, 2011

NUST FALCONS. Team Description for RoboCup Small Size League, 2011 1. Introduction: NUST FALCONS Team Description for RoboCup Small Size League, 2011 Arsalan Akhter, Muhammad Jibran Mehfooz Awan, Ali Imran, Salman Shafqat, M. Aneeq-uz-Zaman, Imtiaz Noor, Kanwar Faraz,

More information

DEFENSE APPLICATIONS IN HYPERSPECTRAL REMOTE SENSING

DEFENSE APPLICATIONS IN HYPERSPECTRAL REMOTE SENSING DEFENSE APPLICATIONS IN HYPERSPECTRAL REMOTE SENSING James M. Bishop School of Ocean and Earth Science and Technology University of Hawai i at Mānoa Honolulu, HI 96822 INTRODUCTION This summer I worked

More information

Photogrammetric System using Visible Light Communication

Photogrammetric System using Visible Light Communication Photogrammetric System using Visible Light Communication Hideaki Uchiyama, Masaki Yoshino, Hideo Saito and Masao Nakagawa School of Science for Open and Environmental Systems, Keio University, Japan Email:

More information

Computer Vision. Howie Choset Introduction to Robotics

Computer Vision. Howie Choset   Introduction to Robotics Computer Vision Howie Choset http://www.cs.cmu.edu.edu/~choset Introduction to Robotics http://generalrobotics.org What is vision? What is computer vision? Edge Detection Edge Detection Interest points

More information

Introduction to DSP ECE-S352 Fall Quarter 2000 Matlab Project 1

Introduction to DSP ECE-S352 Fall Quarter 2000 Matlab Project 1 Objective: Introduction to DSP ECE-S352 Fall Quarter 2000 Matlab Project 1 This Matlab Project is an extension of the basic correlation theory presented in the course. It shows a practical application

More information

STRUCTURE SENSOR QUICK START GUIDE

STRUCTURE SENSOR QUICK START GUIDE STRUCTURE SENSOR 1 TABLE OF CONTENTS WELCOME TO YOUR NEW STRUCTURE SENSOR 2 WHAT S INCLUDED IN THE BOX 2 CHARGING YOUR STRUCTURE SENSOR 3 CONNECTING YOUR STRUCTURE SENSOR TO YOUR IPAD 4 Attaching Structure

More information

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011) Lecture 19: Depth Cameras Kayvon Fatahalian CMU 15-869: Graphics and Imaging Architectures (Fall 2011) Continuing theme: computational photography Cheap cameras capture light, extensive processing produces

More information

A Comparison of Monocular Camera Calibration Techniques

A Comparison of Monocular Camera Calibration Techniques Wright State University CORE Scholar Browse all Theses and Dissertations Theses and Dissertations 2014 A Comparison of Monocular Camera Calibration Techniques Richard L. Van Hook Wright State University

More information

Novel calibration method for structured-light system with an out-of-focus projector

Novel calibration method for structured-light system with an out-of-focus projector Novel calibration method for structured-light system with an out-of-focus projector Beiwen Li, Nikolaus Karpinsky, and Song Zhang* Department of Mechanical Engineering, Iowa State University, Ames, Iowa

More information

Charged Coupled Device (CCD) S.Vidhya

Charged Coupled Device (CCD) S.Vidhya Charged Coupled Device (CCD) S.Vidhya 02.04.2016 Sensor Physical phenomenon Sensor Measurement Output A sensor is a device that measures a physical quantity and converts it into a signal which can be read

More information

6.869 Advances in Computer Vision Spring 2010, A. Torralba

6.869 Advances in Computer Vision Spring 2010, A. Torralba 6.869 Advances in Computer Vision Spring 2010, A. Torralba Due date: Wednesday, Feb 17, 2010 Problem set 1 You need to submit a report with brief descriptions of what you did. The most important part is

More information

ROS Tutorial. Me133a Joseph & Daniel 11/01/2017

ROS Tutorial. Me133a Joseph & Daniel 11/01/2017 ROS Tutorial Me133a Joseph & Daniel 11/01/2017 Introduction to ROS 2D Turtle Simulation 3D Turtlebot Simulation Real Turtlebot Demo What is ROS ROS is an open-source, meta-operating system for your robot

More information

CIS581: Computer Vision and Computational Photography Homework: Cameras and Convolution Due: Sept. 14, 2017 at 3:00 pm

CIS581: Computer Vision and Computational Photography Homework: Cameras and Convolution Due: Sept. 14, 2017 at 3:00 pm CIS58: Computer Vision and Computational Photography Homework: Cameras and Convolution Due: Sept. 4, 207 at 3:00 pm Instructions This is an individual assignment. Individual means each student must hand

More information

Vision Research at. Validation of a Novel Hartmann-Moiré Wavefront Sensor with Large Dynamic Range. Wavefront Science Congress, Feb.

Vision Research at. Validation of a Novel Hartmann-Moiré Wavefront Sensor with Large Dynamic Range. Wavefront Science Congress, Feb. Wavefront Science Congress, Feb. 2008 Validation of a Novel Hartmann-Moiré Wavefront Sensor with Large Dynamic Range Xin Wei 1, Tony Van Heugten 2, Nikole L. Himebaugh 1, Pete S. Kollbaum 1, Mei Zhang

More information

Image Fusion. Pan Sharpening. Pan Sharpening. Pan Sharpening: ENVI. Multi-spectral and PAN. Magsud Mehdiyev Geoinfomatics Center, AIT

Image Fusion. Pan Sharpening. Pan Sharpening. Pan Sharpening: ENVI. Multi-spectral and PAN. Magsud Mehdiyev Geoinfomatics Center, AIT 1 Image Fusion Sensor Merging Magsud Mehdiyev Geoinfomatics Center, AIT Image Fusion is a combination of two or more different images to form a new image by using certain algorithms. ( Pohl et al 1998)

More information