A Mathematical model for the determination of distance of an object in a 2D image

Size: px
Start display at page:

Download "A Mathematical model for the determination of distance of an object in a 2D image"

Transcription

1 A Mathematical model for the determination of distance of an object in a 2D image Deepu R 1, Murali S 2,Vikram Raju 3 Maharaja Institute of Technology Mysore, Karnataka, India rdeepusingh@mitmysore.in 1, murali@mitmysore.in 2, vikramraju17@gmail.com 3 Abstract-Distance measurement in real world has always been one of the challenging tasks to be performed in the field of computer vision. Photographic images are two dimensional depiction of three-dimensional real space. Various methods have been developed, researched and improvised over the years depending on the focus of the object, defocus of the background, the vanishing point of an image and other information derived from images. The impact of various features of the camera are also considered when the above problem is under study like, focal length of the camera, ISO levels, white balance, resolution etc. The effect of the resolution of an object in an image, its actual size, the focal length of the camera, the camera resolution, all contribute in determining the distance between the camera and the object. In this paper we have proposed a model for the relationship between the resolution of an object of interest and the distance from the camera as a growth series. Relationship between the resolution of an object of interest and focal length is also expressed as a growth series model. The object of interest is segmented out and its pixels account to its resolution, which is standardized and used to determine either the distance or the focal length of the camera or the vice versa. This model, thus, very easily helps in tracking the object using a camera with minimal error rate, provided the image is along the optical axis of the camera. Keywords: Resolution, Focal Length, Camera and object distance, growth series, Pixels per inch, Geometric Progression. 1 Introduction The knowledge about the distance of an object in an image has its applications in many fields of science and technology and research, like, human motion estimation, web conferencing, gaming industry, surveillance, security systems, robotics, medical systems and imaging. Etc. The major factors in a photo that cannot be changed once the photo has been clicked are Aperture of the lens, focal length, number of pixels or resolution and shutter speed. A method which allows refocusing is a potential powerful tool for digital image editing. Once the depth map has been obtained, one can de-blur the image in order to acquire an all focus image or blur the image even more to create certain visual effects [1]. The depth map can also apply to the task such as automatic scene segmentation, post exposure refocusing and rerendering of the scene from an alternative view point. By analyzing the depth of field, the coarse depth map of scene can be recovered [2]. Depth from defocus and depth from focus are the two methods to estimate the 3D geometry of the scene by exploiting image focus [3]. In adaptive depth from focus, depth estimation method for images in narrow depth of field setting, the segments produced by mean shift segmentation as windows to analyze the focus measure and employ a hierarchical Markov Random Field to infer from the depth map [4]. Ina non-linear approach, the precise sparse depth map extraction, the noisy focus measurements are replaced with estimated values, which in turn helps to accurately compute 3D shapes while preserving edges of objects [5].In paper [6],authors have proposed a real-time method which can measure distance using a modified camera. The camera s image sensor is inclined by a certain angle. Thus, the image projected on the sensor is defocused differently at different areas. The area where the image is focused best is just at projection plane. The position of the projection plane can be obtained after finding out this area. The distance between the projection plane and the lens of the camera is image distance. Object distance can be obtained by applying image distance to lens formula. Although the results of the above methods are very promising, it is computationally expensive to employ Markov Random Fields or any of the procedures given in above papers. In real world applications it is necessary to provide real-time computations. By studying the model or rate at which features change in an image like size, distance, pixels of an object, focal length, give us a better approximation and a simpler computationally efficient method to determine the relationship among them.

2 The resolution of an image or the number of pixels of an image is a constant, which depends on the camera. Once an image is captured, its focal length is also constant. When an image is captured by a camera, the distance of the image from the camera or the viewer, thus can be determined using the above variables. The resolution also directly depends on the focal length, since it has a proportional relationship with the image size. The camera, like the human eye, works on perspective vision, where, the change in image size with variation of Focal Length and Distance is not linear. To study the relationship among the above, the following methodology is proposed. 2 Method With an objective of designing a robust model, sequence of images at different distances D from the object and for different focal lengths are taken along the optical axis. Every image is of size ix j and it is acquired for objects of various shapes and sizes. The total number of pixels P, in the image is thus given by = (1) The set of the pixels in an object P o,is a subset of P. The image is converted to grey scale using a basic thresholding algorithm to obtain the object of interest and the number of pixels in the object is obtained. This becomes the set of pixels in an object, and stored in a set P o. The behavior of the relationship between the distance with respect to the resolution factor R f and that of the focal length and the resolution factor is sought. Figure 1: Experimental setup varying distance from camera to objects. Once the images are captured, the background is segmented, to get only the pixels of the object of interest. Image preprocessing is necessary at times, when noise is found in the images, when the automatic thresholding algorithm fails to extract the object clearly. Manual image cleaning was used to remove unwanted noise in the image. This can be avoided using a better efficient object extraction algorithm The image segmentation and cleaning was carried out using MATLAB. To build the mathematical model, IBM SPSS was used. To determine the resolution factor, first, the Resolution per Inch, R In, of the object is determined. It is then multiplied by the resolution of the camera and later natural log is taken to scale down the data. This is so done in order to standardize the factor arising due to the resolution of the images and of the objects in them, thus making the method applicable to any kind of camera and of any resolution. In a way, it makes the method independent of the resolution itself. = ln ( ) (2) Figure1 shows the variation of size of the image with the changes in distance from the camera. Figure 2: Test Images (left) and Objects extracted from the image (right)

3 3 Results The results can be analyzed in two folds, one model with resolution factor and distance of object in an image and the other with resolution factor and the focal length of the camera. Following results were tabulated for 3 different objects. Table 1 gives the difference between expected distance and obtained distance for circle object where as Table2 and Tables3 specify the same for square and triangle objects. We can also note that there is a relationship between pixels in an object to distance which is expressed by our model. Table 1:Observed results for circle object of area sq.in Table 4: Model Summary and Parameter Estimates for Distance and Resolution Factor for a focal length of 18mm. Equation Dependent Variable: Distance R Square Model Summary Parameter Estimates F df1 df2 Sig. Constant b1 Growth The R Square value shows a 99.9% significance to the model, which is a good fit for the data set observed. Table 2:Observed results square object of area sq.in Table 3: Observed results triangle object of area 20 sq.in Figure 3: The growth series of distance and resolution factor of object in image The model agrees completely with the growth series and gives us the following equation to determine the distance of the object from the camera: = (3) Part I: Resolution Factor and Distance The distance is considered as the dependent variable in the analysis of the model and the resolution factor is taken as the independent variable. In each case of the focal length the model is tested. Where, D is the distance of the object from the camera, A and B are constants in the growth series, which are determined by a geometric series, and R f is the resolution factor. The r in the geometric ratio is 1.01.The geometric series gives the values of the constants from the above equation.

4 Table 5: The geometric series to determine the constants of the growth series model. Focal Length A B 18mm mm r r 30mm r r 2 36mm r r 3 42mm r r 4 48mm r r 5 54mm r r 6 60mm r r 7 66mm r r 8 The geometric series gives the values of the constants from the above equation. The model was tested on random image samples, which agreed with the growth series model, with very minimal error. The error in determining the distance was in the range of percent. Part II: Resolution Factor and Focal Length The focal length is considered the dependent variable in the analysis of the model and the resolution factor is taken as the independent variable. In each case of the distance of the object from the camera, the model is tested. Table 6: Model Summary and Parameter Estimates for Distance and Resolution Factor at 51 inches from the camera. Equation Dependent Variable: Focal Length R Square Model Summary Parameter Estimates F df1 df2 Sig. Constant b1 Growth The R Square value shows a 99.1% significance to the model, which is a good fit for the data set observed. The model agrees completely with the growth series and gives us the following equation to determine the focal length: = (4) Where, F is the focal length of the camera, A and B are constants in the growth series, which are determined by a geometric series, and R f is the resolution factor. The r in the geometric ratio is Figure 4: The growth series of focal length and resolution factor of object in an image The model was tested on random image samples, which agreed with the growth series model, with very minimal error. The error in determining the distance was in the range of 1-2 percent. Table 7: The geometric series to determine the constants of the growth series model. Distance A B 51in in r 0.543r 71in r r 2 81in r r 3 91in r r 4 101in r r 5 111in r r 6 121in r r 7 131in r r 8 141in r r 9 4 Applications The model finds its use in many applications like obstacle detection in robotics, detection of road humps, navigation systems for the blind, unknown territory exploration, 3D game development are a few to name.

5 5 Future Work The distance of the object from the camera and focal length model is applicable only when the object is head-on or in the line of the optical axis of the camera. Future work can be done on studying the effect of resolution and distance for images not in the axis of the camera, considering the evolution of perspective vision of an image. It can also be used to create 3D scenes by determining the distance from the camera and rebuilding another scene with the same distance. 6 Conclusions The model shows promising enhancement on further development and an efficient model to determine the distance of an object in an image from the camera making use of the data in the resolution of an image, which cannot change once an image is taken. Building the model for varied focal lengths enables the model to work with cameras of any focal length. The obtained growth series enables to work with distance and focal length. With knowing one of the two, the other can be calculated and be easily modeled in the system. Computationally the calculation of the growth series works efficiently than the iterative approach. 7 Acknowledgements The work presented in this paper was supported by VTU research grants (Ref. No. VTU/ Aca./ /A-9/13551). We thank Visvesvaraya Technological University, Belgaum for this. 8 References [1] Potmesil, M., and Chakravarty, I., A lens and aperture camera model for synthetic image generation.in proc. Siggraph, [2] Pentland, A. P., A new sense for depth of field.ieee Trans. Pattern anal. Mach. Intell. 9, 4, [3] Hasinoff S. W., Kutulakos K. N., A layerbased restoration framework for variableaperture photography, F, 2007 [C], IEEE. [4] Bing-Zhong Jing, Daniel S. Yeung, Recovering depth from images using adaptive depth from focus. IEEE /12, [5] Muhammad Tariq Mahmood, Ikhyun Lee, Wook-Jin Choi, Tae-Sun Choi, A non-linear approach for depth from focus for digital cameras. IEEE /11, [6] Liu Xiaoming, Qin Tian, Chen Wanchen, Yin Xingliang, Real Time Distance Measurment Using a Modified Camera.IEEE /09.

Computational Camera & Photography: Coded Imaging

Computational Camera & Photography: Coded Imaging Computational Camera & Photography: Coded Imaging Camera Culture Ramesh Raskar MIT Media Lab http://cameraculture.media.mit.edu/ Image removed due to copyright restrictions. See Fig. 1, Eight major types

More information

A Recognition of License Plate Images from Fast Moving Vehicles Using Blur Kernel Estimation

A Recognition of License Plate Images from Fast Moving Vehicles Using Blur Kernel Estimation A Recognition of License Plate Images from Fast Moving Vehicles Using Blur Kernel Estimation Kalaivani.R 1, Poovendran.R 2 P.G. Student, Dept. of ECE, Adhiyamaan College of Engineering, Hosur, Tamil Nadu,

More information

Dappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing

Dappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing Dappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing Ashok Veeraraghavan, Ramesh Raskar, Ankit Mohan & Jack Tumblin Amit Agrawal, Mitsubishi Electric Research

More information

Coded Computational Photography!

Coded Computational Photography! Coded Computational Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 9! Gordon Wetzstein! Stanford University! Coded Computational Photography - Overview!!

More information

SURVEILLANCE SYSTEMS WITH AUTOMATIC RESTORATION OF LINEAR MOTION AND OUT-OF-FOCUS BLURRED IMAGES. Received August 2008; accepted October 2008

SURVEILLANCE SYSTEMS WITH AUTOMATIC RESTORATION OF LINEAR MOTION AND OUT-OF-FOCUS BLURRED IMAGES. Received August 2008; accepted October 2008 ICIC Express Letters ICIC International c 2008 ISSN 1881-803X Volume 2, Number 4, December 2008 pp. 409 414 SURVEILLANCE SYSTEMS WITH AUTOMATIC RESTORATION OF LINEAR MOTION AND OUT-OF-FOCUS BLURRED IMAGES

More information

Project 4 Results http://www.cs.brown.edu/courses/cs129/results/proj4/jcmace/ http://www.cs.brown.edu/courses/cs129/results/proj4/damoreno/ http://www.cs.brown.edu/courses/csci1290/results/proj4/huag/

More information

A moment-preserving approach for depth from defocus

A moment-preserving approach for depth from defocus A moment-preserving approach for depth from defocus D. M. Tsai and C. T. Lin Machine Vision Lab. Department of Industrial Engineering and Management Yuan-Ze University, Chung-Li, Taiwan, R.O.C. E-mail:

More information

Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring

Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring Ashill Chiranjan and Bernardt Duvenhage Defence, Peace, Safety and Security Council for Scientific

More information

Lenses, exposure, and (de)focus

Lenses, exposure, and (de)focus Lenses, exposure, and (de)focus http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 15 Course announcements Homework 4 is out. - Due October 26

More information

On the Recovery of Depth from a Single Defocused Image

On the Recovery of Depth from a Single Defocused Image On the Recovery of Depth from a Single Defocused Image Shaojie Zhuo and Terence Sim School of Computing National University of Singapore Singapore,747 Abstract. In this paper we address the challenging

More information

Modeling and Synthesis of Aperture Effects in Cameras

Modeling and Synthesis of Aperture Effects in Cameras Modeling and Synthesis of Aperture Effects in Cameras Douglas Lanman, Ramesh Raskar, and Gabriel Taubin Computational Aesthetics 2008 20 June, 2008 1 Outline Introduction and Related Work Modeling Vignetting

More information

Coded Aperture for Projector and Camera for Robust 3D measurement

Coded Aperture for Projector and Camera for Robust 3D measurement Coded Aperture for Projector and Camera for Robust 3D measurement Yuuki Horita Yuuki Matugano Hiroki Morinaga Hiroshi Kawasaki Satoshi Ono Makoto Kimura Yasuo Takane Abstract General active 3D measurement

More information

Why learn about photography in this course?

Why learn about photography in this course? Why learn about photography in this course? Geri's Game: Note the background is blurred. - photography: model of image formation - Many computer graphics methods use existing photographs e.g. texture &

More information

lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response

lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response - application: high dynamic range imaging Why learn

More information

Single Digital Image Multi-focusing Using Point to Point Blur Model Based Depth Estimation

Single Digital Image Multi-focusing Using Point to Point Blur Model Based Depth Estimation Single Digital mage Multi-focusing Using Point to Point Blur Model Based Depth Estimation Praveen S S, Aparna P R Abstract The proposed paper focuses on Multi-focusing, a technique that restores all-focused

More information

Toward Non-stationary Blind Image Deblurring: Models and Techniques

Toward Non-stationary Blind Image Deblurring: Models and Techniques Toward Non-stationary Blind Image Deblurring: Models and Techniques Ji, Hui Department of Mathematics National University of Singapore NUS, 30-May-2017 Outline of the talk Non-stationary Image blurring

More information

NTU CSIE. Advisor: Wu Ja Ling, Ph.D.

NTU CSIE. Advisor: Wu Ja Ling, Ph.D. An Interactive Background Blurring Mechanism and Its Applications NTU CSIE Yan Chih Yu Advisor: Wu Ja Ling, Ph.D. 1 2 Outline Introduction Related Work Method Object Segmentation Depth Map Generation Image

More information

4 STUDY OF DEBLURRING TECHNIQUES FOR RESTORED MOTION BLURRED IMAGES

4 STUDY OF DEBLURRING TECHNIQUES FOR RESTORED MOTION BLURRED IMAGES 4 STUDY OF DEBLURRING TECHNIQUES FOR RESTORED MOTION BLURRED IMAGES Abstract: This paper attempts to undertake the study of deblurring techniques for Restored Motion Blurred Images by using: Wiener filter,

More information

Depth from Focusing and Defocusing. Carnegie Mellon University. Pittsburgh, PA result is 1.3% RMS error in terms of distance

Depth from Focusing and Defocusing. Carnegie Mellon University. Pittsburgh, PA result is 1.3% RMS error in terms of distance Depth from Focusing and Defocusing Yalin Xiong Steven A. Shafer The Robotics Institute Carnegie Mellon University Pittsburgh, PA 53 Abstract This paper studies the problem of obtaining depth information

More information

Linear Gaussian Method to Detect Blurry Digital Images using SIFT

Linear Gaussian Method to Detect Blurry Digital Images using SIFT IJCAES ISSN: 2231-4946 Volume III, Special Issue, November 2013 International Journal of Computer Applications in Engineering Sciences Special Issue on Emerging Research Areas in Computing(ERAC) www.caesjournals.org

More information

Distance Estimation with a Two or Three Aperture SLR Digital Camera

Distance Estimation with a Two or Three Aperture SLR Digital Camera Distance Estimation with a Two or Three Aperture SLR Digital Camera Seungwon Lee, Joonki Paik, and Monson H. Hayes Graduate School of Advanced Imaging Science, Multimedia, and Film Chung-Ang University

More information

Recent Advances in Image Deblurring. Seungyong Lee (Collaboration w/ Sunghyun Cho)

Recent Advances in Image Deblurring. Seungyong Lee (Collaboration w/ Sunghyun Cho) Recent Advances in Image Deblurring Seungyong Lee (Collaboration w/ Sunghyun Cho) Disclaimer Many images and figures in this course note have been copied from the papers and presentation materials of previous

More information

Computational Cameras. Rahul Raguram COMP

Computational Cameras. Rahul Raguram COMP Computational Cameras Rahul Raguram COMP 790-090 What is a computational camera? Camera optics Camera sensor 3D scene Traditional camera Final image Modified optics Camera sensor Image Compute 3D scene

More information

Design of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems

Design of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems Design of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems Ricardo R. Garcia University of California, Berkeley Berkeley, CA rrgarcia@eecs.berkeley.edu Abstract In recent

More information

SUPER RESOLUTION INTRODUCTION

SUPER RESOLUTION INTRODUCTION SUPER RESOLUTION Jnanavardhini - Online MultiDisciplinary Research Journal Ms. Amalorpavam.G Assistant Professor, Department of Computer Sciences, Sambhram Academy of Management. Studies, Bangalore Abstract:-

More information

DEFOCUS BLUR PARAMETER ESTIMATION TECHNIQUE

DEFOCUS BLUR PARAMETER ESTIMATION TECHNIQUE International Journal of Electronics and Communication Engineering and Technology (IJECET) Volume 7, Issue 4, July-August 2016, pp. 85 90, Article ID: IJECET_07_04_010 Available online at http://www.iaeme.com/ijecet/issues.asp?jtype=ijecet&vtype=7&itype=4

More information

Blind Single-Image Super Resolution Reconstruction with Defocus Blur

Blind Single-Image Super Resolution Reconstruction with Defocus Blur Sensors & Transducers 2014 by IFSA Publishing, S. L. http://www.sensorsportal.com Blind Single-Image Super Resolution Reconstruction with Defocus Blur Fengqing Qin, Lihong Zhu, Lilan Cao, Wanan Yang Institute

More information

License Plate Localisation based on Morphological Operations

License Plate Localisation based on Morphological Operations License Plate Localisation based on Morphological Operations Xiaojun Zhai, Faycal Benssali and Soodamani Ramalingam School of Engineering & Technology University of Hertfordshire, UH Hatfield, UK Abstract

More information

MEM: Intro to Robotics. Assignment 3I. Due: Wednesday 10/15 11:59 EST

MEM: Intro to Robotics. Assignment 3I. Due: Wednesday 10/15 11:59 EST MEM: Intro to Robotics Assignment 3I Due: Wednesday 10/15 11:59 EST 1. Basic Optics You are shopping for a new lens for your Canon D30 digital camera and there are lots of lens options at the store. Your

More information

CAMERA BASICS. Stops of light

CAMERA BASICS. Stops of light CAMERA BASICS Stops of light A stop of light isn t a quantifiable measurement it s a relative measurement. A stop of light is defined as a doubling or halving of any quantity of light. The word stop is

More information

Coded photography , , Computational Photography Fall 2018, Lecture 14

Coded photography , , Computational Photography Fall 2018, Lecture 14 Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 14 Overview of today s lecture The coded photography paradigm. Dealing with

More information

International Journal of Innovative Research in Engineering Science and Technology APRIL 2018 ISSN X

International Journal of Innovative Research in Engineering Science and Technology APRIL 2018 ISSN X HIGH DYNAMIC RANGE OF MULTISPECTRAL ACQUISITION USING SPATIAL IMAGES 1 M.Kavitha, M.Tech., 2 N.Kannan, M.E., and 3 S.Dharanya, M.E., 1 Assistant Professor/ CSE, Dhirajlal Gandhi College of Technology,

More information

Colour correction for panoramic imaging

Colour correction for panoramic imaging Colour correction for panoramic imaging Gui Yun Tian Duke Gledhill Dave Taylor The University of Huddersfield David Clarke Rotography Ltd Abstract: This paper reports the problem of colour distortion in

More information

Displacement Measurement of Burr Arch-Truss Under Dynamic Loading Based on Image Processing Technology

Displacement Measurement of Burr Arch-Truss Under Dynamic Loading Based on Image Processing Technology 6 th International Conference on Advances in Experimental Structural Engineering 11 th International Workshop on Advanced Smart Materials and Smart Structures Technology August 1-2, 2015, University of

More information

BASIC OPERATIONS IN IMAGE PROCESSING USING MATLAB

BASIC OPERATIONS IN IMAGE PROCESSING USING MATLAB BASIC OPERATIONS IN IMAGE PROCESSING USING MATLAB Er.Amritpal Kaur 1,Nirajpal Kaur 2 1,2 Assistant Professor,Guru Nanak Dev University, Regional Campus, Gurdaspur Abstract: - This paper aims at basic image

More information

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image Camera & Color Overview Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image Book: Hartley 6.1, Szeliski 2.1.5, 2.2, 2.3 The trip

More information

Simulated Programmable Apertures with Lytro

Simulated Programmable Apertures with Lytro Simulated Programmable Apertures with Lytro Yangyang Yu Stanford University yyu10@stanford.edu Abstract This paper presents a simulation method using the commercial light field camera Lytro, which allows

More information

Name: Date: Math in Special Effects: Try Other Challenges. Student Handout

Name: Date: Math in Special Effects: Try Other Challenges. Student Handout Name: Date: Math in Special Effects: Try Other Challenges When filming special effects, a high-speed photographer needs to control the duration and impact of light by adjusting a number of settings, including

More information

Photo-Consistent Motion Blur Modeling for Realistic Image Synthesis

Photo-Consistent Motion Blur Modeling for Realistic Image Synthesis Photo-Consistent Motion Blur Modeling for Realistic Image Synthesis Huei-Yung Lin and Chia-Hong Chang Department of Electrical Engineering, National Chung Cheng University, 168 University Rd., Min-Hsiung

More information

Motion Estimation from a Single Blurred Image

Motion Estimation from a Single Blurred Image Motion Estimation from a Single Blurred Image Image Restoration: De-Blurring Build a Blur Map Adapt Existing De-blurring Techniques to real blurred images Analysis, Reconstruction and 3D reconstruction

More information

Near-Invariant Blur for Depth and 2D Motion via Time-Varying Light Field Analysis

Near-Invariant Blur for Depth and 2D Motion via Time-Varying Light Field Analysis Near-Invariant Blur for Depth and 2D Motion via Time-Varying Light Field Analysis Yosuke Bando 1,2 Henry Holtzman 2 Ramesh Raskar 2 1 Toshiba Corporation 2 MIT Media Lab Defocus & Motion Blur PSF Depth

More information

To Do. Advanced Computer Graphics. Outline. Computational Imaging. How do we see the world? Pinhole camera

To Do. Advanced Computer Graphics. Outline. Computational Imaging. How do we see the world? Pinhole camera Advanced Computer Graphics CSE 163 [Spring 2017], Lecture 14 Ravi Ramamoorthi http://www.cs.ucsd.edu/~ravir To Do Assignment 2 due May 19 Any last minute issues or questions? Next two lectures: Imaging,

More information

Improving Image Quality by Camera Signal Adaptation to Lighting Conditions

Improving Image Quality by Camera Signal Adaptation to Lighting Conditions Improving Image Quality by Camera Signal Adaptation to Lighting Conditions Mihai Negru and Sergiu Nedevschi Technical University of Cluj-Napoca, Computer Science Department Mihai.Negru@cs.utcluj.ro, Sergiu.Nedevschi@cs.utcluj.ro

More information

Single-Image Shape from Defocus

Single-Image Shape from Defocus Single-Image Shape from Defocus José R.A. Torreão and João L. Fernandes Instituto de Computação Universidade Federal Fluminense 24210-240 Niterói RJ, BRAZIL Abstract The limited depth of field causes scene

More information

Multi Viewpoint Panoramas

Multi Viewpoint Panoramas 27. November 2007 1 Motivation 2 Methods Slit-Scan "The System" 3 "The System" Approach Preprocessing Surface Selection Panorama Creation Interactive Renement 4 Sources Motivation image showing long continous

More information

Coded photography , , Computational Photography Fall 2017, Lecture 18

Coded photography , , Computational Photography Fall 2017, Lecture 18 Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 18 Course announcements Homework 5 delayed for Tuesday. - You will need cameras

More information

Deblurring. Basics, Problem definition and variants

Deblurring. Basics, Problem definition and variants Deblurring Basics, Problem definition and variants Kinds of blur Hand-shake Defocus Credit: Kenneth Josephson Motion Credit: Kenneth Josephson Kinds of blur Spatially invariant vs. Spatially varying

More information

Removing Temporal Stationary Blur in Route Panoramas

Removing Temporal Stationary Blur in Route Panoramas Removing Temporal Stationary Blur in Route Panoramas Jiang Yu Zheng and Min Shi Indiana University Purdue University Indianapolis jzheng@cs.iupui.edu Abstract The Route Panorama is a continuous, compact

More information

Removal of Salt and Pepper Noise from Satellite Images

Removal of Salt and Pepper Noise from Satellite Images Removal of Salt and Pepper Noise from Satellite Images Mr. Yogesh V. Kolhe 1 Research Scholar, Samrat Ashok Technological Institute Vidisha (INDIA) Dr. Yogendra Kumar Jain 2 Guide & Asso.Professor, Samrat

More information

Applications of Flash and No-Flash Image Pairs in Mobile Phone Photography

Applications of Flash and No-Flash Image Pairs in Mobile Phone Photography Applications of Flash and No-Flash Image Pairs in Mobile Phone Photography Xi Luo Stanford University 450 Serra Mall, Stanford, CA 94305 xluo2@stanford.edu Abstract The project explores various application

More information

Focused Image Recovery from Two Defocused

Focused Image Recovery from Two Defocused Focused Image Recovery from Two Defocused Images Recorded With Different Camera Settings Murali Subbarao Tse-Chung Wei Gopal Surya Department of Electrical Engineering State University of New York Stony

More information

Computer Vision Slides curtesy of Professor Gregory Dudek

Computer Vision Slides curtesy of Professor Gregory Dudek Computer Vision Slides curtesy of Professor Gregory Dudek Ioannis Rekleitis Why vision? Passive (emits nothing). Discreet. Energy efficient. Intuitive. Powerful (works well for us, right?) Long and short

More information

fast blur removal for wearable QR code scanners

fast blur removal for wearable QR code scanners fast blur removal for wearable QR code scanners Gábor Sörös, Stephan Semmler, Luc Humair, Otmar Hilliges ISWC 2015, Osaka, Japan traditional barcode scanning next generation barcode scanning ubiquitous

More information

Evolving Measurement Regions for Depth from Defocus

Evolving Measurement Regions for Depth from Defocus Evolving Measurement Regions for Depth from Defocus Scott McCloskey, Michael Langer, and Kaleem Siddiqi Centre for Intelligent Machines, McGill University {scott,langer,siddiqi}@cim.mcgill.ca Abstract.

More information

Computer Vision. Howie Choset Introduction to Robotics

Computer Vision. Howie Choset   Introduction to Robotics Computer Vision Howie Choset http://www.cs.cmu.edu.edu/~choset Introduction to Robotics http://generalrobotics.org What is vision? What is computer vision? Edge Detection Edge Detection Interest points

More information

ME 6406 MACHINE VISION. Georgia Institute of Technology

ME 6406 MACHINE VISION. Georgia Institute of Technology ME 6406 MACHINE VISION Georgia Institute of Technology Class Information Instructor Professor Kok-Meng Lee MARC 474 Office hours: Tues/Thurs 1:00-2:00 pm kokmeng.lee@me.gatech.edu (404)-894-7402 Class

More information

Privacy Preserving Optics for Miniature Vision Sensors

Privacy Preserving Optics for Miniature Vision Sensors Privacy Preserving Optics for Miniature Vision Sensors Francesco Pittaluga and Sanjeev J. Koppal University of Florida Electrical and Computer Engineering Shoham et al. 07, Wood 08, Enikov et al. 09, Agrihouse

More information

Computational Approaches to Cameras

Computational Approaches to Cameras Computational Approaches to Cameras 11/16/17 Magritte, The False Mirror (1935) Computational Photography Derek Hoiem, University of Illinois Announcements Final project proposal due Monday (see links on

More information

Single Camera Catadioptric Stereo System

Single Camera Catadioptric Stereo System Single Camera Catadioptric Stereo System Abstract In this paper, we present a framework for novel catadioptric stereo camera system that uses a single camera and a single lens with conic mirrors. Various

More information

LENSLESS IMAGING BY COMPRESSIVE SENSING

LENSLESS IMAGING BY COMPRESSIVE SENSING LENSLESS IMAGING BY COMPRESSIVE SENSING Gang Huang, Hong Jiang, Kim Matthews and Paul Wilford Bell Labs, Alcatel-Lucent, Murray Hill, NJ 07974 ABSTRACT In this paper, we propose a lensless compressive

More information

MIT CSAIL Advances in Computer Vision Fall Problem Set 6: Anaglyph Camera Obscura

MIT CSAIL Advances in Computer Vision Fall Problem Set 6: Anaglyph Camera Obscura MIT CSAIL 6.869 Advances in Computer Vision Fall 2013 Problem Set 6: Anaglyph Camera Obscura Posted: Tuesday, October 8, 2013 Due: Thursday, October 17, 2013 You should submit a hard copy of your work

More information

Machine Vision for General Cameras for Quality Testing and Dimension Calculations

Machine Vision for General Cameras for Quality Testing and Dimension Calculations Machine Vision for General Cameras for Quality Testing and Dimension Calculations Ashwath Narayan Murali Abstract This paper looks into an economical way to bring machine vision to smartphones and basic

More information

Survey on Impulse Noise Suppression Techniques for Digital Images

Survey on Impulse Noise Suppression Techniques for Digital Images Survey on Impulse Noise Suppression Techniques for Digital Images 1PG Student, Department of Electronics and Communication Engineering, Punjabi University, Patiala, India 2Assistant Professor, Department

More information

Method for out-of-focus camera calibration

Method for out-of-focus camera calibration 2346 Vol. 55, No. 9 / March 20 2016 / Applied Optics Research Article Method for out-of-focus camera calibration TYLER BELL, 1 JING XU, 2 AND SONG ZHANG 1, * 1 School of Mechanical Engineering, Purdue

More information

Defocus Map Estimation from a Single Image

Defocus Map Estimation from a Single Image Defocus Map Estimation from a Single Image Shaojie Zhuo Terence Sim School of Computing, National University of Singapore, Computing 1, 13 Computing Drive, Singapore 117417, SINGAPOUR Abstract In this

More information

Photographing Long Scenes with Multiviewpoint

Photographing Long Scenes with Multiviewpoint Photographing Long Scenes with Multiviewpoint Panoramas A. Agarwala, M. Agrawala, M. Cohen, D. Salesin, R. Szeliski Presenter: Stacy Hsueh Discussant: VasilyVolkov Motivation Want an image that shows an

More information

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK A NEW METHOD FOR DETECTION OF NOISE IN CORRUPTED IMAGE NIKHIL NALE 1, ANKIT MUNE

More information

Light field sensing. Marc Levoy. Computer Science Department Stanford University

Light field sensing. Marc Levoy. Computer Science Department Stanford University Light field sensing Marc Levoy Computer Science Department Stanford University The scalar light field (in geometrical optics) Radiance as a function of position and direction in a static scene with fixed

More information

THE RESTORATION OF DEFOCUS IMAGES WITH LINEAR CHANGE DEFOCUS RADIUS

THE RESTORATION OF DEFOCUS IMAGES WITH LINEAR CHANGE DEFOCUS RADIUS THE RESTORATION OF DEFOCUS IMAGES WITH LINEAR CHANGE DEFOCUS RADIUS 1 LUOYU ZHOU 1 College of Electronics and Information Engineering, Yangtze University, Jingzhou, Hubei 43423, China E-mail: 1 luoyuzh@yangtzeu.edu.cn

More information

Image Deblurring with Blurred/Noisy Image Pairs

Image Deblurring with Blurred/Noisy Image Pairs Image Deblurring with Blurred/Noisy Image Pairs Huichao Ma, Buping Wang, Jiabei Zheng, Menglian Zhou April 26, 2013 1 Abstract Photos taken under dim lighting conditions by a handheld camera are usually

More information

La photographie numérique. Frank NIELSEN Lundi 7 Juin 2010

La photographie numérique. Frank NIELSEN Lundi 7 Juin 2010 La photographie numérique Frank NIELSEN Lundi 7 Juin 2010 1 Le Monde digital Key benefits of the analog2digital paradigm shift? Dissociate contents from support : binarize Universal player (CPU, Turing

More information

Burst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University!

Burst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University! Burst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University! Motivation! wikipedia! exposure sequence! -4 stops! Motivation!

More information

Improving Signal- to- noise Ratio in Remotely Sensed Imagery Using an Invertible Blur Technique

Improving Signal- to- noise Ratio in Remotely Sensed Imagery Using an Invertible Blur Technique Improving Signal- to- noise Ratio in Remotely Sensed Imagery Using an Invertible Blur Technique Linda K. Le a and Carl Salvaggio a a Rochester Institute of Technology, Center for Imaging Science, Digital

More information

Deconvolution , , Computational Photography Fall 2018, Lecture 12

Deconvolution , , Computational Photography Fall 2018, Lecture 12 Deconvolution http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 12 Course announcements Homework 3 is out. - Due October 12 th. - Any questions?

More information

Douglas Photo. Version for iosand Android

Douglas Photo. Version for iosand Android Douglas Photo Calculator Version 3.2.4 for iosand Android Douglas Software 2007-2017 Contents Introduction... 1 Installation... 2 Running the App... 3 Example Calculations... 5 Photographic Definitions...

More information

The ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do?

The ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do? Computational Photography The ultimate camera What does it do? Image from Durand & Freeman s MIT Course on Computational Photography Today s reading Szeliski Chapter 9 The ultimate camera Infinite resolution

More information

A Review over Different Blur Detection Techniques in Image Processing

A Review over Different Blur Detection Techniques in Image Processing A Review over Different Blur Detection Techniques in Image Processing 1 Anupama Sharma, 2 Devarshi Shukla 1 E.C.E student, 2 H.O.D, Department of electronics communication engineering, LR College of engineering

More information

Opto Engineering S.r.l.

Opto Engineering S.r.l. TUTORIAL #1 Telecentric Lenses: basic information and working principles On line dimensional control is one of the most challenging and difficult applications of vision systems. On the other hand, besides

More information

Recent advances in deblurring and image stabilization. Michal Šorel Academy of Sciences of the Czech Republic

Recent advances in deblurring and image stabilization. Michal Šorel Academy of Sciences of the Czech Republic Recent advances in deblurring and image stabilization Michal Šorel Academy of Sciences of the Czech Republic Camera shake stabilization Alternative to OIS (optical image stabilization) systems Should work

More information

Extended depth of field for visual measurement systems with depth-invariant magnification

Extended depth of field for visual measurement systems with depth-invariant magnification Extended depth of field for visual measurement systems with depth-invariant magnification Yanyu Zhao a and Yufu Qu* a,b a School of Instrument Science and Opto-Electronic Engineering, Beijing University

More information

Princeton University COS429 Computer Vision Problem Set 1: Building a Camera

Princeton University COS429 Computer Vision Problem Set 1: Building a Camera Princeton University COS429 Computer Vision Problem Set 1: Building a Camera What to submit: You need to submit two files: one PDF file for the report that contains your name, Princeton NetID, all the

More information

Deconvolution , , Computational Photography Fall 2017, Lecture 17

Deconvolution , , Computational Photography Fall 2017, Lecture 17 Deconvolution http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 17 Course announcements Homework 4 is out. - Due October 26 th. - There was another

More information

To Denoise or Deblur: Parameter Optimization for Imaging Systems

To Denoise or Deblur: Parameter Optimization for Imaging Systems To Denoise or Deblur: Parameter Optimization for Imaging Systems Kaushik Mitra a, Oliver Cossairt b and Ashok Veeraraghavan a a Electrical and Computer Engineering, Rice University, Houston, TX 77005 b

More information

Extended depth-of-field in Integral Imaging by depth-dependent deconvolution

Extended depth-of-field in Integral Imaging by depth-dependent deconvolution Extended depth-of-field in Integral Imaging by depth-dependent deconvolution H. Navarro* 1, G. Saavedra 1, M. Martinez-Corral 1, M. Sjöström 2, R. Olsson 2, 1 Dept. of Optics, Univ. of Valencia, E-46100,

More information

Introduction. Related Work

Introduction. Related Work Introduction Depth of field is a natural phenomenon when it comes to both sight and photography. The basic ray tracing camera model is insufficient at representing this essential visual element and will

More information

International Journal of Advancedd Research in Biology, Ecology, Science and Technology (IJARBEST)

International Journal of Advancedd Research in Biology, Ecology, Science and Technology (IJARBEST) Gaussian Blur Removal in Digital Images A.Elakkiya 1, S.V.Ramyaa 2 PG Scholars, M.E. VLSI Design, SSN College of Engineering, Rajiv Gandhi Salai, Kalavakkam 1,2 Abstract In many imaging systems, the observed

More information

Computational Photography

Computational Photography Computational photography Computational Photography Digital Visual Effects Yung-Yu Chuang wikipedia: Computational photography h refers broadly to computational imaging techniques that enhance or extend

More information

PTC School of Photography. Beginning Course Class 2 - Exposure

PTC School of Photography. Beginning Course Class 2 - Exposure PTC School of Photography Beginning Course Class 2 - Exposure Today s Topics: What is Exposure Shutter Speed for Exposure Shutter Speed for Motion Aperture for Exposure Aperture for Depth of Field Exposure

More information

Light-Field Database Creation and Depth Estimation

Light-Field Database Creation and Depth Estimation Light-Field Database Creation and Depth Estimation Abhilash Sunder Raj abhisr@stanford.edu Michael Lowney mlowney@stanford.edu Raj Shah shahraj@stanford.edu Abstract Light-field imaging research has been

More information

A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung,

A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung, IJCSNS International Journal of Computer Science and Network Security, VOL.11 No.9, September 2011 55 A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang,

More information

How do we see the world?

How do we see the world? The Camera 1 How do we see the world? Let s design a camera Idea 1: put a piece of film in front of an object Do we get a reasonable image? Credit: Steve Seitz 2 Pinhole camera Idea 2: Add a barrier to

More information

Digital Photographic Imaging Using MOEMS

Digital Photographic Imaging Using MOEMS Digital Photographic Imaging Using MOEMS Vasileios T. Nasis a, R. Andrew Hicks b and Timothy P. Kurzweg a a Department of Electrical and Computer Engineering, Drexel University, Philadelphia, USA b Department

More information

Implementation of Image Deblurring Techniques in Java

Implementation of Image Deblurring Techniques in Java Implementation of Image Deblurring Techniques in Java Peter Chapman Computer Systems Lab 2007-2008 Thomas Jefferson High School for Science and Technology Alexandria, Virginia January 22, 2008 Abstract

More information

Performance Evaluation of Different Depth From Defocus (DFD) Techniques

Performance Evaluation of Different Depth From Defocus (DFD) Techniques Please verify that () all pages are present, () all figures are acceptable, (3) all fonts and special characters are correct, and () all text and figures fit within the Performance Evaluation of Different

More information

Noise Reduction Technique in Synthetic Aperture Radar Datasets using Adaptive and Laplacian Filters

Noise Reduction Technique in Synthetic Aperture Radar Datasets using Adaptive and Laplacian Filters RESEARCH ARTICLE OPEN ACCESS Noise Reduction Technique in Synthetic Aperture Radar Datasets using Adaptive and Laplacian Filters Sakshi Kukreti*, Amit Joshi*, Sudhir Kumar Chaturvedi* *(Department of Aerospace

More information

Image formation - Cameras. Grading & Project. About the course. Tentative Schedule. Course Content. Students introduction

Image formation - Cameras. Grading & Project. About the course. Tentative Schedule. Course Content. Students introduction About the course Instructors: Haibin Ling (hbling@temple, Wachman 35) Hours Lecture: Tuesda 5:3-8:pm, TTLMAN 43B Office hour: Tuesda 3: - 5:pm, or b appointment Textbook Computer Vision: Models, Learning,

More information

An Image Matching Method for Digital Images Using Morphological Approach

An Image Matching Method for Digital Images Using Morphological Approach An Image Matching Method for Digital Images Using Morphological Approach Pinaki Pratim Acharjya, Dibyendu Ghoshal Abstract Image matching methods play a key role in deciding correspondence between two

More information

Dynamically Reparameterized Light Fields & Fourier Slice Photography. Oliver Barth, 2009 Max Planck Institute Saarbrücken

Dynamically Reparameterized Light Fields & Fourier Slice Photography. Oliver Barth, 2009 Max Planck Institute Saarbrücken Dynamically Reparameterized Light Fields & Fourier Slice Photography Oliver Barth, 2009 Max Planck Institute Saarbrücken Background What we are talking about? 2 / 83 Background What we are talking about?

More information

Multi-Resolution Estimation of Optical Flow on Vehicle Tracking under Unpredictable Environments

Multi-Resolution Estimation of Optical Flow on Vehicle Tracking under Unpredictable Environments , pp.32-36 http://dx.doi.org/10.14257/astl.2016.129.07 Multi-Resolution Estimation of Optical Flow on Vehicle Tracking under Unpredictable Environments Viet Dung Do 1 and Dong-Min Woo 1 1 Department of

More information

Malaysian Car Number Plate Detection System Based on Template Matching and Colour Information

Malaysian Car Number Plate Detection System Based on Template Matching and Colour Information Malaysian Car Number Plate Detection System Based on Template Matching and Colour Information Mohd Firdaus Zakaria, Shahrel A. Suandi Intelligent Biometric Group, School of Electrical and Electronics Engineering,

More information