A Mathematical model for the determination of distance of an object in a 2D image
|
|
- Archibald Kelly
- 5 years ago
- Views:
Transcription
1 A Mathematical model for the determination of distance of an object in a 2D image Deepu R 1, Murali S 2,Vikram Raju 3 Maharaja Institute of Technology Mysore, Karnataka, India rdeepusingh@mitmysore.in 1, murali@mitmysore.in 2, vikramraju17@gmail.com 3 Abstract-Distance measurement in real world has always been one of the challenging tasks to be performed in the field of computer vision. Photographic images are two dimensional depiction of three-dimensional real space. Various methods have been developed, researched and improvised over the years depending on the focus of the object, defocus of the background, the vanishing point of an image and other information derived from images. The impact of various features of the camera are also considered when the above problem is under study like, focal length of the camera, ISO levels, white balance, resolution etc. The effect of the resolution of an object in an image, its actual size, the focal length of the camera, the camera resolution, all contribute in determining the distance between the camera and the object. In this paper we have proposed a model for the relationship between the resolution of an object of interest and the distance from the camera as a growth series. Relationship between the resolution of an object of interest and focal length is also expressed as a growth series model. The object of interest is segmented out and its pixels account to its resolution, which is standardized and used to determine either the distance or the focal length of the camera or the vice versa. This model, thus, very easily helps in tracking the object using a camera with minimal error rate, provided the image is along the optical axis of the camera. Keywords: Resolution, Focal Length, Camera and object distance, growth series, Pixels per inch, Geometric Progression. 1 Introduction The knowledge about the distance of an object in an image has its applications in many fields of science and technology and research, like, human motion estimation, web conferencing, gaming industry, surveillance, security systems, robotics, medical systems and imaging. Etc. The major factors in a photo that cannot be changed once the photo has been clicked are Aperture of the lens, focal length, number of pixels or resolution and shutter speed. A method which allows refocusing is a potential powerful tool for digital image editing. Once the depth map has been obtained, one can de-blur the image in order to acquire an all focus image or blur the image even more to create certain visual effects [1]. The depth map can also apply to the task such as automatic scene segmentation, post exposure refocusing and rerendering of the scene from an alternative view point. By analyzing the depth of field, the coarse depth map of scene can be recovered [2]. Depth from defocus and depth from focus are the two methods to estimate the 3D geometry of the scene by exploiting image focus [3]. In adaptive depth from focus, depth estimation method for images in narrow depth of field setting, the segments produced by mean shift segmentation as windows to analyze the focus measure and employ a hierarchical Markov Random Field to infer from the depth map [4]. Ina non-linear approach, the precise sparse depth map extraction, the noisy focus measurements are replaced with estimated values, which in turn helps to accurately compute 3D shapes while preserving edges of objects [5].In paper [6],authors have proposed a real-time method which can measure distance using a modified camera. The camera s image sensor is inclined by a certain angle. Thus, the image projected on the sensor is defocused differently at different areas. The area where the image is focused best is just at projection plane. The position of the projection plane can be obtained after finding out this area. The distance between the projection plane and the lens of the camera is image distance. Object distance can be obtained by applying image distance to lens formula. Although the results of the above methods are very promising, it is computationally expensive to employ Markov Random Fields or any of the procedures given in above papers. In real world applications it is necessary to provide real-time computations. By studying the model or rate at which features change in an image like size, distance, pixels of an object, focal length, give us a better approximation and a simpler computationally efficient method to determine the relationship among them.
2 The resolution of an image or the number of pixels of an image is a constant, which depends on the camera. Once an image is captured, its focal length is also constant. When an image is captured by a camera, the distance of the image from the camera or the viewer, thus can be determined using the above variables. The resolution also directly depends on the focal length, since it has a proportional relationship with the image size. The camera, like the human eye, works on perspective vision, where, the change in image size with variation of Focal Length and Distance is not linear. To study the relationship among the above, the following methodology is proposed. 2 Method With an objective of designing a robust model, sequence of images at different distances D from the object and for different focal lengths are taken along the optical axis. Every image is of size ix j and it is acquired for objects of various shapes and sizes. The total number of pixels P, in the image is thus given by = (1) The set of the pixels in an object P o,is a subset of P. The image is converted to grey scale using a basic thresholding algorithm to obtain the object of interest and the number of pixels in the object is obtained. This becomes the set of pixels in an object, and stored in a set P o. The behavior of the relationship between the distance with respect to the resolution factor R f and that of the focal length and the resolution factor is sought. Figure 1: Experimental setup varying distance from camera to objects. Once the images are captured, the background is segmented, to get only the pixels of the object of interest. Image preprocessing is necessary at times, when noise is found in the images, when the automatic thresholding algorithm fails to extract the object clearly. Manual image cleaning was used to remove unwanted noise in the image. This can be avoided using a better efficient object extraction algorithm The image segmentation and cleaning was carried out using MATLAB. To build the mathematical model, IBM SPSS was used. To determine the resolution factor, first, the Resolution per Inch, R In, of the object is determined. It is then multiplied by the resolution of the camera and later natural log is taken to scale down the data. This is so done in order to standardize the factor arising due to the resolution of the images and of the objects in them, thus making the method applicable to any kind of camera and of any resolution. In a way, it makes the method independent of the resolution itself. = ln ( ) (2) Figure1 shows the variation of size of the image with the changes in distance from the camera. Figure 2: Test Images (left) and Objects extracted from the image (right)
3 3 Results The results can be analyzed in two folds, one model with resolution factor and distance of object in an image and the other with resolution factor and the focal length of the camera. Following results were tabulated for 3 different objects. Table 1 gives the difference between expected distance and obtained distance for circle object where as Table2 and Tables3 specify the same for square and triangle objects. We can also note that there is a relationship between pixels in an object to distance which is expressed by our model. Table 1:Observed results for circle object of area sq.in Table 4: Model Summary and Parameter Estimates for Distance and Resolution Factor for a focal length of 18mm. Equation Dependent Variable: Distance R Square Model Summary Parameter Estimates F df1 df2 Sig. Constant b1 Growth The R Square value shows a 99.9% significance to the model, which is a good fit for the data set observed. Table 2:Observed results square object of area sq.in Table 3: Observed results triangle object of area 20 sq.in Figure 3: The growth series of distance and resolution factor of object in image The model agrees completely with the growth series and gives us the following equation to determine the distance of the object from the camera: = (3) Part I: Resolution Factor and Distance The distance is considered as the dependent variable in the analysis of the model and the resolution factor is taken as the independent variable. In each case of the focal length the model is tested. Where, D is the distance of the object from the camera, A and B are constants in the growth series, which are determined by a geometric series, and R f is the resolution factor. The r in the geometric ratio is 1.01.The geometric series gives the values of the constants from the above equation.
4 Table 5: The geometric series to determine the constants of the growth series model. Focal Length A B 18mm mm r r 30mm r r 2 36mm r r 3 42mm r r 4 48mm r r 5 54mm r r 6 60mm r r 7 66mm r r 8 The geometric series gives the values of the constants from the above equation. The model was tested on random image samples, which agreed with the growth series model, with very minimal error. The error in determining the distance was in the range of percent. Part II: Resolution Factor and Focal Length The focal length is considered the dependent variable in the analysis of the model and the resolution factor is taken as the independent variable. In each case of the distance of the object from the camera, the model is tested. Table 6: Model Summary and Parameter Estimates for Distance and Resolution Factor at 51 inches from the camera. Equation Dependent Variable: Focal Length R Square Model Summary Parameter Estimates F df1 df2 Sig. Constant b1 Growth The R Square value shows a 99.1% significance to the model, which is a good fit for the data set observed. The model agrees completely with the growth series and gives us the following equation to determine the focal length: = (4) Where, F is the focal length of the camera, A and B are constants in the growth series, which are determined by a geometric series, and R f is the resolution factor. The r in the geometric ratio is Figure 4: The growth series of focal length and resolution factor of object in an image The model was tested on random image samples, which agreed with the growth series model, with very minimal error. The error in determining the distance was in the range of 1-2 percent. Table 7: The geometric series to determine the constants of the growth series model. Distance A B 51in in r 0.543r 71in r r 2 81in r r 3 91in r r 4 101in r r 5 111in r r 6 121in r r 7 131in r r 8 141in r r 9 4 Applications The model finds its use in many applications like obstacle detection in robotics, detection of road humps, navigation systems for the blind, unknown territory exploration, 3D game development are a few to name.
5 5 Future Work The distance of the object from the camera and focal length model is applicable only when the object is head-on or in the line of the optical axis of the camera. Future work can be done on studying the effect of resolution and distance for images not in the axis of the camera, considering the evolution of perspective vision of an image. It can also be used to create 3D scenes by determining the distance from the camera and rebuilding another scene with the same distance. 6 Conclusions The model shows promising enhancement on further development and an efficient model to determine the distance of an object in an image from the camera making use of the data in the resolution of an image, which cannot change once an image is taken. Building the model for varied focal lengths enables the model to work with cameras of any focal length. The obtained growth series enables to work with distance and focal length. With knowing one of the two, the other can be calculated and be easily modeled in the system. Computationally the calculation of the growth series works efficiently than the iterative approach. 7 Acknowledgements The work presented in this paper was supported by VTU research grants (Ref. No. VTU/ Aca./ /A-9/13551). We thank Visvesvaraya Technological University, Belgaum for this. 8 References [1] Potmesil, M., and Chakravarty, I., A lens and aperture camera model for synthetic image generation.in proc. Siggraph, [2] Pentland, A. P., A new sense for depth of field.ieee Trans. Pattern anal. Mach. Intell. 9, 4, [3] Hasinoff S. W., Kutulakos K. N., A layerbased restoration framework for variableaperture photography, F, 2007 [C], IEEE. [4] Bing-Zhong Jing, Daniel S. Yeung, Recovering depth from images using adaptive depth from focus. IEEE /12, [5] Muhammad Tariq Mahmood, Ikhyun Lee, Wook-Jin Choi, Tae-Sun Choi, A non-linear approach for depth from focus for digital cameras. IEEE /11, [6] Liu Xiaoming, Qin Tian, Chen Wanchen, Yin Xingliang, Real Time Distance Measurment Using a Modified Camera.IEEE /09.
Computational Camera & Photography: Coded Imaging
Computational Camera & Photography: Coded Imaging Camera Culture Ramesh Raskar MIT Media Lab http://cameraculture.media.mit.edu/ Image removed due to copyright restrictions. See Fig. 1, Eight major types
More informationA Recognition of License Plate Images from Fast Moving Vehicles Using Blur Kernel Estimation
A Recognition of License Plate Images from Fast Moving Vehicles Using Blur Kernel Estimation Kalaivani.R 1, Poovendran.R 2 P.G. Student, Dept. of ECE, Adhiyamaan College of Engineering, Hosur, Tamil Nadu,
More informationDappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing
Dappled Photography: Mask Enhanced Cameras for Heterodyned Light Fields and Coded Aperture Refocusing Ashok Veeraraghavan, Ramesh Raskar, Ankit Mohan & Jack Tumblin Amit Agrawal, Mitsubishi Electric Research
More informationCoded Computational Photography!
Coded Computational Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 9! Gordon Wetzstein! Stanford University! Coded Computational Photography - Overview!!
More informationSURVEILLANCE SYSTEMS WITH AUTOMATIC RESTORATION OF LINEAR MOTION AND OUT-OF-FOCUS BLURRED IMAGES. Received August 2008; accepted October 2008
ICIC Express Letters ICIC International c 2008 ISSN 1881-803X Volume 2, Number 4, December 2008 pp. 409 414 SURVEILLANCE SYSTEMS WITH AUTOMATIC RESTORATION OF LINEAR MOTION AND OUT-OF-FOCUS BLURRED IMAGES
More informationProject 4 Results http://www.cs.brown.edu/courses/cs129/results/proj4/jcmace/ http://www.cs.brown.edu/courses/cs129/results/proj4/damoreno/ http://www.cs.brown.edu/courses/csci1290/results/proj4/huag/
More informationA moment-preserving approach for depth from defocus
A moment-preserving approach for depth from defocus D. M. Tsai and C. T. Lin Machine Vision Lab. Department of Industrial Engineering and Management Yuan-Ze University, Chung-Li, Taiwan, R.O.C. E-mail:
More informationImplementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring
Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring Ashill Chiranjan and Bernardt Duvenhage Defence, Peace, Safety and Security Council for Scientific
More informationLenses, exposure, and (de)focus
Lenses, exposure, and (de)focus http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 15 Course announcements Homework 4 is out. - Due October 26
More informationOn the Recovery of Depth from a Single Defocused Image
On the Recovery of Depth from a Single Defocused Image Shaojie Zhuo and Terence Sim School of Computing National University of Singapore Singapore,747 Abstract. In this paper we address the challenging
More informationModeling and Synthesis of Aperture Effects in Cameras
Modeling and Synthesis of Aperture Effects in Cameras Douglas Lanman, Ramesh Raskar, and Gabriel Taubin Computational Aesthetics 2008 20 June, 2008 1 Outline Introduction and Related Work Modeling Vignetting
More informationCoded Aperture for Projector and Camera for Robust 3D measurement
Coded Aperture for Projector and Camera for Robust 3D measurement Yuuki Horita Yuuki Matugano Hiroki Morinaga Hiroshi Kawasaki Satoshi Ono Makoto Kimura Yasuo Takane Abstract General active 3D measurement
More informationWhy learn about photography in this course?
Why learn about photography in this course? Geri's Game: Note the background is blurred. - photography: model of image formation - Many computer graphics methods use existing photographs e.g. texture &
More informationlecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response
lecture 24 image capture - photography: model of image formation - image blur - camera settings (f-number, shutter speed) - exposure - camera response - application: high dynamic range imaging Why learn
More informationSingle Digital Image Multi-focusing Using Point to Point Blur Model Based Depth Estimation
Single Digital mage Multi-focusing Using Point to Point Blur Model Based Depth Estimation Praveen S S, Aparna P R Abstract The proposed paper focuses on Multi-focusing, a technique that restores all-focused
More informationToward Non-stationary Blind Image Deblurring: Models and Techniques
Toward Non-stationary Blind Image Deblurring: Models and Techniques Ji, Hui Department of Mathematics National University of Singapore NUS, 30-May-2017 Outline of the talk Non-stationary Image blurring
More informationNTU CSIE. Advisor: Wu Ja Ling, Ph.D.
An Interactive Background Blurring Mechanism and Its Applications NTU CSIE Yan Chih Yu Advisor: Wu Ja Ling, Ph.D. 1 2 Outline Introduction Related Work Method Object Segmentation Depth Map Generation Image
More information4 STUDY OF DEBLURRING TECHNIQUES FOR RESTORED MOTION BLURRED IMAGES
4 STUDY OF DEBLURRING TECHNIQUES FOR RESTORED MOTION BLURRED IMAGES Abstract: This paper attempts to undertake the study of deblurring techniques for Restored Motion Blurred Images by using: Wiener filter,
More informationDepth from Focusing and Defocusing. Carnegie Mellon University. Pittsburgh, PA result is 1.3% RMS error in terms of distance
Depth from Focusing and Defocusing Yalin Xiong Steven A. Shafer The Robotics Institute Carnegie Mellon University Pittsburgh, PA 53 Abstract This paper studies the problem of obtaining depth information
More informationLinear Gaussian Method to Detect Blurry Digital Images using SIFT
IJCAES ISSN: 2231-4946 Volume III, Special Issue, November 2013 International Journal of Computer Applications in Engineering Sciences Special Issue on Emerging Research Areas in Computing(ERAC) www.caesjournals.org
More informationDistance Estimation with a Two or Three Aperture SLR Digital Camera
Distance Estimation with a Two or Three Aperture SLR Digital Camera Seungwon Lee, Joonki Paik, and Monson H. Hayes Graduate School of Advanced Imaging Science, Multimedia, and Film Chung-Ang University
More informationRecent Advances in Image Deblurring. Seungyong Lee (Collaboration w/ Sunghyun Cho)
Recent Advances in Image Deblurring Seungyong Lee (Collaboration w/ Sunghyun Cho) Disclaimer Many images and figures in this course note have been copied from the papers and presentation materials of previous
More informationComputational Cameras. Rahul Raguram COMP
Computational Cameras Rahul Raguram COMP 790-090 What is a computational camera? Camera optics Camera sensor 3D scene Traditional camera Final image Modified optics Camera sensor Image Compute 3D scene
More informationDesign of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems
Design of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems Ricardo R. Garcia University of California, Berkeley Berkeley, CA rrgarcia@eecs.berkeley.edu Abstract In recent
More informationSUPER RESOLUTION INTRODUCTION
SUPER RESOLUTION Jnanavardhini - Online MultiDisciplinary Research Journal Ms. Amalorpavam.G Assistant Professor, Department of Computer Sciences, Sambhram Academy of Management. Studies, Bangalore Abstract:-
More informationDEFOCUS BLUR PARAMETER ESTIMATION TECHNIQUE
International Journal of Electronics and Communication Engineering and Technology (IJECET) Volume 7, Issue 4, July-August 2016, pp. 85 90, Article ID: IJECET_07_04_010 Available online at http://www.iaeme.com/ijecet/issues.asp?jtype=ijecet&vtype=7&itype=4
More informationBlind Single-Image Super Resolution Reconstruction with Defocus Blur
Sensors & Transducers 2014 by IFSA Publishing, S. L. http://www.sensorsportal.com Blind Single-Image Super Resolution Reconstruction with Defocus Blur Fengqing Qin, Lihong Zhu, Lilan Cao, Wanan Yang Institute
More informationLicense Plate Localisation based on Morphological Operations
License Plate Localisation based on Morphological Operations Xiaojun Zhai, Faycal Benssali and Soodamani Ramalingam School of Engineering & Technology University of Hertfordshire, UH Hatfield, UK Abstract
More informationMEM: Intro to Robotics. Assignment 3I. Due: Wednesday 10/15 11:59 EST
MEM: Intro to Robotics Assignment 3I Due: Wednesday 10/15 11:59 EST 1. Basic Optics You are shopping for a new lens for your Canon D30 digital camera and there are lots of lens options at the store. Your
More informationCAMERA BASICS. Stops of light
CAMERA BASICS Stops of light A stop of light isn t a quantifiable measurement it s a relative measurement. A stop of light is defined as a doubling or halving of any quantity of light. The word stop is
More informationCoded photography , , Computational Photography Fall 2018, Lecture 14
Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 14 Overview of today s lecture The coded photography paradigm. Dealing with
More informationInternational Journal of Innovative Research in Engineering Science and Technology APRIL 2018 ISSN X
HIGH DYNAMIC RANGE OF MULTISPECTRAL ACQUISITION USING SPATIAL IMAGES 1 M.Kavitha, M.Tech., 2 N.Kannan, M.E., and 3 S.Dharanya, M.E., 1 Assistant Professor/ CSE, Dhirajlal Gandhi College of Technology,
More informationColour correction for panoramic imaging
Colour correction for panoramic imaging Gui Yun Tian Duke Gledhill Dave Taylor The University of Huddersfield David Clarke Rotography Ltd Abstract: This paper reports the problem of colour distortion in
More informationDisplacement Measurement of Burr Arch-Truss Under Dynamic Loading Based on Image Processing Technology
6 th International Conference on Advances in Experimental Structural Engineering 11 th International Workshop on Advanced Smart Materials and Smart Structures Technology August 1-2, 2015, University of
More informationBASIC OPERATIONS IN IMAGE PROCESSING USING MATLAB
BASIC OPERATIONS IN IMAGE PROCESSING USING MATLAB Er.Amritpal Kaur 1,Nirajpal Kaur 2 1,2 Assistant Professor,Guru Nanak Dev University, Regional Campus, Gurdaspur Abstract: - This paper aims at basic image
More informationOverview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image
Camera & Color Overview Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image Book: Hartley 6.1, Szeliski 2.1.5, 2.2, 2.3 The trip
More informationSimulated Programmable Apertures with Lytro
Simulated Programmable Apertures with Lytro Yangyang Yu Stanford University yyu10@stanford.edu Abstract This paper presents a simulation method using the commercial light field camera Lytro, which allows
More informationName: Date: Math in Special Effects: Try Other Challenges. Student Handout
Name: Date: Math in Special Effects: Try Other Challenges When filming special effects, a high-speed photographer needs to control the duration and impact of light by adjusting a number of settings, including
More informationPhoto-Consistent Motion Blur Modeling for Realistic Image Synthesis
Photo-Consistent Motion Blur Modeling for Realistic Image Synthesis Huei-Yung Lin and Chia-Hong Chang Department of Electrical Engineering, National Chung Cheng University, 168 University Rd., Min-Hsiung
More informationMotion Estimation from a Single Blurred Image
Motion Estimation from a Single Blurred Image Image Restoration: De-Blurring Build a Blur Map Adapt Existing De-blurring Techniques to real blurred images Analysis, Reconstruction and 3D reconstruction
More informationNear-Invariant Blur for Depth and 2D Motion via Time-Varying Light Field Analysis
Near-Invariant Blur for Depth and 2D Motion via Time-Varying Light Field Analysis Yosuke Bando 1,2 Henry Holtzman 2 Ramesh Raskar 2 1 Toshiba Corporation 2 MIT Media Lab Defocus & Motion Blur PSF Depth
More informationTo Do. Advanced Computer Graphics. Outline. Computational Imaging. How do we see the world? Pinhole camera
Advanced Computer Graphics CSE 163 [Spring 2017], Lecture 14 Ravi Ramamoorthi http://www.cs.ucsd.edu/~ravir To Do Assignment 2 due May 19 Any last minute issues or questions? Next two lectures: Imaging,
More informationImproving Image Quality by Camera Signal Adaptation to Lighting Conditions
Improving Image Quality by Camera Signal Adaptation to Lighting Conditions Mihai Negru and Sergiu Nedevschi Technical University of Cluj-Napoca, Computer Science Department Mihai.Negru@cs.utcluj.ro, Sergiu.Nedevschi@cs.utcluj.ro
More informationSingle-Image Shape from Defocus
Single-Image Shape from Defocus José R.A. Torreão and João L. Fernandes Instituto de Computação Universidade Federal Fluminense 24210-240 Niterói RJ, BRAZIL Abstract The limited depth of field causes scene
More informationMulti Viewpoint Panoramas
27. November 2007 1 Motivation 2 Methods Slit-Scan "The System" 3 "The System" Approach Preprocessing Surface Selection Panorama Creation Interactive Renement 4 Sources Motivation image showing long continous
More informationCoded photography , , Computational Photography Fall 2017, Lecture 18
Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 18 Course announcements Homework 5 delayed for Tuesday. - You will need cameras
More informationDeblurring. Basics, Problem definition and variants
Deblurring Basics, Problem definition and variants Kinds of blur Hand-shake Defocus Credit: Kenneth Josephson Motion Credit: Kenneth Josephson Kinds of blur Spatially invariant vs. Spatially varying
More informationRemoving Temporal Stationary Blur in Route Panoramas
Removing Temporal Stationary Blur in Route Panoramas Jiang Yu Zheng and Min Shi Indiana University Purdue University Indianapolis jzheng@cs.iupui.edu Abstract The Route Panorama is a continuous, compact
More informationRemoval of Salt and Pepper Noise from Satellite Images
Removal of Salt and Pepper Noise from Satellite Images Mr. Yogesh V. Kolhe 1 Research Scholar, Samrat Ashok Technological Institute Vidisha (INDIA) Dr. Yogendra Kumar Jain 2 Guide & Asso.Professor, Samrat
More informationApplications of Flash and No-Flash Image Pairs in Mobile Phone Photography
Applications of Flash and No-Flash Image Pairs in Mobile Phone Photography Xi Luo Stanford University 450 Serra Mall, Stanford, CA 94305 xluo2@stanford.edu Abstract The project explores various application
More informationFocused Image Recovery from Two Defocused
Focused Image Recovery from Two Defocused Images Recorded With Different Camera Settings Murali Subbarao Tse-Chung Wei Gopal Surya Department of Electrical Engineering State University of New York Stony
More informationComputer Vision Slides curtesy of Professor Gregory Dudek
Computer Vision Slides curtesy of Professor Gregory Dudek Ioannis Rekleitis Why vision? Passive (emits nothing). Discreet. Energy efficient. Intuitive. Powerful (works well for us, right?) Long and short
More informationfast blur removal for wearable QR code scanners
fast blur removal for wearable QR code scanners Gábor Sörös, Stephan Semmler, Luc Humair, Otmar Hilliges ISWC 2015, Osaka, Japan traditional barcode scanning next generation barcode scanning ubiquitous
More informationEvolving Measurement Regions for Depth from Defocus
Evolving Measurement Regions for Depth from Defocus Scott McCloskey, Michael Langer, and Kaleem Siddiqi Centre for Intelligent Machines, McGill University {scott,langer,siddiqi}@cim.mcgill.ca Abstract.
More informationComputer Vision. Howie Choset Introduction to Robotics
Computer Vision Howie Choset http://www.cs.cmu.edu.edu/~choset Introduction to Robotics http://generalrobotics.org What is vision? What is computer vision? Edge Detection Edge Detection Interest points
More informationME 6406 MACHINE VISION. Georgia Institute of Technology
ME 6406 MACHINE VISION Georgia Institute of Technology Class Information Instructor Professor Kok-Meng Lee MARC 474 Office hours: Tues/Thurs 1:00-2:00 pm kokmeng.lee@me.gatech.edu (404)-894-7402 Class
More informationPrivacy Preserving Optics for Miniature Vision Sensors
Privacy Preserving Optics for Miniature Vision Sensors Francesco Pittaluga and Sanjeev J. Koppal University of Florida Electrical and Computer Engineering Shoham et al. 07, Wood 08, Enikov et al. 09, Agrihouse
More informationComputational Approaches to Cameras
Computational Approaches to Cameras 11/16/17 Magritte, The False Mirror (1935) Computational Photography Derek Hoiem, University of Illinois Announcements Final project proposal due Monday (see links on
More informationSingle Camera Catadioptric Stereo System
Single Camera Catadioptric Stereo System Abstract In this paper, we present a framework for novel catadioptric stereo camera system that uses a single camera and a single lens with conic mirrors. Various
More informationLENSLESS IMAGING BY COMPRESSIVE SENSING
LENSLESS IMAGING BY COMPRESSIVE SENSING Gang Huang, Hong Jiang, Kim Matthews and Paul Wilford Bell Labs, Alcatel-Lucent, Murray Hill, NJ 07974 ABSTRACT In this paper, we propose a lensless compressive
More informationMIT CSAIL Advances in Computer Vision Fall Problem Set 6: Anaglyph Camera Obscura
MIT CSAIL 6.869 Advances in Computer Vision Fall 2013 Problem Set 6: Anaglyph Camera Obscura Posted: Tuesday, October 8, 2013 Due: Thursday, October 17, 2013 You should submit a hard copy of your work
More informationMachine Vision for General Cameras for Quality Testing and Dimension Calculations
Machine Vision for General Cameras for Quality Testing and Dimension Calculations Ashwath Narayan Murali Abstract This paper looks into an economical way to bring machine vision to smartphones and basic
More informationSurvey on Impulse Noise Suppression Techniques for Digital Images
Survey on Impulse Noise Suppression Techniques for Digital Images 1PG Student, Department of Electronics and Communication Engineering, Punjabi University, Patiala, India 2Assistant Professor, Department
More informationMethod for out-of-focus camera calibration
2346 Vol. 55, No. 9 / March 20 2016 / Applied Optics Research Article Method for out-of-focus camera calibration TYLER BELL, 1 JING XU, 2 AND SONG ZHANG 1, * 1 School of Mechanical Engineering, Purdue
More informationDefocus Map Estimation from a Single Image
Defocus Map Estimation from a Single Image Shaojie Zhuo Terence Sim School of Computing, National University of Singapore, Computing 1, 13 Computing Drive, Singapore 117417, SINGAPOUR Abstract In this
More informationPhotographing Long Scenes with Multiviewpoint
Photographing Long Scenes with Multiviewpoint Panoramas A. Agarwala, M. Agrawala, M. Cohen, D. Salesin, R. Szeliski Presenter: Stacy Hsueh Discussant: VasilyVolkov Motivation Want an image that shows an
More informationINTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY
INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK A NEW METHOD FOR DETECTION OF NOISE IN CORRUPTED IMAGE NIKHIL NALE 1, ANKIT MUNE
More informationLight field sensing. Marc Levoy. Computer Science Department Stanford University
Light field sensing Marc Levoy Computer Science Department Stanford University The scalar light field (in geometrical optics) Radiance as a function of position and direction in a static scene with fixed
More informationTHE RESTORATION OF DEFOCUS IMAGES WITH LINEAR CHANGE DEFOCUS RADIUS
THE RESTORATION OF DEFOCUS IMAGES WITH LINEAR CHANGE DEFOCUS RADIUS 1 LUOYU ZHOU 1 College of Electronics and Information Engineering, Yangtze University, Jingzhou, Hubei 43423, China E-mail: 1 luoyuzh@yangtzeu.edu.cn
More informationImage Deblurring with Blurred/Noisy Image Pairs
Image Deblurring with Blurred/Noisy Image Pairs Huichao Ma, Buping Wang, Jiabei Zheng, Menglian Zhou April 26, 2013 1 Abstract Photos taken under dim lighting conditions by a handheld camera are usually
More informationLa photographie numérique. Frank NIELSEN Lundi 7 Juin 2010
La photographie numérique Frank NIELSEN Lundi 7 Juin 2010 1 Le Monde digital Key benefits of the analog2digital paradigm shift? Dissociate contents from support : binarize Universal player (CPU, Turing
More informationBurst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University!
Burst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University! Motivation! wikipedia! exposure sequence! -4 stops! Motivation!
More informationImproving Signal- to- noise Ratio in Remotely Sensed Imagery Using an Invertible Blur Technique
Improving Signal- to- noise Ratio in Remotely Sensed Imagery Using an Invertible Blur Technique Linda K. Le a and Carl Salvaggio a a Rochester Institute of Technology, Center for Imaging Science, Digital
More informationDeconvolution , , Computational Photography Fall 2018, Lecture 12
Deconvolution http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 12 Course announcements Homework 3 is out. - Due October 12 th. - Any questions?
More informationDouglas Photo. Version for iosand Android
Douglas Photo Calculator Version 3.2.4 for iosand Android Douglas Software 2007-2017 Contents Introduction... 1 Installation... 2 Running the App... 3 Example Calculations... 5 Photographic Definitions...
More informationThe ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do?
Computational Photography The ultimate camera What does it do? Image from Durand & Freeman s MIT Course on Computational Photography Today s reading Szeliski Chapter 9 The ultimate camera Infinite resolution
More informationA Review over Different Blur Detection Techniques in Image Processing
A Review over Different Blur Detection Techniques in Image Processing 1 Anupama Sharma, 2 Devarshi Shukla 1 E.C.E student, 2 H.O.D, Department of electronics communication engineering, LR College of engineering
More informationOpto Engineering S.r.l.
TUTORIAL #1 Telecentric Lenses: basic information and working principles On line dimensional control is one of the most challenging and difficult applications of vision systems. On the other hand, besides
More informationRecent advances in deblurring and image stabilization. Michal Šorel Academy of Sciences of the Czech Republic
Recent advances in deblurring and image stabilization Michal Šorel Academy of Sciences of the Czech Republic Camera shake stabilization Alternative to OIS (optical image stabilization) systems Should work
More informationExtended depth of field for visual measurement systems with depth-invariant magnification
Extended depth of field for visual measurement systems with depth-invariant magnification Yanyu Zhao a and Yufu Qu* a,b a School of Instrument Science and Opto-Electronic Engineering, Beijing University
More informationPrinceton University COS429 Computer Vision Problem Set 1: Building a Camera
Princeton University COS429 Computer Vision Problem Set 1: Building a Camera What to submit: You need to submit two files: one PDF file for the report that contains your name, Princeton NetID, all the
More informationDeconvolution , , Computational Photography Fall 2017, Lecture 17
Deconvolution http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 17 Course announcements Homework 4 is out. - Due October 26 th. - There was another
More informationTo Denoise or Deblur: Parameter Optimization for Imaging Systems
To Denoise or Deblur: Parameter Optimization for Imaging Systems Kaushik Mitra a, Oliver Cossairt b and Ashok Veeraraghavan a a Electrical and Computer Engineering, Rice University, Houston, TX 77005 b
More informationExtended depth-of-field in Integral Imaging by depth-dependent deconvolution
Extended depth-of-field in Integral Imaging by depth-dependent deconvolution H. Navarro* 1, G. Saavedra 1, M. Martinez-Corral 1, M. Sjöström 2, R. Olsson 2, 1 Dept. of Optics, Univ. of Valencia, E-46100,
More informationIntroduction. Related Work
Introduction Depth of field is a natural phenomenon when it comes to both sight and photography. The basic ray tracing camera model is insufficient at representing this essential visual element and will
More informationInternational Journal of Advancedd Research in Biology, Ecology, Science and Technology (IJARBEST)
Gaussian Blur Removal in Digital Images A.Elakkiya 1, S.V.Ramyaa 2 PG Scholars, M.E. VLSI Design, SSN College of Engineering, Rajiv Gandhi Salai, Kalavakkam 1,2 Abstract In many imaging systems, the observed
More informationComputational Photography
Computational photography Computational Photography Digital Visual Effects Yung-Yu Chuang wikipedia: Computational photography h refers broadly to computational imaging techniques that enhance or extend
More informationPTC School of Photography. Beginning Course Class 2 - Exposure
PTC School of Photography Beginning Course Class 2 - Exposure Today s Topics: What is Exposure Shutter Speed for Exposure Shutter Speed for Motion Aperture for Exposure Aperture for Depth of Field Exposure
More informationLight-Field Database Creation and Depth Estimation
Light-Field Database Creation and Depth Estimation Abhilash Sunder Raj abhisr@stanford.edu Michael Lowney mlowney@stanford.edu Raj Shah shahraj@stanford.edu Abstract Light-field imaging research has been
More informationA Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung,
IJCSNS International Journal of Computer Science and Network Security, VOL.11 No.9, September 2011 55 A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang,
More informationHow do we see the world?
The Camera 1 How do we see the world? Let s design a camera Idea 1: put a piece of film in front of an object Do we get a reasonable image? Credit: Steve Seitz 2 Pinhole camera Idea 2: Add a barrier to
More informationDigital Photographic Imaging Using MOEMS
Digital Photographic Imaging Using MOEMS Vasileios T. Nasis a, R. Andrew Hicks b and Timothy P. Kurzweg a a Department of Electrical and Computer Engineering, Drexel University, Philadelphia, USA b Department
More informationImplementation of Image Deblurring Techniques in Java
Implementation of Image Deblurring Techniques in Java Peter Chapman Computer Systems Lab 2007-2008 Thomas Jefferson High School for Science and Technology Alexandria, Virginia January 22, 2008 Abstract
More informationPerformance Evaluation of Different Depth From Defocus (DFD) Techniques
Please verify that () all pages are present, () all figures are acceptable, (3) all fonts and special characters are correct, and () all text and figures fit within the Performance Evaluation of Different
More informationNoise Reduction Technique in Synthetic Aperture Radar Datasets using Adaptive and Laplacian Filters
RESEARCH ARTICLE OPEN ACCESS Noise Reduction Technique in Synthetic Aperture Radar Datasets using Adaptive and Laplacian Filters Sakshi Kukreti*, Amit Joshi*, Sudhir Kumar Chaturvedi* *(Department of Aerospace
More informationImage formation - Cameras. Grading & Project. About the course. Tentative Schedule. Course Content. Students introduction
About the course Instructors: Haibin Ling (hbling@temple, Wachman 35) Hours Lecture: Tuesda 5:3-8:pm, TTLMAN 43B Office hour: Tuesda 3: - 5:pm, or b appointment Textbook Computer Vision: Models, Learning,
More informationAn Image Matching Method for Digital Images Using Morphological Approach
An Image Matching Method for Digital Images Using Morphological Approach Pinaki Pratim Acharjya, Dibyendu Ghoshal Abstract Image matching methods play a key role in deciding correspondence between two
More informationDynamically Reparameterized Light Fields & Fourier Slice Photography. Oliver Barth, 2009 Max Planck Institute Saarbrücken
Dynamically Reparameterized Light Fields & Fourier Slice Photography Oliver Barth, 2009 Max Planck Institute Saarbrücken Background What we are talking about? 2 / 83 Background What we are talking about?
More informationMulti-Resolution Estimation of Optical Flow on Vehicle Tracking under Unpredictable Environments
, pp.32-36 http://dx.doi.org/10.14257/astl.2016.129.07 Multi-Resolution Estimation of Optical Flow on Vehicle Tracking under Unpredictable Environments Viet Dung Do 1 and Dong-Min Woo 1 1 Department of
More informationMalaysian Car Number Plate Detection System Based on Template Matching and Colour Information
Malaysian Car Number Plate Detection System Based on Template Matching and Colour Information Mohd Firdaus Zakaria, Shahrel A. Suandi Intelligent Biometric Group, School of Electrical and Electronics Engineering,
More information