High Performance Imaging Using Large Camera Arrays
|
|
- Daniella Chase
- 5 years ago
- Views:
Transcription
1 High Performance Imaging Using Large Camera Arrays Presentation of the original paper by Bennett Wilburn, Neel Joshi, Vaibhav Vaish, Eino-Ville Talvala, Emilio Antunez, Adam Barth, Andrew Adams, Mark Horowitz, and Marc Levoy. Presented by Jeffrey Sharkey for CS 525.
2 Introduction This paper proposes that large arrays of inexpensive cameras offer better performance and features when compared with an expensive single-camera design. They build on previous research by surveying several existing system designs and pointing out the weaknesses of each. They then design and build a new system that overcomes these weaknesses. After building their new system, they use it to compare results for three specific areas of high-performance photography: (1) high-resolution video, (2) high-resolution HDR video and (3) high-speed video. They also explain how they solved several problems typically encountered when using camera arrays. Motivation From economics, we know that things are often cheaper in bulk. This also holds for most electronic components, such as digital cameras. Regardless of the device complexity, large-volume items are almost always cheaper. Take Google for example, who pioneered the use of simple off-the-shelf computers to build large computing clusters. To buy a single computer with equivalent processing power would have cost hundreds of times what they actually paid. In the same way, the authors of this paper propose that by building an array of inexpensive digital cameras, we can derive performance equal to, or even greater than, a single expensive digital camera. History The idea of using multiple cameras in a single scene was first pioneered in 1996 by Dayton Taylor, who initially applied them in the film industry to create an effect known as virtual camera movement. You ve probably seen this effect used in several movies and commercials, but its best known for its appearance in The Matrix series. By simultaneously triggering a series of still cameras, you can effectively freeze the scene in time. This allows us to move a virtual camera in the space dimension while keeping the time dimension frozen, something not physically possible with a single camera. However, his approach only used static camera images, limiting their motion only to the instant of capture. In 1997, the Virtualized Reality Project improved on Taylor s research by using continuous video streams instead of static images. However, at the time recording massive amounts of video was only feasible by using VCR s for storage. At that time, computers only had enough speed and storage space to record only about 9 seconds. In addition, they required one dedicated PC for every three cameras in the array. Their project experimented with both recording methods, and concluded that computers must improve significantly before becoming a feasible solution. Sharkey 2
3 Design and Build The goal of their project was to build an array of inexpensive digital cameras, along with its supporting infrastructure, to meet and exceed the challenges faced by previous projects. For their system, it was important that they choose off-the-shelf technology. They quickly discovered that most commercially available systems didn t offer enough flexibility for their project. So instead, they chose to build their system from raw components. They maintained a strict goal of leveraging existing standards while trying to stay away from customized hardware. For the actual image sensors, they chose to use CMOS technology over CCD sensors. This choice was mostly because CMOS offers a direct digital output, and better input controls when compared with CCD. The CMOS industry has also historically focused on high-volume markets, yielding lower prices when compared to CCD sensors. Instead of sending the raw video streams directly to a computer, the authors chose to add local processing boards. These boards have multiple camera inputs, include an Sony MPEG-2 encoder, and also an FPGA to help with video preprocessing. Finally, the compressed output is sent over a standard IEEE-1494 Firewire bus to a host computer. These local processing boards are important for two reasons. First, by including an FPGA, we can rapidly perform preprocessing, such as rotation, skew, and color adjustments. Because they are field-programmable, these parameters can be adjusted at each calibration. In addition, FPGA s can work with raw video much faster than a traditional computer. Second, by compressing the video using MPEG-2, we can put more cameras onto each Firewire bus. To ensure that video quality isn t lost, they have a local sanity check that maintains buffers of both raw and compressed frames for comparison purposes. Even at high bitrates, MPEG-2 still offers a 17:1 compression ratio. Finally, they needed to synchronize the timing of each camera in the array. Because Firewire doesn t guarantee timing resolution strict enough for their requirements, they built a second dedicated timing system. Their solution was a master timing clock running Sharkey 3
4 at 27MHz that fed uniform triggers to all local processing boards. They proved that this design has no more than 200-nanoseconds of clock skew, which is far below their minimum timing resolution of 200-microseconds. When finished, they had an array of 100 inexpensive cameras that could capture video at 30 frames per second. The raw video is compressed by a local processing board, and then routed back to a host PC through a tree-like structure of other processing boards. Each root node, which is usually a host PC, can support about 30 cameras. High-performance Applications After building the camera array, they wanted to show how it could be used in three specific high-performance applications. High-resolution Video The first application they examined was high-resolution video. By configuring their array to be pointed at a common center with about 50% overlap between each camera, they created a virtual camera with a field of view of about 30-degrees horizontally and 15-degrees vertically. In addition, most points in the scene are now covered by four cameras, allowing us to compensate for noise by averaging values. The process of stitching (or mosaicing) together each set of frames was automated using a program called Autostitch, which was based on previous research at the University of British Columbia. The program uses scale-invariant features to match frames together, and then uses homographies to stitch them together. Homographies are essentially projection transformations that map points between two planes. In our case, we are mapping the pixels from the individual camera images into the coordinate space of the final image. Sharkey 4
5 One major challenge they had to face was color calibration. Because the array had several inexpensive sensors, each camera had different color responses. They first turned off the automatic white-balancing feature in each camera, which is normally useful, but introduces an unknown variable in our system. They found that the sensors had a roughly linear color response except in cases of extreme values. They showed this was acceptable for their approach, because they only needed uniform results, and were not concerned with exact accuracy. High-resolution HDR Video Because their system had a large number of cameras, they could use varying exposure times to create high-resolution HDR video. High-dynamic-range (HDR) imaging is essentially compressing the dynamic range of an image so that lighter and darker areas are closer together. It can bring out details in dark areas while still showing detail in light areas without over-exposure. Sharkey 5
6 To build HDR images, we need identical images with varying exposure times. This is simple for static scenes where we can take multiple pictures, but can be difficult for video. A common approach is to split a single image from an incoming lens into three or more channels which are then processed by three video cameras at different exposure timings. However, these optics are very expensive, and do not offer very much control over the entire scene. The authors point out that individual cameras in their array can be configured with different exposure timings. Because each camera has an overlapping field of view, we can combine multiple exposures from neighboring cameras to create HDR video. In the paper, they show the results of dynamically choosing between four exposure timings. Based on the overall brightness of the current frame from a camera, its exposure timing is adjusted before capturing the next frame. By comparing pixel intensities against a reference image taken by a high-quality Canon camera, they found that the Canon s resolution was roughly 1.5 times better. However, they countered by pointing out that their system scales nicely, and can increase its resolution by simply adding more cameras. They also found that contrast of their array was worse than the reference Canon. They simply blamed this on the cheap lenses they were using. However, contrast can always be improved with post-processing. High-speed Video Because their system can be precisely timed, it can be used to create high-speed video by staggering the trigger times across the array. By using only 52 cameras that are completely overlapped, they can simulate a 1560 frame per second camera. To their knowledge, there is no non-classified single camera that can meet these specifications. In addition, the speed can be linearly increased with the number of cameras, effectively giving it unlimited resolution in the time domain. Sharkey 6
7 Another feature of using a camera array is that we can integrate each frame for longer than the current frame-rate. This means we can capture more light per unit time than is possible using a single high-speed camera. However, one of the major problems encountered when recording high-speed video from multiple cameras is parallax. The situation is more complex when an object in the scene is moving, in addition to the motion of the observation point. In this case, we can stagger the trigger timings for adjacent cameras, so that it appears as though the object isn t moving. The worst-case parallax occurs at the near and far depth planes. The worst case temporal motion will occur if the object is moving at the maximum velocity on the near-depth plane. We can approximate this temporal motion with a vector along the axis of focus, based on the maximum velocity. Relating this back to the distance between our adjacent cameras, we can find the best timing stagger to counteract this motion. Thus, by staggering our timing triggers, we can increase our sampling rate along the time axis with no additional cost. This is beneficial in scenes with high motion, because it reduces interpolation artifacts. In addition, it doesn t create any addition penalties in low-motion scenes. They conclude that staggered timing should always be used in highspeed video applications. Conclusion The authors have continued existing work by creating a new system that breaks past the limitations of previous research. They provide excellent details into the architecture of their system, along with rationale for their design decisions. Finally, they examine several applications of camera arrays while also handling the unique problems their array approach raises. Sharkey 7
High Performance Imaging Using Large Camera Arrays
High Performance Imaging Using Large Camera Arrays Bennett Wilburn 1 Neel Joshi 2 Vaibhav Vaish 2 Eino-Ville Talvala 1 Emilio Antunez 1 Adam Barth 2 Andrew Adams 2 Mark Horowitz 1 Marc Levoy 2 1 Electrical
More informationFigure 1 HDR image fusion example
TN-0903 Date: 10/06/09 Using image fusion to capture high-dynamic range (hdr) scenes High dynamic range (HDR) refers to the ability to distinguish details in scenes containing both very bright and relatively
More informationSynthetic aperture photography and illumination using arrays of cameras and projectors
Synthetic aperture photography and illumination using arrays of cameras and projectors technologies large camera arrays large projector arrays camera projector arrays Outline optical effects synthetic
More informationAdmin. Lightfields. Overview. Overview 5/13/2008. Idea. Projects due by the end of today. Lecture 13. Lightfield representation of a scene
Admin Lightfields Projects due by the end of today Email me source code, result images and short report Lecture 13 Overview Lightfield representation of a scene Unified representation of all rays Overview
More informationPatents of eye tracking system- a survey
Patents of eye tracking system- a survey Feng Li Center for Imaging Science Rochester Institute of Technology, Rochester, NY 14623 Email: Fxl5575@cis.rit.edu Vision is perhaps the most important of the
More informationHigh Dynamic Range Imaging
High Dynamic Range Imaging 1 2 Lecture Topic Discuss the limits of the dynamic range in current imaging and display technology Solutions 1. High Dynamic Range (HDR) Imaging Able to image a larger dynamic
More informationLight field sensing. Marc Levoy. Computer Science Department Stanford University
Light field sensing Marc Levoy Computer Science Department Stanford University The scalar light field (in geometrical optics) Radiance as a function of position and direction in a static scene with fixed
More informationMidterm Examination CS 534: Computational Photography
Midterm Examination CS 534: Computational Photography November 3, 2015 NAME: SOLUTIONS Problem Score Max Score 1 8 2 8 3 9 4 4 5 3 6 4 7 6 8 13 9 7 10 4 11 7 12 10 13 9 14 8 Total 100 1 1. [8] What are
More informationA Short History of Using Cameras for Weld Monitoring
A Short History of Using Cameras for Weld Monitoring 2 Background Ever since the development of automated welding, operators have needed to be able to monitor the process to ensure that all parameters
More informationBasler. Line Scan Cameras
Basler Line Scan Cameras Next generation CMOS dual line scan technology Up to 140 khz at 2k or 4k resolution, up to 70 khz at 8k resolution Color line scan with 70 khz at 4k resolution High sensitivity
More informationON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES
ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES Petteri PÖNTINEN Helsinki University of Technology, Institute of Photogrammetry and Remote Sensing, Finland petteri.pontinen@hut.fi KEY WORDS: Cocentricity,
More informationColour correction for panoramic imaging
Colour correction for panoramic imaging Gui Yun Tian Duke Gledhill Dave Taylor The University of Huddersfield David Clarke Rotography Ltd Abstract: This paper reports the problem of colour distortion in
More informationCharged Coupled Device (CCD) S.Vidhya
Charged Coupled Device (CCD) S.Vidhya 02.04.2016 Sensor Physical phenomenon Sensor Measurement Output A sensor is a device that measures a physical quantity and converts it into a signal which can be read
More informationImage Formation: Camera Model
Image Formation: Camera Model Ruigang Yang COMP 684 Fall 2005, CS684-IBMR Outline Camera Models Pinhole Perspective Projection Affine Projection Camera with Lenses Digital Image Formation The Human Eye
More informationImproved sensitivity high-definition interline CCD using the KODAK TRUESENSE Color Filter Pattern
Improved sensitivity high-definition interline CCD using the KODAK TRUESENSE Color Filter Pattern James DiBella*, Marco Andreghetti, Amy Enge, William Chen, Timothy Stanka, Robert Kaser (Eastman Kodak
More informationDigital Cameras. Consumer and Prosumer
Digital Cameras Overview While silver-halide film has been the dominant photographic process for the past 150 years, the use and role of technology is fast-becoming a standard for the making of photographs.
More informationCamera Image Processing Pipeline: Part II
Lecture 13: Camera Image Processing Pipeline: Part II Visual Computing Systems Today Finish image processing pipeline Auto-focus / auto-exposure Camera processing elements Smart phone processing elements
More informationWhite Paper High Dynamic Range Imaging
WPE-2015XI30-00 for Machine Vision What is Dynamic Range? Dynamic Range is the term used to describe the difference between the brightest part of a scene and the darkest part of a scene at a given moment
More informationModeling and Synthesis of Aperture Effects in Cameras
Modeling and Synthesis of Aperture Effects in Cameras Douglas Lanman, Ramesh Raskar, and Gabriel Taubin Computational Aesthetics 2008 20 June, 2008 1 Outline Introduction and Related Work Modeling Vignetting
More informationmultiframe visual-inertial blur estimation and removal for unmodified smartphones
multiframe visual-inertial blur estimation and removal for unmodified smartphones, Severin Münger, Carlo Beltrame, Luc Humair WSCG 2015, Plzen, Czech Republic images taken by non-professional photographers
More informationCamera Image Processing Pipeline: Part II
Lecture 14: Camera Image Processing Pipeline: Part II Visual Computing Systems Today Finish image processing pipeline Auto-focus / auto-exposure Camera processing elements Smart phone processing elements
More informationDetermining MTF with a Slant Edge Target ABSTRACT AND INTRODUCTION
Determining MTF with a Slant Edge Target Douglas A. Kerr Issue 2 October 13, 2010 ABSTRACT AND INTRODUCTION The modulation transfer function (MTF) of a photographic lens tells us how effectively the lens
More informationHDR videos acquisition
HDR videos acquisition dr. Francesco Banterle francesco.banterle@isti.cnr.it How to capture? Videos are challenging: We need to capture multiple frames at different exposure times and everything moves
More informationHow to combine images in Photoshop
How to combine images in Photoshop In Photoshop, you can use multiple layers to combine images, but there are two other ways to create a single image from mulitple images. Create a panoramic image with
More informationA Digital Camera Glossary. Ashley Rodriguez, Charlie Serrano, Luis Martinez, Anderson Guatemala PERIOD 6
A Digital Camera Glossary Ashley Rodriguez, Charlie Serrano, Luis Martinez, Anderson Guatemala PERIOD 6 A digital Camera Glossary Ivan Encinias, Sebastian Limas, Amir Cal Ivan encinias Image sensor A silicon
More informationImage stitching. Image stitching. Video summarization. Applications of image stitching. Stitching = alignment + blending. geometrical registration
Image stitching Stitching = alignment + blending Image stitching geometrical registration photometric registration Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2005/3/22 with slides by Richard Szeliski,
More informationCameras. CSE 455, Winter 2010 January 25, 2010
Cameras CSE 455, Winter 2010 January 25, 2010 Announcements New Lecturer! Neel Joshi, Ph.D. Post-Doctoral Researcher Microsoft Research neel@cs Project 1b (seam carving) was due on Friday the 22 nd Project
More informationCoded photography , , Computational Photography Fall 2018, Lecture 14
Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 14 Overview of today s lecture The coded photography paradigm. Dealing with
More informationNoise and ISO. CS 178, Spring Marc Levoy Computer Science Department Stanford University
Noise and ISO CS 178, Spring 2014 Marc Levoy Computer Science Department Stanford University Outline examples of camera sensor noise don t confuse it with JPEG compression artifacts probability, mean,
More informationDual-fisheye Lens Stitching for 360-degree Imaging & Video. Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington
Dual-fisheye Lens Stitching for 360-degree Imaging & Video Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington Introduction 360-degree imaging: the process of taking multiple photographs and
More informationGetting Unlimited Digital Resolution
Getting Unlimited Digital Resolution N. David King Wow, now here s a goal: how would you like to be able to create nearly any amount of resolution you want with a digital camera. Since the higher the resolution
More informationbrief history of photography foveon X3 imager technology description
brief history of photography foveon X3 imager technology description imaging technology 30,000 BC chauvet-pont-d arc pinhole camera principle first described by Aristotle fourth century B.C. oldest known
More informationMulti-sensor Panoramic Network Camera
Multi-sensor Panoramic Network Camera White Paper by Dahua Technology Release 1.0 Table of contents 1 Preface... 2 2 Overview... 3 3 Technical Background... 3 4 Key Technologies... 5 4.1 Feature Points
More informationWhite Paper Focusing more on the forest, and less on the trees
White Paper Focusing more on the forest, and less on the trees Why total system image quality is more important than any single component of your next document scanner Contents Evaluating total system
More informationThe Denali-MC HDR ISP Backgrounder
The Denali-MC HDR ISP Backgrounder 2-4 brackets up to 8 EV frame offset Up to 16 EV stops for output HDR LATM (tone map) up to 24 EV Noise reduction due to merging of 10 EV LDR to a single 16 EV HDR up
More informationPIXPOLAR WHITE PAPER 29 th of September 2013
PIXPOLAR WHITE PAPER 29 th of September 2013 Pixpolar s Modified Internal Gate (MIG) image sensor technology offers numerous benefits over traditional Charge Coupled Device (CCD) and Complementary Metal
More informationLENSES. INEL 6088 Computer Vision
LENSES INEL 6088 Computer Vision Digital camera A digital camera replaces film with a sensor array Each cell in the array is a Charge Coupled Device light-sensitive diode that converts photons to electrons
More informationApplying Automated Optical Inspection Ben Dawson, DALSA Coreco Inc., ipd Group (987)
Applying Automated Optical Inspection Ben Dawson, DALSA Coreco Inc., ipd Group bdawson@goipd.com (987) 670-2050 Introduction Automated Optical Inspection (AOI) uses lighting, cameras, and vision computers
More informationA HIGH PERFORMANCE HARDWARE ARCHITECTURE FOR HALF-PIXEL ACCURATE H.264 MOTION ESTIMATION
A HIGH PERFORMANCE HARDWARE ARCHITECTURE FOR HALF-PIXEL ACCURATE H.264 MOTION ESTIMATION Sinan Yalcin and Ilker Hamzaoglu Faculty of Engineering and Natural Sciences, Sabanci University, 34956, Tuzla,
More informationIntroduction to camera usage. The universal manual controls of most cameras
Introduction to camera usage A camera in its barest form is simply a light tight container that utilizes a lens with iris, a shutter that has variable speeds, and contains a sensitive piece of media, either
More informationTopic 9 - Sensors Within
Topic 9 - Sensors Within Learning Outcomes In this topic, we will take a closer look at sensor sizes in digital cameras. By the end of this video you will have a better understanding of what the various
More informationImage acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor
Image acquisition Digital images are acquired by direct digital acquisition (digital still/video cameras), or scanning material acquired as analog signals (slides, photographs, etc.). In both cases, the
More informationIt should also be noted that with modern cameras users can choose for either
White paper about color correction More drama Many application fields like digital printing industry or the human medicine require a natural display of colors. To illustrate the importance of color fidelity,
More informationReikan FoCal Aperture Sharpness Test Report
Focus Calibration and Analysis Software Test run on: 26/01/2016 17:02:00 with FoCal 2.0.6.2416W Report created on: 26/01/2016 17:03:39 with FoCal 2.0.6W Overview Test Information Property Description Data
More informationImage processing with the HERON-FPGA Family
HUNT ENGINEERING Chestnut Court, Burton Row, Brent Knoll, Somerset, TA9 4BP, UK Tel: (+44) (0)1278 760188, Fax: (+44) (0)1278 760199, Email: sales@hunteng.co.uk http://www.hunteng.co.uk http://www.hunt-dsp.com
More informationF-number sequence. a change of f-number to the next in the sequence corresponds to a factor of 2 change in light intensity,
1 F-number sequence a change of f-number to the next in the sequence corresponds to a factor of 2 change in light intensity, 0.7, 1, 1.4, 2, 2.8, 4, 5.6, 8, 11, 16, 22, 32, Example: What is the difference
More informationpco.edge 4.2 LT 0.8 electrons 2048 x 2048 pixel 40 fps up to :1 up to 82 % pco. low noise high resolution high speed high dynamic range
edge 4.2 LT scientific CMOS camera high resolution 2048 x 2048 pixel low noise 0.8 electrons USB 3.0 small form factor high dynamic range up to 37 500:1 high speed 40 fps high quantum efficiency up to
More informationCameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017
Cameras Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017 Camera Focus Camera Focus So far, we have been simulating pinhole cameras with perfect focus Often times, we want to simulate more
More informationReal-Time Face Detection and Tracking for High Resolution Smart Camera System
Digital Image Computing Techniques and Applications Real-Time Face Detection and Tracking for High Resolution Smart Camera System Y. M. Mustafah a,b, T. Shan a, A. W. Azman a,b, A. Bigdeli a, B. C. Lovell
More informationQuintic Hardware Tutorial Camera Set-Up
Quintic Hardware Tutorial Camera Set-Up 1 All Quintic Live High-Speed cameras are specifically designed to meet a wide range of needs including coaching, performance analysis and research. Quintic LIVE
More informationLecture 30: Image Sensors (Cont) Computer Graphics and Imaging UC Berkeley CS184/284A
Lecture 30: Image Sensors (Cont) Computer Graphics and Imaging UC Berkeley Reminder: The Pixel Stack Microlens array Color Filter Anti-Reflection Coating Stack height 4um is typical Pixel size 2um is typical
More informationThe Big Train Project Status Report (Part 65)
The Big Train Project Status Report (Part 65) For this month I have a somewhat different topic related to the EnterTRAINment Junction (EJ) layout. I thought I d share some lessons I ve learned from photographing
More informationImage Processing for Comets
Image Processing for Comets Page 1 2.5 Surface Today, there are sensors of 768 x 512 pixels up to 8176 x 6132 pixels ( 49,1 mm x 36,8 mm), that's bigger than the old 35mm film. The size of the chip determines
More informationCoded photography , , Computational Photography Fall 2017, Lecture 18
Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 18 Course announcements Homework 5 delayed for Tuesday. - You will need cameras
More informationThe ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do?
Computational Photography The ultimate camera What does it do? Image from Durand & Freeman s MIT Course on Computational Photography Today s reading Szeliski Chapter 9 The ultimate camera Infinite resolution
More informationCS354 Computer Graphics Computational Photography. Qixing Huang April 23 th 2018
CS354 Computer Graphics Computational Photography Qixing Huang April 23 th 2018 Background Sales of digital cameras surpassed sales of film cameras in 2004 Digital Cameras Free film Instant display Quality
More informationLineup for Compact Cameras from
Lineup for Compact Cameras from Milbeaut M-4 Series Image Processing System LSI for Digital Cameras A new lineup of 1) a low-price product and 2) a product incorporating a moving image function in M-4
More informationReikan FoCal Aperture Sharpness Test Report
Focus Calibration and Analysis Software Reikan FoCal Sharpness Test Report Test run on: 26/01/2016 17:14:35 with FoCal 2.0.6.2416W Report created on: 26/01/2016 17:16:16 with FoCal 2.0.6W Overview Test
More informationNOVA S12. Compact and versatile high performance camera system. 1-Megapixel CMOS Image Sensor: 1024 x 1024 pixels at 12,800fps
NOVA S12 1-Megapixel CMOS Image Sensor: 1024 x 1024 pixels at 12,800fps Maximum Frame Rate: 1,000,000fps Class Leading Light Sensitivity: ISO 12232 Ssat Standard ISO 64,000 monochrome ISO 16,000 color
More informationImaging serial interface ROM
Page 1 of 6 ( 3 of 32 ) United States Patent Application 20070024904 Kind Code A1 Baer; Richard L. ; et al. February 1, 2007 Imaging serial interface ROM Abstract Imaging serial interface ROM (ISIROM).
More informationPhase One 190MP Aerial System
White Paper Phase One 190MP Aerial System Introduction Phase One Industrial s 100MP medium format aerial camera systems have earned a worldwide reputation for its high performance. They are commonly used
More informationBe aware that there is no universal notation for the various quantities.
Fourier Optics v2.4 Ray tracing is limited in its ability to describe optics because it ignores the wave properties of light. Diffraction is needed to explain image spatial resolution and contrast and
More informationEarly art: events. Baroque art: portraits. Renaissance art: events. Being There: Capturing and Experiencing a Sense of Place
Being There: Capturing and Experiencing a Sense of Place Early art: events Richard Szeliski Microsoft Research Symposium on Computational Photography and Video Lascaux Early art: events Early art: events
More informationLecture 18: Light field cameras. (plenoptic cameras) Visual Computing Systems CMU , Fall 2013
Lecture 18: Light field cameras (plenoptic cameras) Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today:
More informationLecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)
Lecture 19: Depth Cameras Kayvon Fatahalian CMU 15-869: Graphics and Imaging Architectures (Fall 2011) Continuing theme: computational photography Cheap cameras capture light, extensive processing produces
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 22: Computational photography photomatix.com Announcements Final project midterm reports due on Tuesday to CMS by 11:59pm BRDF s can be incredibly complicated
More informationLecture 11: Clocking
High Speed CMOS VLSI Design Lecture 11: Clocking (c) 1997 David Harris 1.0 Introduction We have seen that generating and distributing clocks with little skew is essential to high speed circuit design.
More informationReikan FoCal Aperture Sharpness Test Report
Focus Calibration and Analysis Software Reikan FoCal Sharpness Test Report Test run on: 10/02/2016 19:57:05 with FoCal 2.0.6.2416W Report created on: 10/02/2016 19:59:09 with FoCal 2.0.6W Overview Test
More informationA new Photon Counting Detector: Intensified CMOS- APS
A new Photon Counting Detector: Intensified CMOS- APS M. Belluso 1, G. Bonanno 1, A. Calì 1, A. Carbone 3, R. Cosentino 1, A. Modica 4, S. Scuderi 1, C. Timpanaro 1, M. Uslenghi 2 1-I.N.A.F.-Osservatorio
More informationTechnologies Explained PowerShot D20
Technologies Explained PowerShot D20 EMBARGO: 7 th February 2012, 05:00 (GMT) HS System The HS System represents a powerful combination of a high-sensitivity sensor and high-performance DIGIC image processing
More informationTonemapping and bilateral filtering
Tonemapping and bilateral filtering http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 6 Course announcements Homework 2 is out. - Due September
More informationReikan FoCal Aperture Sharpness Test Report
Focus Calibration and Analysis Software Reikan FoCal Sharpness Test Report Test run on: 27/01/2016 00:35:25 with FoCal 2.0.6.2416W Report created on: 27/01/2016 00:41:43 with FoCal 2.0.6W Overview Test
More informationImplementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring
Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring Ashill Chiranjan and Bernardt Duvenhage Defence, Peace, Safety and Security Council for Scientific
More informationA new Photon Counting Detector: Intensified CMOS- APS
A new Photon Counting Detector: Intensified CMOS- APS M. Belluso 1, G. Bonanno 1, A. Calì 1, A. Carbone 3, R. Cosentino 1, A. Modica 4, S. Scuderi 1, C. Timpanaro 1, M. Uslenghi 2 1- I.N.A.F.-Osservatorio
More informationReikan FoCal Aperture Sharpness Test Report
Focus Calibration and Analysis Software Test run on: 26/01/2016 17:56:23 with FoCal 2.0.6.2416W Report created on: 26/01/2016 17:59:12 with FoCal 2.0.6W Overview Test Information Property Description Data
More informationIMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics
IMAGE FORMATION Light source properties Sensor characteristics Surface Exposure shape Optics Surface reflectance properties ANALOG IMAGES An image can be understood as a 2D light intensity function f(x,y)
More informationMeasuring the impact of flare light on Dynamic Range
Measuring the impact of flare light on Dynamic Range Norman Koren; Imatest LLC; Boulder, CO USA Abstract The dynamic range (DR; defined as the range of exposure between saturation and 0 db SNR) of recent
More informationPhotographing Long Scenes with Multiviewpoint
Photographing Long Scenes with Multiviewpoint Panoramas A. Agarwala, M. Agrawala, M. Cohen, D. Salesin, R. Szeliski Presenter: Stacy Hsueh Discussant: VasilyVolkov Motivation Want an image that shows an
More informationSMART LASER SENSORS SIMPLIFY TIRE AND RUBBER INSPECTION
PRESENTED AT ITEC 2004 SMART LASER SENSORS SIMPLIFY TIRE AND RUBBER INSPECTION Dr. Walt Pastorius LMI Technologies 2835 Kew Dr. Windsor, ON N8T 3B7 Tel (519) 945 6373 x 110 Cell (519) 981 0238 Fax (519)
More informationLow-Cost, On-Demand Film Digitisation and Online Delivery. Matt Garner
Low-Cost, On-Demand Film Digitisation and Online Delivery Matt Garner (matt.garner@findmypast.com) Abstract Hundreds of millions of pages of microfilmed material are not being digitised at this time due
More informationFocusing and Metering
Focusing and Metering CS 478 Winter 2012 Slides mostly stolen by David Jacobs from Marc Levoy Focusing Outline Manual Focus Specialty Focus Autofocus Active AF Passive AF AF Modes Manual Focus - View Camera
More informationShape sensing for computer aided below-knee prosthetic socket design
Prosthetics and Orthotics International, 1985, 9, 12-16 Shape sensing for computer aided below-knee prosthetic socket design G. R. FERNIE, G. GRIGGS, S. BARTLETT and K. LUNAU West Park Research, Department
More informationWhite Paper: Compression Advantages of Pixim s Digital Pixel System Technology
White Paper: Compression Advantages of Pixim s Digital Pixel System Technology Table of Contents The role of efficient compression algorithms Bit-rate strategies and limits 2 Amount of motion present in
More informationPhotography Help Sheets
Photography Help Sheets Phone: 01233 771915 Web: www.bigcatsanctuary.org Using your Digital SLR What is Exposure? Exposure is basically the process of recording light onto your digital sensor (or film).
More informationproducts PC Control
products PC Control 04 2017 PC Control 04 2017 products Image processing directly in the PLC TwinCAT Vision Machine vision easily integrated into automation technology Automatic detection, traceability
More informationIHV means Independent Hardware Vendor. Example is Qualcomm Technologies Inc. that makes Snapdragon processors. OEM means Original Equipment
1 2 IHV means Independent Hardware Vendor. Example is Qualcomm Technologies Inc. that makes Snapdragon processors. OEM means Original Equipment Manufacturer. Examples are smartphone manufacturers. Tuning
More informationCAMERA BASICS. Stops of light
CAMERA BASICS Stops of light A stop of light isn t a quantifiable measurement it s a relative measurement. A stop of light is defined as a doubling or halving of any quantity of light. The word stop is
More informationWhich equipment is necessary? How is the panorama created?
Congratulations! By purchasing your Panorama-VR-System you have acquired a tool, which enables you - together with a digital or analog camera, a tripod and a personal computer - to generate high quality
More informationEdmonton Camera Club. Introduction to Exposure. and a few other bits!
Edmonton Camera Club Introduction to Exposure and a few other bits! Exposure 3 Variables 1. Aperture how much light 2. Shutter Speed for how long 3. Sensitivity ISO, Film Speed Also cover: Compensation
More informationCamera Overview. Digital Microscope Cameras for Material Science: Clear Images, Precise Analysis. Digital Cameras for Microscopy
Digital Cameras for Microscopy Camera Overview For Materials Science Microscopes Digital Microscope Cameras for Material Science: Clear Images, Precise Analysis Passionate about Imaging: Olympus Digital
More informationPart Number SuperPix TM image sensor is one of SuperPix TM 2 Mega Digital image sensor series products. These series sensors have the same maximum ima
Specification Version Commercial 1.7 2012.03.26 SuperPix Micro Technology Co., Ltd Part Number SuperPix TM image sensor is one of SuperPix TM 2 Mega Digital image sensor series products. These series sensors
More information9/19/16. A Closer Look. Danae Wolfe. What We ll Cover. Basics of photography & your camera. Technical. Macro & close-up techniques.
A Closer Look Danae Wolfe What We ll Cover Basics of photography & your camera Technical Macro & close-up techniques Creative 1 What is Photography? Photography: the art, science, & practice of creating
More informationImage stabilization (IS)
Image stabilization (IS) CS 178, Spring 2009 Marc Levoy Computer Science Department Stanford University Outline what are the causes of camera shake? and how can you avoid it (without having an IS system)?
More informationNote: These sample pages are from Chapter 1. The Zone System
Note: These sample pages are from Chapter 1 The Zone System Chapter 1 The Zones Revealed The images below show how you can visualize the zones in an image. This is NGC 1491, an HII region imaged through
More informationCOLOR CORRECTION METHOD USING GRAY GRADIENT BAR FOR MULTI-VIEW CAMERA SYSTEM. Jae-Il Jung and Yo-Sung Ho
COLOR CORRECTION METHOD USING GRAY GRADIENT BAR FOR MULTI-VIEW CAMERA SYSTEM Jae-Il Jung and Yo-Sung Ho School of Information and Mechatronics Gwangju Institute of Science and Technology (GIST) 1 Oryong-dong
More informationSpheroCam HDR. Image based lighting with. Capture light perfectly SPHERON VR. 0s 20s 40s 60s 80s 100s 120s. Spheron VR AG
Image based lighting with SpheroCam HDR Capture light perfectly 0 60 120 180 240 300 360 0s 20s 40s 60s 80s 100s 120s SPHERON VR high dynamic range imaging Spheron VR AG u phone u internet Hauptstraße
More informationOV7670 Software Application Note
OV7670 Software Application Note Table of Contents OV7670 Software Application Note... 1 1. Select Output format...3 1.1 Backend with full ISP... 3 1.2 Backend with YCbCr ISP... 4 1.3 Backend without ISP...4
More informationDigital camera modes explained: choose the best shooting mode for your subject
Digital camera modes explained: choose the best shooting mode for your subject On most DSLRs, the Mode dial is split into three sections: Scene modes (for doing point-and-shoot photography in specific
More informationImage Sensor and Camera Technology November 2016 in Stuttgart
Image Sensor and Camera Technology 14-15-16 November 2016 in Stuttgart Aphesa organizes an image sensor and camera technology training tour between October 2015 and November 2016. The training sessions
More informationMEM: Intro to Robotics. Assignment 3I. Due: Wednesday 10/15 11:59 EST
MEM: Intro to Robotics Assignment 3I Due: Wednesday 10/15 11:59 EST 1. Basic Optics You are shopping for a new lens for your Canon D30 digital camera and there are lots of lens options at the store. Your
More information