High Performance Imaging Using Large Camera Arrays

Size: px
Start display at page:

Download "High Performance Imaging Using Large Camera Arrays"

Transcription

1 High Performance Imaging Using Large Camera Arrays Presentation of the original paper by Bennett Wilburn, Neel Joshi, Vaibhav Vaish, Eino-Ville Talvala, Emilio Antunez, Adam Barth, Andrew Adams, Mark Horowitz, and Marc Levoy. Presented by Jeffrey Sharkey for CS 525.

2 Introduction This paper proposes that large arrays of inexpensive cameras offer better performance and features when compared with an expensive single-camera design. They build on previous research by surveying several existing system designs and pointing out the weaknesses of each. They then design and build a new system that overcomes these weaknesses. After building their new system, they use it to compare results for three specific areas of high-performance photography: (1) high-resolution video, (2) high-resolution HDR video and (3) high-speed video. They also explain how they solved several problems typically encountered when using camera arrays. Motivation From economics, we know that things are often cheaper in bulk. This also holds for most electronic components, such as digital cameras. Regardless of the device complexity, large-volume items are almost always cheaper. Take Google for example, who pioneered the use of simple off-the-shelf computers to build large computing clusters. To buy a single computer with equivalent processing power would have cost hundreds of times what they actually paid. In the same way, the authors of this paper propose that by building an array of inexpensive digital cameras, we can derive performance equal to, or even greater than, a single expensive digital camera. History The idea of using multiple cameras in a single scene was first pioneered in 1996 by Dayton Taylor, who initially applied them in the film industry to create an effect known as virtual camera movement. You ve probably seen this effect used in several movies and commercials, but its best known for its appearance in The Matrix series. By simultaneously triggering a series of still cameras, you can effectively freeze the scene in time. This allows us to move a virtual camera in the space dimension while keeping the time dimension frozen, something not physically possible with a single camera. However, his approach only used static camera images, limiting their motion only to the instant of capture. In 1997, the Virtualized Reality Project improved on Taylor s research by using continuous video streams instead of static images. However, at the time recording massive amounts of video was only feasible by using VCR s for storage. At that time, computers only had enough speed and storage space to record only about 9 seconds. In addition, they required one dedicated PC for every three cameras in the array. Their project experimented with both recording methods, and concluded that computers must improve significantly before becoming a feasible solution. Sharkey 2

3 Design and Build The goal of their project was to build an array of inexpensive digital cameras, along with its supporting infrastructure, to meet and exceed the challenges faced by previous projects. For their system, it was important that they choose off-the-shelf technology. They quickly discovered that most commercially available systems didn t offer enough flexibility for their project. So instead, they chose to build their system from raw components. They maintained a strict goal of leveraging existing standards while trying to stay away from customized hardware. For the actual image sensors, they chose to use CMOS technology over CCD sensors. This choice was mostly because CMOS offers a direct digital output, and better input controls when compared with CCD. The CMOS industry has also historically focused on high-volume markets, yielding lower prices when compared to CCD sensors. Instead of sending the raw video streams directly to a computer, the authors chose to add local processing boards. These boards have multiple camera inputs, include an Sony MPEG-2 encoder, and also an FPGA to help with video preprocessing. Finally, the compressed output is sent over a standard IEEE-1494 Firewire bus to a host computer. These local processing boards are important for two reasons. First, by including an FPGA, we can rapidly perform preprocessing, such as rotation, skew, and color adjustments. Because they are field-programmable, these parameters can be adjusted at each calibration. In addition, FPGA s can work with raw video much faster than a traditional computer. Second, by compressing the video using MPEG-2, we can put more cameras onto each Firewire bus. To ensure that video quality isn t lost, they have a local sanity check that maintains buffers of both raw and compressed frames for comparison purposes. Even at high bitrates, MPEG-2 still offers a 17:1 compression ratio. Finally, they needed to synchronize the timing of each camera in the array. Because Firewire doesn t guarantee timing resolution strict enough for their requirements, they built a second dedicated timing system. Their solution was a master timing clock running Sharkey 3

4 at 27MHz that fed uniform triggers to all local processing boards. They proved that this design has no more than 200-nanoseconds of clock skew, which is far below their minimum timing resolution of 200-microseconds. When finished, they had an array of 100 inexpensive cameras that could capture video at 30 frames per second. The raw video is compressed by a local processing board, and then routed back to a host PC through a tree-like structure of other processing boards. Each root node, which is usually a host PC, can support about 30 cameras. High-performance Applications After building the camera array, they wanted to show how it could be used in three specific high-performance applications. High-resolution Video The first application they examined was high-resolution video. By configuring their array to be pointed at a common center with about 50% overlap between each camera, they created a virtual camera with a field of view of about 30-degrees horizontally and 15-degrees vertically. In addition, most points in the scene are now covered by four cameras, allowing us to compensate for noise by averaging values. The process of stitching (or mosaicing) together each set of frames was automated using a program called Autostitch, which was based on previous research at the University of British Columbia. The program uses scale-invariant features to match frames together, and then uses homographies to stitch them together. Homographies are essentially projection transformations that map points between two planes. In our case, we are mapping the pixels from the individual camera images into the coordinate space of the final image. Sharkey 4

5 One major challenge they had to face was color calibration. Because the array had several inexpensive sensors, each camera had different color responses. They first turned off the automatic white-balancing feature in each camera, which is normally useful, but introduces an unknown variable in our system. They found that the sensors had a roughly linear color response except in cases of extreme values. They showed this was acceptable for their approach, because they only needed uniform results, and were not concerned with exact accuracy. High-resolution HDR Video Because their system had a large number of cameras, they could use varying exposure times to create high-resolution HDR video. High-dynamic-range (HDR) imaging is essentially compressing the dynamic range of an image so that lighter and darker areas are closer together. It can bring out details in dark areas while still showing detail in light areas without over-exposure. Sharkey 5

6 To build HDR images, we need identical images with varying exposure times. This is simple for static scenes where we can take multiple pictures, but can be difficult for video. A common approach is to split a single image from an incoming lens into three or more channels which are then processed by three video cameras at different exposure timings. However, these optics are very expensive, and do not offer very much control over the entire scene. The authors point out that individual cameras in their array can be configured with different exposure timings. Because each camera has an overlapping field of view, we can combine multiple exposures from neighboring cameras to create HDR video. In the paper, they show the results of dynamically choosing between four exposure timings. Based on the overall brightness of the current frame from a camera, its exposure timing is adjusted before capturing the next frame. By comparing pixel intensities against a reference image taken by a high-quality Canon camera, they found that the Canon s resolution was roughly 1.5 times better. However, they countered by pointing out that their system scales nicely, and can increase its resolution by simply adding more cameras. They also found that contrast of their array was worse than the reference Canon. They simply blamed this on the cheap lenses they were using. However, contrast can always be improved with post-processing. High-speed Video Because their system can be precisely timed, it can be used to create high-speed video by staggering the trigger times across the array. By using only 52 cameras that are completely overlapped, they can simulate a 1560 frame per second camera. To their knowledge, there is no non-classified single camera that can meet these specifications. In addition, the speed can be linearly increased with the number of cameras, effectively giving it unlimited resolution in the time domain. Sharkey 6

7 Another feature of using a camera array is that we can integrate each frame for longer than the current frame-rate. This means we can capture more light per unit time than is possible using a single high-speed camera. However, one of the major problems encountered when recording high-speed video from multiple cameras is parallax. The situation is more complex when an object in the scene is moving, in addition to the motion of the observation point. In this case, we can stagger the trigger timings for adjacent cameras, so that it appears as though the object isn t moving. The worst-case parallax occurs at the near and far depth planes. The worst case temporal motion will occur if the object is moving at the maximum velocity on the near-depth plane. We can approximate this temporal motion with a vector along the axis of focus, based on the maximum velocity. Relating this back to the distance between our adjacent cameras, we can find the best timing stagger to counteract this motion. Thus, by staggering our timing triggers, we can increase our sampling rate along the time axis with no additional cost. This is beneficial in scenes with high motion, because it reduces interpolation artifacts. In addition, it doesn t create any addition penalties in low-motion scenes. They conclude that staggered timing should always be used in highspeed video applications. Conclusion The authors have continued existing work by creating a new system that breaks past the limitations of previous research. They provide excellent details into the architecture of their system, along with rationale for their design decisions. Finally, they examine several applications of camera arrays while also handling the unique problems their array approach raises. Sharkey 7

High Performance Imaging Using Large Camera Arrays

High Performance Imaging Using Large Camera Arrays High Performance Imaging Using Large Camera Arrays Bennett Wilburn 1 Neel Joshi 2 Vaibhav Vaish 2 Eino-Ville Talvala 1 Emilio Antunez 1 Adam Barth 2 Andrew Adams 2 Mark Horowitz 1 Marc Levoy 2 1 Electrical

More information

Figure 1 HDR image fusion example

Figure 1 HDR image fusion example TN-0903 Date: 10/06/09 Using image fusion to capture high-dynamic range (hdr) scenes High dynamic range (HDR) refers to the ability to distinguish details in scenes containing both very bright and relatively

More information

Synthetic aperture photography and illumination using arrays of cameras and projectors

Synthetic aperture photography and illumination using arrays of cameras and projectors Synthetic aperture photography and illumination using arrays of cameras and projectors technologies large camera arrays large projector arrays camera projector arrays Outline optical effects synthetic

More information

Admin. Lightfields. Overview. Overview 5/13/2008. Idea. Projects due by the end of today. Lecture 13. Lightfield representation of a scene

Admin. Lightfields. Overview. Overview 5/13/2008. Idea. Projects due by the end of today. Lecture 13. Lightfield representation of a scene Admin Lightfields Projects due by the end of today Email me source code, result images and short report Lecture 13 Overview Lightfield representation of a scene Unified representation of all rays Overview

More information

Patents of eye tracking system- a survey

Patents of eye tracking system- a survey Patents of eye tracking system- a survey Feng Li Center for Imaging Science Rochester Institute of Technology, Rochester, NY 14623 Email: Fxl5575@cis.rit.edu Vision is perhaps the most important of the

More information

High Dynamic Range Imaging

High Dynamic Range Imaging High Dynamic Range Imaging 1 2 Lecture Topic Discuss the limits of the dynamic range in current imaging and display technology Solutions 1. High Dynamic Range (HDR) Imaging Able to image a larger dynamic

More information

Light field sensing. Marc Levoy. Computer Science Department Stanford University

Light field sensing. Marc Levoy. Computer Science Department Stanford University Light field sensing Marc Levoy Computer Science Department Stanford University The scalar light field (in geometrical optics) Radiance as a function of position and direction in a static scene with fixed

More information

Midterm Examination CS 534: Computational Photography

Midterm Examination CS 534: Computational Photography Midterm Examination CS 534: Computational Photography November 3, 2015 NAME: SOLUTIONS Problem Score Max Score 1 8 2 8 3 9 4 4 5 3 6 4 7 6 8 13 9 7 10 4 11 7 12 10 13 9 14 8 Total 100 1 1. [8] What are

More information

A Short History of Using Cameras for Weld Monitoring

A Short History of Using Cameras for Weld Monitoring A Short History of Using Cameras for Weld Monitoring 2 Background Ever since the development of automated welding, operators have needed to be able to monitor the process to ensure that all parameters

More information

Basler. Line Scan Cameras

Basler. Line Scan Cameras Basler Line Scan Cameras Next generation CMOS dual line scan technology Up to 140 khz at 2k or 4k resolution, up to 70 khz at 8k resolution Color line scan with 70 khz at 4k resolution High sensitivity

More information

ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES

ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES Petteri PÖNTINEN Helsinki University of Technology, Institute of Photogrammetry and Remote Sensing, Finland petteri.pontinen@hut.fi KEY WORDS: Cocentricity,

More information

Colour correction for panoramic imaging

Colour correction for panoramic imaging Colour correction for panoramic imaging Gui Yun Tian Duke Gledhill Dave Taylor The University of Huddersfield David Clarke Rotography Ltd Abstract: This paper reports the problem of colour distortion in

More information

Charged Coupled Device (CCD) S.Vidhya

Charged Coupled Device (CCD) S.Vidhya Charged Coupled Device (CCD) S.Vidhya 02.04.2016 Sensor Physical phenomenon Sensor Measurement Output A sensor is a device that measures a physical quantity and converts it into a signal which can be read

More information

Image Formation: Camera Model

Image Formation: Camera Model Image Formation: Camera Model Ruigang Yang COMP 684 Fall 2005, CS684-IBMR Outline Camera Models Pinhole Perspective Projection Affine Projection Camera with Lenses Digital Image Formation The Human Eye

More information

Improved sensitivity high-definition interline CCD using the KODAK TRUESENSE Color Filter Pattern

Improved sensitivity high-definition interline CCD using the KODAK TRUESENSE Color Filter Pattern Improved sensitivity high-definition interline CCD using the KODAK TRUESENSE Color Filter Pattern James DiBella*, Marco Andreghetti, Amy Enge, William Chen, Timothy Stanka, Robert Kaser (Eastman Kodak

More information

Digital Cameras. Consumer and Prosumer

Digital Cameras. Consumer and Prosumer Digital Cameras Overview While silver-halide film has been the dominant photographic process for the past 150 years, the use and role of technology is fast-becoming a standard for the making of photographs.

More information

Camera Image Processing Pipeline: Part II

Camera Image Processing Pipeline: Part II Lecture 13: Camera Image Processing Pipeline: Part II Visual Computing Systems Today Finish image processing pipeline Auto-focus / auto-exposure Camera processing elements Smart phone processing elements

More information

White Paper High Dynamic Range Imaging

White Paper High Dynamic Range Imaging WPE-2015XI30-00 for Machine Vision What is Dynamic Range? Dynamic Range is the term used to describe the difference between the brightest part of a scene and the darkest part of a scene at a given moment

More information

Modeling and Synthesis of Aperture Effects in Cameras

Modeling and Synthesis of Aperture Effects in Cameras Modeling and Synthesis of Aperture Effects in Cameras Douglas Lanman, Ramesh Raskar, and Gabriel Taubin Computational Aesthetics 2008 20 June, 2008 1 Outline Introduction and Related Work Modeling Vignetting

More information

multiframe visual-inertial blur estimation and removal for unmodified smartphones

multiframe visual-inertial blur estimation and removal for unmodified smartphones multiframe visual-inertial blur estimation and removal for unmodified smartphones, Severin Münger, Carlo Beltrame, Luc Humair WSCG 2015, Plzen, Czech Republic images taken by non-professional photographers

More information

Camera Image Processing Pipeline: Part II

Camera Image Processing Pipeline: Part II Lecture 14: Camera Image Processing Pipeline: Part II Visual Computing Systems Today Finish image processing pipeline Auto-focus / auto-exposure Camera processing elements Smart phone processing elements

More information

Determining MTF with a Slant Edge Target ABSTRACT AND INTRODUCTION

Determining MTF with a Slant Edge Target ABSTRACT AND INTRODUCTION Determining MTF with a Slant Edge Target Douglas A. Kerr Issue 2 October 13, 2010 ABSTRACT AND INTRODUCTION The modulation transfer function (MTF) of a photographic lens tells us how effectively the lens

More information

HDR videos acquisition

HDR videos acquisition HDR videos acquisition dr. Francesco Banterle francesco.banterle@isti.cnr.it How to capture? Videos are challenging: We need to capture multiple frames at different exposure times and everything moves

More information

How to combine images in Photoshop

How to combine images in Photoshop How to combine images in Photoshop In Photoshop, you can use multiple layers to combine images, but there are two other ways to create a single image from mulitple images. Create a panoramic image with

More information

A Digital Camera Glossary. Ashley Rodriguez, Charlie Serrano, Luis Martinez, Anderson Guatemala PERIOD 6

A Digital Camera Glossary. Ashley Rodriguez, Charlie Serrano, Luis Martinez, Anderson Guatemala PERIOD 6 A Digital Camera Glossary Ashley Rodriguez, Charlie Serrano, Luis Martinez, Anderson Guatemala PERIOD 6 A digital Camera Glossary Ivan Encinias, Sebastian Limas, Amir Cal Ivan encinias Image sensor A silicon

More information

Image stitching. Image stitching. Video summarization. Applications of image stitching. Stitching = alignment + blending. geometrical registration

Image stitching. Image stitching. Video summarization. Applications of image stitching. Stitching = alignment + blending. geometrical registration Image stitching Stitching = alignment + blending Image stitching geometrical registration photometric registration Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2005/3/22 with slides by Richard Szeliski,

More information

Cameras. CSE 455, Winter 2010 January 25, 2010

Cameras. CSE 455, Winter 2010 January 25, 2010 Cameras CSE 455, Winter 2010 January 25, 2010 Announcements New Lecturer! Neel Joshi, Ph.D. Post-Doctoral Researcher Microsoft Research neel@cs Project 1b (seam carving) was due on Friday the 22 nd Project

More information

Coded photography , , Computational Photography Fall 2018, Lecture 14

Coded photography , , Computational Photography Fall 2018, Lecture 14 Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 14 Overview of today s lecture The coded photography paradigm. Dealing with

More information

Noise and ISO. CS 178, Spring Marc Levoy Computer Science Department Stanford University

Noise and ISO. CS 178, Spring Marc Levoy Computer Science Department Stanford University Noise and ISO CS 178, Spring 2014 Marc Levoy Computer Science Department Stanford University Outline examples of camera sensor noise don t confuse it with JPEG compression artifacts probability, mean,

More information

Dual-fisheye Lens Stitching for 360-degree Imaging & Video. Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington

Dual-fisheye Lens Stitching for 360-degree Imaging & Video. Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington Dual-fisheye Lens Stitching for 360-degree Imaging & Video Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington Introduction 360-degree imaging: the process of taking multiple photographs and

More information

Getting Unlimited Digital Resolution

Getting Unlimited Digital Resolution Getting Unlimited Digital Resolution N. David King Wow, now here s a goal: how would you like to be able to create nearly any amount of resolution you want with a digital camera. Since the higher the resolution

More information

brief history of photography foveon X3 imager technology description

brief history of photography foveon X3 imager technology description brief history of photography foveon X3 imager technology description imaging technology 30,000 BC chauvet-pont-d arc pinhole camera principle first described by Aristotle fourth century B.C. oldest known

More information

Multi-sensor Panoramic Network Camera

Multi-sensor Panoramic Network Camera Multi-sensor Panoramic Network Camera White Paper by Dahua Technology Release 1.0 Table of contents 1 Preface... 2 2 Overview... 3 3 Technical Background... 3 4 Key Technologies... 5 4.1 Feature Points

More information

White Paper Focusing more on the forest, and less on the trees

White Paper Focusing more on the forest, and less on the trees White Paper Focusing more on the forest, and less on the trees Why total system image quality is more important than any single component of your next document scanner Contents Evaluating total system

More information

The Denali-MC HDR ISP Backgrounder

The Denali-MC HDR ISP Backgrounder The Denali-MC HDR ISP Backgrounder 2-4 brackets up to 8 EV frame offset Up to 16 EV stops for output HDR LATM (tone map) up to 24 EV Noise reduction due to merging of 10 EV LDR to a single 16 EV HDR up

More information

PIXPOLAR WHITE PAPER 29 th of September 2013

PIXPOLAR WHITE PAPER 29 th of September 2013 PIXPOLAR WHITE PAPER 29 th of September 2013 Pixpolar s Modified Internal Gate (MIG) image sensor technology offers numerous benefits over traditional Charge Coupled Device (CCD) and Complementary Metal

More information

LENSES. INEL 6088 Computer Vision

LENSES. INEL 6088 Computer Vision LENSES INEL 6088 Computer Vision Digital camera A digital camera replaces film with a sensor array Each cell in the array is a Charge Coupled Device light-sensitive diode that converts photons to electrons

More information

Applying Automated Optical Inspection Ben Dawson, DALSA Coreco Inc., ipd Group (987)

Applying Automated Optical Inspection Ben Dawson, DALSA Coreco Inc., ipd Group (987) Applying Automated Optical Inspection Ben Dawson, DALSA Coreco Inc., ipd Group bdawson@goipd.com (987) 670-2050 Introduction Automated Optical Inspection (AOI) uses lighting, cameras, and vision computers

More information

A HIGH PERFORMANCE HARDWARE ARCHITECTURE FOR HALF-PIXEL ACCURATE H.264 MOTION ESTIMATION

A HIGH PERFORMANCE HARDWARE ARCHITECTURE FOR HALF-PIXEL ACCURATE H.264 MOTION ESTIMATION A HIGH PERFORMANCE HARDWARE ARCHITECTURE FOR HALF-PIXEL ACCURATE H.264 MOTION ESTIMATION Sinan Yalcin and Ilker Hamzaoglu Faculty of Engineering and Natural Sciences, Sabanci University, 34956, Tuzla,

More information

Introduction to camera usage. The universal manual controls of most cameras

Introduction to camera usage. The universal manual controls of most cameras Introduction to camera usage A camera in its barest form is simply a light tight container that utilizes a lens with iris, a shutter that has variable speeds, and contains a sensitive piece of media, either

More information

Topic 9 - Sensors Within

Topic 9 - Sensors Within Topic 9 - Sensors Within Learning Outcomes In this topic, we will take a closer look at sensor sizes in digital cameras. By the end of this video you will have a better understanding of what the various

More information

Image acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor

Image acquisition. In both cases, the digital sensing element is one of the following: Line array Area array. Single sensor Image acquisition Digital images are acquired by direct digital acquisition (digital still/video cameras), or scanning material acquired as analog signals (slides, photographs, etc.). In both cases, the

More information

It should also be noted that with modern cameras users can choose for either

It should also be noted that with modern cameras users can choose for either White paper about color correction More drama Many application fields like digital printing industry or the human medicine require a natural display of colors. To illustrate the importance of color fidelity,

More information

Reikan FoCal Aperture Sharpness Test Report

Reikan FoCal Aperture Sharpness Test Report Focus Calibration and Analysis Software Test run on: 26/01/2016 17:02:00 with FoCal 2.0.6.2416W Report created on: 26/01/2016 17:03:39 with FoCal 2.0.6W Overview Test Information Property Description Data

More information

Image processing with the HERON-FPGA Family

Image processing with the HERON-FPGA Family HUNT ENGINEERING Chestnut Court, Burton Row, Brent Knoll, Somerset, TA9 4BP, UK Tel: (+44) (0)1278 760188, Fax: (+44) (0)1278 760199, Email: sales@hunteng.co.uk http://www.hunteng.co.uk http://www.hunt-dsp.com

More information

F-number sequence. a change of f-number to the next in the sequence corresponds to a factor of 2 change in light intensity,

F-number sequence. a change of f-number to the next in the sequence corresponds to a factor of 2 change in light intensity, 1 F-number sequence a change of f-number to the next in the sequence corresponds to a factor of 2 change in light intensity, 0.7, 1, 1.4, 2, 2.8, 4, 5.6, 8, 11, 16, 22, 32, Example: What is the difference

More information

pco.edge 4.2 LT 0.8 electrons 2048 x 2048 pixel 40 fps up to :1 up to 82 % pco. low noise high resolution high speed high dynamic range

pco.edge 4.2 LT 0.8 electrons 2048 x 2048 pixel 40 fps up to :1 up to 82 % pco. low noise high resolution high speed high dynamic range edge 4.2 LT scientific CMOS camera high resolution 2048 x 2048 pixel low noise 0.8 electrons USB 3.0 small form factor high dynamic range up to 37 500:1 high speed 40 fps high quantum efficiency up to

More information

Cameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017

Cameras. Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017 Cameras Steve Rotenberg CSE168: Rendering Algorithms UCSD, Spring 2017 Camera Focus Camera Focus So far, we have been simulating pinhole cameras with perfect focus Often times, we want to simulate more

More information

Real-Time Face Detection and Tracking for High Resolution Smart Camera System

Real-Time Face Detection and Tracking for High Resolution Smart Camera System Digital Image Computing Techniques and Applications Real-Time Face Detection and Tracking for High Resolution Smart Camera System Y. M. Mustafah a,b, T. Shan a, A. W. Azman a,b, A. Bigdeli a, B. C. Lovell

More information

Quintic Hardware Tutorial Camera Set-Up

Quintic Hardware Tutorial Camera Set-Up Quintic Hardware Tutorial Camera Set-Up 1 All Quintic Live High-Speed cameras are specifically designed to meet a wide range of needs including coaching, performance analysis and research. Quintic LIVE

More information

Lecture 30: Image Sensors (Cont) Computer Graphics and Imaging UC Berkeley CS184/284A

Lecture 30: Image Sensors (Cont) Computer Graphics and Imaging UC Berkeley CS184/284A Lecture 30: Image Sensors (Cont) Computer Graphics and Imaging UC Berkeley Reminder: The Pixel Stack Microlens array Color Filter Anti-Reflection Coating Stack height 4um is typical Pixel size 2um is typical

More information

The Big Train Project Status Report (Part 65)

The Big Train Project Status Report (Part 65) The Big Train Project Status Report (Part 65) For this month I have a somewhat different topic related to the EnterTRAINment Junction (EJ) layout. I thought I d share some lessons I ve learned from photographing

More information

Image Processing for Comets

Image Processing for Comets Image Processing for Comets Page 1 2.5 Surface Today, there are sensors of 768 x 512 pixels up to 8176 x 6132 pixels ( 49,1 mm x 36,8 mm), that's bigger than the old 35mm film. The size of the chip determines

More information

Coded photography , , Computational Photography Fall 2017, Lecture 18

Coded photography , , Computational Photography Fall 2017, Lecture 18 Coded photography http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 18 Course announcements Homework 5 delayed for Tuesday. - You will need cameras

More information

The ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do?

The ultimate camera. Computational Photography. Creating the ultimate camera. The ultimate camera. What does it do? Computational Photography The ultimate camera What does it do? Image from Durand & Freeman s MIT Course on Computational Photography Today s reading Szeliski Chapter 9 The ultimate camera Infinite resolution

More information

CS354 Computer Graphics Computational Photography. Qixing Huang April 23 th 2018

CS354 Computer Graphics Computational Photography. Qixing Huang April 23 th 2018 CS354 Computer Graphics Computational Photography Qixing Huang April 23 th 2018 Background Sales of digital cameras surpassed sales of film cameras in 2004 Digital Cameras Free film Instant display Quality

More information

Lineup for Compact Cameras from

Lineup for Compact Cameras from Lineup for Compact Cameras from Milbeaut M-4 Series Image Processing System LSI for Digital Cameras A new lineup of 1) a low-price product and 2) a product incorporating a moving image function in M-4

More information

Reikan FoCal Aperture Sharpness Test Report

Reikan FoCal Aperture Sharpness Test Report Focus Calibration and Analysis Software Reikan FoCal Sharpness Test Report Test run on: 26/01/2016 17:14:35 with FoCal 2.0.6.2416W Report created on: 26/01/2016 17:16:16 with FoCal 2.0.6W Overview Test

More information

NOVA S12. Compact and versatile high performance camera system. 1-Megapixel CMOS Image Sensor: 1024 x 1024 pixels at 12,800fps

NOVA S12. Compact and versatile high performance camera system. 1-Megapixel CMOS Image Sensor: 1024 x 1024 pixels at 12,800fps NOVA S12 1-Megapixel CMOS Image Sensor: 1024 x 1024 pixels at 12,800fps Maximum Frame Rate: 1,000,000fps Class Leading Light Sensitivity: ISO 12232 Ssat Standard ISO 64,000 monochrome ISO 16,000 color

More information

Imaging serial interface ROM

Imaging serial interface ROM Page 1 of 6 ( 3 of 32 ) United States Patent Application 20070024904 Kind Code A1 Baer; Richard L. ; et al. February 1, 2007 Imaging serial interface ROM Abstract Imaging serial interface ROM (ISIROM).

More information

Phase One 190MP Aerial System

Phase One 190MP Aerial System White Paper Phase One 190MP Aerial System Introduction Phase One Industrial s 100MP medium format aerial camera systems have earned a worldwide reputation for its high performance. They are commonly used

More information

Be aware that there is no universal notation for the various quantities.

Be aware that there is no universal notation for the various quantities. Fourier Optics v2.4 Ray tracing is limited in its ability to describe optics because it ignores the wave properties of light. Diffraction is needed to explain image spatial resolution and contrast and

More information

Early art: events. Baroque art: portraits. Renaissance art: events. Being There: Capturing and Experiencing a Sense of Place

Early art: events. Baroque art: portraits. Renaissance art: events. Being There: Capturing and Experiencing a Sense of Place Being There: Capturing and Experiencing a Sense of Place Early art: events Richard Szeliski Microsoft Research Symposium on Computational Photography and Video Lascaux Early art: events Early art: events

More information

Lecture 18: Light field cameras. (plenoptic cameras) Visual Computing Systems CMU , Fall 2013

Lecture 18: Light field cameras. (plenoptic cameras) Visual Computing Systems CMU , Fall 2013 Lecture 18: Light field cameras (plenoptic cameras) Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today:

More information

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011) Lecture 19: Depth Cameras Kayvon Fatahalian CMU 15-869: Graphics and Imaging Architectures (Fall 2011) Continuing theme: computational photography Cheap cameras capture light, extensive processing produces

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 22: Computational photography photomatix.com Announcements Final project midterm reports due on Tuesday to CMS by 11:59pm BRDF s can be incredibly complicated

More information

Lecture 11: Clocking

Lecture 11: Clocking High Speed CMOS VLSI Design Lecture 11: Clocking (c) 1997 David Harris 1.0 Introduction We have seen that generating and distributing clocks with little skew is essential to high speed circuit design.

More information

Reikan FoCal Aperture Sharpness Test Report

Reikan FoCal Aperture Sharpness Test Report Focus Calibration and Analysis Software Reikan FoCal Sharpness Test Report Test run on: 10/02/2016 19:57:05 with FoCal 2.0.6.2416W Report created on: 10/02/2016 19:59:09 with FoCal 2.0.6W Overview Test

More information

A new Photon Counting Detector: Intensified CMOS- APS

A new Photon Counting Detector: Intensified CMOS- APS A new Photon Counting Detector: Intensified CMOS- APS M. Belluso 1, G. Bonanno 1, A. Calì 1, A. Carbone 3, R. Cosentino 1, A. Modica 4, S. Scuderi 1, C. Timpanaro 1, M. Uslenghi 2 1-I.N.A.F.-Osservatorio

More information

Technologies Explained PowerShot D20

Technologies Explained PowerShot D20 Technologies Explained PowerShot D20 EMBARGO: 7 th February 2012, 05:00 (GMT) HS System The HS System represents a powerful combination of a high-sensitivity sensor and high-performance DIGIC image processing

More information

Tonemapping and bilateral filtering

Tonemapping and bilateral filtering Tonemapping and bilateral filtering http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 6 Course announcements Homework 2 is out. - Due September

More information

Reikan FoCal Aperture Sharpness Test Report

Reikan FoCal Aperture Sharpness Test Report Focus Calibration and Analysis Software Reikan FoCal Sharpness Test Report Test run on: 27/01/2016 00:35:25 with FoCal 2.0.6.2416W Report created on: 27/01/2016 00:41:43 with FoCal 2.0.6W Overview Test

More information

Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring

Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring Implementation of Adaptive Coded Aperture Imaging using a Digital Micro-Mirror Device for Defocus Deblurring Ashill Chiranjan and Bernardt Duvenhage Defence, Peace, Safety and Security Council for Scientific

More information

A new Photon Counting Detector: Intensified CMOS- APS

A new Photon Counting Detector: Intensified CMOS- APS A new Photon Counting Detector: Intensified CMOS- APS M. Belluso 1, G. Bonanno 1, A. Calì 1, A. Carbone 3, R. Cosentino 1, A. Modica 4, S. Scuderi 1, C. Timpanaro 1, M. Uslenghi 2 1- I.N.A.F.-Osservatorio

More information

Reikan FoCal Aperture Sharpness Test Report

Reikan FoCal Aperture Sharpness Test Report Focus Calibration and Analysis Software Test run on: 26/01/2016 17:56:23 with FoCal 2.0.6.2416W Report created on: 26/01/2016 17:59:12 with FoCal 2.0.6W Overview Test Information Property Description Data

More information

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics IMAGE FORMATION Light source properties Sensor characteristics Surface Exposure shape Optics Surface reflectance properties ANALOG IMAGES An image can be understood as a 2D light intensity function f(x,y)

More information

Measuring the impact of flare light on Dynamic Range

Measuring the impact of flare light on Dynamic Range Measuring the impact of flare light on Dynamic Range Norman Koren; Imatest LLC; Boulder, CO USA Abstract The dynamic range (DR; defined as the range of exposure between saturation and 0 db SNR) of recent

More information

Photographing Long Scenes with Multiviewpoint

Photographing Long Scenes with Multiviewpoint Photographing Long Scenes with Multiviewpoint Panoramas A. Agarwala, M. Agrawala, M. Cohen, D. Salesin, R. Szeliski Presenter: Stacy Hsueh Discussant: VasilyVolkov Motivation Want an image that shows an

More information

SMART LASER SENSORS SIMPLIFY TIRE AND RUBBER INSPECTION

SMART LASER SENSORS SIMPLIFY TIRE AND RUBBER INSPECTION PRESENTED AT ITEC 2004 SMART LASER SENSORS SIMPLIFY TIRE AND RUBBER INSPECTION Dr. Walt Pastorius LMI Technologies 2835 Kew Dr. Windsor, ON N8T 3B7 Tel (519) 945 6373 x 110 Cell (519) 981 0238 Fax (519)

More information

Low-Cost, On-Demand Film Digitisation and Online Delivery. Matt Garner

Low-Cost, On-Demand Film Digitisation and Online Delivery. Matt Garner Low-Cost, On-Demand Film Digitisation and Online Delivery Matt Garner (matt.garner@findmypast.com) Abstract Hundreds of millions of pages of microfilmed material are not being digitised at this time due

More information

Focusing and Metering

Focusing and Metering Focusing and Metering CS 478 Winter 2012 Slides mostly stolen by David Jacobs from Marc Levoy Focusing Outline Manual Focus Specialty Focus Autofocus Active AF Passive AF AF Modes Manual Focus - View Camera

More information

Shape sensing for computer aided below-knee prosthetic socket design

Shape sensing for computer aided below-knee prosthetic socket design Prosthetics and Orthotics International, 1985, 9, 12-16 Shape sensing for computer aided below-knee prosthetic socket design G. R. FERNIE, G. GRIGGS, S. BARTLETT and K. LUNAU West Park Research, Department

More information

White Paper: Compression Advantages of Pixim s Digital Pixel System Technology

White Paper: Compression Advantages of Pixim s Digital Pixel System Technology White Paper: Compression Advantages of Pixim s Digital Pixel System Technology Table of Contents The role of efficient compression algorithms Bit-rate strategies and limits 2 Amount of motion present in

More information

Photography Help Sheets

Photography Help Sheets Photography Help Sheets Phone: 01233 771915 Web: www.bigcatsanctuary.org Using your Digital SLR What is Exposure? Exposure is basically the process of recording light onto your digital sensor (or film).

More information

products PC Control

products PC Control products PC Control 04 2017 PC Control 04 2017 products Image processing directly in the PLC TwinCAT Vision Machine vision easily integrated into automation technology Automatic detection, traceability

More information

IHV means Independent Hardware Vendor. Example is Qualcomm Technologies Inc. that makes Snapdragon processors. OEM means Original Equipment

IHV means Independent Hardware Vendor. Example is Qualcomm Technologies Inc. that makes Snapdragon processors. OEM means Original Equipment 1 2 IHV means Independent Hardware Vendor. Example is Qualcomm Technologies Inc. that makes Snapdragon processors. OEM means Original Equipment Manufacturer. Examples are smartphone manufacturers. Tuning

More information

CAMERA BASICS. Stops of light

CAMERA BASICS. Stops of light CAMERA BASICS Stops of light A stop of light isn t a quantifiable measurement it s a relative measurement. A stop of light is defined as a doubling or halving of any quantity of light. The word stop is

More information

Which equipment is necessary? How is the panorama created?

Which equipment is necessary? How is the panorama created? Congratulations! By purchasing your Panorama-VR-System you have acquired a tool, which enables you - together with a digital or analog camera, a tripod and a personal computer - to generate high quality

More information

Edmonton Camera Club. Introduction to Exposure. and a few other bits!

Edmonton Camera Club. Introduction to Exposure. and a few other bits! Edmonton Camera Club Introduction to Exposure and a few other bits! Exposure 3 Variables 1. Aperture how much light 2. Shutter Speed for how long 3. Sensitivity ISO, Film Speed Also cover: Compensation

More information

Camera Overview. Digital Microscope Cameras for Material Science: Clear Images, Precise Analysis. Digital Cameras for Microscopy

Camera Overview. Digital Microscope Cameras for Material Science: Clear Images, Precise Analysis. Digital Cameras for Microscopy Digital Cameras for Microscopy Camera Overview For Materials Science Microscopes Digital Microscope Cameras for Material Science: Clear Images, Precise Analysis Passionate about Imaging: Olympus Digital

More information

Part Number SuperPix TM image sensor is one of SuperPix TM 2 Mega Digital image sensor series products. These series sensors have the same maximum ima

Part Number SuperPix TM image sensor is one of SuperPix TM 2 Mega Digital image sensor series products. These series sensors have the same maximum ima Specification Version Commercial 1.7 2012.03.26 SuperPix Micro Technology Co., Ltd Part Number SuperPix TM image sensor is one of SuperPix TM 2 Mega Digital image sensor series products. These series sensors

More information

9/19/16. A Closer Look. Danae Wolfe. What We ll Cover. Basics of photography & your camera. Technical. Macro & close-up techniques.

9/19/16. A Closer Look. Danae Wolfe. What We ll Cover. Basics of photography & your camera. Technical. Macro & close-up techniques. A Closer Look Danae Wolfe What We ll Cover Basics of photography & your camera Technical Macro & close-up techniques Creative 1 What is Photography? Photography: the art, science, & practice of creating

More information

Image stabilization (IS)

Image stabilization (IS) Image stabilization (IS) CS 178, Spring 2009 Marc Levoy Computer Science Department Stanford University Outline what are the causes of camera shake? and how can you avoid it (without having an IS system)?

More information

Note: These sample pages are from Chapter 1. The Zone System

Note: These sample pages are from Chapter 1. The Zone System Note: These sample pages are from Chapter 1 The Zone System Chapter 1 The Zones Revealed The images below show how you can visualize the zones in an image. This is NGC 1491, an HII region imaged through

More information

COLOR CORRECTION METHOD USING GRAY GRADIENT BAR FOR MULTI-VIEW CAMERA SYSTEM. Jae-Il Jung and Yo-Sung Ho

COLOR CORRECTION METHOD USING GRAY GRADIENT BAR FOR MULTI-VIEW CAMERA SYSTEM. Jae-Il Jung and Yo-Sung Ho COLOR CORRECTION METHOD USING GRAY GRADIENT BAR FOR MULTI-VIEW CAMERA SYSTEM Jae-Il Jung and Yo-Sung Ho School of Information and Mechatronics Gwangju Institute of Science and Technology (GIST) 1 Oryong-dong

More information

SpheroCam HDR. Image based lighting with. Capture light perfectly SPHERON VR. 0s 20s 40s 60s 80s 100s 120s. Spheron VR AG

SpheroCam HDR. Image based lighting with. Capture light perfectly SPHERON VR. 0s 20s 40s 60s 80s 100s 120s. Spheron VR AG Image based lighting with SpheroCam HDR Capture light perfectly 0 60 120 180 240 300 360 0s 20s 40s 60s 80s 100s 120s SPHERON VR high dynamic range imaging Spheron VR AG u phone u internet Hauptstraße

More information

OV7670 Software Application Note

OV7670 Software Application Note OV7670 Software Application Note Table of Contents OV7670 Software Application Note... 1 1. Select Output format...3 1.1 Backend with full ISP... 3 1.2 Backend with YCbCr ISP... 4 1.3 Backend without ISP...4

More information

Digital camera modes explained: choose the best shooting mode for your subject

Digital camera modes explained: choose the best shooting mode for your subject Digital camera modes explained: choose the best shooting mode for your subject On most DSLRs, the Mode dial is split into three sections: Scene modes (for doing point-and-shoot photography in specific

More information

Image Sensor and Camera Technology November 2016 in Stuttgart

Image Sensor and Camera Technology November 2016 in Stuttgart Image Sensor and Camera Technology 14-15-16 November 2016 in Stuttgart Aphesa organizes an image sensor and camera technology training tour between October 2015 and November 2016. The training sessions

More information

MEM: Intro to Robotics. Assignment 3I. Due: Wednesday 10/15 11:59 EST

MEM: Intro to Robotics. Assignment 3I. Due: Wednesday 10/15 11:59 EST MEM: Intro to Robotics Assignment 3I Due: Wednesday 10/15 11:59 EST 1. Basic Optics You are shopping for a new lens for your Canon D30 digital camera and there are lots of lens options at the store. Your

More information