A Real Time Algorithm for Exposure Fusion of Digital Images
|
|
- Lionel Cross
- 6 years ago
- Views:
Transcription
1 A Real Time Algorithm for Exposure Fusion of Digital Images Tomislav Kartalov #1, Aleksandar Petrov *2, Zoran Ivanovski #3, Ljupcho Panovski #4 # Faculty of Electrical Engineering Skopje, Karpoš II bb, 1000 Skopje, Macedonia 1 kartalov@feit.ukim.edu.mk 3 mars@feit.ukim.edu.mk 4 panovski@feit.ukim.edu.mk * Netcetera, Partizanski Odredi 72a, 1000 Skopje, Republic of Macedonia 2 apetrov@netcetera.com.mk Abstract A real time algorithm for fusion of differently exposed images is proposed in this paper. The algorithm blends the details from two images of high dynamic range scene, acquired with different exposure values, into one output image which can be displayed on low dynamic range devices. The blending is performed in the spatial domain, using pixel by pixel approach, thus eliminating the need for expensive block processing or transform domain coding. The proposed scheme works both on grey and color images. The algorithm shows high efficiency, which make it applicable on low processing power platforms, such as mobile devices. The obtained results are visually comparable with previously published algorithms that are computationally much more expensive. Keywords Image, fusion, dynamic range. I. INTRODUCTION The main objective of the modern electronic devices for digital image and video acquisition is to represent a captured scene as realistic, and as identical to the real eye-observed scene, as possible. With every new model and every new technology, this goal is more achieved. The luminance, the contrast, the saturation of colors, all these parameters of the digital media are continuously getting closer to the psycho physical parameters experienced by human observer in the real scene. However, the photometric quantities existent in the real scene, are differently interpreted by the electronic devices and the human observer, and this still causes in some cases the visual parameters of digital media to be far from visual parameters of the original scene. This especially applies to the real world scenes which have very high ratio between maximum and minimum intensity of the light in the scene, e.g. high dynamic range. Neither image acquisition device, nor display device is manufactured today, which can reproduce the intensity dynamic range that may exist in the real scene, or even the perceived luminance dynamic range of the human eye (brain). This phenomenon often results in loss of detail in the digital reproduction, and overall unpleasant image to the viewer. The detail problem can be addressed by taking multiple snapshots of the scene of interest, using different light sensitivity settings of the capturing device, i.e. different exposure values. That way, multiple digital images can be obtained, in which various segments of the whole light intensity interval are shown. Images taken with longer exposures will reveal the darker objects, while the brighter objects in the scene will be shown on images with shorter exposure values. However, handling multiple images from one scene can be difficult and confusing even if only a still scene is observed. In the case of video material, this approach would be impossible. As a result, a need occurs for combining these multiple snapshots into single understandable image. That image would contain all of the detail in the scene, and would be able of displaying on standard display devices. As pointed above, display devices have lower dynamic range than real world scene, so some dynamic range compression has to take place in the process. Digital image and video community put a lot of effort into optimal solution of the problem of different exposure image fusion, utilizing various concepts and methodologies [1] - [6]. All these proposed solutions use two or more perfectly spatially aligned input images obtained with different exposures, to produce single output image which will contain all the useful parts from the input images. Mann and Picard in [1] propose very complex algorithm which tries to reconstruct point by point the nonlinear response function of an image sensor used in the capturing of the two input images, captured using high and low exposure values, respectively. The selection criteria for combining the luminance of two input images, is based on weighted average, the weights being computed using the previously obtained sensor response function. The luminance values in the result image are mapped to match the low dynamic range of the display devices. In [2], Debevec and Malik use the algorithm for image fusion based on exploiting a physical property of imaging systems, both photochemical and electronic, known as reciprocity. They calculate the characteristic curve of the response of a film to variations in exposure, or Hurter- Driffield curve. That function links the optical density of the film with the logarithm of the exposure to which it has been subjected. The response curve is then used in calculation of the radiance map of the recorded scene, and its mapping on low dynamic range display. Fattal et al. in [3] propose gradient domain dynamic range compression, using the /10/$ IEEE 641
2 property of human visual system to be less sensitive to absolute luminance levels on the retina, but rather responds to local intensity ratio changes. The algorithm is based on spatially variant attenuating mapping to the gradient of the logarithm of the function which represents the luminance range in the scene. The large gradients are more attenuated than the small ones, causing compression of the high dynamic range. In [4], Goshtasby divides the input images on rectangular blocks, calculates the entropy of the luminance within these blocks, and composes the output image from blocks that have highest entropy. In order to solve the imminent occurrence of the tiling effect in the output image, he applies the Gaussian blending function to the image. The size of the blocks and the support of the blending function are iteratively computed on every image, in order to obtain highest entropy of the result, making this algorithm very complex and slow. Using fixed values for block size and blending support speeds up the algorithm, but often results in occurrence of halos around the objects in output image. The authors in [5] and [6] have similar ideas about the method of fusing the images with different exposure values. In the selection part, Mertens in [6] presents more complex scheme for computing the local weights of participants in the output image, while Rubinstein, in [5] uses very simple selection model, producing selection maps as binary, locally choosing only one of the submitted images. The image fusion in both algorithms is performed in a similar manner, using pyramid based image decomposition, as explained in detail in [7] and [8]. Mertens and Rubinstein both employ Gaussian and Laplacian pyramid decomposition of the input images, and of the selection (weight) maps, afterwards building the output image from the useful parts of the pyramids. All explained algorithms perform fairly well, other then being too complex and slow for real time implementation, especially on low power platforms. In this paper, we propose a new algorithm for fusion of the images with different exposures, which visually works comparable with above listed algorithms, and is much less complex in the same time. Our algorithm takes place in spatial domain, in the HSV color space, performs no pre-processing and post processing at all, and every pixel in the output image is calculated only using respective pixels in the input images, thus eliminating the expensive block processing, filtering or blending functions. This paper is organized as follows. In Section II our algorithm is explained in detail, Section III gives some of the obtained experimental results, and the concluding remarks are put in Section IV. Unnumbered sections at the end of the document contain acknowledgments and used references. II. THE ALGORITHM In this paper we propose an algorithm for fusion of multiple images with different exposure values into single image. The design of this algorithm was outlined by the following constraints: The whole procedure should run in low number of operations per pixel, in order to reach the option of embedding this algorithm on low processing power platforms, such as mobile devices. The algorithm should have as lower as possible memory requirement, concerning the usual amount of memory installed on targeted platforms. The algorithm should work both for grayscale and color images. Given the first constraint, we limit our algorithm to two input images, one taken with longer exposure time (overexposed image), and one with shorter exposure time (underexposed image). These two images should be taken consequently by the same device. The period of time between the capturing of the two images should be as short as possible, in order to minimize the changes in the recorded scene. It would be best if the capturing of the two images is automated through the hardware, triggered by single command from the end user. In our algorithm we assume perfect spatial alignment of the overexposed and the underexposed image. The second constraint for low memory consumption is contented by using the pixel by pixel approach, in which only a minimal amount of memory is consumed for processing the pixels with respective places in the overexposed and the underexposed images. Fig. 1. An example of the overexposed (left side), and the underexposed (right side) image taken from the same scene 642
3 Finally, in order to meet the last constraint, this algorithm is performed in the HSV, rather than in the RGB color space. The HSV space allows simple migration between grayscale and color images, in a manner of processing only the V (Value) channel, or all three (Hue, Saturation, Value) channels, respectively. On the other hand, we benefit from the several advantages the HSV space has over RGB space, the most important being the fact that the HSV space is more similar to the psycho physical representation of colors by the human visual system. A. The luminance transfer function The high dynamic range of the light intensity in the real scene cannot be fully represented by a single image with single exposure value. In Fig. 1, an example of one fairly static scene captured with different exposure values is shown. As it can be seen from the figure, the left side image is the overexposed image, and the objects with lower radiance in the scene are clear and rich in detail, while brighter areas and objects can not be observed, due to the white saturation of the image. On the other hand, the right side image, underexposed image, reveal the detail from the objects with higher radiance in the real scene, however the less radiant objects are too dark, and some of them black saturated. If we assume that the normalized average calculated from these two images is a fair representation of the light intensity in the real scene, we can construct a luminance transfer function, Fig. 2. This function shows the manner of mapping the real scene light intensities into image luminance values for the cases of the overexposed and the underexposed images. From Fig. 2 the difference in the global brightness level between the overexposed and the underexposed images is evident, as well as the saturation areas, white for the overexposed and black for the underexposed image. B. The exposure fusion process The main idea in the proposed algorithm is to construct the approximation of the ideal luminance transfer function by translation of certain parts of the transfer functions for the overexposed and the underexposed images. In the following, we explain the procedure for constructing such transfer function, and by itself this procedure is enough to obtain a fused result if the input is grayscale. For color images, few further adjustments should take place, which will be explained in the next subheading. 1) Grayscale images: The processing of a grayscale images is performed using only the V channel of the color image in the HSV color space, or using the whole input image, if it is already grayscale. The procedure for obtaining an ideal luminance transfer function can be understood observing the drawing in Fig. 3. Initially, two threshold values are defined, T H1 and T H2. These values have to enclose the saturation areas of the transfer functions from both images. In our implementation, T H1 is 5% of the maximum possible luminance value in the image, and T H2 equals 95 % of the same value. Then, three different classes of pixel pairs are defined according to the luminance values of the pixels in the respective places from the overexposed and the underexposed images. For every class, different method of constructing the luminance value of the pixel in the fused image is implemented. Luminance Luminance Light intensity Fig. 2. The luminance transfer functions for images shown in Fig. 1 It is clear that in order to get single fused image from the recorded scene, in which all the detail will be present, the ideal luminance transfer function, Fig. 2, must be pursued. In low dynamic range real scenes approximation of the ideal luminance transfer function can be obtained by optimal choice of the exposure value. In high dynamic range scenes, such value does not exist, and the ideal luminance transfer function must be constructed based on the available data in the overexposed and the underexposed images. Light intensity Fig. 3. Approximation of an ideal luminance transfer function The decisions made and the calculations performed during the processing of one pair of pixels are shown in Fig. 4. The two luminance values are read from the respective positions, i- th row and j-th column, in the two input images. This operation is repeated for every pixel in the fused image. The luminance values of the pixels in the fused image are calculated using the formulas (1), (2) and (3). diff f = un+ (1) 4 For the white saturated class, the luminance values f for the pixels in the fused image are obtained with translation of the luminance values in the underexposed image by some portion of the difference between the overexposed and the underexposed image, diff, (1). This is coherent to a slight brightness increase of the underexposed image, in order to match the ideal luminance transfer function. 643
4 The value of the brightness increase should not be too high, because it may impose new white saturation and loss of detail in those areas. Our experiments show that the optimal increase is by quarter of the difference diff. overexposed image. Luminance = ov un < T H1 Calculate the difference diff = ov - un Yes underexposed image. Luminance = un class = black saturated. in the native grayscale images). The equation (3) is constructed so that a smooth luminance transfer function is established. It connects the black saturated and the white saturated classes, and eliminates the abrupt changes in the luminance values. For its border cases, ov = M, or un = 0, the equation (3) reduces into equations (1) or (2), respectively. 2) Color images: In the case when a color output image is required, two input images must be full color images in the HSV color space. The V channel is processed in the same way as explained in the previous subheading. The other two channels are processed as following. The channel S, which represents the color saturation of the observed pixel, imposes a great deal of the subjective quality of a color image, because the standard human observer tends to give higher quality grade to the images with more saturated (pure) colors. To address this fact, in our algorithm we implement the saturation maximization procedure, which is depicted on Fig. 5. No ov > T H2 Yes class = white saturated. overexposed image. Saturation = Sov Hue = Hov underexposed image. Saturation = Sun Hue = Hun No class = weighted average. Pixel (i,j) from the fused image. Luminance = Vfu ov un diff class Calculate the luminance of the pixel (i,j) in the fused image, based on class, diff, ov and un Calculate the visual saturation factor VSFov = Sov(1-Vfu) Calculate the visual saturation factor VSFun = SunVfu Fig. 4. The operations performed on one pair of pixels For the black saturated class, the data from the underexposed image in unusable and the fused image is constructed solely from the pixels in the overexposed image, with slightly lowered luminance values, again by quarter of the difference diff diff f = ov (2) 4 In the regions of the image that are neither black, nor white saturated, the weighted average class, the fused image is constructed as weighted average from overexposed and the underexposed image, using diff diff ov ( M ov) + un + un 4 4 f = (3) M diff where M is the maximum possible luminance value in the images (e.g. 1 for V in the HSV images, or 255 for luminance Sfu = VSFun Hfu = Hun Yes VSFov < VSFun No Sfu = VSFun Hfu = Hov Fig. 5. Saturation maximization and hue selection The information for saturation by itself is not enough to estimate the human observed saturation, because the absolute luminance in the same place is also important. Too bright or too dark regions, although with highly saturated colors, will be perceived as low saturation regions by the human observer. We define the visual saturation factor, VSF, in order to create a measure for the saturation as viewed by the human, taking into account the total amount of luminance too. This factor is calculated differently for the overexposed and for the underexposed image. By comparing the two results and using the higher, visually more saturated colors in the fused image can be achieved. For the saturation of the fused image the VSF values are used rather than the S values from the input images, 644
5 in order to avoid rapid changes in saturation in neighboring pixels, which will create unpleasant noise-like color artifacts in the fused image. The H (Hue) channel from the HSV color images corresponds to the exact tone of the colors in the image, so its value must not be altered by the algorithm, for the fused image to contain the original colors from the scene. Our algorithm is designed simply to choose between two (or in most cases only one) offered values for H, by the overexposed and the underexposed image. The selected value for H is from the pixel with higher VSF value, which contributes to the visual quality of the fused image. III. EXPERIMENTAL RESULTS The proposed algorithm was tested on many images obtained with digital camera using bracketed exposure values. The results were compared to previously published algorithms [4] and [6]. In Fig. 6, the results obtained by processing the images from the example in Fig. 1 are presented. The result using the algorithm [4] with fixed values for block size and blending support for speed is shown in Fig. 6 a). Gradual luminance changes in form of halos are clearly visible around objects and edges in the image. In Fig. 6 b), the result of the algorithm [6] is presented, which has much higher visual quality. The result of the application of our algorithm is shown in Fig 6 c). As it can be observed, our algorithm creates better, or comparable result with the other two algorithms. However, our algorithm is quite less complex, especially compared to the algorithm [6], which calculates two Laplacian decompositions of the input images, and one reconstruction from the fused Laplacian pyramid. The pyramid in the algorithm [6] must be calculated to the level of one pixel, in order to avoid the tiling effect in the output image. The algorithm [4] can produce better visual result if the iterative calculation of the block size and blending support is performed, but in that case this algorithm is even more complex and slow than the algorithm [6]. The moving objects (the car observed through the window) in the scene, are poorly handled with the proposed algorithm. This is due to the fact that this object is not present in one of the input images (see Fig. 1, the car does not exist in the underexposed image) and is present in the other. Mentioned above, our algorithm assumes perfect alignment of the two input images, and there is no procedure implemented that solves this kind of difficulty. On the other hand, for such cases the priority criterion should be defined, because it is unknown which image objects are more important for the end user (e.g. for this case, the car, or the objects hidden by the car). Our algorithm exhibits some sort of priority criterion, in the cases of respective pixels with black and white saturation in the same time, giving priority to the data from the overexposed image, because firstly the black saturation is checked, Fig. 4. This is intentionally designed in such way, in order to obtain generally brighter output image rather than generally darker, which also proves to be more desirable for the visual quality assessment by the human observer. In Fig. 7 few more images are shown depicting the results obtained using the proposed algorithm on different color input images. Alongside the successful luminance fusion, the effect of saturation maximization procedure is evident, resulting in more vivid and colorful fused image. These examples further prove the good performance of the proposed algorithm. IV. CONCLUSIONS In this paper we proposed a new algorithm for fusion of images with different exposure values. The algorithm is fast and simple, and works in the HSV color space. It is designed to process one pixel at a time, operating in very low number of instructions, which makes it a perfect choice for implementation on low memory and low processing power platforms. The algorithm is tested against more complex previously published algorithms on many images. The results show that although simpler, this algorithm produces comparable or sometimes better results than other algorithms. Future work should include development of the procedure for optimal selection of exposure values, targeting the maximization of usable data from the overexposed and the underexposed images. Also, adaptive thresholds T H1 and T H2 should be considered, in order to match the statistical distribution of the luminance values in the given images. Finally, detail or object priority criterion could be implemented, to resolve the non-persistent object problem (like the car in the first example). Of course, all these improvements will make the algorithm more complex and harder to implement on targeted platforms, so further research should be performed in finding the efficient running methods for the intended procedures. ACKNOWLEDGMENT The results were obtained in the course of a research project commissioned and funded by NXP Software B.V., Eindhoven. REFERENCES [1] Steve Mann, Rosalind W. Picard, On being undigital with digital cameras: Extending Dynamic Range by Combining Differently Exposed Pictures, M.I.T. Media Laboratory Perceptual Computing Section Technical Report No. TR-323, 1994, also appears in the Proceedings of the 46th Annual Imaging Science & Technology Conference, Washington D.C., USA, May 1995, pp [2] Paul E. Debevec, Jitendra Malik, Recovering high dynamic range radiance maps from photographs, 24th annual conference on Computer graphics and interactive techniques, SIGGRAPH 97, Los Angeles, California, Aug 1997, pp [3] Raanan Fattal, Dani Lischinski, Michael Werman, Gradient domain high dynamic range compression, Proc. of the 29 th annual conference on Computer graphics and interactive techniques, San Antonio, Texas, July 2002, pp [4] Ardeshir Goshtasby, Fusion of multi-exposure images, Image and Vision Computing, vol. 23, pp , [5] Ron Rubinstein, Fusion of Differently Exposed Images, Technion, Israel Institute of Technology, Final Project Report, Oct [6] Tom Mertens, Jan Kautz, Frank Van Reeth, Exposure Fusion, Proc. of the 15 th Pacific Conference on Computer Graphics and Applications, Maui, Hawaii, Oct/Nov 2007, pp [7] Peter J. Burt, Edward H. Adelson, The Laplacian Pyramid as a Compact Image Code, IEEE Trans. On Communications, vol. com-3l, no. 4, pp , April [8] J. M. Ogden, E. H. Adelson, J. R. Bergen, and P. J. Burt, Pyramidbased computer graphics, RCA Engineer, vol. 30, no. 5,
6 a) Goshtasby, [4] b) Mertens et al, [6] c) The proposed algorithm d) Preview of the input images Fig. 6. Visual comparison of the algorithms applied on grayscale input images Fig. 7. The proposed image fusion algorithm applied on different pairs of color input images 646
A Saturation-based Image Fusion Method for Static Scenes
2015 6th International Conference of Information and Communication Technology for Embedded Systems (IC-ICTES) A Saturation-based Image Fusion Method for Static Scenes Geley Peljor and Toshiaki Kondo Sirindhorn
More informationHigh Dynamic Range Imaging
High Dynamic Range Imaging 1 2 Lecture Topic Discuss the limits of the dynamic range in current imaging and display technology Solutions 1. High Dynamic Range (HDR) Imaging Able to image a larger dynamic
More informationDistributed Algorithms. Image and Video Processing
Chapter 7 High Dynamic Range (HDR) Distributed Algorithms for Introduction to HDR (I) Source: wikipedia.org 2 1 Introduction to HDR (II) High dynamic range classifies a very high contrast ratio in images
More informationDenoising and Effective Contrast Enhancement for Dynamic Range Mapping
Denoising and Effective Contrast Enhancement for Dynamic Range Mapping G. Kiruthiga Department of Electronics and Communication Adithya Institute of Technology Coimbatore B. Hakkem Department of Electronics
More informationRealistic Image Synthesis
Realistic Image Synthesis - HDR Capture & Tone Mapping - Philipp Slusallek Karol Myszkowski Gurprit Singh Karol Myszkowski LDR vs HDR Comparison Various Dynamic Ranges (1) 10-6 10-4 10-2 100 102 104 106
More informationHigh dynamic range imaging and tonemapping
High dynamic range imaging and tonemapping http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 12 Course announcements Homework 3 is out. - Due
More informationContinuous Flash. October 1, Technical Report MSR-TR Microsoft Research Microsoft Corporation One Microsoft Way Redmond, WA 98052
Continuous Flash Hugues Hoppe Kentaro Toyama October 1, 2003 Technical Report MSR-TR-2003-63 Microsoft Research Microsoft Corporation One Microsoft Way Redmond, WA 98052 Page 1 of 7 Abstract To take a
More informationA Short History of Using Cameras for Weld Monitoring
A Short History of Using Cameras for Weld Monitoring 2 Background Ever since the development of automated welding, operators have needed to be able to monitor the process to ensure that all parameters
More informationISSN Vol.03,Issue.29 October-2014, Pages:
ISSN 2319-8885 Vol.03,Issue.29 October-2014, Pages:5768-5772 www.ijsetr.com Quality Index Assessment for Toned Mapped Images Based on SSIM and NSS Approaches SAMEED SHAIK 1, M. CHAKRAPANI 2 1 PG Scholar,
More informationAutomatic Selection of Brackets for HDR Image Creation
Automatic Selection of Brackets for HDR Image Creation Michel VIDAL-NAQUET, Wei MING Abstract High Dynamic Range imaging (HDR) is now readily available on mobile devices such as smart phones and compact
More informationCorrecting Over-Exposure in Photographs
Correcting Over-Exposure in Photographs Dong Guo, Yuan Cheng, Shaojie Zhuo and Terence Sim School of Computing, National University of Singapore, 117417 {guodong,cyuan,zhuoshao,tsim}@comp.nus.edu.sg Abstract
More informationA Kalman-Filtering Approach to High Dynamic Range Imaging for Measurement Applications
A Kalman-Filtering Approach to High Dynamic Range Imaging for Measurement Applications IEEE Transactions on Image Processing, Vol. 21, No. 2, 2012 Eric Dedrick and Daniel Lau, Presented by Ran Shu School
More informationTonemapping and bilateral filtering
Tonemapping and bilateral filtering http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 6 Course announcements Homework 2 is out. - Due September
More informationTone mapping. Digital Visual Effects, Spring 2009 Yung-Yu Chuang. with slides by Fredo Durand, and Alexei Efros
Tone mapping Digital Visual Effects, Spring 2009 Yung-Yu Chuang 2009/3/5 with slides by Fredo Durand, and Alexei Efros Tone mapping How should we map scene luminances (up to 1:100,000) 000) to display
More informationColour correction for panoramic imaging
Colour correction for panoramic imaging Gui Yun Tian Duke Gledhill Dave Taylor The University of Huddersfield David Clarke Rotography Ltd Abstract: This paper reports the problem of colour distortion in
More informationSuper resolution with Epitomes
Super resolution with Epitomes Aaron Brown University of Wisconsin Madison, WI Abstract Techniques exist for aligning and stitching photos of a scene and for interpolating image data to generate higher
More informationMODIFICATION OF ADAPTIVE LOGARITHMIC METHOD FOR DISPLAYING HIGH CONTRAST SCENES BY AUTOMATING THE BIAS VALUE PARAMETER
International Journal of Information Technology and Knowledge Management January-June 2012, Volume 5, No. 1, pp. 73-77 MODIFICATION OF ADAPTIVE LOGARITHMIC METHOD FOR DISPLAYING HIGH CONTRAST SCENES BY
More informationWhite paper. Wide dynamic range. WDR solutions for forensic value. October 2017
White paper Wide dynamic range WDR solutions for forensic value October 2017 Table of contents 1. Summary 4 2. Introduction 5 3. Wide dynamic range scenes 5 4. Physical limitations of a camera s dynamic
More informationBurst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University!
Burst Photography! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 7! Gordon Wetzstein! Stanford University! Motivation! wikipedia! exposure sequence! -4 stops! Motivation!
More informationHDR imaging Automatic Exposure Time Estimation A novel approach
HDR imaging Automatic Exposure Time Estimation A novel approach Miguel A. MARTÍNEZ,1 Eva M. VALERO,1 Javier HERNÁNDEZ-ANDRÉS,1 Javier ROMERO,1 1 Color Imaging Laboratory, University of Granada, Spain.
More informationImage Enhancement for Astronomical Scenes. Jacob Lucas The Boeing Company Brandoch Calef The Boeing Company Keith Knox Air Force Research Laboratory
Image Enhancement for Astronomical Scenes Jacob Lucas The Boeing Company Brandoch Calef The Boeing Company Keith Knox Air Force Research Laboratory ABSTRACT Telescope images of astronomical objects and
More informationA Locally Tuned Nonlinear Technique for Color Image Enhancement
A Locally Tuned Nonlinear Technique for Color Image Enhancement Electrical and Computer Engineering Department Old Dominion University Norfolk, VA 3508, USA sarig00@odu.edu, vasari@odu.edu http://www.eng.odu.edu/visionlab
More informationPhotomatix Light 1.0 User Manual
Photomatix Light 1.0 User Manual Table of Contents Introduction... iii Section 1: HDR...1 1.1 Taking Photos for HDR...2 1.1.1 Setting Up Your Camera...2 1.1.2 Taking the Photos...3 Section 2: Using Photomatix
More informationDigital Radiography using High Dynamic Range Technique
Digital Radiography using High Dynamic Range Technique DAN CIURESCU 1, SORIN BARABAS 2, LIVIA SANGEORZAN 3, LIGIA NEICA 1 1 Department of Medicine, 2 Department of Materials Science, 3 Department of Computer
More informationA Novel Hybrid Exposure Fusion Using Boosting Laplacian Pyramid
A Novel Hybrid Exposure Fusion Using Boosting Laplacian Pyramid S.Abdulrahaman M.Tech (DECS) G.Pullaiah College of Engineering & Technology, Nandikotkur Road, Kurnool, A.P-518452. Abstract: THE DYNAMIC
More informationHigh Dynamic Range Images : Rendering and Image Processing Alexei Efros. The Grandma Problem
High Dynamic Range Images 15-463: Rendering and Image Processing Alexei Efros The Grandma Problem 1 Problem: Dynamic Range 1 1500 The real world is high dynamic range. 25,000 400,000 2,000,000,000 Image
More informationHigh Dynamic Range Images Using Exposure Metering
High Dynamic Range Images Using Exposure Metering 作 者 : 陳坤毅 指導教授 : 傅楸善 博士 Dynamic Range The dynamic range is a ratio between the maximum and minimum physical measures. Its definition depends on what the
More informationFigure 1 HDR image fusion example
TN-0903 Date: 10/06/09 Using image fusion to capture high-dynamic range (hdr) scenes High dynamic range (HDR) refers to the ability to distinguish details in scenes containing both very bright and relatively
More informationCapturing Light in man and machine. Some figures from Steve Seitz, Steve Palmer, Paul Debevec, and Gonzalez et al.
Capturing Light in man and machine Some figures from Steve Seitz, Steve Palmer, Paul Debevec, and Gonzalez et al. 15-463: Computational Photography Alexei Efros, CMU, Fall 2005 Image Formation Digital
More informationUSE OF HISTOGRAM EQUALIZATION IN IMAGE PROCESSING FOR IMAGE ENHANCEMENT
USE OF HISTOGRAM EQUALIZATION IN IMAGE PROCESSING FOR IMAGE ENHANCEMENT Sapana S. Bagade M.E,Computer Engineering, Sipna s C.O.E.T,Amravati, Amravati,India sapana.bagade@gmail.com Vijaya K. Shandilya Assistant
More informationImage Processing by Bilateral Filtering Method
ABHIYANTRIKI An International Journal of Engineering & Technology (A Peer Reviewed & Indexed Journal) Vol. 3, No. 4 (April, 2016) http://www.aijet.in/ eissn: 2394-627X Image Processing by Bilateral Image
More informationExtended Dynamic Range Imaging: A Spatial Down-Sampling Approach
2014 IEEE International Conference on Systems, Man, and Cybernetics October 5-8, 2014, San Diego, CA, USA Extended Dynamic Range Imaging: A Spatial Down-Sampling Approach Huei-Yung Lin and Jui-Wen Huang
More informationNovel Histogram Processing for Colour Image Enhancement
Novel Histogram Processing for Colour Image Enhancement Jiang Duan and Guoping Qiu School of Computer Science, The University of Nottingham, United Kingdom Abstract: Histogram equalization is a well-known
More informationSubjective evaluation of image color damage based on JPEG compression
2014 Fourth International Conference on Communication Systems and Network Technologies Subjective evaluation of image color damage based on JPEG compression Xiaoqiang He Information Engineering School
More informationColor Transformations
Color Transformations It is useful to think of a color image as a vector valued image, where each pixel has associated with it, as vector of three values. Each components of this vector corresponds to
More informationConvolution Pyramids. Zeev Farbman, Raanan Fattal and Dani Lischinski SIGGRAPH Asia Conference (2011) Julian Steil. Prof. Dr.
Zeev Farbman, Raanan Fattal and Dani Lischinski SIGGRAPH Asia Conference (2011) presented by: Julian Steil supervisor: Prof. Dr. Joachim Weickert Fig. 1.1: Gradient integration example Seminar - Milestones
More informationMultiscale model of Adaptation, Spatial Vision and Color Appearance
Multiscale model of Adaptation, Spatial Vision and Color Appearance Sumanta N. Pattanaik 1 Mark D. Fairchild 2 James A. Ferwerda 1 Donald P. Greenberg 1 1 Program of Computer Graphics, Cornell University,
More informationChapter 3 Part 2 Color image processing
Chapter 3 Part 2 Color image processing Motivation Color fundamentals Color models Pseudocolor image processing Full-color image processing: Component-wise Vector-based Recent and current work Spring 2002
More informationFiras Hassan and Joan Carletta The University of Akron
A Real-Time FPGA-Based Architecture for a Reinhard-Like Tone Mapping Operator Firas Hassan and Joan Carletta The University of Akron Outline of Presentation Background and goals Existing methods for local
More informationAutomatic High Dynamic Range Image Generation for Dynamic Scenes
Automatic High Dynamic Range Image Generation for Dynamic Scenes IEEE Computer Graphics and Applications Vol. 28, Issue. 2, April 2008 Katrien Jacobs, Celine Loscos, and Greg Ward Presented by Yuan Xi
More informationCSE 332/564: Visualization. Fundamentals of Color. Perception of Light Intensity. Computer Science Department Stony Brook University
Perception of Light Intensity CSE 332/564: Visualization Fundamentals of Color Klaus Mueller Computer Science Department Stony Brook University How Many Intensity Levels Do We Need? Dynamic Intensity Range
More informationWhite Paper High Dynamic Range Imaging
WPE-2015XI30-00 for Machine Vision What is Dynamic Range? Dynamic Range is the term used to describe the difference between the brightest part of a scene and the darkest part of a scene at a given moment
More informationTesting, Tuning, and Applications of Fast Physics-based Fog Removal
Testing, Tuning, and Applications of Fast Physics-based Fog Removal William Seale & Monica Thompson CS 534 Final Project Fall 2012 1 Abstract Physics-based fog removal is the method by which a standard
More informationEfficient Image Retargeting for High Dynamic Range Scenes
1 Efficient Image Retargeting for High Dynamic Range Scenes arxiv:1305.4544v1 [cs.cv] 20 May 2013 Govind Salvi, Puneet Sharma, and Shanmuganathan Raman Abstract Most of the real world scenes have a very
More informationGraphics and Image Processing Basics
EST 323 / CSE 524: CG-HCI Graphics and Image Processing Basics Klaus Mueller Computer Science Department Stony Brook University Julian Beever Optical Illusion: Sidewalk Art Julian Beever Optical Illusion:
More informationTitle: DCT-based HDR Exposure Fusion Using Multi-exposed Image Sensors. - Affiliation: School of Electronics Engineering,
Title: DCT-based HDR Exposure Fusion Using Multi-exposed Image Sensors Author: Geun-Young Lee, Sung-Hak Lee, and Hyuk-Ju Kwon - Affiliation: School of Electronics Engineering, Kyungpook National University,
More informationA New Auto Exposure System to Detect High Dynamic Range Conditions Using CMOS Technology
15 A New Auto Exposure System to Detect High Dynamic Range Conditions Using CMOS Technology Quoc Kien Vuong, SeHwan Yun and Suki Kim Korea University, Seoul Republic of Korea 1. Introduction Recently,
More informationConcealed Weapon Detection Using Color Image Fusion
Concealed Weapon Detection Using Color Image Fusion Zhiyun Xue, Rick S. Blum Electrical and Computer Engineering Department Lehigh University Bethlehem, PA, U.S.A. rblum@eecs.lehigh.edu Abstract Image
More informationFrequencies and Color
Frequencies and Color Alexei Efros, CS280, Spring 2018 Salvador Dali Gala Contemplating the Mediterranean Sea, which at 30 meters becomes the portrait of Abraham Lincoln, 1976 Spatial Frequencies and
More informationComputer Graphics Fundamentals
Computer Graphics Fundamentals Jacek Kęsik, PhD Simple converts Rotations Translations Flips Resizing Geometry Rotation n * 90 degrees other Geometry Rotation n * 90 degrees other Geometry Translations
More informationImages and Displays. Lecture Steve Marschner 1
Images and Displays Lecture 2 2008 Steve Marschner 1 Introduction Computer graphics: The study of creating, manipulating, and using visual images in the computer. What is an image? A photographic print?
More informationFuzzy Statistics Based Multi-HE for Image Enhancement with Brightness Preserving Behaviour
International Journal of Engineering and Management Research, Volume-3, Issue-3, June 2013 ISSN No.: 2250-0758 Pages: 47-51 www.ijemr.net Fuzzy Statistics Based Multi-HE for Image Enhancement with Brightness
More informationGE 113 REMOTE SENSING. Topic 7. Image Enhancement
GE 113 REMOTE SENSING Topic 7. Image Enhancement Lecturer: Engr. Jojene R. Santillan jrsantillan@carsu.edu.ph Division of Geodetic Engineering College of Engineering and Information Technology Caraga State
More informationNew applications of Spectral Edge image fusion
New applications of Spectral Edge image fusion Alex E. Hayes a,b, Roberto Montagna b, and Graham D. Finlayson a,b a Spectral Edge Ltd, Cambridge, UK. b University of East Anglia, Norwich, UK. ABSTRACT
More informationCorrection of Clipped Pixels in Color Images
Correction of Clipped Pixels in Color Images IEEE Transaction on Visualization and Computer Graphics, Vol. 17, No. 3, 2011 Di Xu, Colin Doutre, and Panos Nasiopoulos Presented by In-Yong Song School of
More informationImproving Image Quality by Camera Signal Adaptation to Lighting Conditions
Improving Image Quality by Camera Signal Adaptation to Lighting Conditions Mihai Negru and Sergiu Nedevschi Technical University of Cluj-Napoca, Computer Science Department Mihai.Negru@cs.utcluj.ro, Sergiu.Nedevschi@cs.utcluj.ro
More informationFast Bilateral Filtering for the Display of High-Dynamic-Range Images
Fast Bilateral Filtering for the Display of High-Dynamic-Range Images Frédo Durand & Julie Dorsey Laboratory for Computer Science Massachusetts Institute of Technology Contributions Contrast reduction
More informationHigh Dynamic Range image capturing by Spatial Varying Exposed Color Filter Array with specific Demosaicking Algorithm
High Dynamic ange image capturing by Spatial Varying Exposed Color Filter Array with specific Demosaicking Algorithm Cheuk-Hong CHEN, Oscar C. AU, Ngai-Man CHEUN, Chun-Hung LIU, Ka-Yue YIP Department of
More informationCAMERA BASICS. Stops of light
CAMERA BASICS Stops of light A stop of light isn t a quantifiable measurement it s a relative measurement. A stop of light is defined as a doubling or halving of any quantity of light. The word stop is
More informationA Multi-resolution Image Fusion Algorithm Based on Multi-factor Weights
A Multi-resolution Image Fusion Algorithm Based on Multi-factor Weights Zhengfang FU 1,, Hong ZHU 1 1 School of Automation and Information Engineering Xi an University of Technology, Xi an, China Department
More informationHigh Dynamic Range (HDR) Photography in Photoshop CS2
Page 1 of 7 High dynamic range (HDR) images enable photographers to record a greater range of tonal detail than a given camera could capture in a single photo. This opens up a whole new set of lighting
More informationThe Dynamic Range Problem. High Dynamic Range (HDR) Multiple Exposure Photography. Multiple Exposure Photography. Dr. Yossi Rubner.
The Dynamic Range Problem High Dynamic Range (HDR) starlight Domain of Human Vision: from ~10-6 to ~10 +8 cd/m moonlight office light daylight flashbulb 10-6 10-1 10 100 10 +4 10 +8 Dr. Yossi Rubner yossi@rubner.co.il
More information! High&Dynamic!Range!Imaging! Slides!from!Marc!Pollefeys,!Gabriel! Brostow!(and!Alyosha!Efros!and! others)!!
! High&Dynamic!Range!Imaging! Slides!from!Marc!Pollefeys,!Gabriel! Brostow!(and!Alyosha!Efros!and! others)!! Today! High!Dynamic!Range!Imaging!(LDR&>HDR)! Tone!mapping!(HDR&>LDR!display)! The!Problem!
More informationFig Color spectrum seen by passing white light through a prism.
1. Explain about color fundamentals. Color of an object is determined by the nature of the light reflected from it. When a beam of sunlight passes through a glass prism, the emerging beam of light is not
More informationContrast Image Correction Method
Contrast Image Correction Method Journal of Electronic Imaging, Vol. 19, No. 2, 2010 Raimondo Schettini, Francesca Gasparini, Silvia Corchs, Fabrizio Marini, Alessandro Capra, and Alfio Castorina Presented
More informationReference Free Image Quality Evaluation
Reference Free Image Quality Evaluation for Photos and Digital Film Restoration Majed CHAMBAH Université de Reims Champagne-Ardenne, France 1 Overview Introduction Defects affecting films and Digital film
More informationHISTOGRAMS. These notes are a basic introduction to using histograms to guide image capture and image processing.
HISTOGRAMS Roy Killen, APSEM, EFIAP, GMPSA These notes are a basic introduction to using histograms to guide image capture and image processing. What are histograms? Histograms are graphs that show what
More informationBrightness Calculation in Digital Image Processing
Brightness Calculation in Digital Image Processing Sergey Bezryadin, Pavel Bourov*, Dmitry Ilinih*; KWE Int.Inc., San Francisco, CA, USA; *UniqueIC s, Saratov, Russia Abstract Brightness is one of the
More informationSelective Detail Enhanced Fusion with Photocropping
IJIRST International Journal for Innovative Research in Science & Technology Volume 1 Issue 11 April 2015 ISSN (online): 2349-6010 Selective Detail Enhanced Fusion with Photocropping Roopa Teena Johnson
More informationLicense Plate Localisation based on Morphological Operations
License Plate Localisation based on Morphological Operations Xiaojun Zhai, Faycal Benssali and Soodamani Ramalingam School of Engineering & Technology University of Hertfordshire, UH Hatfield, UK Abstract
More informationEFFICIENT CONTRAST ENHANCEMENT USING GAMMA CORRECTION WITH MULTILEVEL THRESHOLDING AND PROBABILITY BASED ENTROPY
EFFICIENT CONTRAST ENHANCEMENT USING GAMMA CORRECTION WITH MULTILEVEL THRESHOLDING AND PROBABILITY BASED ENTROPY S.Gayathri 1, N.Mohanapriya 2, B.Kalaavathi 3 1 PG student, Computer Science and Engineering,
More informationHow to reproduce an oil painting with compelling and realistic printed colours (Part 2)
How to reproduce an oil painting with compelling and realistic printed colours (Part 2) Author: ehong Translation: William (4 th June 2007) This document (part 2 or 3) will detail step by step how a digitised
More informationDigital Image Processing
Digital Image Processing IMAGE PERCEPTION & ILLUSION Hamid R. Rabiee Fall 2015 Outline 2 What is color? Image perception Color matching Color gamut Color balancing Illusions What is Color? 3 Visual perceptual
More informationAN INFORMATION-THEORETIC APPROACH TO MULTI-EXPOSURE FUSION VIA STATISTICAL FILTERING USING LOCAL ENTROPY
AN INFORMATION-THEORETIC APPROACH TO MULTI-EXPOSURE FUSION VIA STATISTICAL FILTERING USING LOCAL ENTROPY Johannes Herwig and Josef Pauli Intelligent Systems Group University of Duisburg-Essen Duisburg,
More informationVision Review: Image Processing. Course web page:
Vision Review: Image Processing Course web page: www.cis.udel.edu/~cer/arv September 7, Announcements Homework and paper presentation guidelines are up on web page Readings for next Tuesday: Chapters 6,.,
More informationECC419 IMAGE PROCESSING
ECC419 IMAGE PROCESSING INTRODUCTION Image Processing Image processing is a subclass of signal processing concerned specifically with pictures. Digital Image Processing, process digital images by means
More informationThe Unique Role of Lucis Differential Hysteresis Processing (DHP) in Digital Image Enhancement
The Unique Role of Lucis Differential Hysteresis Processing (DHP) in Digital Image Enhancement Brian Matsumoto, Ph.D. Irene L. Hale, Ph.D. Imaging Resource Consultants and Research Biologists, University
More informationRaymond Klass Photography Newsletter
Raymond Klass Photography Newsletter The Next Step: Realistic HDR Techniques by Photographer Raymond Klass High Dynamic Range or HDR images, as they are often called, compensate for the limitations of
More informationResearch Article Anisotropic Diffusion for Details Enhancement in Multiexposure Image Fusion
Hindawi Publishing Corporation ISRN Signal Processing Volume 213, Article ID 928971, 18 pages http://dx.doi.org/1.1155/213/928971 Research Article Anisotropic Diffusion for Details Enhancement in Multiexposure
More informationA Comparison of the Multiscale Retinex With Other Image Enhancement Techniques
A Comparison of the Multiscale Retinex With Other Image Enhancement Techniques Zia-ur Rahman, Glenn A. Woodell and Daniel J. Jobson College of William & Mary, NASA Langley Research Center Abstract The
More informationLossless Image Watermarking for HDR Images Using Tone Mapping
IJCSNS International Journal of Computer Science and Network Security, VOL.13 No.5, May 2013 113 Lossless Image Watermarking for HDR Images Using Tone Mapping A.Nagurammal 1, T.Meyyappan 2 1 M. Phil Scholar
More informationImage Processing Lecture 4
Image Enhancement Image enhancement aims to process an image so that the output image is more suitable than the original. It is used to solve some computer imaging problems, or to improve image quality.
More informationVU Rendering SS Unit 8: Tone Reproduction
VU Rendering SS 2012 Unit 8: Tone Reproduction Overview 1. The Problem Image Synthesis Pipeline Different Image Types Human visual system Tone mapping Chromatic Adaptation 2. Tone Reproduction Linear methods
More informationColor and More. Color basics
Color and More In this lesson, you'll evaluate an image in terms of its overall tonal range (lightness, darkness, and contrast), its overall balance of color, and its overall appearance for areas that
More informationColor Preserving HDR Fusion for Dynamic Scenes
Color Preserving HDR Fusion for Dynamic Scenes Gökdeniz Karadağ Middle East Technical University, Turkey gokdeniz@ceng.metu.edu.tr Ahmet Oğuz Akyüz Middle East Technical University, Turkey akyuz@ceng.metu.edu.tr
More informationA Vehicle Speed Measurement System for Nighttime with Camera
Proceedings of the 2nd International Conference on Industrial Application Engineering 2014 A Vehicle Speed Measurement System for Nighttime with Camera Yuji Goda a,*, Lifeng Zhang a,#, Seiichi Serikawa
More informationAn Efficient Color Image Segmentation using Edge Detection and Thresholding Methods
19 An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods T.Arunachalam* Post Graduate Student, P.G. Dept. of Computer Science, Govt Arts College, Melur - 625 106 Email-Arunac682@gmail.com
More informationRaw Material Assignment #4. Due 5:30PM on Monday, November 30, 2009.
Raw Material Assignment #4. Due 5:30PM on Monday, November 30, 2009. Part I. Pick Your Brain! (40 points) Type your answers for the following questions in a word processor; we will accept Word Documents
More informationTitle goes Shadows and here Highlights
Shadows Title goes and Highlights here The new Shadows and Highlights command in Photoshop CS (8) is a great new tool that will allow you to adjust the shadow areas of an image while leaving the highlights
More informationAn Approach for Reconstructed Color Image Segmentation using Edge Detection and Threshold Methods
An Approach for Reconstructed Color Image Segmentation using Edge Detection and Threshold Methods Mohd. Junedul Haque, Sultan H. Aljahdali College of Computers and Information Technology Taif University
More information][ R G [ Q] Y =[ a b c. d e f. g h I
Abstract Unsupervised Thresholding and Morphological Processing for Automatic Fin-outline Extraction in DARWIN (Digital Analysis and Recognition of Whale Images on a Network) Scott Hale Eckerd College
More informationColors in Images & Video
LECTURE 8 Colors in Images & Video CS 5513 Multimedia Systems Spring 2009 Imran Ihsan Principal Design Consultant OPUSVII www.opuseven.com Faculty of Engineering & Applied Sciences 1. Light and Spectra
More informationHow to correct a contrast rejection. how to understand a histogram. Ver. 1.0 jetphoto.net
How to correct a contrast rejection or how to understand a histogram Ver. 1.0 jetphoto.net Contrast Rejection or how to understand the histogram 1. What is a histogram? A histogram is a graphical representation
More informationIMAGE PROCESSING: AREA OPERATIONS (FILTERING)
IMAGE PROCESSING: AREA OPERATIONS (FILTERING) N. C. State University CSC557 Multimedia Computing and Networking Fall 2001 Lecture # 13 IMAGE PROCESSING: AREA OPERATIONS (FILTERING) N. C. State University
More informationicam06, HDR, and Image Appearance
icam06, HDR, and Image Appearance Jiangtao Kuang, Mark D. Fairchild, Rochester Institute of Technology, Rochester, New York Abstract A new image appearance model, designated as icam06, has been developed
More informationScientific Working Group on Digital Evidence
Disclaimer: As a condition to the use of this document and the information contained therein, the SWGDE requests notification by e-mail before or contemporaneous to the introduction of this document, or
More informationVLSI Implementation of Impulse Noise Suppression in Images
VLSI Implementation of Impulse Noise Suppression in Images T. Satyanarayana 1, A. Ravi Chandra 2 1 PG Student, VRS & YRN College of Engg. & Tech.(affiliated to JNTUK), Chirala 2 Assistant Professor, Department
More informationIn order to manage and correct color photos, you need to understand a few
In This Chapter 1 Understanding Color Getting the essentials of managing color Speaking the language of color Mixing three hues into millions of colors Choosing the right color mode for your image Switching
More informationHigh dynamic range and tone mapping Advanced Graphics
High dynamic range and tone mapping Advanced Graphics Rafał Mantiuk Computer Laboratory, University of Cambridge Cornell Box: need for tone-mapping in graphics Rendering Photograph 2 Real-world scenes
More informationarxiv: v1 [cs.cv] 29 May 2018
AUTOMATIC EXPOSURE COMPENSATION FOR MULTI-EXPOSURE IMAGE FUSION Yuma Kinoshita Sayaka Shiota Hitoshi Kiya Tokyo Metropolitan University, Tokyo, Japan arxiv:1805.11211v1 [cs.cv] 29 May 2018 ABSTRACT This
More information