Improved Fusing Infrared and Electro-Optic Signals for. High Resolution Night Images

Similar documents
ABSTRACT 1. INTRODUCTION

Performance Evaluation of Edge Detection Techniques for Square Pixel and Hexagon Pixel images

Super Sampling of Digital Video 22 February ( x ) Ψ

Image Deblurring. This chapter describes how to deblur an image using the toolbox deblurring functions.

Evaluation of infrared collimators for testing thermal imaging systems

Improving Signal- to- noise Ratio in Remotely Sensed Imagery Using an Invertible Blur Technique

SUPER RESOLUTION INTRODUCTION

Defense Technical Information Center Compilation Part Notice

Denoising and Effective Contrast Enhancement for Dynamic Range Mapping

A No Reference Image Blur Detection using CPBD Metric and Deblurring of Gaussian Blurred Images using Lucy-Richardson Algorithm

Blurred Image Restoration Using Canny Edge Detection and Blind Deconvolution Algorithm

Vision Review: Image Processing. Course web page:

A Study of Slanted-Edge MTF Stability and Repeatability

Background Adaptive Band Selection in a Fixed Filter System

Imaging-Consistent Super-Resolution

CSC 320 H1S CSC320 Exam Study Guide (Last updated: April 2, 2015) Winter 2015

BASIC OPERATIONS IN IMAGE PROCESSING USING MATLAB

Image Enhancement using Histogram Equalization and Spatial Filtering

Filters. Materials from Prof. Klaus Mueller

Edge-Raggedness Evaluation Using Slanted-Edge Analysis

High resolution images obtained with uncooled microbolometer J. Sadi 1, A. Crastes 2

On Fusion Algorithm of Infrared and Radar Target Detection and Recognition of Unmanned Surface Vehicle

International Journal of Advance Engineering and Research Development CONTRAST ENHANCEMENT OF IMAGES USING IMAGE FUSION BASED ON LAPLACIAN PYRAMID

Ultra-thin Multiple-channel LWIR Imaging Systems

Image Processing for feature extraction

Concealed Weapon Detection Using Color Image Fusion

LAB MANUAL SUBJECT: IMAGE PROCESSING BE (COMPUTER) SEM VII

Investigations on Multi-Sensor Image System and Its Surveillance Applications

Extended depth-of-field in Integral Imaging by depth-dependent deconvolution

Image Quality Assessment for Defocused Blur Images

Sampling Efficiency in Digital Camera Performance Standards

CS6670: Computer Vision Noah Snavely. Administrivia. Administrivia. Reading. Last time: Convolution. Last time: Cross correlation 9/8/2009

Chapter 2 Fourier Integral Representation of an Optical Image

Measurement of Texture Loss for JPEG 2000 Compression Peter D. Burns and Don Williams* Burns Digital Imaging and *Image Science Associates

Lecture Notes 10 Image Sensor Optics. Imaging optics. Pixel optics. Microlens

Comparison of Reconstruction Algorithms for Images from Sparse-Aperture Systems

Image Smoothening and Sharpening using Frequency Domain Filtering Technique

Broadband Optical Phased-Array Beam Steering

Iris Recognition using Histogram Analysis

Comparative Study of Different Wavelet Based Interpolation Techniques

Computer Vision Slides curtesy of Professor Gregory Dudek

Interpolation of CFA Color Images with Hybrid Image Denoising

Study guide for Graduate Computer Vision

Image acquisition. Midterm Review. Digitization, line of image. Digitization, whole image. Geometric transformations. Interpolation 10/26/2016

COMBINING LAPLACIAN AND SOBEL GRADIENT FOR GREATER SHARPENING

INFRARED IMAGING-PASSIVE THERMAL COMPENSATION VIA A SIMPLE PHASE MASK

Multimodal Face Recognition using Hybrid Correlation Filters

Linear Gaussian Method to Detect Blurry Digital Images using SIFT

Edge Detection in SAR Images using Phase Stretch Transform

NOISE REMOVAL TECHNIQUES FOR MICROWAVE REMOTE SENSING RADAR DATA AND ITS EVALUATION

International Journal of Advancedd Research in Biology, Ecology, Science and Technology (IJARBEST)

Adaptive Optimum Notch Filter for Periodic Noise Reduction in Digital Images

T I P S F O R I M P R O V I N G I M A G E Q U A L I T Y O N O Z O F O O T A G E

Target detection in side-scan sonar images: expert fusion reduces false alarms

Introduction to digital image processing

An Application of the Least Squares Plane Fitting Interpolation Process to Image Reconstruction and Enhancement

Filtering and Processing IR Images of PV Modules

1.Discuss the frequency domain techniques of image enhancement in detail.

Last Lecture. photomatix.com

PRACTICAL IMAGE AND VIDEO PROCESSING USING MATLAB

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

Digital Image Processing

Image Processing Final Test

Digital Image Processing 3/e

Exposure schedule for multiplexing holograms in photopolymer films

Blind Single-Image Super Resolution Reconstruction with Defocus Blur

Optical System Case Studies for Speckle Imaging

RESEARCH PAPER FOR ARBITRARY ORIENTED TEAM TEXT DETECTION IN VIDEO IMAGES USING CONNECTED COMPONENT ANALYSIS

New Spatial Filters for Image Enhancement and Noise Removal

Optical transfer function shaping and depth of focus by using a phase only filter

Design of an Efficient Edge Enhanced Image Scalar for Image Processing Applications

Comparison of Fourier transform methods for calculating MTF Joseph D. LaVeigne a, Stephen D. Burks b, Brian Nehring a

Teaching Scheme. Credits Assigned (hrs/week) Theory Practical Tutorial Theory Oral & Tutorial Total

LPCC filters realization as binary amplitude hologram in 4-f correlator: range limitation of hologram pixels representation

Digital Image Processing

Image Segmentation of Color Image using Threshold Based Edge Detection Algorithm in MatLab

MAV-ID card processing using camera images

A Review on Image Enhancement Technique for Biomedical Images

Optimization of Existing Centroiding Algorithms for Shack Hartmann Sensor

DIGITAL IMAGE DE-NOISING FILTERS A COMPREHENSIVE STUDY

restoration-interpolation from the Thematic Mapper (size of the original

Midterm Examination CS 534: Computational Photography

Modeling the MTF and noise characteristics of complex image formation systems

A Spatial Mean and Median Filter For Noise Removal in Digital Images

MTF characteristics of a Scophony scene projector. Eric Schildwachter

Postprocessing of nonuniform MRI

Improving registration metrology by correlation methods based on alias-free image simulation

Extended depth of field for visual measurement systems with depth-invariant magnification

A Comparative Study and Analysis of Image Restoration Techniques Using Different Images Formats

Rotation/ scale invariant hybrid digital/optical correlator system for automatic target recognition

Effective Pixel Interpolation for Image Super Resolution

Impact of Out-of-focus Blur on Face Recognition Performance Based on Modular Transfer Function

Image interpretation and analysis

Transforms and Frequency Filtering

Circular averaging filter (pillbox) Approximates the two-dimensional Laplacian operator. Laplacian of Gaussian filter

Image Filtering. Median Filtering

A Pan-Sharpening Based on the Non-Subsampled Contourlet Transform and Discrete Wavelet Transform

Introduction. Chapter 16 Diagnostic Radiology. Primary radiological image. Primary radiological image

Joint transform optical correlation applied to sub-pixel image registration

Improvement of Satellite Images Resolution Based On DT-CWT

Transcription:

Improved Fusing Infrared and Electro-Optic Signals for High Resolution Night Images Xiaopeng Huang, a Ravi Netravali, b Hong Man, a and Victor Lawrence a a Dept. of Electrical and Computer Engineering, Stevens Institute of Technology, Hoboken, NJ 07030, USA; b Dept. of Electrical Engineering, Columbia University, New York, NY 10027, USA ABSTRACT Electro-optic (EO) images exhibit the properties of high resolution and low noise level, while it is a challenge to distinguish objects with infrared (IR), especially for objects with similar temperatures. In earlier work, we proposed a novel framework for IR image enhancement based on the information (e.g., edge) from EO images. Our framework superimposed the detected edges of the EO image with the corresponding transformed IR image. Obviously, this framework resulted in better resolution IR images that help distinguish objects at night. For our IR image system, we used the theoretical point spread function (PSF) proposed by Russell C. Hardie et al., which is composed of the modulation transfer function (MTF) of a uniform detector array and the incoherent optical transfer function (OTF) of diffraction-limited optics. In addition, we designed an inverse filter based on the proposed PSF to transform the IR image. In this paper, blending the detected edge of the EO image with the corresponding transformed IR image and the original IR image is the principal idea for improving the previous framework. This improved framework requires four main steps: (1) inverse filter-based IR image transformation, (2) image edge detection, (3) images registration, and (4) blending of the corresponding images. Simulation results show that blended IR images have better quality over the superimposed images that were generated under the previous framework. Based on the same steps, the simulation result shows a blended IR image of better quality when only the original IR image is available. Keywords: Theoretical PSF, IR image transformation, EO image edge detection, image blending. 1. INTRODUCTION Infrared (IR) imaging systems depend on thermal contrast between target and background to generate real-time images. These systems create images by utilizing the infrared energy emitted by the objects as a result of their temperature difference with background and emissivity. Unfortunately, the created IR images have low resolution and high noise level. Therefore, accurately distinguishing objects at night is a challenging topic that has perplexed scientists and engineers for a long time. Fusing IR and electro-optic (EO) images is an effective way to solve this challenge. It combines images from two different sources to obtain a single composite image with improved resolution, low noise, and ability to see clearly at night.. Up to now, scientists have developed many efficient image fusion algorithms, such as the expectation maximization (EM) fusing algorithm [1], the discrete wavelet transform (DWT) fusing algorithm [2], and the Laplacian pyramid fusing algorithm [3]. Additionally, much literature focuses on presenting techniques for Infrared Imaging Systems: Design, Analysis, Modeling, and Testing XXIII, edited by Gerald C. Holst, Keith A. Krapels, Proc. of SPIE Vol. 8355, 835517 2012 SPIE CCC code: 0277-786X/12/$18 doi: 10.1117/12.950905 Proc. of SPIE Vol. 8355 835517-1

estimating a high-resolution IR image via the optimized IR image system [4-5], e.g., with reduced aliasing. However, based on the existing approaches, it is still impossible to accurately distinguish the edges of different objects at night when they have a similar temperature and background. We have proposed a novel framework to solve this challenging problem in [6]. Since EO images exhibit properties of high resolution and low noise level, the principle idea of our proposed framework was to utilize the high resolution property of EO images to help us reconstruct the corresponding edges of different objects. Assuming we had a pair of EO and IR images, then four necessary steps were required to complete this framework: (1) transform the original IR image into an IR image via the established system point spread function (PSF) and designed inverse filter; (2) detect a clean edge map of the high resolution EO image; (3) register the transformed IR image and the detected clean edge map; and (4) superimpose the detected edge map of the EO image onto the transformed IR image. In that framework, we adopted a theoretical PSF for the IR image system, which consisted of the modulation transfer function (MTF) of a uniform detector array and the incoherent optical transfer function (OTF) of diffraction-limited optics. Final simulation results showed that, with the help of the superimposed edge map, we could distinguish between objects in the transformed IR image and even small parts of a single object. Moreover, the performance was independent of the object s temperature and background. Therefore, that proposed framework could be regarded as a breakthrough with regard to night-time distinguishing of objects. In this paper, we propose improvement on our previous framework. This new approach blends the edge-detected EO image with the transformed IR image and the original IR image. The improved framework requires four main steps: (1) inverse filter-based IR image transformation, (2) EO image edge detection, (3) image registration, and (4) blending of the corresponding images. Here, step 3) is not required when only an IR image is available. Simulation results will show that images generated using our improved framework have edges that are clearly visible and seamlessly matched with the corresponding objects. In particular, we compare the results generated by the use of EO and IR images to those generated by the use of only an IR image. The following questions will be addressed in this paper: (1) Are there some things we can see now that we could not see before? (2) How do the results look if we use only IR images for processing? Can we derive the best image with only IR images? (3) Are there things we can only see partially now? (4) Are there false alarms e.g., things that should not be seen that are now seen falsely? Could this include some extraneous patterns? (5) In summary, what are the improvements in picture quality using our framework? The remainder of this paper is organized as follows. Section 2 is the IR image transformation process based on the previously proposed theoretical PSF and designed inverse filter. Section 3 presents the edge detection of EO images and the related image registration process. The blended image results and discussion are provided in Section 4. Finally, some conclusions and future work are addressed in Section 5. Proc. of SPIE Vol. 8355 835517-2

2. IR IMAGE TRANSFORMATION 2.1 Theoretical PSF We adopt the Russell C. Hardie et al. proposed theoretical PSF [7] for IR image systems, which consists of the MTF of a uniform detector array and the incoherent OTF of diffraction-limited optics. The primary contributor is the finite detector size, and this effect is spatially invariant for a uniform detector array. We begin by considering an infrared system with this uniform detector array. We can model the effect of the integration of light intensity over the span of the detectors as a linear convolution operation with a PSF determined by the geometry of a single detector. The second contributor is the optics, and we assume an isoplanatic model for optics. Theoretical descriptions of this infrared imaging system can be found in [6-8]. We consider a particular IR imaging system as an example [7], the typical system considered is the forward looking infrared (FLIR) imager. Here, we adopt the same parameter settings as that in [7]. This systems has square detectors of size a=b=0.040mm, the imager is equipped with 100mm f 3 optics, the center wavelength =0.004mm and the cutoff frequency 83.3 cycles/mm are used for the OTF calculation. In Figure 1, Figure 1(a) shows the effective MTF of the detectors, D (u, v), and figure 1(b) shows the diffraction-limited OTF for the optics, H (u, v). The overall system MTF is shown in Figure 1(c), and the continuous system PSF is shown in Figure 1(d). 2.2 Inverse filtering and IR image transformation Usually direct inverse filtering is the simplest approach we can take to restoring a degraded image, which ignores the noise term in the model and forms an estimator in the form of [9] F u, v G, H, (1) where G (u, v) is the degraded image and H (u, v) is the system PSF. Then, we obtain the corresponding estimate of the image by taking the inverse Fourier transform of F u, v. In this paper, we assume the proposed theoretical PSF to be H (u, v) in the designed inverse filter. If we let the original IR image pass through the designed inverse filter, then we can obtain a transformed IR image with temperature information of the objects. Figure 2 shows two examples of the IR image transformation based on the designed inverse filter. 3. IMAGE EDGE DECTION AND REGISTRATION 3.1 Image edge detection Image edge detection refers to the process of identifying and locating sharp two-dimensional discontinuities in an image. The discontinuities are abrupt changes in pixel intensity that characterize boundaries of objects in a scene. By far, edge detection is the most common approach for detecting meaningful discontinuities in intensity values. There are many edge operators to perform edge detection, e.g., Sobel, Prewitt, Roberts, Laplacian of a Gaussian (LoG), Zero crossings and Canny. So far the Canny edge detection algorithm [10] has been known as the optimal edge detector. The detailed description of the Canny edge detector can be found in [10], and the syntax for the Canny edge detector in Matlab is [g, t] = edge (f, canny, T, sigma) (2) Proc. of SPIE Vol. 8355 835517-3

. (a) Effective MTF of the detectors in the FLIR imagers (b) Diffraction-limited OTF of the optics (c) Overall system MTF (d) Overall continuous system PSF (right side figure is the enlarged PSF) Figure 1. FLIR system parameters. Proc. of SPIE Vol. 8355 835517-4

(a) Original IR Image 1 (b) Transformed IR Image 1 via the inverse filter (c) Original IR Image 2 (public image) (d) Transformed IR Image 2 via the inverse filter Figure 2. Examples of the IR image based on the designed inverse filter. where f is the original EO image, T is a vector, T = [T1, T2], containing the two thresholds of the preceding procedure, and sigma is the standard deviation of the smoothing filter. We can change these parameters so as to produce clean edge maps. Here T1=0.04 and T2=0.09. Figure 3 shows two edge detection results of EO images via the Canny edge operator, where sigma is chosen to be the default value 1. Seen from the obtained results, we can find out that the edge-detected images are clean, and results cover all essential information in original EO images. 3.2 Image pair registration Image registration methods, which seek to align two or more images of the same scene, generally consist of the following basic steps: 1) detect features, 2) match corresponding features, 3) infer geometric transformation, and 4) use the geometric transformation to align one image with the other. In this paper, we adopt the cp2tform function from the Matlab image processing toolbox to do the manual image pair registration. The more complicated image registration Proc. of SPIE Vol. 8355 835517-5

(a) Original EO Image 1 (b) Detected edge of EO Image 1 (c) Original EO Image 2 (d) Detected edge of EO Image 2 (c) Detected edge of IR Image 2 Figure 3. Two edge detection results of EO images via the Canny edge operator, where sigma is the default value 1. Proc. of SPIE Vol. 8355 835517-6

processes (e.g., image registration involves shift, rotation and transformation), are beyond the scope of our paper. In some particular cases, we could complete a partial image registration rather than the whole image registration. We could even adopt other methods to help obtain the exactly aligned images (e.g., image warping). 4. BLENDED IMAGE RESULTS The last step of our framework is to blend the edge-detected EO image with the transformed IR image and the original IR image. On one hand, in order to make the blended edge seamlessly match with its corresponding object, we let the edge-detected image pass through a small size low pass filter (e.g., 2 by 2). On the other hand, in order to make sure the superimposed edge is not prominent in the blended image, we set the pixel value of the detected edge to be 0.4 rather than its default value 1. In this paper, we adopt the alpha blending process [11] to achieve the fourth step. Two images before blending are read in the variable a and b, respectively. Then, the two images are blended and stored in the variable c using the formula c 1 α a α b (3) During this process, we first blend the transformed IR image with the original IR image, and then we blend the obtained result with the edge-detected image. Here, we fix the blending fraction s values to be 0.8 and 0.01, respectively. Figure 4 shows two blended images that are generated with both EO and IR images. Based on the same steps, Figure 5 shows a blended image that is generated when only original IR image is available. Seen from the obtained results, we can clearly see the seamless edge in the blended images and accurately distinguish objects at night via the superimposed edge, especially for objects with a similar temperature, and tiny parts of any object, which are extremely difficult or even impossible based on traditional object distinguishing approaches. Now, we are able to answer the questions in Section 1 as follows: 1) We can clearly see the edge of each object, which will help distinguish objects at night. We also can see the edge of a tiny part of any object, which seems to be impossible using the previous methods. Moreover, we can see the edge of any color gas (e.g., smoke) or liquid and, in some particular cases, we can even see the edge of any colorless liquid or gas. Blending the transformed IR image (temperature information) with the original IR image makes the final IR image look sharper than that of the original IR image only. 2) Generally speaking, if only IR images are available, we can use the same parameter settings to complete the proposed three-step framework (without registration) and obtain similar blended results as those of available EO images. However, the detected edge of an object, such as the edge of its shadow area rather than the real object, is not accurate, and it is impossible to detect the shape of smoke (gas) only based on the original IR image only (see results of OCTEC image). 3) There is nothing we can see partially if both EO and IR images are available. However, if only IR images are available, we can only see partially gas/smoke e.g., we can only see a light point from blended OCTEC IR image. 4) If we have both EO and IR images, there are no false alarms. However, if only IR images are available, sometimes there are false alarms e.g., we detected the edge of its shadow area rather than the house in Proc. of SPIE Vol. 8355 835517-7

the OCTEC IR image, and we could only see the light point of a fire but could not see the edge of its smoke. We think the shadow area belongs to one kind of the extraneous pattern. 5) Comparing the results of both EO and IR images with that of only the IR image, our proposed framework can help accurately distinguish any object in the IR image. In addition, the improved framework can make the superimposed edge seamlessly match with the original objects, and the reduced edge s value can ensure that the superimposed edge is not prominent in the blended image. 5. CONCLUSIONS AND FUTURE WORK It is a challenge to accurately distinguish objects with the IR image, especially objects with a similar temperature or tiny parts of any object. Therefore, in contrast to the traditional approaches used in image fusing, we blended the edgedetected EO image with the corresponding transformed IR image and the original IR image. This was achieved by the alpha blending process. Simulation results showed that we could clearly distinguish objects, and even small parts of a single object, regardless of their temperature. Generally speaking, if only IR images are available, our framework could achieve similar performances to those of available EO images, except for some special objects or objects that have some extraneous patterns. On the contrary, we will consider how to utilize merits of IR images to improve EO images in the future, e.g., include objects temperature information in EO images. REFERENCES [1] G. Liu, S. Sun and Z. Jing, Image fusion based on an expectation maximization algorithm, Opt. Eng., 44:7, July 2005. [2] Y. Zheng, E. A. Essock, and B. C. Hansen, Advanced discrete wavelet transform fusing algorithm and its optimization by using the metric of image quality index, Opt. Eng., 44:3, Mar. 2004. [3] Y. Zheng, Multi-scale fusion algorithm comparisons: Pyramid, DWT and iterative DWT, 12th International Conference on fusion, 2009, pp. 1060 1067. [4] J. C. Gillette, T. M. Stadtmiller, and R. C. Hardie, Reduction of aliasing in staring infrared imagers utilizing subpixel techniques, Opt. Eng., 34:11, Nov. 1995, pp. 3130 3137. [5] E. A. Watson, R. A. Muse, and F. P. Blommel, Aliasing and blurring in microscanned imagery, in Proceedings of SPIE Infrared Imaging Systems: Design, Analysis, Modeling and Testing III, vol. 1689, Apr. 1992, pp.242 250. [6] X. Huang, R. Netravali, H. Man and V. Lawrence, Fusing electro-optic and infrared signals for high resolution night images, in Proceedings of SPIE Image Processing: Algorithms and Systems X, vol. 8295, Feb. 2012. [7] R. C. Hardie, K. J. Barnard, J. G. Bognar, E. E. Armstrong and E. A. Watson, High-resolution image reconstruction from a sequence of rotated and translated frames and its application to an infrared imaging system, Opt. Eng., 37:1, 1998. [8] M. S. Alam, J. G. Bognar, R. C. Hardie, and B. J. Yasuda, Infrared image registration and high-resolution reconstruction using multiple translationally shifted aliased video frames, IEEE Trans. on Instrumentation and Measurement, 49:.5, Oct. 2000, pp.915 923. Proc. of SPIE Vol. 8355 835517-8

[9] R. C. Gonzalez, R. E. Woods, and S. L. Eddins, Digital Image Processing Using Matlab, 2 nd ed., Gatesmark Publishing, 2009. [10] R. Maini and H. Aggarwal, Study and comparison of various image edge detection techniques, International Journal of Image Processing, 3:1, 2009, pp. 1 12. [11] S. Jayaraman, S. Esakkirajan and T. Veerakumar, Digital Image Processing, McGraw Hill, 2009. (a) Blended result of IR Image 1 (b) Blended result of IR Image 2 Figure 4. Blended images generated with both EO and IR images. Proc. of SPIE Vol. 8355 835517-9

Figure 5. Blended result of IR image 2 when only original IR image is available. Proc. of SPIE Vol. 8355 835517-10