An Analysis of Image Denoising and Restoration of Handwritten Degraded Document Images

Similar documents
Binarization of Historical Document Images Using the Local Maximum and Minimum

Remove Noise and Reduce Blurry Effect From Degraded Document Images Using MATLAB Algorithm

Recovery of badly degraded Document images using Binarization Technique

Robust Document Image Binarization Techniques

Robust Document Image Binarization Technique for Degraded Document Images

A Robust Document Image Binarization Technique for Degraded Document Images

IJSRD - International Journal for Scientific Research & Development Vol. 4, Issue 05, 2016 ISSN (online):

PHASE PRESERVING DENOISING AND BINARIZATION OF ANCIENT DOCUMENT IMAGE

Image binarization techniques for degraded document images: A review

Efficient Document Image Binarization for Degraded Document Images using MDBUTMF and BiTA

[More* et al., 5(8): August, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116

An Analysis of Binarization Ground Truthing

Effect of Ground Truth on Image Binarization

Document Recovery from Degraded Images

BINARIZATION TECHNIQUE USED FOR RECOVERING DEGRADED DOCUMENT IMAGES

An Improved Binarization Method for Degraded Document Seema Pardhi 1, Dr. G. U. Kharat 2

Contrast adaptive binarization of low quality document images

Image Restoration and De-Blurring Using Various Algorithms Navdeep Kaur

Er. Varun Kumar 1, Ms.Navdeep Kaur 2, Er.Vikas 3. IJRASET 2015: All Rights are Reserved

MAJORITY VOTING IMAGE BINARIZATION

ICFHR2014 Competition on Handwritten Document Image Binarization (H-DIBCO 2014)

Colored Rubber Stamp Removal from Document Images

Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition

Quantitative Analysis of Local Adaptive Thresholding Techniques

An Improved Bernsen Algorithm Approaches For License Plate Recognition

Restoration of Degraded Historical Document Image 1


Study and Analysis of various preprocessing approaches to enhance Offline Handwritten Gujarati Numerals for feature extraction

Method for Real Time Text Extraction of Digital Manga Comic

Implementation of Block based Mean and Median Filter for Removal of Salt and Pepper Noise

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

Chapter 6. [6]Preprocessing

Automatic Enhancement and Binarization of Degraded Document Images

Restoration of Motion Blurred Document Images

Processing and Enhancement of Palm Vein Image in Vein Pattern Recognition System

` Jurnal Teknologi IDENTIFICATION OF MOST SUITABLE BINARISATION METHODS FOR ACEHNESE ANCIENT MANUSCRIPTS RESTORATION SOFTWARE USER GUIDE.

International Conference on Computer, Communication, Control and Information Technology (C 3 IT 2009) Paper Code: DSIP-024

Keywords: Image segmentation, pixels, threshold, histograms, MATLAB

VLSI Implementation of Impulse Noise Suppression in Images

Extraction of Newspaper Headlines from Microfilm for Automatic Indexing

RESEARCH PAPER FOR ARBITRARY ORIENTED TEAM TEXT DETECTION IN VIDEO IMAGES USING CONNECTED COMPONENT ANALYSIS

Enhanced Binarization Technique And Recognising Characters From Historical Degraded Documents

Fast Inverse Halftoning

Keywords Fuzzy Logic, ANN, Histogram Equalization, Spatial Averaging, High Boost filtering, MSE, RMSE, SNR, PSNR.

A Study on Image Enhancement and Resolution through fused approach of Guided Filter and high-resolution Filter

A new seal verification for Chinese color seal

Contrast Enhancement Techniques using Histogram Equalization: A Survey

Image Segmentation of Historical Handwriting from Palm Leaf Manuscripts

A Proficient Roi Segmentation with Denoising and Resolution Enhancement

International Journal of Computer Science Trends and Technology (IJCST) Volume 4 Issue 2, Mar - Apr 2016

Improving the Quality of Degraded Document Images

Removal of High Density Salt and Pepper Noise through Modified Decision based Un Symmetric Trimmed Median Filter

Color Image Segmentation Using K-Means Clustering and Otsu s Adaptive Thresholding

Extraction and Recognition of Text From Digital English Comic Image Using Median Filter

Paper Sobel Operated Edge Detection Scheme using Image Processing for Detection of Metal Cracks

Guided Image Filtering for Image Enhancement

Neighborhood Window Pixeling for Document Image Enhancement

IMPROVEMENT USING WEIGHTED METHOD FOR HISTOGRAM EQUALIZATION IN PRESERVING THE COLOR QUALITIES OF RGB IMAGE

MAV-ID card processing using camera images

A Novel Method for Enhancing Satellite & Land Survey Images Using Color Filter Array Interpolation Technique (CFA)

Advanced Maximal Similarity Based Region Merging By User Interactions

Fig 1 Complete Process of Image Binarization Through OCR 2016, IJARCSSE All Rights Reserved Page 213

A Review of Optical Character Recognition System for Recognition of Printed Text

Performance Evaluation of Edge Detection Techniques for Square Pixel and Hexagon Pixel images

Automatic Segmentation of Fiber Cross Sections by Dual Thresholding

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

Computing for Engineers in Python

A Study On Preprocessing A Mammogram Image Using Adaptive Median Filter

A Comparative Analysis of Different Edge Based Algorithms for Mobile/Camera Captured Images

Automatic Detection Of Optic Disc From Retinal Images. S.Sherly Renat et al.,

Classification of Road Images for Lane Detection

IEEE Signal Processing Letters: SPL Distance-Reciprocal Distortion Measure for Binary Document Images

Contrast Enhancement for Fog Degraded Video Sequences Using BPDFHE

EFFICIENT CONTRAST ENHANCEMENT USING GAMMA CORRECTION WITH MULTILEVEL THRESHOLDING AND PROBABILITY BASED ENTROPY

Historical Document Preservation using Image Processing Technique

HISTOGRAM BASED AUTOMATIC IMAGE SEGMENTATION USING WAVELETS FOR IMAGE ANALYSIS

REALIZATION OF VLSI ARCHITECTURE FOR DECISION TREE BASED DENOISING METHOD IN IMAGES

Multispectral Image Restoration of Historical Document Images

A Survey Based on Region Based Segmentation

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

Binarization of Color Document Images via Luminance and Saturation Color Features

A New Character Segmentation Approach for Off-Line Cursive Handwritten Words

Example Based Colorization Using Optimization

COLOR IMAGE SEGMENTATION USING K-MEANS CLASSIFICATION ON RGB HISTOGRAM SADIA BASAR, AWAIS ADNAN, NAILA HABIB KHAN, SHAHAB HAIDER

An Adaptive Kernel-Growing Median Filter for High Noise Images. Jacob Laurel. Birmingham, AL, USA. Birmingham, AL, USA

Study of Various Image Enhancement Techniques-A Review

Multilevel Rendering of Document Images

IJRASET 2015: All Rights are Reserved

Automatic Morphological Segmentation and Region Growing Method of Diagnosing Medical Images

International Journal of Computer Science and Mobile Computing

Direction based Fuzzy filtering for Color Image Denoising

International Journal of Advanced Research in Computer Science and Software Engineering

Automatics Vehicle License Plate Recognition using MATLAB

Global and Local Quality Measures for NIR Iris Video

Keyword: Morphological operation, template matching, license plate localization, character recognition.

Segmentation using Saturation Thresholding and its Application in Content-Based Retrieval of Images

Vehicle Number Plate Recognition with Bilinear Interpolation and Plotting Horizontal and Vertical Edge Processing Histogram with Sound Signals

Implementation of License Plate Recognition System in ARM Cortex A8 Board

COMPARITIVE STUDY OF IMAGE DENOISING ALGORITHMS IN MEDICAL AND SATELLITE IMAGES

][ R G [ Q] Y =[ a b c. d e f. g h I

Transcription:

Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 12, December 2014, pg.590 594 REVIEW ARTICLE ISSN 2320 088X An Analysis of Image Denoising and Restoration of Handwritten Degraded Document Images Sarika Jain 1, Pankaj Parihar 2 ¹Department of computer science and Engineering, Institute of Technology and Management, Bhilwara, India ²Department of computer science and Engineering, Institute of Technology and Management, Bhilwara, India 1 sarikajain03@gmail.com; 2 pankajsinghparihar2002@gmail.com Abstract The restoration of a blurry or noisy image is commonly performed with a MAP estimator, which maximizes a posterior probability to reconstruct a clean image from a degraded image. A MAP estimator, when used with a sparse gradient image prior, reconstructs piecewise smooth images and typically removes textures that are important for visual realism. The three public datasets that were used in the recent. Document Image Binarization Contest (DIBCO) 2009 & 2011 and Handwritten Document Image Binarization Contest (H-DIBCO) 2010 and achieves different accuracies. Experiments on the Bickley diary dataset that consists of several challenging bad quality document images also show the superior performance in image binarization technique which is compared with different techniques. The general objective is to identify current advantages in document image binarization using established evaluation performance measures. Keywords Image Processing, Pixel Classification, Degraded Document, Image Binarization, Adaptive Image Contrast I. INTRODUCTION Degradations in document images result from poor quality of paper, the printing process, ink blot and fading, document aging, extraneous marks, noise from scanning, etc. The goal of document restoration is to remove some of these artifacts and recover an image that is close to what one would obtain under ideal printing and imaging conditions. The ability to restore a degraded document image to its ideal condition would be highly useful in a variety of fields such as document recognition, search and retrieval, historic document analysis, law enforcement, etc. The emergence of large collections of scanned books in digital libraries [1, 10] has introduced an imminent need for such restorations that will aid their recognition or ability to search. Images with certain known noise models can be restored using traditional image restoration techniques such as Median filtering, Weiner filtering, etc. [9]. However, in practice, degradations arising from phenomena such as document aging or ink bleeding cannot be described using popular image noise models. Document processing algorithms improve upon the generic methods by incorporating document specific degradation models [20] and text specific content models [2, 16]. Document image binarization is an important step in the document image analysis. It s aims to segment the foreground text from the document background. A fast and accurate document image binarization technique is important for the ensuing document image processing tasks such as optical character recognition (OCR). As illustrated in Figure 1, the handwritten text within the degraded documents often shows a certain amount of variation in terms of the stroke width, stroke brightness, stroke connection, and document background. In addition, historical documents are often degraded by the bleed-through as illustrated in Figure 1(a) and (c) where the ink of the other side seeps through to the front. In addition, historical documents 2014, IJCSMC All Rights Reserved 590

are often degraded by different types of imaging artifacts as illustrated in Figure 1(e). These different types of document degradations tend to induce the document thresholding error and make degraded document image binarization a big challenge to most state-of-the-art techniques. The recent Document Image Binarization Contest (DIBCO) [1], [2] held under the framework of the International Conference on Document Analysis and Recognition (ICDAR) 2009 & 2011 and the Handwritten Document Image Binarization Contest(H- DIBCO)[3]held under the framework of the International Conference on Frontiers in Handwritten Recognition show recent efforts on this issue. (a) (b) (c) (d) 2014, IJCSMC All Rights Reserved 591

(e) Fig.1 Five degraded document image examples taken from DIBCO, H-DIBCO and Bickley diary datasets. II. RELATED WORK A number of thresholding techniques [3], [6], [8], [11] have been reported for document image binarization. As many degraded documents do not have a clear bimodal pattern, global thresholding [10], [11], [12], [13] is usually not a suitable approach for the degraded document binarization. Adaptive thresholding [11],[12], which estimates a local threshold for each document image pixel, is often a better approach to deal with different variations within degraded document images. For example, the early window-based adaptive thresholding techniques [8], [9] estimate the local threshold by using the mean and the standard variation of image pixels within a local neighborhood window. The main drawback of these window-based thresholding techniques is that the thresholding performance depends heavily on the window size and hence the character stroke width. Other approaches have also been reported, including background subtraction [4] texture analysis, recursive method decomposition method and combination of binarization techniques.these methods combine different types of image information and domain knowledge and are often complex. The local image contrast and the local image gradient are very useful features for segmenting the text from the document background because the document text usually has certain image contrast to the neighboring document background. They are very effective and have been used in many document image binarization techniques [5], [9],[15], [16]. In Bernsen s paper [14], the local contrast is defined as follows: C(i, j) = Imax(i, j) Imin(i, j)...(1) where C(i, j) denotes the contrast of an image pixel (i, j), Imax(i, j) and Imin(i, j) denote the maximum and minimum intensities within a local neighborhood windows of (i, j), respectively. If the local contrast C(i, j) is smaller than a threshold, the pixel is set as background directly. Otherwise it will be classified into text or background by comparing with the mean of Imax(i, j) and Imin(i, j). Bernsen s method is simple, but cannot work properly on degraded document images with a complex document background. In a novel document image binarization method [5] by using the local image contrast that is evaluated as follows [14] C(i, j)=[imax(i, j) Imin(i, j)]/[imax(i, j)+imin(i, j)+e].(2) where e is a positive but infinitely small number that is added in case the local maximum is equal to 0. Compared with Bernsen s contrast in Equation 1, the local image contrast in Equation 2 introduces a normalization factor (the denominator) to compensate the image variation within the document background. III. METHODS This section describes the document image binarization techniques. Given a degraded document image, an adaptive contrast map is first constructed and the text stroke edges are then detected through the combination of the binarized adaptive contrast map and the canny edge map. The text is then segmented based on the local threshold that is estimated from the detected text stroke edge pixels. A. Image Contrast Construction The image gradient has been widely used for edge detection [2] and it can be used to detect the text stroke edges of the document images effectively that have a uniform document background. On the other hand, it often detects many non-stroke edges from the background of degraded document that often contains certain image 2014, IJCSMC All Rights Reserved 592

variations due to noise, uneven lighting, bleed-through, etc. To extract only the stroke edges properly, the image gradient needs to be normalized to compensate the image variation within the document background. In our earlier method [5], The local contrast evaluated by the local image maximum and minimum is used to suppress the background variation as described in Equation 2. B. Pixel Edge Detection The purpose of the contrast image construction is to detect the stroke edge pixels of the document text properly. The constructed contrast image has a clear bi-modal pattern [5], where the adaptive image contrast computed at text stroke edges is obviously larger than that computed within the document background. We therefore detect the text stroke edge pixel candidate by using Otsu s global thresholding method. The binary map can be further improved through the combination with the edges by Canny s edge detector, because Canny s edge detector has a good localization property that it can mark the edges close to real edge locations in the detecting image. C. Threshold Estimation The text can then be extracted from the document background pixels once the high contrast stroke edge pixels are detected properly. Two characteristics can be observed from different kinds of document images [5]: First, the text pixels are close to the detected text stroke edge pixels. Second, there is a distinct intensity difference between the high contrast stroke edge pixels and the surrounding background pixels. IV. EVALUATION MEASURES For the evaluation, the measures used comprise an ensemble of measures that have been widely used for evaluation purposes. These measures consist of: Definitions: (i) F-Measure (ii) Negative Rate Metric and (iii) Misclassification Penalty Metric (i) F-Measure: (1) Where Recall= Precision = (2) TP, FP, FN denote the True positive, False positive and False Negative values, respectively. (ii) Negative Rate Metric (NRM): The negative rate metric NRM is based on the pixelwise mismatches between the GT and prediction. It combines the false negative rate NRFN and the false positive rate NRFP. It is denoted as follows: NRM = Where NR FN = NR F P = (3) 2014, IJCSMC All Rights Reserved 593

N TP denotes the number of true positives, N FP denotes the number of false positives, N TN denotes the number of true negatives, N FN denotes the number of false negatives. In contrast to F-Measure and PSNR, the binarization quality is better for lower NRM. (iii) Misclassification penalty metric (MPM): The Misclassification penalty metric MPM evaluates the prediction against the Ground Truth (GT) on object -by-object basis. Misclassification pixels are penalized by their distance from the ground truth object s border. MPM= (4) Where MP FN =, MP FP = d i FN and d j FP denote the distance of the i th false negative and the j th false positive pixel from the contour of the GT segmentation. The normalization factor D is the sum over all the pixel to-contour distances of the GT object. A low MPM score denotes that the algorithm is good at identifying an object s boundary. V. CONCLUSION AND FUTURE SCOPE This paper concludes the different methods that have been tested on various datasets. The DIBCO 2009 Document Image Binarization Contest attracted 35 research groups that are currently active in document image analysis. The increased interest in this competition is a two-fold proof: first, it shows the importance of binarization as a step towards effective document image recognition and second, the need for pursuing a benchmark that will lead to a meaningful and objective evaluation. REFERENCES [1] Digital library of India. http://dli.iiit.ac.in/. [2] E. Borenstein and S. Ullman. Combined top-down/bottomup segmentation. IEEE Trans. Pattern Anal. Mach. Intell., 30(12):2109 2125, 2008. [3] H. Cao and V. Govindaraju. Handwritten carbon form preprocessing based on markov random field. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2007. [4] K. Donaldson and G. K.Myers. Bayesian super-resolution of text in video with a text-specific bimodal prior. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2005. [5] P. F. Felzenszwalb and D. P. Huttenlocher. Efficient belief propagation for early vision. International Journal of Computer Vision, 70(1):41 54, 2006. [6] W. T. Freeman, T. R. Jones, and E. C. Pasztor. Examplebased super-resolution. IEEE Comput. Graph. Appl., 22(2):56 65, 2002. [7] M. Sezgin and B. Sankur, Survey over image thresholding techniques and quantitative performance evaluation, Journal of Electronic Imaging, vol. 13, no. 1, pp. 146 165, 2004. [8] O. D. Trier and A. K. Jain, Goal-directed evaluation of binarization methods, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 17, no. 12, pp. 1191 1201, 1995. [9] O. D. Trier and T. Taxt, Evaluation of binarization methods for document images, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 17, no. 3, pp. 312 315, 1995. [10] A. Brink, Thresholding of digital images using two-dimensional entropies, Pattern Recognition, vol. 25, no. 8, pp. 803 808, 1992. [11] J. Kittler and J. Illingworth, On threshold selection using clustering criteria, IEEE transactions on Systems, Man, and Cybernetics,vol. 15, pp. 652 655, 1985. [12] N. Otsu, A threshold selection method from gray level histogram, IEEE Transactions on System, Man, Cybernetics, vol. 19, no. 1, pp. 62 66, January 1978. [13] N. Papamarkos and B. Gatos, A new approach for multithreshold selection, Computer Vision Graphics and Image Processing, vol. 56, no. 5, pp. 357 370, 1994. [14] H. Lu, A. Kot, and Y. Shi. Distance-reciprocal distortion measure for binary document images. IEEE Signal Processing Letters, 11(2):228 231, 2004. [15] H. Q. Luong and W. Philips. Robust reconstruction of lowresolution document images by exploiting repetitive character behaviour. International Journal of Document Analysis and Recognition, 11(1):39 51, 2008. [16] G. Myers and K. Donaldson. Bayesian super-resolution of text in video with a text-specific bimodal prior. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 1188 1195, 2005. 2014, IJCSMC All Rights Reserved 594