Efficient Document Image Binarization for Degraded Document Images using MDBUTMF and BiTA

Similar documents
Robust Document Image Binarization Technique for Degraded Document Images

Recovery of badly degraded Document images using Binarization Technique

A Robust Document Image Binarization Technique for Degraded Document Images

Robust Document Image Binarization Techniques

An Analysis of Image Denoising and Restoration of Handwritten Degraded Document Images

IJSRD - International Journal for Scientific Research & Development Vol. 4, Issue 05, 2016 ISSN (online):

Image binarization techniques for degraded document images: A review

PHASE PRESERVING DENOISING AND BINARIZATION OF ANCIENT DOCUMENT IMAGE

Binarization of Historical Document Images Using the Local Maximum and Minimum

[More* et al., 5(8): August, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116

An Improved Binarization Method for Degraded Document Seema Pardhi 1, Dr. G. U. Kharat 2

Contrast adaptive binarization of low quality document images

Document Recovery from Degraded Images

Er. Varun Kumar 1, Ms.Navdeep Kaur 2, Er.Vikas 3. IJRASET 2015: All Rights are Reserved

Removal of High Density Salt and Pepper Noise through Modified Decision based Un Symmetric Trimmed Median Filter


Quantitative Analysis of Local Adaptive Thresholding Techniques

An Improved Bernsen Algorithm Approaches For License Plate Recognition

Document Image Binarization Technique For Enhancement of Degraded Historical Document Images

BINARIZATION TECHNIQUE USED FOR RECOVERING DEGRADED DOCUMENT IMAGES

Chapter 6. [6]Preprocessing

An Efficient Noise Removing Technique Using Mdbut Filter in Images

A Novel Method for Enhancing Satellite & Land Survey Images Using Color Filter Array Interpolation Technique (CFA)

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition

An Analysis of Binarization Ground Truthing

Effect of Ground Truth on Image Binarization

A Fast Median Filter Using Decision Based Switching Filter & DCT Compression

Extraction of Newspaper Headlines from Microfilm for Automatic Indexing

Study and Analysis of various preprocessing approaches to enhance Offline Handwritten Gujarati Numerals for feature extraction

International Conference on Computer, Communication, Control and Information Technology (C 3 IT 2009) Paper Code: DSIP-024

A Review on Image Enhancement Technique for Biomedical Images

Neighborhood Window Pixeling for Document Image Enhancement

Hardware implementation of Modified Decision Based Unsymmetric Trimmed Median Filter (MDBUTMF)

OTSU Guided Adaptive Binarization of CAPTCHA Image using Gamma Correction

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

MAJORITY VOTING IMAGE BINARIZATION

EFFICIENT CONTRAST ENHANCEMENT USING GAMMA CORRECTION WITH MULTILEVEL THRESHOLDING AND PROBABILITY BASED ENTROPY

Keyword: Morphological operation, template matching, license plate localization, character recognition.

Noise Removal and Binarization of Scanned Document Images Using Clustering of Features

` Jurnal Teknologi IDENTIFICATION OF MOST SUITABLE BINARISATION METHODS FOR ACEHNESE ANCIENT MANUSCRIPTS RESTORATION SOFTWARE USER GUIDE.

Automatic Licenses Plate Recognition System

Restoration of Degraded Historical Document Image 1

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi

Historical Document Preservation using Image Processing Technique

An Efficient DTBDM in VLSI for the Removal of Salt-and-Pepper Noise in Images Using Median filter

Decision Based Median Filter Algorithm Using Resource Optimized FPGA to Extract Impulse Noise

I. INTRODUCTION II. EXISTING AND PROPOSED WORK

Removal of High Density Salt and Pepper Noise along with Edge Preservation Technique

FPGA IMPLEMENTATION OF RSEPD TECHNIQUE BASED IMPULSE NOISE REMOVAL

High density impulse denoising by a fuzzy filter Techniques:Survey

Published by: PIONEER RESEARCH & DEVELOPMENT GROUP ( 1

Colored Rubber Stamp Removal from Document Images

An Effective Method for Removing Scratches and Restoring Low -Quality QR Code Images

Improving the Quality of Degraded Document Images

Differentiation of Malignant and Benign Masses on Mammograms Using Radial Local Ternary Pattern

COMPARITIVE STUDY OF IMAGE DENOISING ALGORITHMS IN MEDICAL AND SATELLITE IMAGES

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

A new seal verification for Chinese color seal

Image Restoration and De-Blurring Using Various Algorithms Navdeep Kaur

A Proficient Roi Segmentation with Denoising and Resolution Enhancement

A Modified Non Linear Median Filter for the Removal of Medium Density Random Valued Impulse Noise

CoE4TN4 Image Processing. Chapter 3: Intensity Transformation and Spatial Filtering

NON UNIFORM BACKGROUND REMOVAL FOR PARTICLE ANALYSIS BASED ON MORPHOLOGICAL STRUCTURING ELEMENT:

MAV-ID card processing using camera images

Method for Real Time Text Extraction of Digital Manga Comic

An Enhancement of Images Using Recursive Adaptive Gamma Correction

Design of Hybrid Filter for Denoising Images Using Fuzzy Network and Edge Detecting

Improvement in image enhancement using recursive adaptive Gamma correction

Fast pseudo-semantic segmentation for joint region-based hierarchical and multiresolution representation

Image Enhancement using Histogram Equalization and Spatial Filtering

Extraction and Recognition of Text From Digital English Comic Image Using Median Filter

Enhanced Binarization Technique And Recognising Characters From Historical Degraded Documents

Bi-Level Weighted Histogram Equalization with Adaptive Gamma Correction

A Study On Preprocessing A Mammogram Image Using Adaptive Median Filter

Open Access An Improved Character Recognition Algorithm for License Plate Based on BP Neural Network

A Novel Approach to Image Enhancement Based on Fuzzy Logic

Remove Noise and Reduce Blurry Effect From Degraded Document Images Using MATLAB Algorithm

LAB MANUAL SUBJECT: IMAGE PROCESSING BE (COMPUTER) SEM VII

A.P in Bhai Maha Singh College of Engineering, Shri Muktsar Sahib

Chapter 17. Shape-Based Operations

A Comparative Analysis of Different Edge Based Algorithms for Mobile/Camera Captured Images

Impulse Noise Removal and Detail-Preservation in Images and Videos Using Improved Non-Linear Filters 1

ICFHR2014 Competition on Handwritten Document Image Binarization (H-DIBCO 2014)

DIGITAL IMAGE PROCESSING Quiz exercises preparation for the midterm exam

Study of Various Image Enhancement Techniques-A Review

Multilevel Rendering of Document Images

Automatic Counterfeit Protection System Code Classification

Moving Object Detection for Intelligent Visual Surveillance

Contrast Enhancement Using Bi-Histogram Equalization With Brightness Preservation

An Efficient Nonlinear Filter for Removal of Impulse Noise in Color Video Sequences

C. Efficient Removal Of Impulse Noise In [7], a method used to remove the impulse noise (ERIN) is based on simple fuzzy impulse detection technique.

IDENTIFICATION OF FISSION GAS VOIDS. Ryan Collette

PERFORMANCE ANALYSIS OF LINEAR AND NON LINEAR FILTERS FOR IMAGE DE NOISING

Intelligent Nighttime Video Surveillance Using Multi-Intensity Infrared Illuminator

Detection of Defects in Glass Using Edge Detection with Adaptive Histogram Equalization

Automatic Enhancement and Binarization of Degraded Document Images

A Review of Optical Character Recognition System for Recognition of Printed Text

Virtual Restoration of old photographic prints. Prof. Filippo Stanco

Multispectral Image Restoration of Historical Document Images

Transcription:

RESEARCH ARTICLE OPEN ACCESS Efficient Document Image Binarization for Degraded Document Images using MDBUTMF and BiTA Leena.L.R, Gayathri. S2 1 Leena. L.R,Author is currently pursuing M.Tech (Information Technology) in Vins Christian College of Engineering, e-mail:leenalr699@gmail.com. 2 Gayathri.S, M.E., Asst.Prof.,Deparment of Information Technology, Vins Christian college of Engineering. Abstract: It is a challenging task for segmenting the foreground text from the background of the document. Salt and pepper noise can corrupt the images in which the corrupted pixels will take either maximum or minimum grey level. In the proposed technique, Modified Decision Based Unsymmetric Trimmed Median Filter algorithm processes the corrupted images by detecting the impulse noise. Salt and pepper noise is removed by the MDBUTMF algorithm. Global contrast enhancement is required to reveal hidden details in dark and bright regions. In addition to enhancing regions with extremely high or low luminance, proposed BiTA also significantly stretches the contrast in mid tone regions. The adaptive contrast map for degraded document is then constructed and binarized. Adaptive image contrast is the combination of both local image contrast and gradient. It is then combined with Canny s edge detection method to identify the text stroke edge pixels. The local threshold is estimated from the output of Canny s edge detection method. Based on those intensities, the text is segmented from the background of the degraded document accurately. The proposed method is simple, efficient and has minimum parameter tuning. Keywords Adaptive image contrast, MDBUTMF, BiTA, degraded document image binarization, pixel classification. I. INTRODUCTION Segmentation of the foreground text from the background of the document is very difficult due to the inter intra variations between the foreground and background in the historical degraded documents. The thresholding of the degraded document image is an unsolved problem. The handwritten documents show variations in stroke width, stroke brightness and document background. Historical documents are often affected by ink bleed issues [1]. These problems will induce errors during the thresholding and makes segmentation of the foreground text from the back ground a quiet difficult task. This method presents the binarization method that extends the previous maxima minima method. The proposed method is simple, efficient and includes minimum parameter tuning. It makes use of local image contrast and local image gradient. In local maximum minimum method the image contrast alone is used to construct the contrast map. In r the proposed system both local image contrast and local image gradient is used to construct the adaptive contrast map. Compared with the image gradient local image contrast is more capable of detecting the high contrast image pixels. Document image binarization divides the document image in to two classes. They are (1) foreground text (2) document background. It is usually performed on the preprocessing stage of document analysis. Compared with the printed documents the historical documents have the variations in the foreground and the background. So, fast and accurate document image binarization method is required. Document image binarization methods uses a variety of scanned machine printed and hand written documents. There is currently a substantial and growing interest in the field of document image understanding where several research groups are developing and designing systems to automatically process and extract relevant information from documents such as drawings, maps, magazines, newspapers, forms and mail envelopes. Approaches dealing with the document image binarization are 52 P a g e

traditionally differentiated into global and local approaches. Global method uses a single computed threshold value to classify image pixels. The pre and post processing normally consist of speckle noise removal, hole filling and in some cases the merging of broken characters using heuristic rules. For some poor quality document images with variable and inhomogeneous document background intensity like shadows, smear or smudge, complex background patterns and signaldependent noise, a practical problem is that no thresholding algorithm could work well for all kind of document images. The aim of post processing is to remove the noise and false print to improve the image quality. The rest of the paper is organized as follows. Section II first reviews existing technologies. Section III describes the proposed system and the major work. Section IV presents the experimental results that shows the superior performance of the system when coared with other systems. II. RELATED WORKS Shijian et al. [2] proposed document image binarization using background estimation and stroke edges. This is a method in which estimates the document background first through an iterative polynomial smoothening procedure. Different types of document degradation are then compensated by using the estimated document back ground surface. The text stroke edge is further detected from the compensated document image by using L1-norm image gradient. The document text is segmented by a local threshold that is estimated based on the detected stroke edges. Text stroke edges is used to estimate the local threshold and accordingly overcomes the limitation of many existing adaptive thresholding methods.l1-norm image gradient is often more suitable for the edge detector and edge profile used for the text stroke detection and based on empirical observations. Yi Huang et al. [3] distinguished ink bleed reduction from the existing techniques by its allowance for user interaction and our dual layer MRF formulation. Ink bleed characteristics, namely that ink bleed intensity is lighter than the foreground or that the grey scale intensity values of ink bleed, foreground, and background follow a parametric distribution. Rajat gupta et al. [4] proposed the concept of matched wavelet filters to develop the globally matched wavelet filters specifically adapted for the text and non text region. These filters are used for detecting text regions in scene images and for segmentation of document images into text, picture and back ground. M.Cheriet et al. [5] proposed a new approach that goes beyond segmenting only one bright object from an image in to an approach that recursively segments the brightest object at each recursion, leaving the darkest object in the given digitized image. In digitized images, the uniformity of objects plays a significant role in separating these objects from the background. Ostu s method in thresholding grey scale images is efficient on the basis of uniformity measure between the two classes C0 and C1 that should be segmented. Eric Saund et al. [6] designed and implemented a customized pixel level labeling tool, for document images. Document images are characterized by certain kinds of structure that lend themselves to the use of accelerated user interface commands so that large collections of like label pixels can be tagged together. Oivind Due Trier et al [7] evaluated binarization methods requiring manual inspection of sub images and fine tuning of parameters were executed. This is because the binarization algorithms are intended to be party of an automatic digitizing system for maps, being able to process a large number of maps per times unit. III. PROPOSED SCHEME In the proposed system, when a degraded document is given as input noise reduction is performed by MDBUTMF [8] and then it is enhanced using bilateral tone adjustment. The enhanced image is then used for adaptive contrast map construction and then it is binarized and Canny s edge detection is done. Then the foreground text is segmented from the background. A. Noise Removal Every pixel of image is checked for salt and pepper noise. If the selected window contains salt/pepper noise as processing pixel and neighboring pixel values contains all pixels that adds salt and pepper noise to the image: 53 P a g e

0 0 () 0 0 (1) In the matrix (1), is the processing pixel. The mean of the processing window is found and processing pixel is replaced by the mean value. If the selected window contains salt or pepper noise as processing pixel and neighboring pixel values contains some pixels that adds salt and pepper noise to the image: 78 120 97 90 (0) 0 73 (2) In matrix (2),the processing pixel is 0. Now eliminate the salt and pepper noise from the selected window. Here the median value is replaced by median value. If the selected window contains a noise free pixel as a processing pixel: 43 55 85 67 (90) 81 70 79 66 (3) In matrix (3), 90 is the processing pixel. Since it is a noise free pixel it does not require further processing. B. Image Enhancement It improves the visibility of details in dark and bright regions using the bilateral gamma adjustment [9]. G ( L) ( G G a d ( L) G b d 1/ L G ) 2 b ( L) 1 (1 L) 1/ (4) (5) (6) Where L is the input luminance and γ is a user specified variable that indicates the degree of enhancement. Note that Gd is a convex function for enhancing dark regions and Gb is a concave function for bright regions. The final function of gamma adjustment Ga is the average of Gd and Gb. Substantially, Ga is an inverse S curve which reveals hidden details in dark and bright regions. Although Ga can enhance the contrast in dark and bright regions, it unavoidably reduces the contrast in midtone regions. C.Adaptive Contrast Image Construction To overcome the over normalization problem the local image contrast is combined with the local image gradient and adaptive local image contrast is calculated in eqn (7) C ( i, C( i, (1 )( I max( i, ( I min ( i, ) Where C(i, denotes the local contrast. I ( i, ( I ( i, denotes the image gradient. ( max min (a) (b) Fig 1.a. degraded document image Fig b. adaptive contrast map D.Canny s Edge Detection The purpose of the contrast image construction is to detect the stroke edge pixels of the document properly. The constructed contrast image has a clear bi-modal pattern where the adaptive image contrast computed at text stroke edges is obviously larger than that computed within the document background. It therefore detects the text stroke edge pixel candidate by using Otsu s global thresholding method. For the contrast images a binary map by Otsu s algorithm that extracts the stroke edge pixels properly. As the local image contrast and the local image gradient are evaluated by the difference between the maximum and minimum intensity in a local window, the pixels at both sides of the text stroke will be selected as the high contrast pixels. The binary map can be further improved through the combination with the edges by Canny s edge detector because Canny s edge detector has a good localization property that it can mark the edges close to real edge locations in the detecting image. In addition, canny edge detector uses two adaptive 54 P a g e (7)

thresholds and is more tolerant to different imaging artifacts such as shading. It should be noted that Canny s edge detector by itself often extracts a large amount of non-stroke edges without tuning the parameter manually. In the combined map, it keep only pixels that appear within both the high contrast image pixel map and canny edge map. Algorithm 1 : Edge Width Estimation Require : The Document Image I and Binary Text Stroke Edge Image Edg Step 1: Get the width and height of I Step 2: for Each Row i = 1 to height in Edg do Step 3: Scan from left to right to find edge pixels that meet the Following criteria: a) Its label is 0 (background) b) The next pixel is labeled as 1(edge) Step 4: Examine the intensities in I of those pixels selected in Step 3, and remove those pixels that have a lower intensity than the following pixel next to it in the same row of I Step 5: Match the remaining adjacent pixels in the same row into pairs, and calculate the distance between the two pixels in pair. Step 6: end for Step 7: Construct a histogram of calculated distances. Step 8: Use the frequently occurring distance as the estimated stroke edge width EW E. Threshold Estimation And Segmentation The text can then be extracted from the document background pixels once the high contrast stroke edge pixels are detected. Two characteristics can be observed from the document images: First, the text pixels are close to the detected text stroke edge pixels. Second, there is a distinct intensity difference between the high contrast stroke edge pixels and the surrounding background pixels. The neighborhood window should be at least larger than the stroke width in order to contain stroke edge pixels. So the size of the neighborhood window W can be set based on the stroke width of the document image under study, EW, which can be estimated from the detected stroke as stated in Algorithm Algorithm 2:Post Processing Procedure Require : The Input Document Image I, Initial Binary Result B and Corresponding Binary Text Stroke Edge Image Edg Ensure : The Final Binary Result B f Step 1: Find out all the connected components of the stroke edge pixels in Edg. Step 2: Remove those pixels that will not connect with other pixels. Step 3: for Each remaining edge pixels (i, j ): do Step 4: Get its neighborhood pairs: (i 1, j ) and (i + 1, j );(i, j 1) and (i, j + 1) Step 5: if the pixels in the same pairs belong to the same class (both text and background) then Step 6: Assign the pixel with lower intensity to foreground class (text), and the other to background class. Step 7: end if Step 8: end for Step 9: Remove single pixel artifacts along the boundaries after the document thresholding Step 10: Store the new binary result to Bf. IV. EXPERIMENTAL RESULTS It quantitatively compare the proposed method with other state of the art techniques on DIBCO 2009, H-DIBCO 2010 and DIBCO 2011 datasets. These methods include Otsu s method (OTSU), Sauvola s method (SAUV) [18], Niblack s method (NIBL), Bernsen s method (BERN), Gatos et al. s method (GATO), and the previous methods (LMM [5], BE). The three datasets are composed of the same series of document images that suffer from several common document degradations such as smear and smudge. The DIBCO 2009 dataset contains ten testing images that consist of five degraded handwritten documents and five degraded printed documents. The DIBCO 2011 dataset contains eight degraded handwritten documents and eight degraded printed documents. In total, we have 36 degraded document images with ground truth. 55 P a g e

(a) (b) (b) (d) Fig. 2.(a) Input image (b)ostu (c)sauv (d)proposed The superior performance of the proposed method can be explained by several factors. First, the proposed method combines the local image contrast and the local image gradient that help to suppress the background variation and avoid the overnormalization of document images with less variation. Second, the combination with edge map helps to produce a precise text stroke edge map. Third, the proposed method makes use of the text stroke edges that help to extract the foreground text from the document background accurately. TABLE I- EVALUATION RESULTS methods F - PSNR DRD MPM Rank measure Score OSTU 82.22 15.77 8.72 15.64 412 SAVU 82.54 15.78 8.09 9.20 403 BERN 47.28 7.92 82.28 136.54 664 GATO 82.11 16.04 5.42 7.13 353 LMM 82.56 16.75 6.02 6.42 516 BE 81.67 15.59 11.24 11.40 376 SNUS 85.2 17.16 15.66 9.07 279 HOWE 88.74 17.14 5.37 8.64 299 proposed 87.8 17.56 4.84 5.17 307 V. CONCLUSION This paper presents an adaptive image contrast based document image binarization technique that is tolerant to different types of document degradation such as uneven illumination and document smear. The proposed technique is simple and robust, only few parameters are involved. Moreover, it works for different kinds of degraded document images. The proposed technique makes use of the local image contrast that is evaluated based on the local maximum and minimum. The proposed method has been tested on the various datasets. Experiments show that the proposed method outperforms most reported document binarization methods in term of the F-measure, pseudo F-measure, PSNR, NRM, MPM and DRD. In future work it can explore the proposed algorithm with noise reduction for of noise in degraded documents and image enhancement to improve accuracy. In the future enhancement, we have planned to improve the segmentation quality by improving the accuracy. REFERENCES [1] H. Yi, M. S. Brown, and X. Dong, Userassisted ink-bleed reduction, IEEE Trans. Image Process., vol. 19, no. 10, pp. 2646 2658, Oct. 2010. [2] S. Lu, B. Su, and C. L. Tan, Document image binarization using background estimation and stroke edges, Int. J. Document Anal. Recognit. vol. 13, no. 4, pp. 303 314, Dec. 2010. [3] H. Yi, M. S. Brown, and X. Dong, Evaluation of Image binarization, IEEE Trans. Image Process., vol. 19, no. 10, pp. 2646 2658, Oct. 2011. [4] S. Kumar, R. Gupta, N. Khanna, S. Chaudhury, and S. D. Joshi, Text extraction and document image segmentation using matched wavelets and MRF model, IEEE Trans. Image Process., vol. 16, no. 8,pp. 2117 2128, Aug. 2007 [5] M. Cheriet, J. N. Said, and C. Y. Suen, A recursive thresholding technique for image segmentation, in Proc. IEEE Trans. Image Process., Jun. 1998, pp. 918 921. [6] E. Saund, J. Lin, and P. Sarkar, Pixlabeler: User interface for pixel-level labeling of elements in document images, in Proc. Int. Conf. Document Anal. Recognit., Jul. 2009, pp. 646 650. [7] O. D. Trier and T. Taxt, Evaluation of binarization methods for document images, IEEE Trans. Pattern Anal. Mach. Intell., vol. 17, no. 3, pp. 312 315, Mar. 1995. 56 P a g e

[8] J. Parker, C. Jennings, and A. Salkauskas, Thresholding using an illumination model, in Proc. Int. Conf. Doc. Anal. Recognit., Oct. 1993, pp. 270 273. [9] J. Sauvola and M. Pietikainen, Adaptive document image binarization, Pattern Recognit., vol. 33, no. 2, pp. 225 236, 2000. [10] S. Lu, B. Su, and C. L. Tan, comparison Document image binarization techniques, Int. J. Document Anal. Recognit. vol. 13, no. 4, pp. 303 314, Dec. 2010. [11] H. Yi, M. S. Brown, and X. Dong,,Document image binarization IEEE Trans. Image Process., vol. 19, no. 10, pp. 2646 2658, Oct. 2010. [12] M. Nilsson, M. Dahl, and I. Claesson, Grayscale image enhancement using the SMQT, in Proc. IEEE Int. Conf. Image Process., vol. 1. Sep. 2005, pp. 933 936. 57 P a g e