Contrast adaptive binarization of low quality document images

Similar documents
An Improved Binarization Method for Degraded Document Seema Pardhi 1, Dr. G. U. Kharat 2

Quantitative Analysis of Local Adaptive Thresholding Techniques

[More* et al., 5(8): August, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116

Image binarization techniques for degraded document images: A review

Efficient Document Image Binarization for Degraded Document Images using MDBUTMF and BiTA

Binarization of Historical Document Images Using the Local Maximum and Minimum

Recovery of badly degraded Document images using Binarization Technique


Robust Document Image Binarization Technique for Degraded Document Images

An Analysis of Image Denoising and Restoration of Handwritten Degraded Document Images

Robust Document Image Binarization Techniques

APPLICATION OF THRESHOLD TECHNIQUES FOR READABILITY IMPROVEMENT OF JAWI HISTORICAL MANUSCRIPT IMAGES

IJSRD - International Journal for Scientific Research & Development Vol. 4, Issue 05, 2016 ISSN (online):

Extraction of Newspaper Headlines from Microfilm for Automatic Indexing

An Analysis of Binarization Ground Truthing

PHASE PRESERVING DENOISING AND BINARIZATION OF ANCIENT DOCUMENT IMAGE

Colored Rubber Stamp Removal from Document Images

Real Time Word to Picture Translation for Chinese Restaurant Menus

` Jurnal Teknologi IDENTIFICATION OF MOST SUITABLE BINARISATION METHODS FOR ACEHNESE ANCIENT MANUSCRIPTS RESTORATION SOFTWARE USER GUIDE.

International Conference on Computer, Communication, Control and Information Technology (C 3 IT 2009) Paper Code: DSIP-024

An Improved Bernsen Algorithm Approaches For License Plate Recognition

MAJORITY VOTING IMAGE BINARIZATION

Document Recovery from Degraded Images

Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition

Multilevel Rendering of Document Images

AUTOMATIC LICENSE PLATE RECOGNITION USING PYTHON

Enhanced MLP Input-Output Mapping for Degraded Pattern Recognition

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi

Effective Pixel Interpolation for Image Super Resolution

High-speed Micro-crack Detection of Solar Wafers with Variable Thickness

License Plate Localisation based on Morphological Operations

A new seal verification for Chinese color seal

Libyan Licenses Plate Recognition Using Template Matching Method

Histogram Equalization: A Strong Technique for Image Enhancement

Image Processing Lecture 4

Content Based Image Retrieval Using Color Histogram

A Robust Document Image Binarization Technique for Degraded Document Images

Image Rendering for Digital Fax

MAV-ID card processing using camera images

Number Plate Recognition Using Segmentation

Restoration of Degraded Historical Document Image 1

Automatic Licenses Plate Recognition System

Keywords: Image segmentation, pixels, threshold, histograms, MATLAB

OTSU Guided Adaptive Binarization of CAPTCHA Image using Gamma Correction

Restoration of Motion Blurred Document Images

Methods of Bitonal Image Conversion for Modern and Classic Documents

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

Fig 1 Complete Process of Image Binarization Through OCR 2016, IJARCSSE All Rights Reserved Page 213

Processing and Enhancement of Palm Vein Image in Vein Pattern Recognition System

A Chinese License Plate Recognition System

Implementation of global and local thresholding algorithms in image segmentation of coloured prints

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

Study and Analysis of various preprocessing approaches to enhance Offline Handwritten Gujarati Numerals for feature extraction

VEHICLE LICENSE PLATE DETECTION ALGORITHM BASED ON STATISTICAL CHARACTERISTICS IN HSI COLOR MODEL

10mW CMOS Retina and Classifier for Handheld, 1000Images/s Optical Character Recognition System

Implementation of Barcode Localization Technique using Morphological Operations

Method for Real Time Text Extraction of Digital Manga Comic

MEASUREMENT OF ROUGHNESS USING IMAGE PROCESSING. J. Ondra Department of Mechanical Technology Military Academy Brno, Brno, Czech Republic

Analytical Review of Preprocessing Techniques for Offline Handwritten Character Recognition

EFFICIENT CONTRAST ENHANCEMENT USING GAMMA CORRECTION WITH MULTILEVEL THRESHOLDING AND PROBABILITY BASED ENTROPY

Chapter 6. [6]Preprocessing

Er. Varun Kumar 1, Ms.Navdeep Kaur 2, Er.Vikas 3. IJRASET 2015: All Rights are Reserved

An Algorithm for Fingerprint Image Postprocessing

Image Enhancement using Histogram Equalization and Spatial Filtering

Research on 3-D measurement system based on handheld microscope

CHAPTER-4 FRUIT QUALITY GRADATION USING SHAPE, SIZE AND DEFECT ATTRIBUTES

A Histogram based Algorithm for Denoising Images Corrupted with Impulse Noise

A simple Technique for contrast stretching by the Addition, subtraction& HE of gray levels in digital image

Effective and Efficient Fingerprint Image Postprocessing

COMPARITIVE STUDY OF IMAGE DENOISING ALGORITHMS IN MEDICAL AND SATELLITE IMAGES

Binarization of Color Document Images via Luminance and Saturation Color Features

An Efficient Noise Removing Technique Using Mdbut Filter in Images

ECC419 IMAGE PROCESSING

Thresholding Technique for Document Images using a Digital Camera

Extraction and Recognition of Text From Digital English Comic Image Using Median Filter

Decision Based Median Filter Algorithm Using Resource Optimized FPGA to Extract Impulse Noise

A Method of Multi-License Plate Location in Road Bayonet Image

BINARIZATION TECHNIQUE USED FOR RECOVERING DEGRADED DOCUMENT IMAGES

Contents 1 Introduction Optical Character Recognition Systems Soft Computing Techniques for Optical Character Recognition Systems

Improving the Quality of Degraded Document Images

Multispectral Image Restoration of Historical Document Images

A Study On Preprocessing A Mammogram Image Using Adaptive Median Filter

IEEE Signal Processing Letters: SPL Distance-Reciprocal Distortion Measure for Binary Document Images

AN EFFICIENT IMAGE ENHANCEMENT ALGORITHM FOR SONAR DATA

An Autonomous Vehicle Navigation System using Panoramic Machine Vision Techniques

Review and Analysis of Image Enhancement Techniques

A Study on Single Camera Based ANPR System for Improvement of Vehicle Number Plate Recognition on Multi-lane Roads

Implementation of License Plate Recognition System in ARM Cortex A8 Board

Background Pixel Classification for Motion Detection in Video Image Sequences

IMAGE TYPE WATER METER CHARACTER RECOGNITION BASED ON EMBEDDED DSP

DENSE-CLUSTER BASED VOTING APPROACH FOR LICENSE PLATE IDENTIFICATION

Integrated Digital System for Yarn Surface Quality Evaluation using Computer Vision and Artificial Intelligence

Improved Hybrid Binarization based on Kmeans for Heterogeneous document processing

NON UNIFORM BACKGROUND REMOVAL FOR PARTICLE ANALYSIS BASED ON MORPHOLOGICAL STRUCTURING ELEMENT:

Automatic Counterfeit Protection System Code Classification

Open Access An Improved Character Recognition Algorithm for License Plate Based on BP Neural Network

A Comparative Analysis of Different Edge Based Algorithms for Mobile/Camera Captured Images

Open Access The Application of Digital Image Processing Method in Range Finding by Camera

Exercise questions for Machine vision

Local prediction based reversible watermarking framework for digital videos

Transcription:

Contrast adaptive binarization of low quality document images Meng-Ling Feng a) and Yap-Peng Tan b) School of Electrical and Electronic Engineering, Nanyang Technological University, Nanyang Avenue, Singapore 639798, Singapore a) FENG0010@ntu.edu.sg b) eyptan@ntu.edu.sg Abstract: We propose in this paper an improved method for binarizing document images by adaptively exploiting the local image contrast. The proposed method aims to overcome the common problems encountered in low quality images, such as uneven illumination, low contrast, and random noise. Experiments have been conducted and the results are presented to show the effectiveness of the proposed method. Keywords: document image analysis, image binarization, optical character recognition, OCR Classification: Science and engineering for electronics References [1] M. Pilu and S. Polland, A light-weight text image processing method for handheld embedded cameras, Hewlett-Packard Laboratories, Bristol, England, March 2002. [2] H.-S. Don, A noise attribute thresholding method for document image binarization, Third International Conference on Document Analysis and Recognition, pp. 231 234, 1995. [3] N. Otsu, A threshold selection method from gray-level histograms, IEEE Trans. Syst., Man, Cybern., vol. 9, no. 1, pp. 62 66, March 1979. [4] O. D. Trier and T. Taxt, Evaluation of binarization methods for document images, IEEE Trans. Pattern Anal. Machine Intell., vol. 17, no. 3, pp. 312 315, March 1995. [5] W. Niblack, An introduction to digital image processing, Englewood Cliffs, Prentice Hall, N.J., pp. 115 116, 1986. [6] J. Sauvola and M. Pietikainen, Adaptive document image binarization, Patt. Recogn., vol. 33, pp. 225 236, 2000. [7] C. Wolf and J.-M. Jolion, Extraction and recognition of artificial text in multimedia documents, Patt. Anal. Appl., vol. 6, no. 4, pp. 309 326, Feb. 2004. [8] Transym Optical Character Recognition (TOCR) [Available online] http://www.sorcery.demon.co.uk. 501

1 Introduction We have witnessed the migration of small and low-cost imaging devices into portable electronic gadgets, such as personal digital assistants (PDAs) and mobile phones, over the past few years [1]. This revolutionary integration has made possible a range of new and useful applications, e.g., quick capture and analysis of document images for faxing, auto-note taking, foreign text translation, etc. However, the integration also brings new technical challenges to document image analysis owing to the size and cost constraints imposed by these imaging devices. The problems commonly seen in the document images captured by these imaging devices include: (1) poor contrast due to the lack of sufficient or controllable lighting, (2) non-uniform image background intensity due to uneven illumination, and (3) immoderate amount of random noises due to limited sensitivity of imaging sensor and lack of adjustable camera s exposure time and aperture size. To extract text information from these low quality document images, an image binarization method that is robust against these problems is therefore indispensable. This paper proposes such a binarization method. 2 Related Work Image binarization, a common first step to document image analysis, converts the gray values of document images into two-level representations for text and non-text regions. A large number of binarization techniques have been proposed during the past decade [1]-[7]. These techniques can be generally classified into two categories: one is based on global thresholding; the other, local thresholding. Global thresholding uses only a single threshold value, which is generally chosen based on some heuristics or the statistics of some global image attributes, to classify image pixels into text (foreground) and non-text (background) pixels [3]. The main drawback of global thresholding is that it cannot adapt well to uneven illumination and random noises, and hence performs unsatisfactorily for low quality document images. In local thresholding, the threshold value is varied based on the local content of the document image. In comparison with global thresholding, local thresholding generally performs better for low quality images, especially in classifying pixels near text and object boundaries. One of the well performing local thresholding methods was proposed by Niblack [5]. The main idea of Niblack s method is to build a threshold surface, based on the local mean, m, and standard deviation, s, of gray values computed over a small neighborhood around each pixel in the form of T = m + k s, (1) where k is a negative constant set to 0.2. This method, however, tends to produce a large amount of binarization noises in non-text image regions [e.g., see Fig. 3 (b)]. As a result, computationally intensive post-processing is 502

required to reduce or remove the noises prior to the subsequent OCR (Optical Character Recognition) analysis. Sauvola et al. [6] improve Niblack s method by imposing a hypothesis on the gray values of text and non-text pixels text pixels have near 0 gray values and non-text pixels have near 255 gray values and compute the local threshold value as ( ( T = m 1 k 1 s )), (2) R where k is a constant set to 0.5, and R denotes the dynamic range of image gray-value standard deviation and is fixed to 128. Sauvola s method could outperform Niblack s for well-scanned document images, but it faces difficulties in dealing with images that do not agree well with the hypothesis (e.g., documents images captured at insufficient illumination, especially when the gray values of text and non-text pixels are close to each other), as illustrated in Fig. 3 (c). To overcome this problem, Christian et al. [7] propose to determine the local threshold value by normalizing the contrast and the mean gray value of the image as follows T =(1 k) m + k M + k s (m M), (3) R where k is fixed to 0.5, M is the minimum gray value of the image, and R is set to the maximum gray-value standard deviation obtained from all local image neighborhoods. This approach achieves the best binarization results among the three local thresholding methods considered. However, the performance of Christian s method degrades considerably when there is a notable change of background s gray values across the image, as all the local threshold values are determined based on the minimum gray value and the maximum standard deviation of local gray values computed from the whole image. 3 Proposed Method To extract useful text information from low quality document images, we have devised a new and reliable local thresholding method to perform binarization based on image local gray-value contrast. Specifically, to determine the local threshold value, the proposed method first computes the local mean, m, the minimum, M, and the standard deviation, s, of gray values in a primary local window (as shown in Fig. 1) with size large enough to cover 1-2 characters. To compensate for the negative effect of uneven illumination, the dynamic range of gray-value standard deviation R S is calculated in a larger local window, the secondary local window, instead of the whole image, as also illustrated in Fig. 1. Our study shows that the size of this secondary local window can be properly set to tolerate different degrees of illumination unevenness. To this end, we propose to calculate the local threshold value as: ( ) s T =(1 α 1 ) m + α 2 (m M)+α 3 M, (4) R S 503

Fig. 1. Primary local window and secondary local window of the proposed binarization method. where α 2 = k 1 ( s R S ) γ, α 3 = k 2 ( s R S ) γ, and α 1, γ, k 1 and k 2 are positive constants. The proposed local threshold as computed in (4) consists of three components, and three coefficients (α 1, α 2,andα 3 ) are defined to allow more flexible and adaptive weighting of these three components. Furthermore, coefficients α 2 and α 3 are adaptively set based on the normalized local standard deviation, s/r S, for the reason that local windows consisting of texts normally assume a larger gray-value standard deviation (and hence a larger s/r S ratio) than those without covering any texts. Consequently, setting the values of coefficients α 2 and α 3 basedonthes/r S ratio enables the local threshold so calculated to separate text regions from non-text regions more effectively. Moreover, we found from our empirical study that the power of s/r S ratio affects, to a certain extent, the binarization performance as well. Hence, an additional exponential parameter (γ) is introduced to reflect this effect. In our implementation of the proposed method, a 5 5 median filter is also applied to remove, as many as possible, the random noises from the image under consideration. To minimize the computational load, the local threshold values calculated at the centers of the primary local windows are bilinearly interpolated to obtain the threshold values for all the pixels. Based on our experiments, setting γ to 2 and the values of α 1, k 1 and k 2 in the ranges of 0.1-0.2, 0.15-0.25 and 0.01-0.05, respectively, can generally obtain good binarization results for a large variety of document images. 4 Experimental Results We have applied the proposed method, with parameters γ, α 1, k 1 and k 2 set to 2, 0.12, 0.25 and 0.04, to a set of document images with various sizes and qualities. Fig. 2 shows the thresholds determined by the three existing local thresholding methods and the proposed adaptive method along a scanline of a test document image; the image bar on top of the figure shows the enlarged content of the scanline, where dark regions denote text pixels and gray regions represent non-text pixels. Compared with Niblack s and Christian s methods, 504

Fig. 2. Local thresholds determined by different binarization methods along an image scanline. it is evident that the local thresholds obtained by our proposed method can adapt better to the unevenness of illumination. In particular, our method could successfully classify the right portion of the scanline (from pixel 140 onwards) as non-text region, while the other two wrongly classified part of it as text region. For Sauvola s method, owing to the constraint imposed by its additional hypothesis, the threshold values turned out to be near zero and failed to binarize the scanline satisfactorily. Fig. 3 (a)-(e) show the results of applying the methods under comparison to three test images captured by a low-resolution handheld camera. The results show that Niblack s method can segment the text characters well, but it also produces excessive noises in non-text regions. On the other hand, Sauvola s method is suitable for good quality images, but it does not perform well for images that do not agree with its additional hypothesis. Both Christian s method and the proposed method produce superior results. Nevertheless, Christian s method suffers from the effect of uneven illumination, resulting in some noises in non-text regions and broken text strokes. Our proposed method is supreme in terms of robustness to uneven illumination, suppression of noises in non-text regions, and completeness of binarized text characters. More experimental results are shown in Fig. 3 (f)-(h) to illustrate the superiority of our proposed method as compared with Christian s method. In conjunction with the Transom Optical Character Recognizer (TOCR) [8], we have also performed OCR on a large set of test document images. The binarization results of our proposed method can achieve an average 90.8% correct character recognition rate, while that of Christian s method can only attain 85.4%. 505

Fig. 3. Binarization results. (a) Original images. (b) Niblack s method. (c) Sauvola s method. (d) Christian s method. (e) Proposed method. (f) Additional images. (g) Christian s method. (h) Proposed method. 5 Conclusion We have presented an improved local thresholding method for binarizing document images by adaptively exploiting the local image contrast. Compared with other existing local thresholding methods, our proposed method performs better for low quality document images, especially those with uneven illumination, low contrast, and random noises. Experimental results obtained from various challenging document images have also been presented to testify the effectiveness and superiority of our proposed method. 506