Intelligent Traffic Sign Detector: Adaptive Learning Based on Online Gathering of Training Samples

Similar documents
Recognition of very low-resolution characters from motion images captured by a portable digital camera

Background Pixel Classification for Motion Detection in Video Image Sequences

SCIENCE & TECHNOLOGY

Face Detection System on Ada boost Algorithm Using Haar Classifiers

Hand Waving Gesture Detection using a Far-infrared Sensor Array with Thermo-spatial Region of Interest

Effects of the Unscented Kalman Filter Process for High Performance Face Detector

A VIDEO CAMERA ROAD SIGN SYSTEM OF THE EARLY WARNING FROM COLLISION WITH THE WILD ANIMALS

Face Detection: A Literature Review

Real-Time Face Detection and Tracking for High Resolution Smart Camera System

Experiments with An Improved Iris Segmentation Algorithm

A Real Time Static & Dynamic Hand Gesture Recognition System

Estimation of Folding Operations Using Silhouette Model

Recognition Of Vehicle Number Plate Using MATLAB

Student Attendance Monitoring System Via Face Detection and Recognition System

A Chinese License Plate Recognition System

Near Infrared Face Image Quality Assessment System of Video Sequences

Multi-Resolution Estimation of Optical Flow on Vehicle Tracking under Unpredictable Environments

License Plate Localisation based on Morphological Operations

Vehicle Detection using Images from Traffic Security Camera

An Improved Bernsen Algorithm Approaches For License Plate Recognition

An Efficient Method for Vehicle License Plate Detection in Complex Scenes

Linear Gaussian Method to Detect Blurry Digital Images using SIFT

Super resolution with Epitomes

tsushi Sasaki Fig. Flow diagram of panel structure recognition by specifying peripheral regions of each component in rectangles, and 3 types of detect

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

Conglomeration for color image segmentation of Otsu method, median filter and Adaptive median filter

Detection of AIBO and Humanoid Robots Using Cascades of Boosted Classifiers

中国科技论文在线. An Efficient Method of License Plate Location in Natural-scene Image. Haiqi Huang 1, Ming Gu 2,Hongyang Chao 2

EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION

Classification of Road Images for Lane Detection

Enhanced Method for Face Detection Based on Feature Color

Intelligent Nighttime Video Surveillance Using Multi-Intensity Infrared Illuminator

Real-Time Tracking via On-line Boosting Helmut Grabner, Michael Grabner, Horst Bischof

Driver Assistance for "Keeping Hands on the Wheel and Eyes on the Road"

Speed Traffic-Sign Recognition Algorithm for Real-Time Driving Assistant System

IMAGE PROCESSING TECHNIQUES FOR CROWD DENSITY ESTIMATION USING A REFERENCE IMAGE

Distributed Vision System: A Perceptual Information Infrastructure for Robot Navigation

Face detection, face alignment, and face image parsing

True Color Distributions of Scene Text and Background

FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM

VEHICLE LICENSE PLATE DETECTION ALGORITHM BASED ON STATISTICAL CHARACTERISTICS IN HSI COLOR MODEL

Implementation of Face Detection System Based on ZYNQ FPGA Jing Feng1, a, Busheng Zheng1, b* and Hao Xiao1, c

Wheeler-Classified Vehicle Detection System using CCTV Cameras

CS4670 / 5670: Computer Vision Noah Snavely

Face Detection using 3-D Time-of-Flight and Colour Cameras

Multi-Image Deblurring For Real-Time Face Recognition System

OPEN CV BASED AUTONOMOUS RC-CAR

Main Subject Detection of Image by Cropping Specific Sharp Area

Autonomous Localization

CROWD ANALYSIS WITH FISH EYE CAMERA

UM-Based Image Enhancement in Low-Light Situations

A Method for Estimating Meanings for Groups of Shapes in Presentation Slides

TRAFFIC SIGN DETECTION AND ANALYSIS: RECENT STUDIES AND EMERGING TRENDS

Published in: IEEE 17th International Conference on Intelligent Transportation Systems (ITSC), 2014

Automatic Ground Truth Generation of Camera Captured Documents Using Document Image Retrieval

PHASE PRESERVING DENOISING AND BINARIZATION OF ANCIENT DOCUMENT IMAGE

Automated Driving Car Using Image Processing

Traffic Sign Detection for U.S. Roads: Remaining Challenges and a case for Tracking

An Analysis of Image Denoising and Restoration of Handwritten Degraded Document Images

Keyword: Morphological operation, template matching, license plate localization, character recognition.

Urban Road Network Extraction from Spaceborne SAR Image

Automatic Licenses Plate Recognition System

Image Recognition for PCB Soldering Platform Controlled by Embedded Microchip Based on Hopfield Neural Network

Reliable Classification of Partially Occluded Coins

Design of Temporally Dithered Codes for Increased Depth of Field in Structured Light Systems

Vehicle License Plate Recognition System Using LoG Operator for Edge Detection and Radon Transform for Slant Correction

An Un-awarely Collected Real World Face Database: The ISL-Door Face Database

Moving Object Detection for Intelligent Visual Surveillance

Image Processing Based Vehicle Detection And Tracking System

IMPLEMENTATION METHOD VIOLA JONES FOR DETECTION MANY FACES

MAV-ID card processing using camera images

II. LITERATURE SURVEY

Target detection in side-scan sonar images: expert fusion reduces false alarms

Removal of Haze in Color Images using Histogram, Mean, and Threshold Values (HMTV)

Finding Text Regions Using Localised Measures

An Algorithm and Implementation for Image Segmentation

An Automatic System for Detecting the Vehicle Registration Plate from Video in Foggy and Rainy Environments using Restoration Technique

Spatially Adaptive Algorithm for Impulse Noise Removal from Color Images

Dynamic Data-Driven Adaptive Sampling and Monitoring of Big Spatial-Temporal Data Streams for Real-Time Solar Flare Detection

Automatic Segmentation of Fiber Cross Sections by Dual Thresholding

Today. CS 395T Visual Recognition. Course content. Administration. Expectations. Paper reviews

The Classification of Gun s Type Using Image Recognition Theory

ECC419 IMAGE PROCESSING

Target Recognition and Tracking based on Data Fusion of Radar and Infrared Image Sensors

Night-time pedestrian detection via Neuromorphic approach

Learning-based Face Detection by Adaptive Switching of Skin Color Models and AdaBoost under Varying Illumination

Method Of Defogging Image Based On the Sky Area Separation Yanhai Wu1,a, Kang1 Chen, Jing1 Zhang, Lihua Pang1

PAPER Grayscale Image Segmentation Using Color Space

Differentiation of Malignant and Benign Masses on Mammograms Using Radial Local Ternary Pattern

Privacy-Protected Camera for the Sensing Web

Video Synthesis System for Monitoring Closed Sections 1

A Novel Algorithm for Hand Vein Recognition Based on Wavelet Decomposition and Mean Absolute Deviation

CS231A Final Project: Who Drew It? Style Analysis on DeviantART

On Observer-based Passive Robust Impedance Control of a Robot Manipulator

Extraction and Recognition of Text From Digital English Comic Image Using Median Filter

Introduction to Video Forgery Detection: Part I

Improving Image Quality by Camera Signal Adaptation to Lighting Conditions

Development of Hybrid Image Sensor for Pedestrian Detection

Speed Traffic-Sign Number Recognition on Low Cost FPGA for Robust Sign Distortion and Illumination Conditions

Photo Quality Assessment based on a Focusing Map to Consider Shallow Depth of Field

Transcription:

2011 IEEE Intelligent Vehicles Symposium (IV) Baden-Baden, Germany, June 5-9, 2011 Intelligent Traffic Sign Detector: Adaptive Learning Based on Online Gathering of Training Samples Daisuke Deguchi, Mitsunori Shirasuna, Keisuke Doman, Ichiro Ide and Hiroshi Murase Graduate School of Information Science, Nagoya University Furo-cho Chikusa-ku Nagoya, Aichi 464 8601, Japan ddeguchi@is.nagoya-u.ac.jp Abstract This paper proposes an intelligent traffic sign detector using adaptive learning based on online gathering of training samples from in-vehicle camera image sequences. To detect traffic signs accurately from in-vehicle camera images, various training samples of traffic signs are needed. In addition, to reduce false alarms, various background images should also be prepared before constructing the detector. However, since their appearances vary widely, it is difficult to obtain them exhaustively by manual intervention. Therefore, the proposed method simultaneously obtains both traffic sign images and background images from in-vehicle camera images. Especially, to reduce false alarms, the proposed method gathers background images that were easily mis-detected by a previously constructed traffic sign detector, and re-trains the detector by using them as negative samples. By using retrospectively tracked traffic sign images and background images as positive and negative training samples, respectively, the proposed method constructs a highly accurate traffic sign detector automatically. Experimental results showed the effectiveness of the proposed method. I. INTRODUCTION In recent years, ITS (Intelligent Transport System) technologies play an important role to make our driving environment safe and comfortable. In ITS, in-vehicle camera is now widely used for recognition and understanding of the road environment. There are various objects in our road environment, and it is necessary to develop methods to detect and to recognize them accurately. Since traffic signs are very important information for driving, several research groups proposed methods to detect them automatically [1], [2], [3]. Nowadays, such traffic sign detection systems are also commercially available (BMW, Mercedes-Benz, etc.). However, they can only detect limited types of traffic signs, such as speed limit and overtaking restrictions. One of the successful methods for detecting objects from an image is the method employing a cascaded AdaBoost classifier that is widely used for face detection [4]. Bahlmann et al. used this approach for traffic sign detection from invehicle camera images, and they showed that accurate traffic sign detection could be achieved. Although this approach is quite accurate and fast enough, it requires a large number of traffic sign images for training the classifier. Usually, these traffic sign images are manually obtained by specifying their clipping rectangle in in-vehicle camera images. Since this task is very expensive and time consuming, it is difficult to apply this approach to detect various traffic signs. To solve this problem, Doman et al. proposed a method for construct- Fig. 1. Examples of various appearances of traffic signs. ing the AdaBoost classifier by generating various traffic sign images based on image degradation models [3]. Although this method can greatly reduce the cost for preparing traffic sign images, it is difficult to generate various appearances of traffic signs actually observed in a real environment (Fig. 1). To construct a traffic sign detector easily and accurately, it is necessary to obtain a large number of training samples from a real environment without manual intervention. Wöhler et al. [5] and our group [6] tried to solve these problems by obtaining training samples automatically from in-vehicle camera images. These approaches are effective to construct an accurate detector with less manual intervention. However, these are methods for obtaining positive training samples only, such as pedestrians and traffic signs. Therefore, these methods still require a large number of negative samples, that is, background images. To construct an accurate traffic sign detector, the negative samples should have various appearances, which however, is difficult and time consuming to obtain manually. Therefore, this paper proposes a method for constructing an accurate traffic sign detector automatically by obtaining positive and negative samples simultaneously from in-vehicle camera images. This paper focuses on regulatory signs with a red ring surrounding them, such as those in Fig. 2, because they are very common and important when driving a vehicle in Japan. Section II describes the details of the proposed method. Then, section III shows the experimental setup and results using in-vehicle camera images. The results are discussed in section IV. Finally, we will conclude this paper in section V. 978-1-4577-0889-3/11/$26.00 2011 IEEE 72

(a) Distant traffic signs (b) Middle traffic signs (c) Close traffic signs Fig. 3. Appearances observed at distant, middle and close traffic signs from a vehicle. Fig. 2. Target traffic signs. II. METHOD This paper proposes a method for constructing a highly accurate traffic sign detector automatically by gathering positive and negative training samples simultaneously from invehicle camera images. To construct an accurate traffic sign detector, various traffic sign images (positive samples) and background images (negative samples) should be prepared, and they must include various appearances observed in a real environment, such as size, illumination, and texture changes. However, it is very difficult to obtain such samples comprehensively. Also, since backgrounds vary widely, gathering of those images is an especially difficult problem. To solve these problems, by extending the idea we proposed in [6], this paper proposes a method for constructing an accurate traffic sign detector by gathering trafficsign imagesand background images simultaneously from in-vehicle camera images. A. Basic idea As shown in Fig. 3(a), it is difficult to detect and segment small traffic sign images (distant traffic signs) accurately and automatically. On the other hand, large traffic sign images (close traffic signs) can be segmented accurately (Fig. 3(c)), and it is comparatively stable to detect them automatically. Also, appearances of traffic signs change gradually in successive frames. Based on these characteristics, the proposed method obtains traffic sign images and background images simultaneously, under the following assumptions: (i) Large traffic signs can be detected stably in successive frames. (ii) Large traffic signs can be detected even if the position of the detection window changes slightly. (iii) If the position of a large traffic sign is given, it is easy to track back smaller traffic signs retrospectively from it. From the assumptions (ii) and (iii), traffic sign images can be obtained automatically by applying two steps: (1) find large traffic signs (high resolution), and (2) retrospectively track from a large traffic sign to smaller ones. On the other hand, from the assumptions (i) and (ii), background images can be obtained as a large image that is not detected in successive frames nor in adjacent detection windows. Therefore, the proposed method defines a likelihood of a background by measuring the overlap-ratio of the detected traffic signs temporally and spatially. From the fact that these are false alarms obtained by the detector, it is expected that they should become good negative samples for training to reduce false alarms of the detector we train it again. Therefore, the proposed method uses them as negative samples for re-training the detector. Finally, the proposed method constructs a traffic sign detector by using traffic sign images and background images obtained automatically as positive and negative training samples, respectively. Fig. 4 shows the flowchart of our proposed method. The proposed method consists of three parts: (1) retrospective gathering of traffic sign images from in-vehicle camera images, (2) gathering of background images by evaluating temporal and spatial overlap-ratio of the detected traffic signs, and (3) construction of a traffic sign detector by using both traffic sign images and background images. By repeating this process, the proposed method improves the accuracy of the traffic sign detector iteratively. The following sections describe details of these two parts. B. Retrospective gathering of traffic signimages The proposed method gathers traffic sign images as positive samples for training the traffic sign detector. This step consists of three parts: (1) detection of traffic sign candidates, (2) detection of a large traffic sign, and (3) retrospective tracking of traffic signs. In the proposed method, a nested 73

Image sequence,,,... Traffic sign detection by detector Detected traffic signs Retrospective gathering Background gathering Detected traffic signs Detector,,... Re-training of the detector Positives New samples Negatives Fig. 4. Flowchart of the proposed method. Fig. 5. 90ο 75 ο 60 ο 45 ο 30 ο 15 ο 0 ο Edge detection of a traffic sign. cascade of a Real AdaBoost classifier is used to detect traffic signs [10], [11]. The details of the detector are explained in section II-D. The proposed method uses this detector for obtaining large traffic signs. Then, small traffic signs are obtained automatically by retrospective tracking. The following sections describe details of these steps. 1) Detection of traffic sign candidates: This step searches traffic sign candidates from an in-vehicle camera image. First of all, traffic sign candidates are searched by scanning detection windows inside the entire region of an in-vehicle camera image. Each detection window is evaluated by using a previously constructed traffic sign detector H k, and decided whether it is a traffic sign or not. Then, the detected traffic sign candidates are merged by Mean shift clustering [7]. 2) Selection of a large traffic sign: From the assumption (ii), multiple traffic sign candidates can be obtained around a large traffic sign. Based on this characteristic, the proposed method removes false positives (i.e. not a traffic sign) by evaluating the number of merged candidates. Finally, large traffic signs are obtained by selecting large size detected candidates, and these are used as initial positions of the traffic sign in the retrospective tracking. 3) Retrospective tracking of traffic signs: To obtain smaller traffic sign images, this step tracks back in the image sequence from the position of the large traffic sign detected in the previous step. This process is formulated as a process of finding the position and the size of the t-th traffic signby using those of the (t +1)-th one. Before tracking traffic signs, the proposed method enhances red components of an input image by calculating the red value of normalized color (Eq.(3)), since our target traffic signs have red edges as shown in Fig. 2. Then, we obtain an image F t by applying a Gaussian filter to the image, and use this image F t for tracking smaller traffic signs. First, by increasing l, the proposed method finds edge pixels of the t-th traffic sign as a pixel that satisfies F t (x t+1 + lδx) Δx < 0, (1) where F t (x) is the gradient of the image intensity at position x, and is the inner product of vectors. As shown in Fig. 5, this search process is repeated with rotating the search direction Δx. Finally, the proposed method computes the position and the size of the traffic sign at time t by fitting a circle to those edge pixels [8]. Here, RANSAC approach is employed to avoid the outliers of inappropriate edge pixels. By repeating these steps as t t 1, the proposed method tracks back smaller traffic signs in the image sequence. C. Gathering of background images This step gathers background images as negative samples for training the traffic sign detector. From the assumptions (i) and (ii), large traffic signs can be found as a large image that is detected both in successive frames and in adjacent detection windows. That is, a large image, not detected in successive frames nor in adjacent detection windows, can be considered as a background. Based on this idea, the proposed method obtains background images by evaluating the following criteria. First, the proposed method applies the step II-B.1) to an in-vehicle camera image sequence, and obtains traffic sign candidates. Then, the proposed method evaluates two types of overlapping criteria N 1 and N 2. N 1 is related to the assumption (i), and it is calculated as the number of overlapped candidates between successive frames. Meanwhile, N 2 corresponds to the assumption (ii), and this is calculated by the number of merged candidates in the step II-B.1). By using N 1 and N 2, background images are obtained as a candidate that satisfies 1 1+e N 1 1 1+e N <T 1. (2) 2 74

(a) Input (b) Gray (c) Red (d) Green TABLE I DETECTION RATE OF THE ITERATIVELY CONSTRUCTED DETECTORS H 0, H 1,...,H 4 BY THE PROPOSED METHOD (M =1). Detector Precision Recall F-measure H 0 0.139 0.917 0.242 H 1 0.447 0.868 0.590 H 2 0.745 0.903 0.816 H 3 0.848 0.904 0.875 H 4 0.856 0.901 0.878 (e) Blue (f) Eq.(3) (g) Eq.(4) (h) Eq.(5) Fig. 6. Examples of color feature images for computing LRP features. D. Construction of a traffic sign detector In this paper, a nested cascade of a Real AdaBoost classifier [10], [11] is used as the traffic sign detector. The classifier is trained by using positive and negative samples obtained in the steps II-B and II-C, respectively. LRP (Local Rank Pattern) features [9] are used for training the Real AdaBoost classifier. Here, as used in the previous works [2], [3], we use seven types of color feature images for computing LRP features. Fig. 6 shows examples of these color feature images. Pixel values of these images are calculated as gray scale value (f 1 ), RGB values (f 2 f 4 ), and normalized RGB values (f 5 f 7 ). Here, f 5 (x) = r(x) r(x)+g(x)+b(x), (3) f 6 (x) = g(x) r(x)+g(x)+b(x), (4) f 7 (x) = b(x) r(x)+g(x)+b(x), (5) where r(x), g(x) and b(x) represent red, green and blue values at a pixel x, respectively. III. EXPERIMENT To evaluate the accuracy of the proposed method, we used in-vehicle camera images. This section shows details of the data, experimental setup, and evaluation and results. A. Data We captured in-vehicle camera images by using SANYO Xacti DMX-HD2 mounted on the windshield. The image was captured in the size of 640 480 pixels (30 fps). The captured images were divided into five image sequences (A 0, A 1, A 2, A 3,andA 4 ), and the total number of images were 3,907. These images were used as inputs of the proposed method. On the other hand, we prepared another 2,967 images containing at least one traffic sign between 15 15 pixels and 45 45 pixels, and these images were used for evaluation. Here, this evaluation data contained 4,886 traffic signs. Also, we prepared M (= 1or 10) images (640 480 pixels) containing no traffic sign, and these images were used for constructing an initial traffic sign detector H 0. B. Experimental setup In this experiment, five traffic sign detectors were constructed one by one according to the following steps. At first, thirteen large traffic signs were manually selected from the image sequence A 0, and 500 traffic sign images were simply generated by changing their clipping positions and rotation angles. Then, an initial traffic sign detector H 0 was constructed by using these images. Here, negative samples were obtained from M in-vehicle camera images containing no traffic sign by changing their clipping positions randomly, and 5,000 negative samples were used for training classifiers in each stage of the cascade. Next, according to the processes described in sections II-B and II-C, traffic sign images and background images were gathered simultaneously from the image sequence A 1 by using the constructed detector H 0. Then, images used in H 0 and images obtained in the above step were used for constructing a second traffic sign detector H 1. H 2, H 3,andH 4 were also constructed iteratively by the same manner as the above. C. Evaluation To evaluate the simultaneous gathering of training samples proposed in this paper, we compared the proposed method with the previous method [6]. We performed this evaluation in two conditions by changing the number of initial invehicle camera images used for obtaining negative samples. The first condition used only one in-vehicle camera image (M = 1), and the second one increased it to ten images (M = 10). We used T 1 = 0.4 for gathering background images described in section II-C. The experiment shown in section III-B was performed ten times, and we evaluated the accuracy of the constructed detectors in precision and recall rates with corresponding F-measures. Table I shows the results of the iteratively constructed detectors H 0, H 1,...,H 4 of the proposed method in precision and recall rates with corresponding F-measures (M =1). Here, F-measure was calculated as 2 Recall Precision F-measure = Recall + Precision. (6) Fig. 7 shows the comparison of the performance between the proposed method and the previous method in precision, recall rates and F-measures. Fig. 8 shows examples of the detected traffic signs by the proposed method and the previous method. 75

Precision 1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 Proposed method (M =1) Previous method (M =1) Recall 1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 Previous method (M =1) Proposed method (M =1) F-measure 1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 Proposed method (M =10) Previous method (M =10) Proposed method (M =1) Previous method (M =1) 0.2 0.2 0.2 0.1 0.1 0.1 0.0 0.0 0.0 (a) Precision rate (b) Recall rate (c) F-measure Fig. 7. Results of detectors H 0 H 4 iteratively constructed by the proposed method and the previous method in precision, recall rates and F-measure. Here, M is the number of in-vehicle camera images not containing traffic signs that were input initially for obtaining negative samples. (a) Detection results of the proposed method. (b) Detection results of the previous method [6]. Fig. 8. Examples of detection results using the detector H 4 constructed by the proposed method and the previous method. (a) shows results of the proposed method, and (b) shows results of the previous method. The previous method does not gather background images as negative samples for training the detector. Therefore, many mis-detections occur in comparison with the results of the proposed method. 76

IV. DISCUSSIONS If we want to make use of a traffic sign detector in a real road environment, it is required that both precision and recall of the detector should be high. That is, F-measure calculated by Eq.(6) should be high. From this point of view, as seen in Fig.7(c), the proposed method could achieve high F-measure by increasing the number of image sequences. Especially, since the proposed method retrains the detector by using background images that were easily mis-detected by a previously constructed detector, the proposed method could obtain F-measure of 0.878 at maximum. In contrast, the previous method could not improve the F-measure even if the number of image sequences increased. Since the previous method does not employ the mechanism of background image gathering, it obtains negative samples from only initial in-vehicle camera images containing no traffic sign, to train the detector. Consequently, the previous method could not train enough variations of backgrounds, especially in the case of M =1. Although the previous method could keep high recall rate by using retrospective gathering for obtaining traffic sign images of positive samples, its precision did not improve. Therefore, it is confirmed that the proposed method could improve the accuracy by obtaining both traffic sign images and background images simultaneously. We compared the accuracy of the detectors when increasing the number of initial in-vehicle camera images not containing traffic signs. Fig.7(c) shows results of the proposed method and the previous method in F-measure when using one or ten in-vehicle camera images for initial inputs. As seen in Fig.7(c), the proposed method achieved higher F- measure than the previous method, even if we used only one in-vehicle camera image not containing traffic signs as an initial input (M =1). Since the proposed method automatically obtains negative samples from in-vehicle camera image sequences, the proposed method can obtain various negative samples that are not included in the initial training samples, as shown in Fig. 9. Therefore, the proposed method could improve the accuracy of the traffic sign detector iteratively by taking in-vehicle camera image sequences. Since the traffic sign detector is usually used in an unknown environment, the proposed method will become a strong tool for adapting the detector to that environment. However, the proposed method sometimes gathered inappropriate background images, such as images partially containing a traffic sign. This degrades the accuracy of the traffic sign detector. We intend to improve the mechanism of background gathering for avoiding the above problem in our future work. V. CONCLUSIONS This paper proposed an intelligent traffic sign detector using adaptive learning based on online gathering of various traffic sign images and background images. The proposed method applies retrospective tracking for obtaining smaller traffic sign images. At the same time, background images are obtained by evaluating temporal and spatial overlapratio of detected traffic signs. Finally, the proposed method constructs a traffic sign detector by using them. We evaluated Fig. 9. Examples of background images obtained by the proposed method. the accuracy and the effectiveness of the proposed method by applying it to actual in-vehicle camera images. Experimental results showed that the proposed method could improve the accuracy of the traffic sign detector satisfactorily, even if we use only one in-vehicle camera image not containing traffic signs as an initial input for obtaining negative samples. Future works include: (i) improvement of the retrospective tracking especially for small traffic signs, (ii) improvement of the mechanism for gathering background images, and (iii) evaluation by applying the method to a larger dataset. VI. ACKNOWLEDGMENTS Parts of this research were supported by a Grant-in-Aid for Young Scientists from MEXT, a Grant-In-Aid for Scientific Research from MEXT, and a Core Research for Evolutional Science and Technology (CREST) project of JST. MIST library (http://mist.murase.m.is.nagoya-u.ac.jp/) was used for developing the proposed method. REFERENCES [1] S. Maldonado-Bascón, S. Lafuente-Arroyo, P. Gil-Jiménez, H. Gómez- Moreno, and F. López-Ferreras, Road-sign detection and recognition based on support vector machines, IEEE Transactions on Intelligent Transportation Systems, Vol.8, No.2, 2007, pp.264 278. [2] C. Bahlmann, Y. Zhu, V. Ramesh, M. Pellkofer, and T. Koehler, A system for traffic sign detection, tracking, and recognition using color, shape, and motion information, Proceedings of IEEE Intelligent Vehicles Symposium, 2005, pp.255 260. [3] K. Doman, D. Deguchi, T. Takahashi, Y. Mekada, I. Ide, and H. Murase, Construction of cascaded traffic sign detector using generative learning, Proceedings of International Conference on Innovative Computing, Information and Control, 2009, ICICIC-2009-1362. [4] P. Viola and M. Jones, Robust real-time face detection, International Journal of Computer Vision, Vol.57, No.2, 2004, pp.137 154. [5] C. Wöhler, Autonomous in situ training of classification modules in real-time vision systems and its application to pedestrian recognition, Pattern Recognition Letters, Vol.23, No.11, 2002, pp.1263 1270. [6] D. Deguchi, K. Doman, I. Ide, and H. Murase, Improvement of a traffic sign detector by retrospective gathering of training samples from in-vehicle camera image sequences, Proceedings of ACCV2010 Workshop on Computer Vision in Vehicle Technology: From Earth to Mars, 2010, pp.1 10. [7] Y. Cheng, Mean shift, mode seeking, and clustering, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol.17, No.8, 2005, pp.790 799. [8] I. D. Coope, Circle fitting by linear and nonlinear least squares, Journal of Optimization Theory and Applications, Vol.76, No.2, 1993, pp.381 388. [9] M. Hradis, A. Herout, and P. Zemcik, Local rank patterns novel features for rapid object detection, Proceedings of the International Conference on Computer Vision and Graphics, Lecture Notes in Computer Science, Vol.5337, 2008, pp.239 248. [10] R. E. Schapire and Y. Singer, Improved boosting algorithms using confidence-rated predictions, Machine Learning, Vol.37, No.3, 1999, pp.297 336. [11] C. Huang, H. Ai, B. Wu, and S. Lao, Boosting nested cascade detector for multi-view face detection, Proceedings of the International Conference on Pattern Recognition, Vol.2, 2004, pp.415 418. 77