車載カメラにおける信号機認識および危険運転イベント検知 Traffic Light Recognition and Detection of Dangerous Driving Events from Surveillance Video of Vehicle Camera

Similar documents
アルゴリズムの設計と解析. 教授 : 黄潤和 (W4022) SA: 広野史明 (A4/A8)

研究開発評価に関する国際的な視点や国際動向

Installation Manual WIND TRANSDUCER

Delivering Business Outcomes

Ⅲ. 研究成果の刊行に関する一覧表 発表者氏名論文タイトル名発表誌名巻号ページ出版年. lgo/kourogi_ pedestrian.p df. xed and Augmen ted Reality

Decisions in games Minimax algorithm α-β algorithm Tic-Tac-Toe game

D80 を使用したオペレーション GSL システム周波数特性 アンプコントローラー設定. Arc 及びLine 設定ラインアレイスピーカーを2 から7 までの傾斜角度に湾曲したアレイセクションで使用する場合 Arcモードを用います Lineモード

U N I T. 1. What are Maxine and Debbie talking about? They are talking about. 2. What doesn t Maxine like? She doesn t like. 3. What is a shame?

レーダー流星ヘッドエコー DB 作成グループ (murmhed at nipr.ac.jp) 本規定は レーダー流星ヘッドエコー DB 作成グループの作成した MU レーダー流星ヘッド エコーデータベース ( 以下 本データベース ) の利用方法を定めるものである

Title inside of Narrow Hole by Needle-Typ. Issue Date Journal Article. Text version author.

Omochi rabbit amigurumi pattern

Infineon 24GHz Radar Solution. May 2017 PMM RSF DM PMM Business development

JSPS Science Dialog Program Kofu Higashi High School

CER7027B / CER7032B / CER7042B / CER7042BA / CER7052B CER8042B / CER8065B CER1042B / CER1065B CER1242B / CER1257B / CER1277B

TED コーパスを使った プレゼンにおける効果的な 英語表現の抽出

Lesson 5 What The Last Supper Tells Us

Title of the body. Citation. Issue Date Conference Paper. Text version author. Right

(Osaka Industrial Technology - Platform)

The Current State of Digital Healthcare

Final Product/Process Change Notification Document # : FPCN22191XD1 Issue Date: 24 January 2019

L1 Cultures Go Around the World

相関語句 ( 定型のようになっている語句 ) の表現 1. A is to B what C is to D. A と B の関係は C と D の関係に等しい Leaves are to the plant what lungs are to the animal.

ジェスチャ併用型 Voice-to-MIDI システムの提案 第五回知識創造支援システムシンポジウム報告書 : 本著作物の著作権は著者に帰属します

[ 言語情報科学論 A] 統計的言語モデル,N-grams

Intermediate Conversation Material #10

情Propagation Characteristics of 700MHz Band V2X Wireless Communication*

Chronicle of a Disaster: Understand

Effects and Problems Coming in Sight Utilizing TRIZ for Problem Solving of Existing Goods

[1] 大橋和也, 森拓哉, 古関隆章 運転整理時における乗車率に応じた旅客行動の変化のモデル化 電気学会論文誌 D,J-Rail 2013 特集,2015,pp

Multi-Band CMOS Low Noise Amplifiers Utilizing Transformers

The seven pillars of Data Science

特集 米国におけるコンシューマ向けブロードバンド衛星サービスの現状

Private Equity: where should you invest today? P&I Global Pension Symposium, Tokyo

Immersive and Non-Immersive VR Environments: A Preliminary EEG Investigation 没入型および非没入型 VR 環境 :EEG の比較. Herchel Thaddeus Machacon.

128 Dental Materials Journal 10(2): , 1991

ITU-R WP5D 第 9 回会合報告書

修士 / 博士課程専門課題 Ⅱ 試験問題

超小型 Very small (L:3.2 W:1.5 t:0.4mm) 高利得 High gain 無指向性 Omini-directional. < 用途 Applications> PHS 機器 DECT 機器 その他 PHS & DECT systems, etc ST01

科学研究費助成事業 ( 科学研究費補助金 ) 研究成果報告書

Local Populations Facing Long- Term Consequences of Nuclear Accidents: Lessons learned from Chernobyl and Fukushima

Present Status of SMEs I

P (o w) P (o s) s = speaker. w = word. Independence bet. phonemes and pitch. Insensitivity to phase differences. phase characteristics

Magellan Systems Japan, Inc.

Keio University Global Innovator Accelera6on Program 2015 Day 7 Design Process Exercise

Standardization of Data Transfer Format for Scanning Probe Microscopy

Indonesian Printing Industry Trends, Current Technology, and Future Development

TDK-Lambda A C 1/27

Call for a Pro-Innovation

Page No. 原文 リライト EDITOR'S NOTES 1 4 NATURAL ART

Studies on Modulation Classification in Cognitive Radios using Machine Learning

CPM6018RA Datasheet 定電流モジュール. Constant-current Power Modules. TAMURA CORPORATION Rev.A May, / 15

GDC2009 ゲーム AI 分野オーバービュー

Simple Utility-Interactive Photovoltaic Power Conditioners Installed in Veranda of Apartment Houses

Study on Multipath Propagation Modeling and Characterization in Advanced MIMO Communication Systems. Yi Wang

HARD LOCK Technical Reports

修士論文の和文要旨 研究科 専攻大学院情報理工学研究科情報. 通信工学専攻博士前期課程

CG Image Generation of Four-Dimensional Origami 4 次元折り紙の CG 画像生成

SanjigenJiten : Game System for Acquiring New Languages Visually 三次元辞典 : 第二言語学習のためのゲームシステム. Robert Howland Emily Olmstead Junichi Hoshino

科学技術 学術審議会大型プロジェクト作業部会 2015 年 12 月 22 日 永野博

Omni LED Bulb. Illustration( 实际安装, 설치사례, 設置事例 ) Bulb, Downlight OBB. OBB-i15W OBB-i20W OBB-i25W OBB-i30W OBB-i35W. Omni LED.

PH75A280-* RELIABILITY DATA 信頼性データ

SS 32. October 12, Yuji NAKAKITA. Land Mobile Communications Division, Radio Department Ministry of Internal Affairs and Communications, Japan

Sirichai Tammaruckwattana. A thesis submitted in partial fulfilment for the requirements of. the degree of Ph.D. (by Research) at the

Understanding User Acceptance of Electronic Information Resources:

XG PARAMETER CHANGE TABLE

NINJA LASER INNOVATORS BY DESIGN SINCE 1770


Development of XML and IP Based Distributed Ground Station System for Pico / Small Satellite

Preparation and Properties of Retted Kenaf Bast Fiber Pulp and Evaluation as Substitute for Manila Hemp Pulp

宇宙飛行生物学 (Bioastronautics( 宇宙飛行生物学 (Bioastronautics) の大学院教育への利用. Astrobiology)? 宇宙生物学 (Astrobiology( 宇宙生物学 カリキュラム詳細

Assessing Avian Predators of Japanese Murrelets on Birojima

Wideband Compact Antennas for MIMO Wireless Communications Dinh Thanh Le

The Bright Side of Urban Shrinkage: Steps toward Restructuring Cities

国際宇宙ステーションからの雷放電と 高高度発光現象観測 (GLIMS) の概要と 現状

9 th TRIZ symposium Meltex, Inc. Tajima. QFD and TRIZ Case Study in Surface Treatment Chemical

国際会議 ACM CHI ( ) HCI で生まれた研究例 2012/10/3 人とコンピュータの相互作用 WHAT IS HCI? (Human-Computer Interaction (HCI)

次の対話の文章を読んで, あとの各問に答えなさい ( * 印の付いている単語 語句には, 本文のあとに 注 がある )

Global Lightning and Sprite Measurements from International Space Station - GLIMS mission -

Global Standards Collaboration (GSC) 14. ITS Radiocommunication Systems in Japan

Multi-bit Sigma-Delta TDC Architecture for Digital Signal Timing Measurement

Big thank you from Fukushima Friends UK (FF)

Supporting Communications in Global Networks. Kevin Duh & 歐陽靖民

アナログ RF 回路の先端設計技術動向. Akira Matsuzawa. Department of Physical Electronics Tokyo Institute of Technology A.

Development of a pixel sensor based on SOI technology for the ILC vertex detector

Study in Patent Risk and Countermeasures Related to Open Management in Interaction Design

Evaluation of Spatial Recognition in a Virtual Environment

Developing Visual Information Processing Technology through Human Exploration

4. Contact arrangement 回路形式 1 poles 1 throws 1 回路 1 接点 (Details of contact arrangement are given in the assembly drawings 回路の詳細は製品図による )

Specifications characterize the warranted performance of the instrument under the stated operating conditions.

Toward The Organisational Innovation Study: A Critical Study of Previous Innovation Research

TDK Lambda C /35

Finding Near Optimal Solutions for Complex Real-world Problems

Instruction Manual. Model IB100 Interface Box. IM 12B06J09-01E-E 2nd edition. IM 12B06J09-01E-E_ed02.indd 1 01/12/16 15:52

IMPORTANT SAFETY INSTRUCTIONS Regulatory Safety Information

Onboard Antenna for 700 MHz Band V2X Communication *

Implementation as a Trickle-down Process of Knowledge and Technology to a Local Community

Future Perspectives of Science, Technology and Innovation

Season 15: GRAND FINAL PLAYER GUIDE. ver.2019/1/10

NI PXI/PCI-5411/5431 Specifications

Stealth dicer DISCO DFL 7340 user manual

Transcription:

車載カメラにおける信号機認識および危険運転イベント検知 Traffic Light Recognition and Detection of Dangerous Driving Events from Surveillance Video of Vehicle Camera * 関海克 * 笠原亮介 * 矢野友章 Haike GUAN Ryosuke KASAHARA Tomoaki YANO 要旨 赤信号無視や急発進などの危険運転を検知, ドライバーに警告し, 安全運転を支援するニーズが高まっている. ドライブレコーダーの録画解析によりドライバーの運転性向を診断するには, 信号機認識と危険運転イベント検知技術は不可欠である. 本稿は車載カメラにおける高精度の信号機認識, 危険運転イベントを検知する方法を提案する.YUV 色特徴により信号機の色を認識, 信号候補領域を作成, 面積の少ない領域で信号機の形状認識を行う. 信号機の周囲特徴を用いて, 誤認識を削除する. 複雑な撮影背景の昼シーンと明かり外乱の多い夜シーンで共に誤認識の少ない高い認識率が得られる. 信号機認識結果と車の加速度, 速度の情報を組み合わせ, 赤信号無視, 急発進の危険運転イベントを検知する. 日本とアメリカで実車テストをすることにより信号機認識性能評価を行った. その結果, 高い認識率と低い誤認識率を得られ, 提案手法の有効性を確認した. ドライブレコーダーの動画データで, 危険運転イベント検知実験を行い, 提案手法の有効性を確認した. ABSTRACT The surveillance videos of vehicle cameras are widely used to support safe driving on the part of drivers, especially taxi and truck drivers. Long-term videos are often manually inspected by human operators in order to identify dangerous driving events, which is tedious and time-consuming work. A new method is proposed to detect dangerous driving events (such as rocket starts or an ignored red light) from the surveillance videos automatically. In the proposed method, traffic light recognition is performed first and then the speed and acceleration data of the car is used in conjunction with the traffic light recognition results to detect dangerous driving events. We propose using color, shape, and context features to recognize traffic lights more accurately. Vehicle road tests were performed in both the USA and Japan in order to determine the effectiveness of the proposed method. Surveillance videos taken by driving recorder cameras were used to conduct traffic light recognition and dangerous driving event detection experiments. * 研究開発本部リコー ICT 研究所基盤 ICT 研究室 Core Information and Communication Technology Research Department, Ricoh Institute of Information and Communication Technology, Research and Development Division 本稿は,IS&T Electronic Imaging 学会に帰属の著作権の利用許諾を受け,IS&T Electronic Imaging 学会誌,IS&T International Symposium on Electronic Imaging 2017, Surveillance: Applications and Algorithms, pp. 3-11 (2017) に掲載した論文を基に作成した. Ricoh Technical Report No.43 16 FEBRUARY, 2018

1. Introduction The surveillance videos of vehicle cameras are widely used to support safe driving on the part of drivers, especially taxi and truck drivers. If dangerous driving events are identified in the videos, drivers are advised to improve their driving style and skill. Insurance companies also use these surveillance videos to look into a driver s driving record, and if he/she drives safely, the company will reduce his/her insurance fee. Long-term videos are often manually inspected by human operators in order to identify dangerous driving events, which is tedious and time-consuming work. Detecting dangerous driving events automatically can help inspectors check the surveillance videos more easily. Detecting dangerous driving events in real-time can also alert the driver to prevent traffic accidents in time. We propose a new method to automatically detect dangerous driving events from the surveillance video of vehicle cameras. Traffic light recognition is performed first and then the motion information of speed and acceleration data of the car is used in conjunction with the traffic light recognition results to detect dangerous driving events. 1-1 Related Work 1-1-1 Detection of Dangerous Driving Events In conventional methods, the sensor data of a car has been used to detect dangerous events. For example, the data of a vehicle s inertial sensors from the controller area network (CAN) bus of car can be used to recognize a particular driver s driving style 1). Vehicle velocity and acceleration data recorded by drive recorders have been used to detect dangerous driving 2), and sensor data from smartphones have been used to detect driving events such as turns, U-turns, acceleration, braking, and excessive speed. The driving styles of aggressive and nonaggressive drivers can be differentiated by examining the features of detected events data 3). An aggressive driving style is often associated with dangerous driving. The detection of dangerous driving events by smartphone sensors and cameras has been proposed 4). Dangerous driving is detected in order to alert drivers that they are in potential danger. Cameras are used to track the road conditions that may lead to dangerous driving events. Tailgating, lane weaving and drifting, and careless lane changes are detected by examining the camera s video stream. The inertial sensor data of smartphones is also used to detect dangerous events. Although conventional methods can detect dangerous driving events such as sudden braking and sudden accelerating, it is difficult to determine precisely why these events happen because these methods use only the car s sensor data. Dangerous driving events that have a relation to traffic lights, such as rocket starts and ignored red lights, have not been addressed in these conventional methods. 1-1-2 Traffic Light Recognition To detect dangerous driving events by traffic light recognition results, it is necessary to recognize traffic lights with a high recognition rate and a very low false positive rate, which can be quite difficult and challenging. Figures 1 and 2 show examples of daytime traffic lights in Japan and the USA, respectively. The color and shape of traffic lights are different in different countries and areas. For example, in Japan, most traffic lights are placed horizontally, while in the USA, they are mostly placed vertically. Size and brightness are also different; for example, traffic lights in the USA are often smaller and brighter than those in Japan. Ricoh Technical Report No.43 17 FEBRUARY, 2018

Fig. 4 Daytime traffic light in Japan at different frames. Fig. 1 Daytime traffic light in Japan. Fig. 2 Daytime traffic light in the USA. The color and shape of traffic lights are also different at different times. Figure 3 shows a nighttime traffic light in Japan, where the color and shape are different from that in Figure 1, which shows a traffic light in the daytime. In the night scene, because the gain of the camera is increased, the lighting area has been saturated to white, and the traffic light color extends to the surrounding area, making recognition more difficult. The traffic light color can also differ depending on the light source, e.g., LED light and incandescent lamp light. Fig. 3 Nighttime traffic light in Japan. Moreover, the color and shape of traffic lights are different in different frames. For example, the color and shape of the red traffic light in Figure 4 is different in different frames because the phase and frequency of the light are not synchronized with that of the vehicle camera. The traffic lights blink in the vehicle camera videos. Complex backgrounds also make traffic light recognition difficult. In the daytime, objects such as colored signboards and billboards that may have the same color as traffic lights make it easy for false positive detection to occur. In the nighttime, neon lights and the tail lamps of vehicles also make it easy for false positives to occur. Conventional methods have been proposed to detect traffic lights, but due to the difficulties mentioned above, these methods have not been able to ensure high enough recognition accuracy for detecting rocket starts and ignored red lights. In conventional methods, color features are used for traffic lights recognition: specifically, the (R, G, B) color spaces have been used to detect traffic light color 5-8). In contrast, we propose using the (Y, U, V) color space, which can separate traffic light color from that of the background more effectively than (R, G, B). Moreover, the conventional methods use the circle feature to recognize traffic lights 5-8), and because this is a simple feature, false detection happens easily. For example, a circular signboard and the tail lamp of a car can both easily be miss-detected as a traffic light. Overall, it is difficult for conventional methods to detect traffic lights with a high recognition rate and a low false positive rate. In addition, although these methods are designed to detect traffic lights, they do not consider using the detection results to identify dangerous driving events such as rocket starts and ignored red lights. We propose using the circle feature along with context information around traffic lights to obtain a much lower false positive rate than the conventional methods. The traffic light recognition results and sensor data of the car (speed, acceleration, etc.) are used to detect dangerous driving events in our proposed method. Ricoh Technical Report No.43 18 FEBRUARY, 2018

2. Proposed Method We propose a method to detect dangerous driving events such as rocket starts and ignored red lights from surveillance video automatically. In the proposed method, first, traffic light recognition is performed, and next, the speed and acceleration data of the car are used along with traffic light recognition results to detect dangerous driving events. To detect the events, motion information pertaining to speed and acceleration that is recorded by a driving recorder or obtained by sensors in real-time is used. Color, shape, and context features are used to recognize traffic lights more accurately. Vehicle road tests in both the USA and Japan were performed in order to determine the effectiveness of the proposed method. Real-time processing recognition experiments were done using vehicle camera video streams. Surveillance videos taken by driving recorder cameras were also used for experiments on traffic light recognition and the detection of dangerous driving events. 2-1 Traffic Light Recognition In our method, traffic lights are recognized first, and the red and green traffic light recognition results are used to detect dangerous driving events. The color of the light is recognized by referencing dictionary parameters learned from traffic light image samples. (Y, U, V) color space features are used for color recognition. Then, the candidate area for shape recognition is calculated from traffic light image pixels recognized by color. Shape recognition is done by using the circle feature of the traffic light, and context features around the traffic light area are used to reduce false positive recognition. Daytime and nighttime scenes are recognized automatically from the image by using a support vector machine (SVM) learning method. A traffic light recognition dictionary is selected and parameters are separately set for daytime or nighttime. A flow chart of the proposed method is shown in Figure 5. The algorithm is described in detail in the following sections. Input image Color recognition Candidate area calculation Shape recognition Traffic light recognition Dangerous driving events detection Output results Fig. 5 Flow chart of proposed traffic light recognition and dangerous driving events detection. 2-1-1 Color Recognition In the proposed method, the color space of (Y, U, V) is used to detect the color of the traffic light. The color space of (R, G, B) is transformed into (Y, U, V) by Y 0.299 U = 0.147 V 0.615 0.587 0.289 0.515 Day/night scene recognition Input dictionary Input context information Input speed and acceleration 0.114 R 0.436 G 0.100 B. (1) Image samples of traffic lights are collected to obtain the recognition dictionary parameter, and pixels of the lighting area are picked up and collected from the traffic light images. The collected pixels of the lighting area from sample images are transformed into the (Y, U, V) color space for learning the dictionary parameters. Ricoh Technical Report No.43 19 FEBRUARY, 2018

The traffic light image samples are separated into various groups (Japan daytime, Japan nighttime, USA daytime, and USA nighttime). Dictionary parameters are learned separately for each group by finding the minimum and maximum values in the (Y, U, V) color space. Y < Y < (2) min Y max U < U < (3) min U max Dictionary parameters V < V < (4) min V max Y min, Y max, U min, U max, V min, and V max in Equations (2) (4) are obtained from the distribution of the traffic light lighting area pixels by finding the minimum and maximum values in the (Y, U, V) space. The traffic light color area is recognized by using the dictionary parameters. The (Y, U, V) color space can separate traffic light colors more effectively than the (R, G, B) and (L*, a*, b*) color spaces used by conventional methods 5-8). Because of saturation and noise, the detected color area is smaller than the traffic light lighting area. 2-1-2 Identifying a Candidate Area A candidate area is identified for the shape recognition from the recognized color areas by means of expansion. If one pixel is recognized as a traffic light color, N N pixels around the recognized pixel are picked up from the original image. Figure 7 illustrates expanded red traffic light candidate areas that were made by picking 7 7 pixels around each recognized traffic light color pixel from the original image in Figure 6. The shapes of the traffic lights can be recognized easily by the candidate areas. Fig. 6 Example of daytime red traffic light image in Japan. Fig. 7 Example of red traffic light candidate area obtained from Fig. 6 by color recognition and expansion. 2-1-3 Traffic Light Shape Recognition Shape recognition of the traffic light is performed in the candidate areas of Figure 7. The circle features of the traffic light are extracted by Hough transform and used for the recognition. The circle is expressed by 2 2 2 ( x a) + ( y b) = r, (5) where a and b are the center axis of the circle and r is the radius of the circle. Hough transform is used to obtain the parameters of (a, b, r). Figure 8 shows the circle recognition results by Hough transform, with the results indicated by red rectangles around the traffic light circles. Figure 9 shows Ricoh Technical Report No.43 20 FEBRUARY, 2018

the recognized results of the red traffic light in the original image. area of the traffic light is used as a feature to reduce false detections. Context intensity patterns are also used. (1) Area of Traffic Light The area of a traffic light is used as a feature to reduce false positives. The area of traffic light lighting is limited to a certain range. When a vehicle is near the traffic light, the traffic light area size has a maximum value, and when the vehicle is far away, it has a minimum value. Area feature parameters are obtained from traffic light image samples. The candidate areas in Figure 8 are used to obtain these area feature Fig. 8 Shape recognition results of red traffic light in the daytime. parameters. Candidate area size minimum value A min and maximum value A max are obtained from traffic light image samples. If A, the area size of a traffic light candidate area, does not satisfy Equation (6), it is a false positive detection. A < A< (6) min A max For different times and places, different area feature parameters are obtained. Different area feature parameters are used for daytime, nighttime, USA, and Japan scenes. Image samples of a traffic light are collected to obtain the area feature parameters specific Fig. 9 Recognition results of red traffic lights. to that scene. 2-1-4 Context Information for Reducing False Positive Rate In daytime scenes, false detection frequently occurs because of complex backgrounds. For example, colored signboards and billboards are easily miss-recognized as traffic lights. In nighttime scenes, street lights and vehicle tail lamps are easily miss-recognized as traffic lights. Figure 9 shows an example in which the tail lamp of a car has been miss-recognized as a red traffic light, which is a false positive recognition. To reduce the number of false detections, context information is used. Specifically, the (2) Context Intensity Pattern In the proposed method, the context intensity pattern around the traffic light area is used to reduce the number of false detections. Figure 10 shows an example of a daytime red traffic light, where (b) is an image sample of a red traffic light and (a) illustrates the image parts of (b). The red circle in the rectangle area represents the lighting area of the red traffic light. The black circle in the rectangle area labeled A is a rectangle area at the left side of the red lighting traffic light area. Ricoh Technical Report No.43 21 FEBRUARY, 2018

(a) (b) Fig. 10 Context intensity pattern of daytime red traffic light in Japan. The rectangle areas labeled A, B, C, and D refer to areas around the lighting red traffic light area and have the same size as the red lighting area. The average intensity of each rectangle area is used as an intensity pattern. Figure 10 shows the context intensity pattern of a daytime red traffic light in Japan and Figure 11 shows that of one in nighttime Japan. (a) (b) Fig. 11 Context intensity pattern of nighttime red traffic light in Japan. The average intensities I pa, I pb, I pc and I pd of the A, B, C, and D areas among the image samples are used as parameters of the context area information. In Figure 10, A is a dark area and B, C, and D are comparatively bright areas. In Figure 11, A, B, C, and D are all dark areas. A A C B C B D D If a candidate traffic light has a context feature pattern matching that of a traffic light, it is a traffic light; otherwise, it is a false positive detection. In test image of Figure10, average intensities I ta, I tb, I tc and I td, of the A, B, C and D areas are calculated. If average intensity differences between parameter data I pi and test image data I ti (I = A, B, C, D), are all less than a threshold value T avi, the context feature pattern matches that of a traffic light. In Figure 11, traffic light image of nighttime, red light image pixels spread to wider distance than that of Figure 10. Positions of A, B, C and D areas are different from that of Figure 10. In Figure 11, A, B, C and D areas are set at dark areas around red light area. Different context intensity patterns are used for different times and places (e.g., daytime, nighttime, USA, and Japan). 2-1-5 Day and Night Scene Recognition Daytime and nighttime scenes are recognized by an input image. The average intensity of the input image is used as a feature for recognition. The image is divided into blocks, and N by N blocks are made. The number of black blocks is used as a recognition feature. A black block is defined by the average intensity of the block. If the average intensity is lower than a certain threshold, the block is a black block. A support vector machine (SVM) is used to recognize daytime and nighttime scenes automatically. We selected SVM for this task because in this case, only a very few data samples are needed to make the SVM recognition dictionary. The recognition dictionary is obtained by using the daytime and nighttime image sample data. 2-2 Detection of Dangerous Driving Events The detection of dangerous driving events is performed by using the traffic light recognition results in conjunction with speed and acceleration motion information. Rocket starts and ignored red lights are detected. 2-2-1 Rocket Start Event Detection When a traffic light changes from red to green and a driver suddenly accelerates his/her car, it is considered a dangerous driving event called a rocket start. The traffic Ricoh Technical Report No.43 22 FEBRUARY, 2018

light recognition results and motion information pertaining to the vehicle s acceleration and speed are used to detect such events. When a red traffic light is recognized and it then changes to green, if the acceleration becomes higher than a certain threshold within a short period time, it is detected as a rocket start. The rocket start level is defined by the time between the time at which the traffic light changes to green and the time at which the acceleration value increases to more than a threshold value. A short time indicates a strong level of rocket start. 2-2-2 Ignored Red Light Event Detection When a traffic light is red, traffic law dictates that the driver should stop his/her car. If the driver does not stop, it is considered a very dangerous driving event called an ignored red light. In the proposed method, this dangerous driving event is detected by using a combination of red traffic light recognition and motion information pertaining to the speed and acceleration of the car. If a red traffic light is recognized and the speed of the car is faster than a certain threshold, it is the event candidate. It is also the event candidate if the speed is not fast but the acceleration is higher than a certain threshold. If the car is far from the traffic light, it is not considered a dangerous driving event; it is only an event candidate. To reduce the number of false positive detections, the size and position features of the red traffic light in an image frame are used. If the car is far away from the red traffic light, it is not considered a dangerous driving event. If the car is far away, the red light has small size in the image, position of the red light is lower from the upper edge of the image. Image samples are collected in which the size and position of the red traffic lights are used to obtain size and position parameters. Candidate events are evaluated to determine if they are false positives or not by the learned dictionary parameters. When the car is near the red traffic light, the size of the traffic light becomes bigger and is situated closer to the upper end of the image frame. False positives are reduced by using the dictionary parameters of size and position from the upper end of the image. In Japan, there are green direction arrow traffic lights, as shown in Figure 12. When the car turns right, if there is a direction green arrow sign, as shown in Figure 12 (a), it is not a dangerous driving event. When a red traffic light is recognized, the green arrow direction light is detected. If a red light and a turn right green direction arrow light are both detected, and the car is moving straight forward, it is recognized as a dangerous driving event, as shown in Figure 12 (a). If a forward direction green arrow traffic light is on, even if a red traffic light is detected, it is not a dangerous event, as shown in Figure 12 (b). The green arrow is detected by finding green pixels in a specific area next to the red traffic light area. In Figure 12 (a), a green arrow is directly under the red traffic light. If green pixels are found in this area, it is detected as a green arrow. The same green color as a green light is used to detect green arrows. (a) (b) Fig. 12 Red traffic light and green arrow direction traffic light. (a) Turn right; (b) Straight forward. Ricoh Technical Report No.43 23 FEBRUARY, 2018

3. Experiment Results Vehicle road testing experiments were performed in both Japan and the USA using a prototype vehicle camera. The vehicle camera used in the experiment is shown in Figure 13 (a) and a prototype of the vehicle camera recognition system is shown in (b). (a) (b) Fig. 13 Prototype for real-time traffic light dangerous driving events detection. (a) Vehicle camera; (b) Prototype of vehicle camera recognition system. and ignored red light events. All of the dangerous driving events were detected accurately. We mainly evaluated our dangerous driving events detection algorithm by driving recorder video data. 3-1 Traffic Light Recognition Evaluation Traffic light recognition was evaluated by data taken from vehicle road tests performed in Japan and the USA. Part of the image frames were used to learn recognition dictionary parameters. Japan daytime and nighttime and USA daytime data were used to learn the recognition dictionary. (In the USA, only a daytime vehicle road test was performed.) For the Japan daytime data, 60,365 images were used for evaluation. An example of the recognition results for a Japan daytime red traffic light is shown in Figure 14. Table 1 lists the data of the recognition results. For the USA daytime data, 45,815 images were used for testing. An example of the recognition results for a green traffic light is shown in Figure 15. Table 2 lists the data of the testing results. For the Japan nighttime data, 8,293 images were used for testing. An example of the recognition results for the Japan nighttime green traffic light is shown in Figure 16. Table 3 lists the data of the testing results. A video stream is input from the camera by a controller box. Image size is 1280 720 pixels. Frame rate is 30 fps. Speed data was input from CAN data of the car. Acceleration data was input by an acceleration sensor on the controller box. The traffic light recognition and the dangerous driving events detection algorithm were running on the PC of the prototype. The traffic light recognition algorithm was evaluated by examining the vehicle road testing experimental data, as discussed in the next section. Dangerous driving events detection was also tested using a vehicle road test. Because it is dangerous to create such events on purpose, we only tested a few rocket starts Fig. 14 Example of recognition results for Japan daytime red traffic light. Table 1 Testing results of daytime Japan traffic light. Color Red Green Recognition rate (%) 91.1 95.1 False positive rate (%) 0.04 0.04 Ricoh Technical Report No.43 24 FEBRUARY, 2018

3-2 Dangerous Driving Events Detection Evaluation Fig. 15 Recognition result of USA daytime green traffic light. The driving recorder used to take surveillance videos for performing the traffic light recognition and dangerous driving event detection experiments is shown in Figure 17. The camera of the driving recorder has a 160-degree view angle. The image size is 1280 720 pixels. Frame rate is 30 fps. Speed and acceleration data were recorded simultaneously by the driving recorder sensors. Table 2 Testing results of daytime USA traffic light. Color Red Green Recognition rate (%) 85.7 100 False positive rate (%) 0.0 0.06 Fig. 17 Driving recorder used in dangerous driving events recognition experiment. Fig. 16 Recognition result of Japan nighttime green traffic light. Table 3 Testing results of nighttime Japan traffic light. Color Red Green Recognition rate (%) 90.0 100 False positive rate (%) 0.64 0.02 The processing time was tested using a PC environment consisting of a 3.5 GHz CPU, 15.4 GB of memory, and the Windows 7 OS. Processing times are listed in Table 4. The real-time processing speed of 30 fps was obtained. Table 4 Processing time of traffic light recognition. Data Processing time (ms) Japan daytime Japan nighttime USA daytime 33.9 23.6 31.0 A total of 68 test cars were used to collect surveillance video data, with each car equipped with one driving recorder. Videos were recorded in SD memory cards that were collected every three weeks. Once the SD cards were collected, the surveillance video data were processed to obtain dangerous driving events. We collected the SD cards three times. A total of 204 SD cards were collected. Because of the large amount of video data, it took several hours to process one SD card. The result of each detected dangerous driving event was recorded on a short video that could be checked later. We calculated the recognition rates among the dangerous driving events detected. A total of 21 rocket starts were detected, among which the detection rate was about 81%, and 68 ignored red lights were detected, among which detection rate was about 34%. Because dangerous driving events happen rarely, the number of detected events was low. Red traffic light false positive recognition sometime caused false positive detection of the dangerous driving events, which decreased the overall recognition rate. Ricoh Technical Report No.43 25 FEBRUARY, 2018

3-2-1 Rocket Starts We performed an experiment where, when the traffic light color changes from red to green, if the acceleration data increases to a value higher than a threshold, it is considered a rocket start. The rocket start level is defined by the rocket start times in Table 5. A short rocket start time means a strong rocket level. An example of the detection results for rocket starts is shown in Figure 18. In Figure 18 (b), the speed of the car suddenly increased to 18 km/hour. The acceleration value increased extremely, as illustrated by the vertical color bar. 3-2-2 Ignored Red Traffic Lights The ignored red traffic light dangerous event detection experiment was performed using recorded vehicle camera surveillance video. An example of the results is shown in Figure 19. The traffic light is red, the turn right green arrow traffic light is on, and the car speed is 52 km/hour. Table 5 Rocket start level. Rocket start level Rocked start time (s) 1 1.0 Fig. 19 Detection results of ignored red traffic light. 2 2.0 3 3.0 4 4.0 (a) (b) Fig. 18 Rocket start detection results. (a) Recognized red traffic light; (b) Rocket start when traffic light color changes from red to green and acceleration is higher than threshold value. 4. Conclusion A new method has been developed to detect dangerous driving events automatically from the surveillance video of a vehicle camera. Events such as rocket starts and ignored red lights can be detected. In the method, traffic light recognition is made first, and then the speed and acceleration data of the car, the traffic light recognition results, and multimodal features are used to detect dangerous driving events. It is difficult for conventional methods to obtain both a high recognition rate and a low false positive rate. We proposed using color, shape, and context features to recognize traffic lights more accurately. Vehicle road tests in both the USA and Japan were performed in order to determine the effectiveness of the proposed method. Real-time processing recognition experiments were performed in a car using a vehicle camera video stream. Surveillance videos taken by a driving recorder camera were also used to perform traffic light recognition and dangerous driving events detection experiments. Ricoh Technical Report No.43 26 FEBRUARY, 2018

A total of 68 test cars equipped with driving recorders were used to collect surveillance videos for the experiments on the detection of dangerous driving events. Dangerous driving events were detected and recorded as short videos, which are easy to inspect. Results showed that red traffic light false positive recognition sometimes caused the false positive detection of dangerous driving events, which decreased the overall event detection rate. Improvement of the traffic light recognition rate remains a future work. 7) M. Diaz-Cabrera, P. Cerri: Suspended traffic lights detection and distance estimation using color features, 15th International IEEE Conference on intelligent transportation system, pp. 1315-1320 (2012). 8) M. Diaz-Cabrera, P. Cerri, P. Medici: Robust realtime traffic light detection and distance estimation using a single camera, Expert Systems with Applications, pp. 3911-3923 (2015). References 1) M. V. Ly, S. Martin, M. M. Trivedi: Driver Classification and Driving Style Recognition using Inertial Sensors, IEEE Intelligent Vehicles Symposium (IV), pp. 1040-1045 (2013). 2) A. Naito et al.: Driver evaluation based on classification of rapid decelerating patterns, IEEE International conference on Vehicular Electronics and Safety (ICVES), pp. 108-112 (2009). 3) D. A. Johnson, M. M. Trivedi: Driving Style Recognition Using a Smartphone as a Sensor Platform, 14th International IEEE Conference on Intelligent Transportation Systems, pp. 1609-1615 (2011). 4) C. You et al.: CarSafe App: Alerting Drowsy and Distracted Drivers using Dual Cameras on Smartphones, Proceeding of the 11th annual international conference on Mobile systems, applications, and services, pp. 13-26 (2013). 5) M. Omachi, S. Omachi: Traffic light detection with color and edge information, International conference on computer science and information technology, pp. 284-287 (2009). 6) M. Omachi, S. Omachi: Detection of traffic light using structural information, 10th international conference on signal processing, pp. 809-893 (2010). Ricoh Technical Report No.43 27 FEBRUARY, 2018