Advances in Vehicle Periphery Sensing Techniques Aimed at Realizing Autonomous Driving

Similar documents
Intelligent Technology for More Advanced Autonomous Driving

White paper on CAR28T millimeter wave radar

Development of a 24 GHz Band Peripheral Monitoring Radar

Choosing the Optimum Mix of Sensors for Driver Assistance and Autonomous Vehicles

White paper on CAR150 millimeter wave radar

P1.4. Light has to go where it is needed: Future Light Based Driver Assistance Systems

23270: AUGMENTED REALITY FOR NAVIGATION AND INFORMATIONAL ADAS. Sergii Bykov Technical Lead Machine Learning 12 Oct 2017

Development of 24 GHz-band High Resolution Multi-Mode Radar

Development of Gaze Detection Technology toward Driver's State Estimation

Evaluation of Connected Vehicle Technology for Concept Proposal Using V2X Testbed

THE EXPANSION OF DRIVING SAFETY SUPPORT SYSTEMS BY UTILIZING THE RADIO WAVES

Silicon radars and smart algorithms - disruptive innovation in perceptive IoT systems Andy Dewilde PUBLIC

Applications of Millimeter-Wave Sensors in ITS

Global Image Sensor Market with Focus on Automotive CMOS Sensors: Industry Analysis & Outlook ( )

FLASH LiDAR KEY BENEFITS

Measurement of Surface Profile and Layer Cross-section with Wide Field of View and High Precision

Tsuyoshi Sato PIONEER CORPORATION July 6, 2017

Situational Awareness A Missing DP Sensor output

OPEN CV BASED AUTONOMOUS RC-CAR

Visione per il veicolo Paolo Medici 2017/ Visual Perception

Bluetooth Low Energy Sensing Technology for Proximity Construction Applications

A Winning Combination

Current Technologies in Vehicular Communications

SIS63-Building the Future-Advanced Integrated Safety Applications: interactive Perception platform and fusion modules results

Moving from legacy 24 GHz to state-of-the-art 77 GHz radar

Deliverable D1.6 Initial System Specifications Executive Summary

A Method of Measuring Distances between Cars. Using Vehicle Black Box Images

An Information Fusion Method for Vehicle Positioning System

76-GHz High-Resolution Radar for Autonomous Driving Support

ITS Radiocommunications in Japan Progress report and future directions

Sony Releases the Industry's Highest Resolution Effective Megapixel Stacked CMOS Image Sensor for Automotive Cameras

White paper on SP25 millimeter wave radar

Digital Radio in the car in 10 years

Development of Hybrid Image Sensor for Pedestrian Detection

Automotive 77GHz; Coupled 3D-EM / Asymptotic Simulations. Franz Hirtenfelder CST /AG

Humans and Automated Driving Systems

Positioning Challenges in Cooperative Vehicular Safety Systems

Evaluation of Actuated Right Turn Signal Control Using the ITS Radio Communication System

CMOS Image Sensors in Cell Phones, Cars and Beyond. Patrick Feng General manager BYD Microelectronics October 8, 2013

Honda R&D Americas, Inc.

Ricoh's Machine Vision: A Window on the Future

Effective Collision Avoidance System Using Modified Kalman Filter

Ultra-small, economical and cheap radar made possible thanks to chip technology

[Overview of the Consolidated Financial Results]

Autonomous driving technology and ITS

HAVEit Highly Automated Vehicles for Intelligent Transport

Circularly Polarized Post-wall Waveguide Slotted Arrays

interactive IP: Perception platform and modules

By Pierre Olivier, Vice President, Engineering and Manufacturing, LeddarTech Inc.

DOWNLOAD OR READ : 24 GHZ RADAR SENSOR EMPIRE XPU PDF EBOOK EPUB MOBI

Systems characteristics of automotive radars operating in the frequency band GHz for intelligent transport systems applications

AUTODRIVE PROJECT. Kleber Moreti de Camargo Rodrigo Diniz FATEC Itapetininga

CS686: High-level Motion/Path Planning Applications

TECHNOLOGY DEVELOPMENT AREAS IN AAWA

II. ADVANTAGES AND DISADVANTAGES

MATLAB 및 Simulink 를이용한운전자지원시스템개발

ADAS Development using Advanced Real-Time All-in-the-Loop Simulators. Roberto De Vecchi VI-grade Enrico Busto - AddFor

Vehicle-to-X communication using millimeter waves

Speed Traffic-Sign Recognition Algorithm for Real-Time Driving Assistant System

Exercise questions for Machine vision

$FWLYH DQG 3DVVLYH &DU 6DIHW\ $Q,QWHJUDWHG $SSURDFK WR 5HGXFLQJ $FFLGHQWV

DEVELOPMENT OF PASSIVE SURVEILLANCE RADAR

THE SCHOOL BUS. Figure 1

Virtual Homologation of Software- Intensive Safety Systems: From ESC to Automated Driving

ITS radiocommunications toward automated driving systems in Japan

Message points from SARA Active Safety through Automotive UWB Short Range Radar (SRR)

Automated Driving Car Using Image Processing

Revision of the EU General Safety Regulation and Pedestrian Safety Regulation

DEVELOPMENT OF A MICROSCOPIC TRAFFIC SIMULATION MODEL FOR INTERACTIVE TRAFFIC ENVIRONMENT

EG 1 Millimeter-wave & Integrated Antennas

Making Vehicles Smarter and Safer with Diode Laser-Based 3D Sensing

HiFi Radar Target. Kristian Karlsson (RISE)

EM Simulation of Automotive Radar Mounted in Vehicle Bumper

An Autonomous Vehicle Navigation System using Panoramic Machine Vision Techniques

Introduction of SIP-adus FOT SIP-adus ー Mobility bringing everyone a smile ー

DENSO www. densocorp-na.com

Using Vision-Based Driver Assistance to Augment Vehicular Ad-Hoc Network Communication

DENSO

Evaluation of Roadside Wrong-Way Warning Systems with Different Types of Sensors

DATE: 17/08/2006 Issue No 2 e-plate Operation Overview

Inter- and Intra-Vehicle Communications

Detection and Tracking of the Vanishing Point on a Horizon for Automotive Applications

Digital Sensing Technology for Visualizing Sites

Technical Datasheet. Blaxtair is an intelligent cameraa with the ability to generate alarms when a pedestrian is detected

The GATEway Project London s Autonomous Push

Current Status of ITS Radiocommunications in Japan

A Study on Single Camera Based ANPR System for Improvement of Vehicle Number Plate Recognition on Multi-lane Roads

Passive Components around ADAS Applications By Ron Demcko, AVX Fellow, AVX Corporation

A NEW NEUROMORPHIC STRATEGY FOR THE FUTURE OF VISION FOR MACHINES June Xavier Lagorce Head of Computer Vision & Systems

GPS-Based Navigation & Positioning Challenges in Communications- Enabled Driver Assistance Systems

MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS

Digital Engines for Smart and Connected Cars By Bob O Donnell, TECHnalysis Research Chief Analyst

International Journal of Scientific & Engineering Research Volume 8, Issue 7, July-2017 ISSN

Keysight Technologies Achieving Accurate E-band Power Measurements with E8486A Waveguide Power Sensors. Application Note

IS STANDARDIZATION FOR AUTONOMOUS CARS AROUND THE CORNER? By Shervin Pishevar

16. Sensors 217. eye hand control. br-er16-01e.cdr

Machine Vision Beyond the Factory. Jeff Burnstein President October 18, 2012 Beijing

Driver Assistance and Awareness Applications

A Vehicle Speed Measurement System for Nighttime with Camera

Israel Railways No Fault Liability Renewal The Implementation of New Technological Safety Devices at Level Crossings. Amos Gellert, Nataly Kats

Transcription:

FEATURED ARTICLES Autonomous Driving Technology for Connected Cars Advances in Vehicle Periphery Sensing Techniques Aimed at Realizing Autonomous Driving Progress is being made on vehicle periphery sensing, the greatest challenge to overcome for autonomous driving. This sensing can be broadly divided into the detection while driving of potential obstacles (such as other vehicles or pedestrians) and of lane markings (such as the white lines that indicate the lanes on a road). Detection of obstacles in front of the vehicle is performed using a stereo camera and millimeter-wave radars, while detection of obstacles in the immediate vicinity or nearby is done using SurroundEye, a system that uses four monocular cameras mounted on the edges of the vehicle. Lane marking detection uses the stereo camera and SurroundEye. Information about what is present in the region centered on the vehicle (which is essential for autonomous driving) can then be obtained by combining the results of this sensing using a sensor fusion technique and determining the vehicle position by matching to a map. This article describes the progress of development for these key technologies. Makoto Kudo, Ph.D. Shinji Kakegawa Takeshi Shima Akira Kuriyama Yoshitaka Uchida Shigenori Hayase 1. Introduction The industry has been competing to develop advanced driver-assistance systems (ADAS) since the announcement of Subaru Corporation s EyeSight * (ver.2) in 2010, with this competition becoming even more intense in recent years as the focus has shifted to autonomous driving. To coincide with sports events in 2020, the Japanese government is pursuing development in key areas required for automated driving * EyeSight is a registered trademark of Subaru Corporation. systems, with the aim of developing a next-generation traffic system in collaboration with the Tokyo Metropolitan Government. Hitachi is also proceeding with research and development based on a similar roadmap. Driving automation level is classified into six levels (levels 0 to 5) based on the role that the system plays in driving the vehicle. Levels 1 and 2 correspond to ADAS functions, with the driver remaining responsible for driving. Responsibility for driving shifts to the system at levels 3 and higher, and it is these levels that correspond to autonomous driving as the term is commonly used. 52.

Host vehicle Figure 1 Vehicle Periphery Sensing and Localization Autonomous driving requires not only the recognition of nearby vehicles and white lines, but also the use of map-matching for localization (to formulate a trajectory plan based on the estimated vehicle position). Recognition of nearby objects Recognition of lane Map matching Autonomous driving requires that the following three processes be executed rapidly and repetitively. (1) Sensing of vehicle periphery and localization (see Figure 1) (2) Trajectory planning (3) Vehicle control Obviously, safe autonomous driving cannot be achieved if the area around the vehicle monitored by (1) is too small, regardless of how well the second two processes are performed. This article describes the progress of technology development at Hitachi in relation to vehicle periphery sensing and localization, the greatest challenge to more advanced autonomous driving. Specifically, the article describes improvements in stereo camera performance, ultra-miniaturization of millimeter-wave radar, further development of SurroundEye, and a localization technique that uses lane markings and map matching. 2. Improvements to Stereo Camera Performance In addition to providing gray-scale images, a stereo camera can also serve as a distance sensor. The parallax (disparity between the left and right camera views) can be measured and the distances calculated using triangulation. The ability to combine recognition of gray-scale patterns and three-dimensional shapes at the level of individual pixels and at high resolution is an advantage over systems with only monocular cameras or with monocular cameras and millimeterwave radars. Recognizing the importance of a vehicle-mounted stereo camera for front sensing, Hitachi Automotive Systems, Ltd. is putting a lot of effort into its development. With a view to autonomous driving, recent years have seen particular demand for sensing functions that can identify the drivable area. The term drivable area used here means not only the vehicle s lane, as identified by conventional lane sensing, but also the physical region within which the vehicle is able to drive based on detecting the edge of the road. Identification of the drivable area is essential for autonomous driving so that the vehicle can plan its own trajectory. Given the performance factors described above, this function can be realized using a single stereo camera unit because the characteristics of the stereo camera are suitable for space recognition at high resolution. For these reasons, Hitachi has developed a technique for using a stereo camera to detect the road edge and drivable area. Figure 2 shows the road edges and drivable areas detected by the technique. This shows the many different types of road edges that can be identified as well as the drivable area on the road. One of the challenges was a tendency for shapes on the road boundary to be indistinct, such as the presence of low features such as curbstones or intermittent ones like poles. To overcome this, Hitachi s technique defines the calculation of boundary shapes as an optimal path problem that works by performing a comprehensive search for candidate boundary shapes within the space and identifying the shapes that maximize boundary likelihood (1). This boundary likelihood is defined in terms of a feature value based on height above road surface and continuity in the depth direction. In this way, reliable estimation even in the presence of features with indistinct boundaries was achieved by adopting a technique based on solving optimal path problems. Road edge detection was also further improved by developing a Hitachi Review Vol. 67, No. 1 052 053 53.

Figure 2 Detection of Road Edges and Drivable Areas The red lines indicate the detected road edges. The technique can detect a wide variety of different types of road edges, including guardrails, banks, curbstones, poles, bushes, sidewalls, grass, and the sides of moving and parked vehicles. The green regions indicate the detected drivable area. Guardrail Bank Sidewall Curbstone Curbstone Curbstone Grass Moving vehicle Bush Pole Parked vehicle Parked vehicle road surface estimation method that served as a base technique for quickly and accurately determining the height of boundary regions (2). Figure 3 shows an overview of road edge detection. Figure 3 (a) shows an elevation map, which indicates the height of boundary regions above road surface, generated from distance measurements made by a stereo camera and road surface estimation. Figure 3 (b) shows a boundary feature map generated by transforming the elevation map that indicates the likelihood of a road boundary being present (defined as Figure 3 Overview of Road Edge Detection Respectively, images (a), (b), and (c) show the elevation map (map of height above road surface displayed on a bird s eye view), the boundary feature map (map of likelihood of road boundary being present displayed on a bird s eye view), and the results of road edge detection. Image horizontal Disparity the boundary above a certain height that is closest, as seen from the vehicle). Figure 3 (c) shows the results of road edge detection in the form of the optimal paths calculated through the area of the boundary feature map projected onto an image of the road. The drivable area is identified by calculating the region enclosed by the left, right, forward, and backward boundaries. Two new vehicle control applications were developed using this technique and demonstrated on actual vehicles both inside and outside the company (1), (3). (1) Road departure prevention An application for predictive safety that provides steering control assistance when there is a high likelihood of a collision with an obstacle in the horizontal direction. (2) Automatic obstacle avoidance function An application for use in autonomous driving that automatically avoids obstacles on the road and selects the path through the drivable area. (a) Image horizontal Disparity Extract boundary feature Find optimal path 3. Miniaturization and Cost Reduction for Millimeter-wave Radar (b) (c) Image horizontal Image vertical Road edge Optimal path Project onto image Road edge Because it uses radio waves, millimeter-wave radar is less subject to environmental factors such as rain, fog, darkness, or reflected light. Unfortunately, the way it works makes identification of detected objects difficult. These characteristics mean that, in autonomous driving, it is most commonly used in tandem with camera-based sensors, leading to demand for smaller size and lower cost. 54.

FEATURED ARTICLES (a) 47 mm 19 mm 45 mm (b) 28 mm Patch antenna Lens Horn Figure 4 Conventional and Newly Developed Antenna Design Respectively, diagrams (a), (b), and (c) show a conventional microstrip patch antenna array, the side view of a horn lens antenna design, and the side view of the new low-profile horn lens antenna design. The latter two antenna substrate designs reduce costs because they reduce the size of the bottom surface opening for the horn contained in each patch. (c) 45 mm 14 mm Patch antenna One way to make long-distance millimeter-wave radar smaller and less expensive is to shrink the size of the antenna substrate that accounts for several tenths of the cost. Hitachi sought to make the antenna substrate considerably smaller without compromising its beam profile by replacing the conventional microstrip patch array antenna [see Figure 4 (a)] with a horn and lens antenna [see Figure 4 (b)]. Unfortunately, the height of this antenna design proved to be a problem. In this configuration, the patch antenna that serves as the radiative source is located close to the focal point of the lens. As the focal length is approximately proportional to the lens diameter, Hitachi split the lens in two as shown in Figure 4 (c). The results of measurements indicated that this provides an antenna efficiency approximately 1.5 times that of the previous antenna (or twice that of a microstrip patch antenna) while halving its height. It was estimated that this would enable the antennas to have less than half the volume of competing products from other vendors at approximately 30% less cost than for a microstrip patch array antenna (4). 4. SurroundEye Enhancements SurroundEye is the name of a function for augmenting the driver s view of the area around the vehicle by synthesizing a single bird s eye view of this area from images captured by cameras mounted on the front, rear, and sides (such as on the front grill, rear bumper, or side mirrors) (see Figure 5). 360-degree sensing White line detection in parking spaces Warning of rear blind spots Warning of moving objects Warning of road departure Parking assist Lane-keeping assistance system Communications integration ADAS integration Automatic parking Information-based driving assistance Remote parking Automatic valet parking ADAS safety systems Restricted automated driving at low speed Vehicle-control-based driving assistance Visibility and warning assistance Figure 5 SurroundEye Enhancements Improvements are planned for the functions and performance of SurroundEye because of its role as a vital sensor for autonomous driving, including use as a system for providing better visibility when parking. SurroundEye Visibility enhancement 2010 2015 2020 ADAS: advanced driver assistance system Hitachi Review Vol. 67, No. 1 054 055 55.

Figure 6 Combination of Sensing Results and Map-based Lane Boundaries This technique automatically augments vehicle location, using the location of lane boundaries to improve horizontal localization and matching with corner layout to improve localization in the direction of travel. Record of lane boundary detection results Vehicle position Results of lane boundary detection by a stereo camera Results of lane boundary detection by SurroundEye Lane boundaries on map Already available on the market, SurroundEye has come to be seen as an essential part of the human machine interface (HMI). Clarion Co., Ltd. has developed a new display format called SurroundEye 3D View, and a warning/ assistance system that uses SurroundEye image recognition techniques to perform 360-degree sensing of the surrounding area to detect moving and other objects. The latter system is intended to meet demand from the automotive camera market for providing better visibility. A contribution has also been made to intelligent parking assist (IPA) systems that use parking space detection in car parks. Recently, industry, government, and academia have been working on the development of automatic parking systems such as valet parking as well as on automated driving systems to provide the public with safe and secure driving conditions, including a reduction in vehicle accidents. In addition to working on developments that will facilitate advances in automatic parking systems (such as remote parking) by using communication functions for interoperation with infrastructure, Hitachi also intends, in the future, to consider participating in automated driving systems by expanding the scope of sensing to develop technologies for things like the recognition of lane markings, road signs, and road markings, and for identifying and avoiding collisions with pedestrians at intersections (5). 5. Localization Maps are essential for autonomous driving. For example, information from maps is needed to preemptively identify intersections or places where roads diverge or converge so as to determine the correct lane to drive in, and to control vehicle speed based on the nature of the upcoming road (such as the presence of curves or hills). However, to make good use of the map, the vehicle must be able to determine precisely where on the map it is currently located. Unfortunately, position information obtained by the global positioning system (GPS) commonly used on vehicles is only accurate to within a few meters, with additional error arising when GPS signal reception is poor. In order to determine the precise vehicle location under these conditions, Hitachi has developed a technique for estimating vehicle location by comparing information obtained by external sensors against the map (see Figure 6). The improvement in localization accuracy was demonstrated in trials conducted on a test course. Whereas GPS had an error of about 4 m, the map-matching technique improved accuracy to within about 1 m. In the future, Hitachi plans to investigate using the technique for dealing with sharp corners where lane boundary detection is difficult, and for estimating vehicle movement based on changes in the camera images (6) (8). 6. Conclusions This article has described the latest work being undertaken to enhance the vehicle periphery sensing techniques needed for autonomous driving. In the future, Hitachi intends to continue combining these technologies to achieve advances in autonomous driving. 56.

FEATURED ARTICLES References 1) S. Kakegawa et al., Road Edge Detection using Dynamic Programming with Application to Free Space Computation, 25th Aachen Colloquium Automobile and Engine Technology (Oct. 2016). 2) S. Kakegawa et al., Road Surface Segmentation based on Vertically Local Disparity Histogram for Stereo Camera, International Journal of Intelligent Transportation Systems Research (Mar. 2017). 3) T. Sugawara et al., Applications of Road Edge Information for Advanced Driver Assistance Systems and Autonomous Driving, 21st International Forum on Advanced Microsystems for Automotive Applications (Sep. 2017). 4) T. Shima et al., Fundamental Technologies Driving the Evolution of Autonomous Driving, Hitachi Review, 65, pp. 426 432 (Nov. 2016). 5) T. Ogata et al., Development of Autonomous Parking System Based on Wide View Cameras, Journal of Society of Automotive Engineers of Japan, Vol. 71, No. 2, pp. 55 60 (Feb. 2017) in Japanese. 6) Y. Duan et al., Lane-level Vehicle Positioning by Integrating Digital Map and Visual Sensor Data, FISITA2016 F2016- ACVA-003 (Sep. 2016). 7) Y. Tanaka et al., Lane-level Vehicle Positioning by Integrating Digital Map and Visual Sensor Data, 2017 JSAE Annual Congress (Spring) 20175030 (May 2017) in Japanese. 8) Hitachi News Release, Hitachi Automotive Systems Develops Vehicle Localization Estimation Technology that Increases the Stability of Autonomous Driving through the Integration of GPS Positional Information, Camera Image Information, and the Industry s First Function Addressing Loss of Sight of Lane, (Sep. 2016), http://www.hitachi.com/ New/cnews/month/2016/09/160921b.html Authors Makoto Kudo, Ph.D. Advanced Sensing Technology Development Department, Advanced Development Center, Technology Development Division, Hitachi Automotive Systems, Ltd. Current work and research: Development of autonomous driving technology. Society memberships: Society of Automotive Engineers of Japan, Inc. (JSAE), and The Institute of Electronics, Information and Communication Engineers (IEICE). Shinji Kakegawa Smart Systems Research Department, Center for Technology Innovation Controls, Research & Development Group, Hitachi, Ltd. Current work and research: Development of image recognition technology for autonomous driving. Takeshi Shima Smart Systems Research Department, Center for Technology Innovation Controls, Research & Development Group, Hitachi, Ltd. Current work and research: Development of automated driving systems and in-vehicle image recognition technology. Akira Kuriyama Advanced Sensing Technology Development Department, Advanced Development Center, Technology Development Division, Hitachi Automotive Systems, Ltd. Current work and research: Development of sensor fusion systems for advanced driving assistant systems and autonomous driving. Society memberships: IEICE. Yoshitaka Uchida Safety & Information System Development Department, Safety & Information Systems Business Promotion Division, Clarion Co., Ltd. Current work and research: Development of auto parking and automated driving systems, and in-vehicle camera system technology. Shigenori Hayase Advanced Sensing Technology Development Department, Advanced Development Center, Technology Development Division, Hitachi Automotive Systems, Ltd. Current work and research: Development of sensor fusion systems for advanced driving assistant systems and autonomous driving. Society memberships: IEICE. Hitachi Review Vol. 67, No. 1 056 057 57.