Computer vision, wearable computing and the future of transportation

Similar documents
A Winning Combination

Challenges and Solutions for Bundling Multiple DAS Applications on a Single Hardware platform.

Electronic Costing & Technology Experts

Choosing the Optimum Mix of Sensors for Driver Assistance and Autonomous Vehicles

The Building Blocks of Autonomous Control. Phil Magney, Founder & Principal Advisor July 2016

VSI Labs The Build Up of Automated Driving

Deep Learning for Autonomous Driving

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen

A.I in Automotive? Why and When.

CS6700: The Emergence of Intelligent Machines. Prof. Carla Gomes Prof. Bart Selman Cornell University

Teaching icub to recognize. objects. Giulia Pasquale. PhD student

23270: AUGMENTED REALITY FOR NAVIGATION AND INFORMATIONAL ADAS. Sergii Bykov Technical Lead Machine Learning 12 Oct 2017

Today I t n d ro ucti tion to computer vision Course overview Course requirements

HAVEit Highly Automated Vehicles for Intelligent Transport

Neural Networks The New Moore s Law

Embedding Artificial Intelligence into Our Lives

KÜNSTLICHE INTELLIGENZ JOBKILLER VON MORGEN?

Global Image Sensor Market with Focus on Automotive CMOS Sensors: Industry Analysis & Outlook ( )

The GATEway Project London s Autonomous Push

Digital Engines for Smart and Connected Cars By Bob O Donnell, TECHnalysis Research Chief Analyst

Driving Using End-to-End Deep Learning

Automated Testing of Autonomous Driving Assistance Systems

Deep Learning Overview

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. ECE 289G: Paper Presentation #3 Philipp Gysel

White paper on CAR150 millimeter wave radar

SIS63-Building the Future-Advanced Integrated Safety Applications: interactive Perception platform and fusion modules results

AI Application Processing Requirements

GPU ACCELERATED DEEP LEARNING WITH CUDNN

Artificial Intelligence and Deep Learning

INTRODUCTION TO DEEP LEARNING. Steve Tjoa June 2013

Introduction. BIL719 Computer Vision Pinar Duygulu Hacettepe University

Human-Centric Trusted AI for Data-Driven Economy

Virtual Homologation of Software- Intensive Safety Systems: From ESC to Automated Driving

arxiv: v1 [cs.lg] 2 Jan 2018

Perception platform and fusion modules results. Angelos Amditis - ICCS and Lali Ghosh - DEL interactive final event

Recognition: Overview. Sanja Fidler CSC420: Intro to Image Understanding 1/ 83

Data-Starved Artificial Intelligence

ROAD RECOGNITION USING FULLY CONVOLUTIONAL NEURAL NETWORKS

Technology trends in the digitalization era. ANSYS Innovation Conference Bologna, Italy June 13, 2018 Michele Frascaroli Technical Director, CRIT Srl

Executive summary. AI is the new electricity. I can hardly imagine an industry which is not going to be transformed by AI.

Large-scale Incremental Machine Learning for Robotics

Sketch-a-Net that Beats Humans

P1.4. Light has to go where it is needed: Future Light Based Driver Assistance Systems

Australian Institute for Machine Learning: Catching the wave of the next industrial revolution

Proposers Day Workshop

Virtual Worlds for the Perception and Control of Self-Driving Vehicles

Learning Pixel-Distribution Prior with Wider Convolution for Image Denoising

Biologically Inspired Computation

Physics Based Sensor simulation

Artificial Intelligence Machine learning and Deep Learning: Trends and Tools. Dr. Shaona

Sensing, Computing, Communication

Making Vehicles Smarter and Safer with Diode Laser-Based 3D Sensing

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu

[Overview of the Consolidated Financial Results]

Automotive and Discrete Group (ADG)

Domain Adaptation & Transfer: All You Need to Use Simulation for Real

Revision of the EU General Safety Regulation and Pedestrian Safety Regulation

ADAS/AD Challenge. Copyright 2017, dspace GmbH

ADAS COMPUTER VISION AND AUGMENTED REALITY SOLUTION

Jeff Bezos, CEO and Founder Amazon

Radio Deep Learning Efforts Showcase Presentation

Minimizing Distraction While Adding Features

ADAS Development using Advanced Real-Time All-in-the-Loop Simulators. Roberto De Vecchi VI-grade Enrico Busto - AddFor

Machine Learning for Intelligent Transportation Systems

Sensing, Computing, Communication

Driver Assistance and Awareness Applications

FLASH LiDAR KEY BENEFITS

Semantic Localization of Indoor Places. Lukas Kuster

Transer Learning : Super Intelligence

Sensing, Computing, Communication

What We Talk About When We Talk About AI

Application Areas of AI Artificial intelligence is divided into different branches which are mentioned below:

Final Report Non Hit Car And Truck

Artificial intelligence: past, present and future

THE AI REVOLUTION. How Artificial Intelligence is Redefining Marketing Automation

Monte Carlo Tree Search

Addressing the Uncertainties in Autonomous Driving

Transformation to Artificial Intelligence with MATLAB Roy Lurie, PhD Vice President of Engineering MATLAB Products

Behaviour-Based Control. IAR Lecture 5 Barbara Webb

Automotive In-cabin Sensing Solutions. Nicolas Roux September 19th, 2018

DRIVING TO GET AHEAD WITH AI.

Intelligent driving TH« TNO I Innovation for live

OPEN CV BASED AUTONOMOUS RC-CAR

Author s Name Name of the Paper Session. DYNAMIC POSITIONING CONFERENCE October 10-11, 2017 SENSORS SESSION. Sensing Autonomy.

Autonomous driving made safe

Unlock the power of location. Gjermund Jakobsen ITS Konferansen 2017

ECE 599/692 Deep Learning Lecture 19 Beyond BP and CNN

Making a Difference in 2017 IPOs & CES

Deep Learning. Dr. Johan Hagelbäck.

What Is And How Will Machine Learning Change Our Lives. Fair Use Agreement

Artificial intelligence, made simple. Written by: Dale Benton Produced by: Danielle Harris

The Virtues of Virtual Reality Artur Filipowicz and Nayan Bhat Princeton University May 18th, 2017

Speed Traffic-Sign Number Recognition on Low Cost FPGA for Robust Sign Distortion and Illumination Conditions

Development of a 24 GHz Band Peripheral Monitoring Radar

Industrial Keynotes. 06/09/2018 Juan-Les-Pins

신경망기반자동번역기술. Konkuk University Computational Intelligence Lab. 김강일

What we are expecting from this presentation:

How Machine Learning and AI Are Disrupting the Current Healthcare System. Session #30, March 6, 2018 Cris Ross, CIO Mayo Clinic, Jim Golden, PwC

Advances in Vehicle Periphery Sensing Techniques Aimed at Realizing Autonomous Driving

Digital image processing vs. computer vision Higher-level anchoring

Transcription:

Computer vision, wearable computing and the future of transportation Amnon Shashua Hebrew University, Mobileye, OrCam 1

Computer Vision that will Change Transportation Amnon Shashua Mobileye 2

Computer Vision Making Computer See and Understand What they See Major branch of A.I. goes together with Machine Learning. Major progress in the last decade. Human level perception is achievable in some narrow domains (face recognition, object detection). Camera: lowest cost sensor with highest information density. 3

Avoiding Collisions: Under the Hood 4

Technology: Machine Perception & System-on-Chip Lane Detection Lane Departure Warning Lane Keeping and Support Vehicle Detection Forward Collision Warning Adaptive Cruise Control Traffic Jam Assistant Emergency Braking Pedestrian Detection Collision Warning Emergency Braking Traffic Sign Recognition Intelligent High Beam Control Autonomous Driving Free-space Estimation Environmental Model Holistic Path Planning General Object Detection Road Profile Reconstruction Traffic Light Detection Surround Vision (Hyper-AVM) Multi-focal configurations 360 awareness # 5

Computer Vision Disruption EyeQ1 EyeQ1 EyeQ2 EyeQ2 1X 6X 48X EyeQ3 180nm 180nm 90nm 90nm 40nm 2007 2008 2010-11 2013 2015 First Camera/Radar Fusion First Bundling of LDW, IHC, TSR First Pedestrian AEB First Camera Only ACC and Traffic Jam Assistant First Camera Only Full Auto Braking (AEB) First Camera Only FCW First Camera Only AEB (partial braking) 6

The Camera Disruption The functional territory taken by the camera is rapidly increasing: WHY? 2011: warning against collisions 2013: ACC, partial brake AEB, TJA 2015: full brake AEB Richest source of raw data about the scene - only sensor that can reflect the true complexity of the scene. The lowest cost sensor - nothing can beat it, not today and not in the future. Cameras are getting better - higher dynamic range, higher resolution Radars/Lidar/Ultrasonic: for redundancy, robustness 7

EyeQx Vision Application Processor 2.5T Flop Processor 256G Flop Processor 4 x 64 MAC/cycle @ 0.5GHZ utilization (CNN): 0.9 effective MAC/s: 115G 6 x 76 MAC/cycle @ 1GHZ 2 x 384 MAC/cycle @ 1GHZ 2 x 8 MAC/cycle @ 1GHZ PMA-Program Macro Array VMP-Vector Microcode Processor MPC-Multi-Thread Processor Cluster 8

EyeQ System on Chip Roadmap Performance of EyeQ Chip has Increased Rapidly over Time Performance 384X EyeQ4 28nm PD-SOI 48X EyeQ3 40nm 6X 1X EyeQ1 180nm EyeQ2 90nm Production Year 9

Market Drivers Two Major Trends Evolution New Safety Rating Regulations Revolution Autonomous Driving Megatrend 10

11 Euro NCAP. All rights reserved. This content is excluded from our Creative Common license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 11

12 Euro NCAP. All rights reserved. This content is excluded from our Creative Common license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 12

Mobileye in Numbers 2007-2012: 1,000,000 EyeQ 2013: 1,300,000 EyeQ 2014: 2,700,000 EyeQ H1 2015: ~2.5M EyeQ 2010: 36 car models, 7 auto-makers 2014: 160 car models, 18 auto-makers 2016: 240 car models, 25 auto-makers 13

Increasing Awareness: Hyundai Super Bowl 2014 Commercial Still running on Times Square (as of 5/2015) Hyundai Motor Company. All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 14

Autonomous Emergency Braking Volvo S60 - launched 5/2010 - tests by Polish warriors AutomotiveBlog.pl. All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 15

ADAS 2016-2020 16

Two Paradigms for Achieving Autonomous Driving Pre-drive recording of 360 surround 3D HD Maps Sparse recording No Recording Google Mobileye Store & Align Sense & Understand 17

Leap I: Human Level Perception Human Level Perception is possible - already achieved in narrow domains. ADAS -> HLP requires: Extending list of objects (cars at all angles, general objects, ~1000 traffic signs, traffic lights, ) Using context to predict path ( holistic path planning ) Detailed Road interpretation: free-space, curbs, barriers, guard rails, construction, highway exits, Deep Layered Networks is the tool required for the leap. 18

The Need for Context (the rise of Deep Layered Networks) Path planning: fuse all the information available from the image, not only lane marks. Environmental Model: ultimately a category label for every pixel in the image 3D Model for Vehicles (VD at any angle, Viewed from any angle). Scene Recognition : Stop-line, Bumps, Road Surface 19

Deep Networks 20

Convolutional Neural Network Krizhevsky, A., Sutskever, I., & Hinton, G. (2012) 60M parameters 832M MAC ops Zeiler & Fergus, 2013 upper diagram A. Krizhevsky et al (2012 NIPS conference); lower diagram Matthew Zeiler and Rob Fergus (2014 ECCV conference). All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 21

Breakthroughs in object recognition Jia Deng et al (2009 CVPR conference). All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. Imagenet: 1000 classes, 1.2M images Top 5 err. 2011 Top 5 err, 2012 25.8%. 1.Krizhevsky-et-al 16.4% 2. ISI 26.2% 22

Deep Nets in computer vision: Follow-up Imagenet object recognition competitions Recent results: Number of Deep Net approaches / Total Top 5 error (%) Winning Team, Year Top 5 err. (%) Team/Company 1 / 6 16.4 Supervision Kizhevsky et al, 2012 17 / 24 11.7 Clarifai Zeiler & Fergus, 2013 31 / 32 6.66 GoogLe Net Szegedy et al, 2014 6.8 VGG,Simonyan 14 5.98 Baidu, Wu 15 4.94 Microsoft, He 15 4.82 Google, Ioffe 15 Human: 5.1% (estim) Note: Error was 25.8% in 2011! All subsequent years : DNN solutions Wide adoption in industry: Google, Microsoft, Baidu, Apple, Nuance, Mobileye, etc integrate deep network solutions 23

Human Level Face Recognition Labeled Faces in the Wild LFW benchmark 99.70% 99.62% NUS-LV* Baidu* 99.50% Face++, Megvii 99.47% DeepId2+, CUHK Deep Nets 97.35% DeepFace, Facebook, 2014 91.37% LBP/SVM *Newest results, no publications yet Human performance ~97.5% upper image, Taigman et al (2014 CVPR conference); lower image Wang et al (2009 CVPR conference). All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 24

Breakthrough s in Speech and NLP Baidu Research. All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 25

Potential Impact of DNN for Automotive Networks are at their best for multi-class problems - enables a rich vocabulary of objects (vehicles, pedestrians, types-of, traffic signs, etc.) Networks are very good at using context - holistic perception. Case in point: Path Planning. Network design is ideal for pixel-level labeling - objects that do not fit into a bounding-box. Examples, barriers, curbs, guard-rails, Case in point: Semantic Free-space. Networks can be used for Sensor Integration and Control decisions. The classical control point can be determined using a holistic process. 26

Challenges for using DNN for Automotive Networks are very large ~1.5B parameters Require huge training sets Not real-time driven Success for easy problems: Object detection. Academic research on higher-level perception (like pixel-level labeling) are sketchy. 27

DNNs at Mobileye 28

The Need for Context (the rise of Deep Layered Networks) Path planning: fuse all the information available from the image, not only lane marks. Environmental Model: ultimately a category label for every pixel in the image 3D Model for Vehicles (VD at any angle, Viewed from any angle). Scene Recognition : Stop-line, Bumps, Road Surface 29

Holistic Path Planning (HPP) 30

Path Planning using Holistic Cues 31

Path Planning using Holistic Cues 32

HPP: Increasing Availability of Road 33

Semantic Free Space (SFS) 34

Free Space through Pixel Labeling 35

Free Space through Pixel Labeling 36

Free Space through Pixel Labeling 37

Free Space through Pixel Labeling 38

Free Space through Pixel Labeling 39

Free Space through Pixel Labeling 40

Free Space through Pixel Labeling 41

SFS from Various Viewpoints and Fields of View 42

SFS from Various Angles 43

SFS from Various Angles 44

SFS from Various Angles 45

SFS from Various Angles 46

3D Modeling of Vehicles (3DVD) 47

3DVD 48

3DVD 49

Scene Recognition 50

Bump Detection (non-geometric) 51

Bump Detection (non-geometric) 52

Long Range Stop Line 53

Lane Assignment 54

Road Surface Recognition 55

Traffic Light Detection 56

TFL: main building blocks Detect Traffic Lights (some are country specific) Decide Relevancy for each TFL in a Junction Detect Stop Line Detect Road Markings Decide on Lane Assignment Scene Recognition: detection junctions in general (as a Prior) 57

TFL: main building blocks 58

Multiple Cameras 59

360 degrees sensing 60

Hardware Architecture 61

Automated Driving Reading Rainbow. All rights reserved. This content is excluded from our Creative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/. 62

Impact of Autonomous Driving Hands-free on Highways (no lane change) - Now on Tesla, 2016 GM, Audi, Driver has primary responsibility (and Alert) Highway to Highway: on and off-ramps executed autonomously. Early 2016. Driver has primary responsibility (and Alert) ~2018-2020: Driver responsible but not alert. Driver is attendant (transition from primary responsibility to Monitoring - like in Aviation). The beginnings of disruption. ~2020-2022: Driverless cars without passengers. Big disruption. ~2025-2030: No driver. Transformative. 63

Automated Driving 64

MIT OpenCourseWare https://ocw.mit.edu Resource: Brains, Minds and Machines Summer Course Tomaso Poggio and Gabriel Kreiman The following may not correspond to a particular c ou rse o n MIT OpenCourseWare, but has been provided by the author as an individual learning resource. For information about citing these materials or our Terms of Use, visit: https://ocw.mit.edu/terms.