Deep Learning for Autonomous Driving

Similar documents
REINFORCEMENT LEARNING (DD3359) O-03 END-TO-END LEARNING

신경망기반자동번역기술. Konkuk University Computational Intelligence Lab. 김강일

23270: AUGMENTED REALITY FOR NAVIGATION AND INFORMATIONAL ADAS. Sergii Bykov Technical Lead Machine Learning 12 Oct 2017

Tutorial of Reinforcement: A Special Focus on Q-Learning

Reinforcement Learning for CPS Safety Engineering. Sam Green, Çetin Kaya Koç, Jieliang Luo University of California, Santa Barbara

TUD Poker Challenge Reinforcement Learning with Imperfect Information

Creating an Agent of Doom: A Visual Reinforcement Learning Approach

Computer vision, wearable computing and the future of transportation

VISUAL ANALOGIES BETWEEN ATARI GAMES FOR STUDYING TRANSFER LEARNING IN RL

ARGUING THE SAFETY OF MACHINE LEARNING FOR HIGHLY AUTOMATED DRIVING USING ASSURANCE CASES LYDIA GAUERHOF BOSCH CORPORATE RESEARCH

Reinforcement Learning in Games Autonomous Learning Systems Seminar

CS 380: ARTIFICIAL INTELLIGENCE MONTE CARLO SEARCH. Santiago Ontañón

ADAS Development using Advanced Real-Time All-in-the-Loop Simulators. Roberto De Vecchi VI-grade Enrico Busto - AddFor

Radio Deep Learning Efforts Showcase Presentation

A.I in Automotive? Why and When.

Stanford Center for AI Safety

Reinforcement Learning Agent for Scrolling Shooter Game

Reinforcement Learning Simulations and Robotics

CS221 Project Final Report Deep Q-Learning on Arcade Game Assault

DeepMind Self-Learning Atari Agent

TD-Leaf(λ) Giraffe: Using Deep Reinforcement Learning to Play Chess. Stefan Lüttgen

William Milam Ford Motor Co

Online Interactive Neuro-evolution

AI Application Processing Requirements

11/13/18. Introduction to RNNs for NLP. About Me. Overview SHANG GAO

Heads-up Limit Texas Hold em Poker Agent

Success Stories of Deep RL. David Silver

CandyCrush.ai: An AI Agent for Candy Crush

Decision Making in Multiplayer Environments Application in Backgammon Variants

Intelligent Technology for More Advanced Autonomous Driving

Proactive Indoor Navigation using Commercial Smart-phones

The next level of intelligence: Artificial Intelligence. Innovation Day USA 2017 Princeton, March 27, 2017 Michael May, Siemens Corporate Technology

Learning via Delayed Knowledge A Case of Jamming. SaiDhiraj Amuru and R. Michael Buehrer

Proposers Day Workshop

VSI Labs The Build Up of Automated Driving

Game Playing for a Variant of Mancala Board Game (Pallanguzhi)

Background Pixel Classification for Motion Detection in Video Image Sequences

Playing CHIP-8 Games with Reinforcement Learning

CS 354R: Computer Game Technology

46.1 Introduction. Foundations of Artificial Intelligence Introduction MCTS in AlphaGo Neural Networks. 46.

Computer Go: from the Beginnings to AlphaGo. Martin Müller, University of Alberta

AI for Autonomous Ships Challenges in Design and Validation

ECE 517: Reinforcement Learning in Artificial Intelligence

Multimedia Forensics

Policy Teaching. Through Reward Function Learning. Haoqi Zhang, David Parkes, and Yiling Chen

Inteligência Artificial. Arlindo Oliveira

The Three Laws of Artificial Intelligence

V2X-Locate Positioning System Whitepaper

CS440/ECE448 Lecture 11: Stochastic Games, Stochastic Search, and Learned Evaluation Functions

Monte Carlo Tree Search and AlphaGo. Suraj Nair, Peter Kundzicz, Kevin An, Vansh Kumar

Mutliplayer Snake AI

AI Agent for Ants vs. SomeBees: Final Report

Artificial Intelligence and Deep Learning

Attention-based Multi-Encoder-Decoder Recurrent Neural Networks

Research Seminar. Stefano CARRINO fr.ch

Deep Learning. Dr. Johan Hagelbäck.

CS6700: The Emergence of Intelligent Machines. Prof. Carla Gomes Prof. Bart Selman Cornell University

Swing Copters AI. Monisha White and Nolan Walsh Fall 2015, CS229, Stanford University

Convolutional Networks Overview

Agent. Pengju Ren. Institute of Artificial Intelligence and Robotics

10703 Deep Reinforcement Learning and Control

Mini Project 3: GT Evacuation Simulation

Kernels and Support Vector Machines

Physics Based Sensor simulation

A Winning Combination

2048: An Autonomous Solver

An Approach to Semantic Processing of GPS Traces

Introduction to Machine Learning

Lecture 1 What is AI?

Philosophy. AI Slides (5e) c Lin

CS 4700: Foundations of Artificial Intelligence

Automatic Maneuver Recognition in the Automobile: the Fusion of Uncertain Sensor Values using Bayesian Models

Chapter 2 Distributed Consensus Estimation of Wireless Sensor Networks

An Artificially Intelligent Ludo Player

DYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION

Cellular automata applied in remote sensing to implement contextual pseudo-fuzzy classication - The Ninth International Conference on Cellular

Autonomous Automobile Behavior through Context-based Reasoning

Announcements. Homework 1. Project 1. Due tonight at 11:59pm. Due Friday 2/8 at 4:00pm. Electronic HW1 Written HW1

Deep Neural Network Architectures for Modulation Classification

Convolutional neural networks

Generating an appropriate sound for a video using WaveNet.

Semantic Segmentation on Resource Constrained Devices

Random Administrivia. In CMC 306 on Monday for LISP lab

COS 402 Machine Learning and Artificial Intelligence Fall Lecture 1: Intro

Transer Learning : Super Intelligence

Deep Learning for Human Activity Recognition: A Resource Efficient Implementation on Low-Power Devices

Plan for the 2nd hour. What is AI. Acting humanly: The Turing test. EDAF70: Applied Artificial Intelligence Agents (Chapter 2 of AIMA)

Applying Modern Reinforcement Learning to Play Video Games

Augmenting Self-Learning In Chess Through Expert Imitation

Presentation Overview. Bootstrapping from Game Tree Search. Game Tree Search. Heuristic Evaluation Function

DENSO www. densocorp-na.com

Learning to Play Love Letter with Deep Reinforcement Learning

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

arxiv: v1 [cs.lg] 2 Jan 2018

GESTURE RECOGNITION FOR ROBOTIC CONTROL USING DEEP LEARNING

A Spatiotemporal Approach for Social Situation Recognition

REAL TIME DIGITAL SIGNAL PROCESSING

Hierarchical Controller for Robotic Soccer

Adversarial Attacks on Face Detectors using Neural Net based Constrained Optimization

MSc(CompSc) List of courses offered in

Transcription:

Deep Learning for Autonomous Driving Shai Shalev-Shwartz Mobileye IMVC dimension, March, 2016 S. Shalev-Shwartz is also affiliated with The Hebrew University Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 1 / 23

Autonomous Driving Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 2 / 23

Autonomous Driving Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 3 / 23

Major Sub-Problems Sensing: Static objects: Road edge, curbs, guard rails,... Moving objects: Cars, pedestrians,... Semantic information: Lanes, traffic signs, traffic lights,... Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 4 / 23

Major Sub-Problems Sensing: Static objects: Road edge, curbs, guard rails,... Moving objects: Cars, pedestrians,... Semantic information: Lanes, traffic signs, traffic lights,... Mapping: Take me home Foresight Robustness Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 4 / 23

Major Sub-Problems Sensing: Static objects: Road edge, curbs, guard rails,... Moving objects: Cars, pedestrians,... Semantic information: Lanes, traffic signs, traffic lights,... Mapping: Take me home Foresight Robustness Driving Policy: Planning: e.g. Change lane now because you need to take a highway exit soon Slow down because someone is likely to cut into your lane Negotiation: e.g. Merge into traffic Roundabouts, 4-way stops Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 4 / 23

Challenges Everything should run in real time Difficult driving conditions Robustness: No margin for severe errors Unpredictable behavior of other drivers/pedestrians Beyond bounding box : need to understand the entire image and must utilize contextual information Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 5 / 23

Example: Free Space Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 6 / 23

Example: Free Space Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 7 / 23

Why Deep Learning? Why Learning? Manual engineering is not powerful enough to solve complex problems Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 8 / 23

Why Deep Learning? Why Learning? Manual engineering is not powerful enough to solve complex problems Why Deep Learning? To solve hard problems, we must use powerful models Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 8 / 23

Why Deep Learning? Why Learning? Manual engineering is not powerful enough to solve complex problems Why Deep Learning? To solve hard problems, we must use powerful models Why Are Deep Networks Powerful? Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 8 / 23

Why Deep Learning? Why Learning? Manual engineering is not powerful enough to solve complex problems Why Deep Learning? To solve hard problems, we must use powerful models Why Are Deep Networks Powerful? Theorem: Any function that can be implemented by a Turing machine in T steps can also be expressed by a T -depth network Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 8 / 23

Why Deep Learning? Why Learning? Manual engineering is not powerful enough to solve complex problems Why Deep Learning? To solve hard problems, we must use powerful models Why Are Deep Networks Powerful? Theorem: Any function that can be implemented by a Turing machine in T steps can also be expressed by a T -depth network Generalization: Deep networks are both expressive and generalizing (meaning that the learned model works well on unseen examples) Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 8 / 23

Additional Benefits of Deep Learning Hierarchical representations for every pixel ( pooling ) Spatial sharing of computation ( convolutions ) Accelerate computation by dedicated hardware ( lego ) Development language : by designing architectures and loss functions Modeling of complex spatial-temporal structures (using RNNs) Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 9 / 23

Is Deep Learning the Answer for Everything? Current algorithms fail for some trivial problems Parity of more than 30 bits Multiplication of large numbers Modeling of piece-wise curves... Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 10 / 23

Is Deep Learning the Answer for Everything? Current algorithms fail for some trivial problems Parity of more than 30 bits Multiplication of large numbers Modeling of piece-wise curves... Main reason: Training a deep network is computationally hard, and understanding when and why it works is a great scientific mystery Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 10 / 23

Is Deep Learning the Answer for Everything? Current algorithms fail for some trivial problems Parity of more than 30 bits Multiplication of large numbers Modeling of piece-wise curves... Main reason: Training a deep network is computationally hard, and understanding when and why it works is a great scientific mystery In practice: Deep learning is useful only when it is combined with smart modeling/engineering In practice: Domain knowledge is very helpful In practice: Architectural transfer only works for similar problems In practice: Standard training algorithms are not always satisfactory for automotive applications Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 10 / 23

Example: Typical vs. Rare Cases Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 11 / 23

Typical vs. Rare Cases Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 12 / 23

Failures of Existing Methods for Rare Cases State-of-the-art training methods are variants of Stochastic Gradient Descent (SGD) SGD is an iterative procedure At each iteration, a random training example is picked The random sample is used to estimate an update direction The weights of the network are updated based on this direction Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 13 / 23

Failures of Existing Methods for Rare Cases 10 0 objective 10 1 10 2 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 # of gradients 10 7 SGD finds an o.k. solution very fast, but significantly slows down at the end. Why? Rare mistakes: Suppose all but 1% of the examples are correctly classified. SGD will now waste 99% of its time on examples that are already correct by the model High variance, even close to the optimum Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 14 / 23

Requires Novel Algorithms 0.35 0.3 SGD FOL 0.25 % error 0.2 0.15 0.1 5 10 2 0 10 3 10 4 10 5 10 6 10 7 Iteration Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 15 / 23

Deep Learning for Driving Policy Input: Detailed semantic environmental modeling Output: Where to drive and an what speed Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 16 / 23

Reinforcement Learning Goal: Learn a policy, mapping from states to actions Learning Process: For t = 1, 2,... Agent observes state s t Agent decides on action a t based on the current policy Environment provides reward r t Environment moves the agent to next state s t+1 Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 17 / 23

Reinforcement Learning vs. Supervised Learning In SL, actions do not effect the environment, therefore we can collect training examples in advance, and only then search for a policy In SL, the effect of actions is local, while in RL, actions have long-term effect In SL we are given the correct answer, while in RL we only observe a reward Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 18 / 23

Reinforcement Learning: Existing Approaches Most algorithms rely on Markovity Next state only depends on current state and action Yields a Markov Decision Process (MDP) Can couple all the future into the so-called Q function Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 19 / 23

Reinforcement Learning: Existing Approaches Most algorithms rely on Markovity Next state only depends on current state and action Yields a Markov Decision Process (MDP) Can couple all the future into the so-called Q function Inadequate for driving policy Next state depends on other drivers Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 19 / 23

A Decomposable Approach for Reinforcement Learning Decompose the problem into 1 Supervised Learning problems Predict the near future Predict the intermediate reward 2 and then explicitly optimize over the policy using Recurrent Neural Network Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 20 / 23

A Decomposable Approach for Reinforcement Learning ˆrt DNNr at DNNN ŝt+1 + st st+1 t+1 Simulatort Simulatort+1 Simulatort+2 Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 21 / 23

Illustration Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 22 / 23

Summary The Deep Learning Revolution: Stunning empirical success in hard AI tasks Existing deep Learning algorithms fail for some trivial problems Prior knowledge is still here, it just shifted its shape A deeper theoretical understanding of deep learning is the most important open problem in machine learning... Shai Shalev-Shwartz (MobilEye) DL for Autonomous Driving IMVC 16 23 / 23