a realistic simulation system for environmental understanding of … · 2018-11-06 · driving...

1
I am grateful to Dr. John M. Dolan for his mentorship and guidance throughout the summer, and to Zhiqian (Carolyn) Qiao and other lab members for their help. Huge thanks to Rachel Burcin for everything shes done for us and making this amazing RISS program happen, and to the enre RISS cohort for immense support. Lastly, I would like to thank TOBITATE! Young Ambassador Program for supporng me. Introducon Problem: The behavior of autonomous vehicles tends to be unpredictable, which could affect the behavior of other vehicles around them. Goal: Make autonomous vehicles drive as human do using Inverse Reinforcement Learning (IRL). Need for a realisc simulaon system - A simulator is required to collect hu- man driving data and to simulate human driving policies. - We target urban areas for this work. Collecng human driving data Collect human driving data by leng sub- jects drive in simulaon with steering wheel and pedals. Figure 1: Driving in simulaon Method Data Collecon Figure 2: (a) - (c) are images obtained from three cameras on the front of the car, (b) and (c) are images converted to a more human readable palee of colors [2], (d) shows speed and acceleraon of the hu- man-driven vehicle corresponding to throle/brake pedal input against me. We collected the following data at each step: three types of images in Figure 2, sensor data from a ray-cast based rotang LIDAR, and driving data of both the human-driven vehicle and vehicles around it. The depth map image and LIDAR sensor tell us how far the objects are, and semanc segmentaon tells us what the objects are. Acknowledgments References Future Work [1] Dorsa Sadigh, Shanker Sastry, Sanjit A. Seshia, An- ca D. Dragan. Planning for Autonomous Cars that Leverage Effects on Human Acons. In Proceedings of the Robocs: Science and Systems Conference (RSS), June 2016. [2] Alexey Dosovitskiy, German Ros, Felipe Codevilla, Antonio López, and Vladlen Koltun. CARLA: An open urban driving simulator. In Conference on Robot Learning (CoRL), 2017. (a) Normal vision camera (c) Semanc segmentaon camera (b) Depth map camera (d) Human driving data - Collect more human driving data - Generalize human drivers behavior model - Find preferred reward funcon - Mimic human drivers behavior for autonomous vehicles A Realisc Simulaon System for Environmental Understanding of Autonomous Driving Vehicles Akari Minami, Zhiqian Qiao, John M. Dolan Kyushu University, Carnegie Mellon University, Carnegie Mellon University Images & Sensor Data Vehicle Detecon & Tracking Human Driving Data Learn Human Driving Behavior Time step

Upload: others

Post on 13-Aug-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A Realistic Simulation System for Environmental Understanding of … · 2018-11-06 · driving policies. - We target urban areas for this work. ollecting human driving data ollect

I am grateful to Dr. John M. Dolan for his mentorship

and guidance throughout the summer, and to Zhiqian

(Carolyn) Qiao and other lab members for their help.

Huge thanks to Rachel Burcin for everything she’s

done for us and making this amazing RISS program

happen, and to the entire RISS cohort for immense

support. Lastly, I would like to thank TOBITATE!

Young Ambassador Program for supporting me.

Introduction

Problem: The behavior of autonomous

vehicles tends to be unpredictable, which

could affect the behavior of other vehicles

around them.

Goal: Make autonomous vehicles drive

as human do using Inverse Reinforcement

Learning (IRL).

➡ Need for a realistic simulation system

- A simulator is required to collect hu-

man driving data and to simulate human

driving policies.

- We target urban areas for this work.

➡ Collecting human driving data

Collect human driving data by letting sub-

jects drive in simulation with

steering wheel and pedals.

Figure 1: Driving in simulation

Method

Data Collection

Figure 2: (a) - (c) are images obtained from three cameras on the front of the car, (b) and (c) are images

converted to a more human readable palette of colors [2], (d) shows speed and acceleration of the hu-

man-driven vehicle corresponding to throttle/brake pedal input against time.

We collected the following data at each step: three types of images in Figure

2, sensor data from a ray-cast based rotating LIDAR, and driving data of both

the human-driven vehicle and vehicles around it. The depth map image and

LIDAR sensor tell us how far the objects are, and semantic segmentation tells

us what the objects are.

Acknowledgments

References

Future Work

[1] Dorsa Sadigh, Shanker Sastry, Sanjit A. Seshia, An-

ca D. Dragan. Planning for Autonomous Cars that

Leverage Effects on Human Actions. In Proceedings

of the Robotics: Science and Systems Conference

(RSS), June 2016.

[2] Alexey Dosovitskiy, German Ros, Felipe Codevilla,

Antonio López, and Vladlen Koltun. CARLA: An open

urban driving simulator. In Conference on Robot

Learning (CoRL), 2017.

(a) Normal vision camera

(c) Semantic segmentation camera

(b) Depth map camera

(d) Human driving data

- Collect more human driving data

- Generalize human driver’s behavior

model

- Find preferred reward function

- Mimic human driver’s behavior for

autonomous vehicles

A Realistic Simulation System for Environmental Understanding

of Autonomous Driving Vehicles

Akari Minami, Zhiqian Qiao, John M. Dolan

Kyushu University, Carnegie Mellon University, Carnegie Mellon University

Images & Sensor Data ➡ Vehicle Detection & Tracking

Human Driving Data ➡ Learn Human Driving Behavior

Time step