a literature review of steering angle prediction ... · for self-driving cars aatiq oussama * and...

1
A literature review of steering angle prediction algorithms for Self-driving cars Aatiq Oussama* and Talea Mohamed Information Processing Laboratory, Ben M’Sik Faculty of Sciences, Hassan 2 Casablanca University, Morocco [email protected], [email protected] A crucial re qu irem ent for inte ll ige nt, drive rles s cars is to man euv er with out mov in g out of its driv ab le r eg io n of th e ro ad. It is we ll k no wn th at ste erin g ang le cal cul atio n p lay s an im porta nt rol e in m ai ntai ni ng th e veh icl e in th e ce nter of the ro ad or wit hin th e bo un dar y lan es to me et safety critic al r eq uir eme nts. This w ork pr ese nts a rev iew of auto no mou s steeri ng tec hn iq ues for se lf driv in g cars w hic h is a r elat ivel y un exp lor ed task i n the fi el ds of comp uter vis io n, rob otics a nd m achi ne le arni ng. Our r ese arch i nve stig atio ns le ad us to conc lud e th at the c om bin ati on of Res Net 50 de ep arch itectur e w ith eve nt cam eras ca n b e assum ed t o g ive better pre dicti on of the du e whe el an gl e. An overv ie w of fut ure r ese arc h direction and applications is also given. qComputer vision based approach v Umamaheswari et al.’s method (2015) Umam ah esw ari et al. hav e us ed E ucl ide an m eth od to ca lcu late th e steer in g a ngl e. How ev er, it also seems to be lim ited to the cas e wh ere bot h of the bou nd arie s or eithe r one of them is visible, also to slow and smooth turns rather than sharp turns. A review of the Literature Abstract It would be our future researc h subject to study the adaptation of the output of event sensors with ‘’ ST-Conv + ConvLSTM + LSTM’’ model since it gives com petit ive r esults with traditional cameras. Architecture Input images EVA R MSE PilotNet Grayscale 0.161 9.02° CNN-LSTM Grayscale 0.300 8.19° ResNet50 architecture: (ImageNet initialisation) Events 0.826 4.10° A compar ais on m ad e bet we en 3 dee p n etwor ks (Table 2) shows that event ca mer as c omb in ed with state-of-the-art ResNet50 network gives a good results (RMSE = 4.10° ) Concluding Remarks and Future Work Computer vision based approach Neural network based approach It follows two steps: -Road boundary extraction (road region extraction) - Steering angle computation. End-to-End learning approach that requires: - Lots of data to train the network to make good predictions - High cost computing requirements qNeural network based approach v ALVINN (1989) ALVINN (1989) was am ong the very first attempts to us e a neural network for auto no mou s vehicle navigation. The a ppr oac h was c om pris ed of a s ha llo w netw ork th at pr edi cts acti ons fr om p ixe l i np uts applied to simple driving scenarios with few obstacles. Obtain Frame from the Camera Extract drivable road region using GMM-EM Extract Road Boundaries using Canny Edge No of boundaries extracted CASE 1: Both the boundaries are visible and it has been extracted using the mentioned road region extraction process. From the extracted boundaries, the slope of the boundaries are calculated. CASE 2: Only one o f the boundaries is visible within the field o f viewof the camera. In tha t case,the o ther boundary is usually taken to be the las t column o f pixels on the other side of the image. CASE 3: When both the boundaries are no t present, the vehicle is made to move straigh t along i ts pa th til l any one of the boundaries is identified 1 2 3 Cancel spurious angle transition Estimate Steering Angle The req uir ed ste eri ng a ng le w ill b e the d evi atio n of the p oint of int ersecti on of the boundaries from the orientation of the vehicle (midline of the image) as shown in Fig.4. The root-mean-square d error (RMSE) value is found to be 2.598. Fig 3:Road Region Extraction (a)Input Images (b) GMM-EM Fig 4: Steering angle calculation Steering angle prediction algorithms are divided into two categories as shown in Fig. 1. Fig 1: The two main approaches for steering angle prediction v PilotNet (2016) Bojarski et al. proposed a solution called PilotNet (itself inspired from ALVINN model (1989)). It is a co nvo luti on al ne ural net work ( CN N) bas ed ap pro ach th at ma ps r aw pix els fro m a s in gl e front-facing camera directly to steering commands (see Figs. 5 and 6). Fig. 5: PilotNet block diagram for training. Fig. 6: PilotNet CNN architecture. Steering angle prediction approaches vEvent camera – ResNet50 network (2018) Maq ue da et a l. intr od uce d a n ew ap pro ach th at pre dicts ste eri ng wh eel c omm an ds from a forward-looking DVS sensor mounted on a car. It was done by adapting state-of-the-art conv ol utio na l arch itectur es to the output of event sensors as shown in Fig. 7. Fig. 7: Block diagram of the event cameras-based approach . Fig 2: System architecture for steering angle prediction (Ranjith Rochan et al.) Table 2 : Performance analysis between ResNet50 deep architecture tha tprocesses event frames against two state-of-the-art learning approaches using grayscale frames. v Deep-Steering (2017) Deep-Steering (Chi and Mu) is a method that predicts steer ing angle in a stateful proc ess. It takes int o acco unt i nstant ane ou s mon ocu lar c amer a obs erv atio ns an d ve hicl e’s h istor ica l states. Experimentals results ( see Table 1) Model R MSE PilotNet 0.1604 Deep-Steering: ST-Conv + ConvLSTM + LSTM 0.0637 Table 1: Deep-Steering vs PilotNet performance. v Ranjith Rochan et al.’s method (2018) Ran jith R och an et al. (2 01 8) pro pos ed a no vel m etho d for steeri ng a ngl e calc ul atio n for auto nom ous veh icl es u sin g c omp uter v isi on tec hn iq ues of re lativ ely lo wer c om puti ng c ost (see Fig.2.)

Upload: others

Post on 21-Jul-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A literature review of steering angle prediction ... · for Self-driving cars Aatiq Oussama * and Talea Mohamed Information Processing Laboratory, Ben M’SikFaculty of Sciences,

A literature review of steering angle prediction algorithms for Self-driving cars

Aatiq Oussama* and Talea MohamedInformation Processing Laboratory, Ben M’Sik Faculty of Sciences, Hassan 2 Casablanca University, Morocco

[email protected], [email protected]

A crucial requirem ent for inte ll igent, driverles s cars is to maneuv er without mov ingout of its driv able r eg ion of the road. It is well k nown that steering angle calculation play san im portant role in m aintaining the vehicle in the center of the road or wit hin the boundar ylanes to meet safety critic al r equir ements. This work pr esents a rev iew of autonomoussteering tec hniques for se lf driv ing cars whic h is a r elat ively unexplor ed task in the fields ofcomputer vis ion, robotics and m achine learning. Our r esearch investigations lead us toconc lude that the c om bination of Res Net 50 deep arch itectur e with event cam eras can beassum ed t o give better prediction of the due wheel angle. An overv iew of fut ure r esearc hdirection and applications is also given.

qComputer vision based approach

v Umamaheswari et al.’s method (2015)

Umam aheswari et al. hav e us ed E ucl idean m ethod to calcu late the steer ing angle.Howev er, it also seems to be lim ited to the cas e where bot h of the boundaries or eitherone of them is visible, also to slow and smooth turns rather than sharp turns.

A review of the Literature

Abstract

It would be our future researc h subject to study the adaptation of the output of event sensorswith ‘’ST-Conv + ConvLSTM + LSTM’’ model since it g ives com petit ive r esults with traditionalcameras.

Architecture Input images EVA RMSEPilotNet Grayscale 0.161 9.02°

CNN-LSTM Grayscale 0.300 8.19°ResNet50 architecture: (ImageNet initialisation) Events 0.826 4.10°

A compar ais on m ade bet ween 3 deep networ ks (Table 2) shows that event cameras c ombinedwith state-of-the-art ResNet50 network gives a good results (RMSE = 4.10°)

Concluding Remarks and Future Work

Computer vision based

approach

Neural network based approach

It follows two steps:-Road boundary extraction (road region extraction) - Steering angle computation.

End-to-End learning approach that requires:- Lots of data to train the network to make good predictions- High cost computing requirements

qNeural network based approach

v ALVINN (1989)

ALVINN (1989) was am ong the very first attempts to us e a neura l network for autonomousvehicle navigation.The appr oac h was c om pris ed of a s hallow network that pr edicts actions fr om pixe l inputsapplied to simple driving scenarios with few obstacles.

Obtain Frame from the Camera

Extract drivable road regionusing GMM-EM

Extract Road Boundaries usingCanny Edge

No of boundariesextracted

CASE 1: Both the boundaries arevis ible and i t has beenextracted using thementioned road regionextraction process. From theextracted boundaries, theslope of the boundaries arecalculated.

CASE 2: Only one o f the boundariesis vis ible within the field o fv iew of the camera. In tha tcase, the o ther boundary isusually taken to be the las tcolumn o f pixels on theother side of the image.

CASE 3: When both theboundaries are no tpresent, the vehic le ismade to move straigh talong i ts pa th til l any oneof the boundaries isidentified

1

2

3

Cancel spurious angle transition

Estimate Steering Angle

The requir ed steering angle will be the deviation of the point of int ersection of theboundaries from the orientation of the vehicle (midline of the image) as shown in Fig.4.The root-mean-squared error (RMSE) value is found to be 2.598.

Fig 3:Road Region Extraction (a)Input Images (b) GMM-EM

Fig 4: Steering angle calculation

Steering angle prediction algorithms are divided into two categories as shown in Fig. 1.

Fig 1: The two main approaches for steering angle prediction

v PilotNet (2016)

Bojarski et al. proposed a solution called PilotNet (itself inspired from ALVINN model (1989)).It is a convolutional neural net work ( CNN) bas ed approach that maps r aw pix els from a s inglefront-facing camera directly to steering commands (see Figs. 5 and 6).

Fig. 5: PilotNet block diagram for training. Fig. 6: PilotNet CNN architecture.

Steering angle prediction approaches

vEvent camera – ResNet50 network (2018)

Maqueda et a l. intr oduced a new approach that predicts steering wheel c omm ands from aforward-looking DVS sensor mounted on a car.It was done by adapting state-of-the-art conv olutional arch itectur es to the output of eventsensors as shown in Fig. 7.

Fig. 7: Block diagram of the event cameras-based approach .

Fig 2: System architecture for steering angle prediction (Ranjith Rochan et al.)

Table 2 : Performance analysis between ResNet50 deep architecture tha tprocesses event frames against twostate-of-the-art learning approaches using grayscale frames.

v Deep-Steering (2017)

Deep-Steering (Chi and Mu) is a method that predicts steer ing angle in a statefu l proc ess. Ittakes int o account instant aneous monocular c amer a obs erv ations and vehicle’s h istor ica lstates.Experimentals results ( see Table 1)

Model RMSEPilotNet 0.1604

Deep-Steering: ST-Conv + ConvLSTM + LSTM 0.0637

Table 1: Deep-Steering vs PilotNet performance.

v Ranjith Rochan et al.’s method (2018)

Ranjith Rochan et al. (2018) propos ed a novel m ethod for steering angle calc ulation forautonom ous vehicles using c omputer v ision tec hniques of re lativ ely lower c om puting c ost(see Fig.2.)