department of electrical and computer engineering | university of...

6
Unsupervised Learning of Motion Patterns of Rear Surrounding Vehicles Brendan Morris and Mohan Trivedi Computer Vision and Robotics Research Laboratory University of California, San Diego La Jolla, California 92093-0434 Email: {b1morris, mtrivedi}@ucsd.edu Abstract—Awareness to a vehicle’s surrounding is necessary for safe driving. Current surround technologies focus on the detection of obstacles in hard-to-view places but may neglect temporal information. This paper seeks the causes of dangerous situations by examining surround behavior. A general hierar- chical learning framework is introduced to automatically learn surround behaviors. By observing motion trajectories during natural driving, models of rear vehicle behaviors are obtained in an unsupervised fashion. The extracted behaviors are shown to correspond to typical driving scenarios, vehicle overtake and surround overtake, demonstrating the effectiveness of the learning framework. I. I NTRODUCTION The automobiles is ubiquitous in modern society but the access and mobility it grants comes at a staggering price. According to the 2009 World Health Organization’s Global Status Report on Road Safety [1], road crashes cause over 1.27 million fatalities a year and between 20 and 50 million non-fatal injuries. It is estimated that this will increase to 2.4 million fatalities a year by 2030, ranking road traffic injuries as the fifth leading cause of death. While not as bad as less developed countries, there were still 2.5 million injury accidents and 41 thousand motor vehicle related fatalities in the United States in 2008 [2]. These alarming statistics highlight the critical need for advanced vehicle safety systems. The US National Highway Traffic Safety Administration (NHTSA) report Traffic Safety Facts 2006 [3] found in pas- senger car crashes that the initial point of impact was the front 49.6%, either the right or left side 27.8% (left 14.2, right 13.6), or in the rear 21.2% of the time. Therefore, hightened surround awareness can directly affect safe driving and manuvering of an automobile. Successful systems currently in vehicles are active cruise control (ACC), which adapts vehicle speed to maintain a safe following distance, and lane assist, which can keep a vehicle in a lane or warn a driver when drifting. Unfortunately, such advanced systems exist only in the front of vehicles. The rest of the surround is usually covered by more basic detectors such as back-up cameras, sonar parking assist, and radar/vision blind spot warnings which just give an indication or an impeding object. This work seeks to aid the development of more advanced surround based safety systems which can give advance warning of impending situations. By predicting the behavior of surrounding vehicles, a driver is able to prepare earlier to ensure safer conditions. Recently, researchers have been investigating intent predic- tion of a driver using a looking-in and looking-out framework (LiLo) [4], where sensors simultaneous capture the surround- ing environment outside a vehicle, the vehicle dynamic state through onboard sensors, and the internal activities and state of a driver and other cockpit occupants. With careful integration of these varied knowledge sources, a driver’s intent to brake [5], turn [6], or change lanes [7], [8] can be assessed seconds in advance. This same type of prediction must be done for surrounding vehicles to understand what actions they might perform in the near future as well. An active surround safety system could then use the surround predictions warn a driver of possible dangerous situations for increased awareness. Unfor- tunately, the LiLo framework is not applicable because a driver can only observe a surround vehicle from the outside and does not have any internal indicators for intent. This work intends to learn surround patterns automatically to extract typical highway behaviors observed during natural driving. Rather than manually specifying expected behaviors of interest, a data-driven approach is utilized to learn behavior patterns in an unsupervised fashion from the surround trajectories of detected objects. II. HIERARCHICAL TRAJECTORY LEARNING Intelligent driver support systems rely on a number of differ- ent sensors to to observe the vehicle surround. Typical sensors are cameras, thermal infrared imagers, RADAR, LIDAR, and ultrasonic sensors. While each particular sensor has its own strengths and unique set of requirements for effective use, the end goal is the same, to detect obstacles. Once vehicles are detected they can be tracked to maintain a history of temporal motion. In order to be independent of sensor type, the learning framework we propose relies on low level motion information contained in tracks [9]. The main inputs for the learning engine are motion tra- jectories (Fig. 3). A trajectory F = {f 1 ,...,f T } compactly represents object motion during the time T it is observed. The flow vector f t represents tracked parameters at time t. In this work, f t =[x, y, u, v] T represents the xy position and velocity in either direction. This work mainly focuses on camera and RADAR based sensing for vehicle detection. In 2006, Sun et al. presented a review of vision based detection [10] for front looking cameras while techniques for top mounted 360

Upload: others

Post on 13-Aug-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Department of Electrical and Computer Engineering | University of …b1morris/docs/morris_icves2009.pdf · 2010-04-29 · Title: Untitled Created Date: 11/9/2009 1:20:34 PM

Unsupervised Learning of Motion Patterns

of Rear Surrounding Vehicles

Brendan Morris and Mohan Trivedi

Computer Vision and Robotics Research Laboratory

University of California, San Diego

La Jolla, California 92093-0434

Email: {b1morris, mtrivedi}@ucsd.edu

Abstract—Awareness to a vehicle’s surrounding is necessaryfor safe driving. Current surround technologies focus on thedetection of obstacles in hard-to-view places but may neglecttemporal information. This paper seeks the causes of dangeroussituations by examining surround behavior. A general hierar-chical learning framework is introduced to automatically learnsurround behaviors. By observing motion trajectories duringnatural driving, models of rear vehicle behaviors are obtainedin an unsupervised fashion. The extracted behaviors are shownto correspond to typical driving scenarios, vehicle overtakeand surround overtake, demonstrating the effectiveness of thelearning framework.

I. INTRODUCTION

The automobiles is ubiquitous in modern society but the

access and mobility it grants comes at a staggering price.

According to the 2009 World Health Organization’s Global

Status Report on Road Safety [1], road crashes cause over

1.27 million fatalities a year and between 20 and 50 million

non-fatal injuries. It is estimated that this will increase to 2.4

million fatalities a year by 2030, ranking road traffic injuries

as the fifth leading cause of death. While not as bad as

less developed countries, there were still 2.5 million injury

accidents and 41 thousand motor vehicle related fatalities

in the United States in 2008 [2]. These alarming statistics

highlight the critical need for advanced vehicle safety systems.

The US National Highway Traffic Safety Administration

(NHTSA) report Traffic Safety Facts 2006 [3] found in pas-

senger car crashes that the initial point of impact was the front

49.6%, either the right or left side 27.8% (left 14.2, right 13.6),

or in the rear 21.2% of the time. Therefore, hightened surround

awareness can directly affect safe driving and manuvering of

an automobile. Successful systems currently in vehicles are

active cruise control (ACC), which adapts vehicle speed to

maintain a safe following distance, and lane assist, which

can keep a vehicle in a lane or warn a driver when drifting.

Unfortunately, such advanced systems exist only in the front

of vehicles. The rest of the surround is usually covered by

more basic detectors such as back-up cameras, sonar parking

assist, and radar/vision blind spot warnings which just give an

indication or an impeding object. This work seeks to aid the

development of more advanced surround based safety systems

which can give advance warning of impending situations. By

predicting the behavior of surrounding vehicles, a driver is

able to prepare earlier to ensure safer conditions.

Recently, researchers have been investigating intent predic-

tion of a driver using a looking-in and looking-out framework

(LiLo) [4], where sensors simultaneous capture the surround-

ing environment outside a vehicle, the vehicle dynamic state

through onboard sensors, and the internal activities and state of

a driver and other cockpit occupants. With careful integration

of these varied knowledge sources, a driver’s intent to brake

[5], turn [6], or change lanes [7], [8] can be assessed seconds

in advance. This same type of prediction must be done for

surrounding vehicles to understand what actions they might

perform in the near future as well. An active surround safety

system could then use the surround predictions warn a driver of

possible dangerous situations for increased awareness. Unfor-

tunately, the LiLo framework is not applicable because a driver

can only observe a surround vehicle from the outside and does

not have any internal indicators for intent. This work intends

to learn surround patterns automatically to extract typical

highway behaviors observed during natural driving. Rather

than manually specifying expected behaviors of interest, a

data-driven approach is utilized to learn behavior patterns in an

unsupervised fashion from the surround trajectories of detected

objects.

II. HIERARCHICAL TRAJECTORY LEARNING

Intelligent driver support systems rely on a number of differ-

ent sensors to to observe the vehicle surround. Typical sensors

are cameras, thermal infrared imagers, RADAR, LIDAR, and

ultrasonic sensors. While each particular sensor has its own

strengths and unique set of requirements for effective use, the

end goal is the same, to detect obstacles. Once vehicles are

detected they can be tracked to maintain a history of temporal

motion. In order to be independent of sensor type, the learning

framework we propose relies on low level motion information

contained in tracks [9].

The main inputs for the learning engine are motion tra-

jectories (Fig. 3). A trajectory F = {f1, . . . , fT } compactly

represents object motion during the time T it is observed. The

flow vector ft represents tracked parameters at time t. In this

work, ft = [x, y, u, v]T represents the xy position and velocity

in either direction. This work mainly focuses on camera and

RADAR based sensing for vehicle detection. In 2006, Sun

et al. presented a review of vision based detection [10] for

front looking cameras while techniques for top mounted 360◦

Page 2: Department of Electrical and Computer Engineering | University of …b1morris/docs/morris_icves2009.pdf · 2010-04-29 · Title: Untitled Created Date: 11/9/2009 1:20:34 PM

Fig. 1. LISAQ collects video from 8 different cameras which capture theinterior and surround of the vehicle.

omni-directional cameras was presented by Gandhi and Trivedi

[11]. Fig. 1 gives example video to monitor both the inside

and outside of the ego-vehicle obtained while driving.

A diagram for the learning framework is presented in Fig.

2. On the left are 4 levels to abstract increasing information

while the right side shows the corresponding learning steps.

An activity is more completely described and modeled moving

down the levels of the diagram. The lowest activity level (Level

1) of goals are regions in space where vehicles tend to reside

and can be learned based on detections. Level 2 incorporates

sequential spatial information to create links between the

previous goals. In this stage, the spatial measurements from

trajectories are used to build clusters of typical routes. The

third learning phase combines dynamics and temporal char-

acteristics (Level 3 and 4) in one step using hidden Markov

models (HMMs). The HMM formulation conveniently decom-

poses activities into distinct atomic actions with a specified

duration and incorporates vehicle speed. The resulting models

provide the vocabulary to describe scene activity which allows

behavior identification, prediction, and anomaly detection.

III. LEARNING GOALS

Goals are places of interest (POI) in an image and indicate

start and end destinations. In addition, the POI are used to filter

noise trajectories that arise from incorrect tracking because

successfully tracked vehicles begin and end in clearly defined

goals. Only trajectories that pass the POI filtering are retained

for activity learning.

There are three different types of POI in a scene, entry,

exit, and stop. The entry and exit POI are the locations where

objects either appear or disappear, such as the edge of a

camera field of view, and correspond to the first, f1, and

last, fT , tracking point respectively. The stop POI indicate

where vehicles tend to idle or remain stationary. Stop points

are defined in two different ways, either as samples with zero

speed or those that are within a circle of radius R for more

than τ seconds [12]. The distribution of points from a given

POI type defines an interest zone.

A B

A B

A B

A B

POI

Routes

Activity

HMM

GMM

LCSS

Spectral Clustering

Route Merge

HMM Training

Fig. 2. Hierarchical learning framework for activity modeling. In the firstlevel the goals indicate noteworthy destinations and provide a mechanism forhigh level reasoning of behaviors. Spatially different routes to reach the samegoal describe different behaviors in the second level. Similar routes are furtherdistinguished based on dynamic information such as speed and decomposedinto a set of sequential actions to provide the finest description granularity inthe third level.

The zones are learned through a 2D Gaussian mixture mod-

eling (GMM) procedure. A zone∑Z−1

i=0 ziG([x, y]T , µi,Σi)is composed of Z Gaussians G([x, y]T , µ,Σ) of [x, y]T co-

ordinates learned using expectation maximization (EM) [13].

Using a density criterion [14], the number of zones in a

scene can be estimated without a priori knowledge. Tight

mixtures with high density indicate a true POI while low

density mixtures imply tracking noise (they are not localized).

Trajectories that do not begin and end in a POI are removed

from the training database and considered a tracking error.

An example of POI goals learned through this procedure

is shown in Fig. 5(a). The start zones, colored green, indicate

regions where vehicles are first detected. The stop zones in red,

are the areas where tracking ends. Stop zones are denoted with

yellow ellipses. The black ellipses indicate low density noise

mixtures which can be ignored as false POI.

IV. LEARNING ROUTES

The second level of activity understanding calls for the

definition of spatial routes which separate different ways to get

between goals. These routes can be learned in unsupervised

fashion through track clustering. Clustering automatically ex-

tracts the most typical trajectory patterns and only relies on

the definition of similarity between tracks.

In this work, spatial routes are learned by first finding the

distance between training trajectory pairs using a measure

derived from the longest common sub-sequence (LCSS). A

similarity matrix is formed from the pairwise distances and

spectrally decomposed to generate clusters. Since the number

of clusters is not known a priori, the count is estimated through

an agglomerative merge procedure.

A. LCSS Distance

LCSS is an alignment tool for unequal length data that is

robust to noise and outliers because not all points need to

Page 3: Department of Electrical and Computer Engineering | University of …b1morris/docs/morris_icves2009.pdf · 2010-04-29 · Title: Untitled Created Date: 11/9/2009 1:20:34 PM

LCSS(Fi, Fj) =

0 Ti = 0 | Tj = 0

1 + LCSS(FTi−1i , F

Tj−1j ) dE(fi,Ti

, fj,Tj) < ǫ & |Ti − Tj | < δ

max (LCSS(FTi−1i , F

Tj

j ), LCSS(FTi , F

Tj−1j )) otherwise

(1)

−15 −10 −5 0 5 10 15−70

−60

−50

−40

−30

−20

−10

0

10

Fig. 3. Trajectories seen from behind a vehicle collected over 100 minutesof natural driving.

be matched. This property is well suited for trajectory data

because their size varies depending on the amount of time

spent in view of sensors. Instead of a one-to-one mapping

between points, a point with no good match can be ignored

to prevent unfair biasing. The LCSS distance suggested by

Vlachos et al. [15] is defined as

DLCSS(Fi, Fj) = 1 −LCSS(Fi, Fj)

min(Ti, Tj), (2)

where the LCSS(Fi, Fj) value (1) specifies the number of

matching points between two trajectories. F t = {f1, . . . , ft}denotes all the flow vectors in trajectory F up to time t. The

LCSS can be efficiently computed using dynamic program-

ming.

B. Spectral Clustering

Spectral clustering has become popular technique recently

because it can be efficiently computed and has improved per-

formance over more traditional clustering algorithms. Spectral

methods do not make assumptions on the distribution of data

points and instead relies on eigen decomposition of a similarity

matrix which approximates an optimal graph partition [16].

The similarity matrix S = {sij}, which represents a fully

connected graph, is constructed from the LCSS trajectory

distances using a Gaussian kernel function

sij = e−D2

LCSS(Fi,Fj)/2σ2

∈ [0, 1]. (3)

where the parameter σ describes the trajectory neighborhood.

Large values of σ cause trajectories to have a higher similarity

score while small values lead to a more sparse similarity

matrix (more entries will be very small).

We use the normalized spectral decomposition presented by

Ng et al. [16] with Laplacian

L = I − D−1/2SD−1/2 (4)

followed by the final clustering of eigenvectors with fuzzy

C means (FCM). D is a diagonal matrix with elements the

sum of the same row in S. The Laplacian is transformed

using eigen decomposition to find the K largest eigenvectors

and placed as columns of a matrix U . The rows of U , each

corresponding to a trajectory, are finally clustered using FCM

after normalization using the orthogonal initialization method

proposed by Hu et al. [17].

The advantage of FCM clustering in the last spectral stage

are soft class assignment that minimizes the effects of outliers

and the cluster membership values uik which indicate the

quality of sample i. The membership variable uik ∈ [0, 1]indicates the confidence for assignment of trajectory i to

cluster k. High membership value means little route ambiguity,

or a typical realization of an activity process. This value is later

used to improve clustering results. Fig. 5(b) shows 10 typical

routes found through clustering.

C. Estimating Typical Activity Count

The number of paths, Np, in a scene must be estimated be-

cause it is not known a priori. Initially, the spectral clustering

selects a large number of clusters, Nc > Np then refines to a

smaller number of routes by merging similar clusters.

Trajectories are resampled [18] and a single iteration of

FCM is performed to produce route prototypes {rk}, k =1, . . . , Nc. Two route clusters, rm and rn, are considered

similar if each consecutive point is within a small radius,

dl(rm, rn) =√

(xml − xn

l )2 + (yml − yn

l )2 < ǫd ∀l, (5)

or if the total distance between tracks is small enough,

D =

L∑

l=1

dl < ǫD = Lǫd. (6)

The threshold value ǫd controls how far apart trajectories can

lie and is chosen experimentally. A cluster correspondence list

is created from these pairwise similarities, forming similarity

groups {Vs}. Each correspondence group is reduced to a single

route

r(Vs) = argminρ∈Vs

N∑

i=1

|uik(ρ) − uik(ρ)| (7)

by retaining only the prototype that would cause maximal

change in training membership if removed. The membership

uik(ρ) represents the (re-normalized) membership when pro-

totype ρ is removed from the correspondence set and uik(ρ)is the recomputed membership when ρ is removed (FCM with

starting membership uik(ρ)). Fig. 5(c) depicts the 7 remaining

routes after the merge procedure.

Page 4: Department of Electrical and Computer Engineering | University of …b1morris/docs/morris_icves2009.pdf · 2010-04-29 · Title: Untitled Created Date: 11/9/2009 1:20:34 PM

(a) (b)

Fig. 4. Example data collected from the instrumented vehicle. Trajectories are collected from both the front and rear of the vehicle. (a) On overtake on thepassenger side is seen in pink. (b) High density tracks during congestion have little motion and are long.

V. DYNAMICS AND TEMPORAL AUGMENTATION

Not only do we need to know where objects are located but

also the manner in which they travel along a given path. The

dynamics are needed to completely characterize a behavior.

Using HMMs, the spatio-temporal properties of every path

is encoded probabilistically. The HMM definition naturally

compares trajectories of different length and elegantly learns

temporal dependencies between activity actions.

A. Activity HMM

Hidden Markov models are convenient to represent activities

because it provides a means to incorporate both dynamic and

temporal information in a single principled learning step. An

activity route is decomposed into a sequence of smaller events

each having its own spatial, dynamic, and temporal charac-

teristics by with the use of both position and velocity. Each

activity HMM is compactly represented as λk = (Aλ, Bj , π0)and is designed to have Q states. The parameters π0 represents

the prior probability of an activity to start with state (event)

q. This is naturally defined to favor events that occur earlier

in an activity

π0(j) = e−αpj j = 1, . . . , Q (8)

where αp controls which states and activity is allowed to

begin in. The activity parameters Aλ and Bj are learned from

track data. The transition matrix Aλ describes the sequential

evolution of events and models their duration with an expo-

nential distribution across time. The observation distribution

Bj describes both the spatial and temporal characteristics of

an event (HMM states {qj}Qj=1). The states are modeled as

Gaussian mixtures with unknown mean and covariance,

qj ∼M∑

m=1

G(f, µjm,Σjm) (9)

with M specifying the number of activities in each path (all

experiments use M = 1 which assumes a single behavior for

each path).

B. Activity HMM Learning

A HMM is trained for each path by dividing the training

set into Np disjoint sets, D =⋃Np

k=1 Dk. The set Dk is

the collection of trajectories belonging to route rk based on

membership

r∗i = argmaxk

uik ∀i. (10)

Only those trajectories with membership uir∗i

> 0.9 are

retained when creating a path training set because they are

typical trajectories and can be confidently placed into route r∗i .

By using only highly confident tracks, HMM training is eased

an made more precise because of outlier removal. Using path

training set Dk, the Np HMMs can be efficiently learned using

standard methods such as the Baum-Welch method and EM

[19]. Unlike the route learning stage, when learning an activity

HMM a full trajectory, including all tracked information e.g.

velocity and acceleration, is used as the training input.

VI. EXPERIMENTAL STUDY

The activity learning scheme is tested on driving data

collected from an instrumented vehicle. Vehicles are tracked

with respect to the moving platform and represent differential

motion. This work only considers natural driving patterns

along a highway.

A. LISA testbeds

The Laboratory for Intelligent and Safe Automobiles (LISA)

is a mobile computing platform capable of monitoring the

vehicle surround, inside the cockpit, and the state of the

vehicle itself. The different test environments provide adapt-

able experimental testbeds for the evaluation of different

Page 5: Department of Electrical and Computer Engineering | University of …b1morris/docs/morris_icves2009.pdf · 2010-04-29 · Title: Untitled Created Date: 11/9/2009 1:20:34 PM

−15 −10 −5 0 5 10 15−70

−60

−50

−40

−30

−20

−10

0

10

enter

exit

stop

(a)

−15 −10 −5 0 5 10 15−70

−60

−50

−40

−30

−20

−10

0

10

Route 01

Route 02

Route 03

Route 04

Route 05

Route 06

Route 07

Route 08

Route 09

Route 10

(b)

−15 −10 −5 0 5 10 15−70

−60

−50

−40

−30

−20

−10

0

10

Route 01

Route 02

Route 03

Route 04

Route 05

Route 06

Route 07

(c)

−15 −10 −5 0 5 10 15−70

−60

−50

−40

−30

−20

−10

0

10

(d)

Fig. 5. (a) Entry/Exit/Stop zones: notice the stop zone behind the vehicle at -25 meters. (b) Initial set of spatial routes. (c) Remaining routes after mergeestimation. (d) Abnormal trajectories seem to arise from lane changes. Notice the activities on the driver side (left side −x axis) correspond to the ego-vehiclebeing overtaken by faster moving cars while the passenger side (+x axis) shows the opposite with the ego-vehicle overtaking slower vehicles.

sensor combinations and associated safety algorithms. A LISA

vehicle is able to synchronously capture and process data

from all sensor systems. Data from manufacturer installed

onboard sensors is recorded by tapping into the controller area

network (CAN) bus, cameras are mounted inside the vehicle to

view driver behavior using face, hands, and feet cameras, and

the surround is monitored with an array of sensors including

monocular cameras, omnidirectional cameras, stereo camera

pairs, and radars. A visualization of sensor data is presented

in Fig. 4. In this setup, cameras monitor the driver, front, and

rear or the ego-vehicle and radar sensors help with surround

detection.

B. Examining Rear Trajectories

The behavior learning framework was tested on trajectories

obtained from the rear of a car. Monitoring of this area is much

less prevalent than the front or blind spots event though a high

percentage of accidents have the rear as a point of contact. The

trajectories do not correspond to a fixed world location but are

measured with respect to the instrumented vehicle. Because

both the surrounding vehicles and the instrumented vehicle

can move independently, the trajectories are the sum of two

motion components.

The 170 trajectories collected looking out the rear of the

car over 100 minutes of natural driving are shown in Fig.

3. These trajectories correspond to vehicles that were tracked

for a minimum of 5 seconds (5 ∗ 30 fps = 150 samples)

along highway segments. Notice some of the trajectories

appear to move horizontally. This occurs when the ego-vehicle

changes lanes. This added motion is not explicitly modeled or

compensated for during the learning process but most of these

trajectories are removed with POI filtering because they do

not have much support.

The results from activity learning for the moving platform

are shown in Fig. 5. The scene goals can be viewed in Fig.

5(a). The regions where tracking begin (Enter) are marked

with green, the locations when tracking ends (Exit) is marked

with red, and locations where vehicles remain for an extended

Page 6: Department of Electrical and Computer Engineering | University of …b1morris/docs/morris_icves2009.pdf · 2010-04-29 · Title: Untitled Created Date: 11/9/2009 1:20:34 PM

time (Stop) are marked with yellow. Although there seems

to be a large number of randomly distributed points due

to tracking errors, the resulting zones are pleasing because

they show surround vehicles appearing and disappearing either

in the rear or on the passenger side of the ego-vehicle. In

addition, there is a stop zone directly behind the ego-vehicle at

approximately 25 meters. This following distance corresponds

to approximately 0.8 second time gap at a typical highway

speed of 70 mph. (This is closer than is generally considered

safe, the time gap should be at least 1 second).

Fig. 5(b) displays the 10 dominant routes automatically

learned from the trajectory training set. The refined routes are

shown in Fig. 5(c) after three similar activities are removed

during the merge procedure. The routes seem to indicate

mostly overtake maneuvers e.g. Route 2 and Route 6 in Fig.

5(c). During an ego-overtake (Route 2) vehicle tend to appear

in the rear (green circle) and travel alongside the ego-vehicle in

the adjacent lane before exiting sensor view on the driver side

(red x). Conversely, during an overtake (Route 6), vehicles

appear on the passenger side and move further away until

disappearing far behind the ego-vehicle indicating overtake of

a slow moving vehicle. This corresponds with typical driving

conventions where the faster lanes are toward the inside of

the highway on the driver side (left) while slower vehicles

tend to travel on the passenger side. This generally results

in a fast moving vehicle passing on the left side of a slow

moving vehicle in the right-hand lane. Finally, in Fig. 5(d)

some abnormal trajectories that were automatically extracted

are displayed (More details for abnormality detection can be

found in previous publications [18], [20]).

VII. CONCLUDING REMARKS

This work presents the first attempt to automatically learn

behaviors of surround vehicles based on natural observation

during driving. A hierarchical learning framework is presented

to describe behaviors at different resolutions with accompa-

nying unsupervised techniques. Trajectories of surrounding

obstacles obtained from the rear or an instrumented vehicle are

examined and key behaviors which correspond to observable

phenomenon, driver overtake and surround vehicle overtake,

are automatically discovered demonstrating the value of this

data driven approach.

REFERENCES

[1] (2009) Global status report on road safety. World Health Organization.[Online]. Available: http://www.who.int/violence injury prevention/road safety status/2009/en/index.html

[2] (2009) FARS encyclopedia. U.S. Department of Transportation -National Highway Traffic Safety Administration. [Online]. Available:http://www-fars.nhtsa.dot.gov/Main/index.aspx

[3] (2006) Traffic safety facts 2006. U.S. National Highway TrafficSafety Administration. [Online]. Available: http://www-nrd.nhtsa.dot.gov/Pubs/TSF2006FE.PDF

[4] M. M. Trivedi, T. Gandhi, and J. McCall, “Looking-in and looking-out of a vehicle: Computer-vision-based enhanced vehicle safety,” IEEE

Trans. Intell. Transp. Syst., vol. 8, no. 1, pp. 108–120, Mar. 2007.

[5] J. C. McCall and M. M. Trivedi, “Driver behavior and situation awarebrake assistance for intelligent vehicles,” Proc. IEEE, vol. 95, no. 2, pp.374–387, Feb. 2007.

[6] S. Y. Cheng and M. M. Trivedi, “Turn-intent analysis using body pose forintelligent driver assistance,” IEEE Pervasive Computing, vol. 5, no. 4,pp. 28–37, Oct-Dec 2006.

[7] J. C. McCall, D. Wipf, M. M. Trivedi, and B. Rao, “Lane change intentanalysis using robust operators and sparse bayesian learning,” IEEE

Trans. Intell. Transp. Syst., vol. 8, no. 3, pp. 431–440, Sep. 2007.[8] A. Doshi and M. M. Trivedi, “On the roles of eye gaze and head pose in

predicting driver’s intent to change lanes,” IEEE Trans. Intell. Transp.

Syst., Sep. 2009, forthcoming.[9] B. T. Morris and M. M. Trivedi, “A survey of vision-based trajectory

learning and analysis for surveillance,” IEEE Trans. Circuits Syst. Video

Technol., vol. 18, no. 8, pp. 1114–1127, Aug. 2008, Special Issue onVideo Surveillance.

[10] Z. Sun, G. Bebis, and R. Miller, “On-road vehicle detection: A review,”IEEE Trans. Intell. Transp. Syst., vol. 28, no. 5, pp. 694–711, May 2006.

[11] T. Gandhi and M. M. Trivedi, “Vehicle surround capture: Surveyof techniques and a novel omni-video-based approach for dynamicpanoramic surround maps,” IEEE Trans. Intell. Transp. Syst., vol. 7,no. 3, pp. 293–308, Sep. 2006.

[12] N. Brandle, D. Bauer, and S. Seer, “Track-based finding of stoppingpedestrians - a practical approach for analyzing a public infrastructure,”in Proc. IEEE Conf. Intell. Transport. Syst., Toronto, Canada, Sep. 2006,pp. 115–120.

[13] A. P. Dempster, N. M. Laird, and D. B. Rubin, “Maximum likelihoodfrom incomplete data via the EM algorithm,” J. Royal Stat. Soc., vol. 39,pp. 1–38, 1977.

[14] D. Makris and T. Ellis, “Learning semantic scene models from observingactivity in visual surveillance,” IEEE Trans. Syst., Man, Cybern. B,vol. 35, no. 3, pp. 397–408, Jun. 2005.

[15] M. Vlachos, G. Kollios, and D. Gunopulos, “Discovering similar mul-tidimensional trajectories,” in Proc. IEEE Conf. on Data Engineering,Feb. 2002, pp. 673–684.

[16] A. Y. Ng, M. I. Jordan, and Y. Weiss, “On spectral clustering: Analysisand an algorithm,” in Advances in Neural Information Processing

Systems, vol. 14, sep 2002, pp. 849–856.[17] W. Hu, D. Xie, Z. Fu, W. Zeng, and S. Maybank, “Semantic-based

surveillance video retrieval,” IEEE Trans. Image Process., vol. 16, no. 4,pp. 1168–1181, Apr. 2007.

[18] B. Morris and M. Trivedi, “Learning and classification of trajectories indynamic scenes: A general framework for live video analysis,” in Proc.

IEEE International Conference on Advanced Video and Signal based

Surveillance, Santa Fe, New Mexico, Sep. 2008.[19] L. R. Rabiner, “A tutorial on hidden markov models and selected

applications in speech recognition,” Proc. IEEE, vol. 77, no. 2, pp. 257–286, Feb. 1989.

[20] B. T. Morris and M. M. Trivedi, “Learning, modeling, and classificationof vehicle track patterns from live video,” IEEE Trans. Intell. Transp.

Syst., vol. 9, no. 3, pp. 425–437, Sep. 2008.