lifelong visual localization for automated vehicles -...

74
D O C T O R A L T H E S I S | Halmstad University Dissertations No. 12 Supervisors: Dr. Roland Philippsen Professor Dr. Thorsteinn Rögnvaldsson Examiners: Professor Dr. Tom Duckett Associate Prof. Dr. Hedvig Kjellström Assistant Prof. Dr. Davide Scaramuzza Associate Prof. Dr. Thomas Hellström Lifelong Visual Localization for Automated Vehicles Peter Mühlfellner

Upload: lamminh

Post on 12-Mar-2018

213 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

D O C T O R A L T H E S I S | Halmstad University Dissertations No. 12

Supervisors: Dr. Roland PhilippsenProfessor Dr. Thorsteinn Rögnvaldsson

Examiners: Professor Dr. Tom DuckettAssociate Prof. Dr. Hedvig KjellströmAssistant Prof. Dr. Davide ScaramuzzaAssociate Prof. Dr. Thomas Hellström

Lifelong Visual Localization for Automated Vehicles

Peter Mühlfellner

Page 2: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Lifelong Visual Localization for Automated Vehicles© Peter MühlfellnerHalmstad University Dissertations No. 12ISBN 978-91-87045-27-1 (printed)ISBN 978-91-87045-26-4 (pdf)Publisher: Halmstad University Press, 2015 | www.hh.se/hupPrinter: Media-Tryck, Lund

Page 3: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Abstract

Automated driving can help solve the current and future problems of indi-

vidual transportation. Automated valet parking is a possible approach to help

with overcrowded parking areas in cities and make electric vehicles more ap-

pealing. In an automated valet system, drivers are able to drop off their vehicle

close to a parking area. The vehicle drives to a free parking spot on its own,

while the driver is free to perform other tasks — such as switching the mode

of transportation. Such a system requires the automated car to navigate un-

structured, possibly three dimensional areas. This goes beyond the scope of

the tasks performed in the state of the art for automated driving.

This thesis describes a visual localization system that provides accurate

metric pose estimates. As sensors, the described system uses multiple monoc-

ular cameras and wheel-tick odometry. This is a sensor set-up that is close

to what can be found in current production cars. Metric pose estimates with

errors in the order of tens of centimeters enable maneuvers such as parking

into tight parking spots. This system forms the basis for automated navigation

in the EU-funded V-Charge project.

Furthermore, we present an approach to the challenging problem of life-

long mapping and localization. Over long time spans, the visual appearance of

the world is subject to change due to natural and man-made phenomena. The

effective long-term usage of visual maps requires the ability to adapt to these

changes. We describe a multi-session mapping system, that fuses datasets into

i

Page 4: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

ii

a single, unambiguous, metric representation. This enables automated navi-

gation in the presence of environmental change. To handle the growing com-

plexity of such a system we propose the concept of Summary Maps, which

contain a reduced set of landmarks that has been selected through a combi-

nation of scoring and sampling criteria. We show that a Summary Map with

bounded complexity can achieve accurate localization under a wide variety of

conditions.

Finally, as a foundation for lifelong mapping, we propose a relational database

system. This system is based on use-cases that are not only concerned with

solving the basic mapping problem, but also with providing users with a better

understanding of the long-term processes that comprise a map. We demon-

strate that we can pose interesting queries to the database, that help us gain a

better intuition about the correctness and robustness of the created maps. This

is accomplished by answering questions about the appearance and distribution

of visual landmarks that were used during mapping.

This thesis takes on one of the major unsolved challenges in vision-based

localization and mapping: long-term operation in a changing environment. We

approach this problem through extensive real world experimentation, as well

as in-depth evaluation and analysis of recorded data. We demonstrate that

accurate metric localization is feasible both during short term changes, as

exemplified by the transition between day and night, as well as longer term

changes, such as due to seasonal variation.

Page 5: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Dedication

To my mother, my brothers and my sister.

iii

Page 6: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved
Page 7: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Acknowledgments

The past three years were an interesting journey involving many different

places. Sometimes it was a wild ride, not without uncertainty. Without Roland

Philippsen, Paul Furgale and Wojtek Derendarz I would have never made it

through. Thank you! You taught me and showed me the way.

Similarly, I would like to thank Mathias Buerki, Fabian Pucks as well as

Thomas Holleis and Bastian Bücheler—for working so closely with me, for the

rough patches we got through and for the good things we shared.

My thanks go out to Denni Rögnvaldsson, for having all the patience in the

world and for enabling our long-distance collaboration. Furthermore, I would

like to thank Veronica Gaspes and Josef Bigun who provided guidance for my

studies.

I am thankful to my colleagues at Volkswagen Group Research. To Navid

Nourani for helping me reacquire a taste for real food and being a good buddy,

Nghia Nguyen for the friendship, general wisdom and pragmatism. Sebastian

Grysczyk, Maike Beenken, Mahyar Mousavi, Timur Aminev and Sebastian

Weise for the time we shared. Jan Aue, for always looking out that I remained

on the right track. My colleagues RenÃľ Waldmann, Stefan Wonneberger,

Martin Stellmacher, Stefan Brüning, Bastian Schmidt, Simon Steinmeyer, Sven

Horstmann, Jan Effertz, Stefan Brosig, Lutz Junge, Michael Darms and many

more—for being great to work alongside. You were always there for discussions,

with help and with advice.

v

Page 8: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

vi

I also consider myself lucky to have been given a chance to contribute to

the V-Charge project. This gave me the chance to work together with Mike

Bosse, spend many hours experimenting in the car with Ulrich Schwesinger

and be welcomed to the labs in Zurich by Gim Hee Lee, Lionel Heng and

Christian Hähne and the other members of ASL and CVG. I am also glad I

got to know Lina Maria Paz, Hugo Grimmet, Keiji Mimura, Benjamin Völz

and all the other people from the V-Charge team.

Despite visiting infrequently, I always felt welcome at the lab in Halmstad

because of the people I got know there. Thank you Saeed Gholami Shahbandi

for the nice welcome during summer vacation and Slawomir Novaczyk for let-

ting me stay at your place. I am also thankful to Sid Khandelwal, Björn Å

strand, Stefan Karlsson, Antanas Verikas, Roland Thörner, Karl Iagnemma,

Jens Lundström, Jennifer David, Tommy Salomonsson and all the others! Es-

pecially, I would also like to thank Stefan Byttner and Eva Nestius with for

being patient and understanding with my organizational blind spots.

Finally, I would like to thank my family and friends. I am lucky to share

my life with you.

The work in this thesis was supported in part by the European Community’s Seventh

Framework Programme (FP7/2007-2013) under grant #269916 (V-Charge).

Page 9: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Disclaimer

The views published in this thesis do not necessarily represent the position

taken by the Volkswagen Group.

vii

Page 10: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved
Page 11: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Contents

1 Introduction 1

2 Synopsis 9

2.1 Fundamentals of Visual Localization and Mapping . . . . . . . 9

2.2 Simple Multi-Camera Localization . . . . . . . . . . . . . . . . 15

2.3 Lifelong Localization . . . . . . . . . . . . . . . . . . . . . . . . 23

2.4 Database Design . . . . . . . . . . . . . . . . . . . . . . . . . . 31

3 List of Appended Papers 37

3.1 Paper A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

3.2 Paper B . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

3.3 Paper C . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

4 Discussion 41

4.1 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

4.1.1 Extensions of the Mapping Pipeline . . . . . . . . . . . 44

4.1.2 Improvements to the Database Architecture . . . . . . . 46

5 Conclusion 49

References 51

Appended Papers 63

ix

Page 12: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved
Page 13: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Chapter 1

Introduction

Cars are as ubiquitous as the problems that arise from their use: pollution,

accidents, traffic-congestion and the overflowing of cities with parked vehicles.

The increasing scarcity of fossil fuels will force us to adapt different power

sources, such as electric batteries, that further inconvenience users and put

additional demands on infrastructure. To mitigate these problems and enable

individual transportation in the future, we will need more intelligent vehicles

that relieve drivers of dangerous, difficult or arduous tasks. When performing

these tasks, the systems on board the vehicle need to be able to perceive,

interpret and interact with the same situations that were previously handled

by humans.

As an example, consider the combined problem of charging and park-

ing electrical vehicles. High-current quick-charging stations are expensive and

therefore limited in number; electrical vehicles in public spaces will need co-

ordinated charging. Chargers might not be available on-demand, and waiting

times can be quite lengthy: even fast high-current charging takes 30-120 min-

utes (IEA, 2013). Conversely, charging stations should be freed as quickly as

possible and not used as permanent parking spots. Automated vehicles can re-

1

Page 14: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2 CHAPTER 1. INTRODUCTION

lieve drivers of this task, by navigating between holding positions and charging

stations.

Such an automated valet parking system is a very attractive application for

introducing fully automated driving to a consumer market. Navigation is per-

formed in a privately owned area, which provides an opportunity to facilitate

the introduction of adequate regulatory and actuarial procedures. Further-

more, the areas are limited in size and navigable routes can be determined

beforehand.

This thesis investigates long-term visual robotic mapping and localization,

for the purpose of performing tasks such as automated valet parking. Under

the paradigm of map-based navigation, a map that contains information about

the area of automated operation is provided to the vehicle. For a valet scenario,

this map can bundle together information about the road network, as well as

the locations of parking spots and charging stations. The car is then instructed

to navigate to different mission goals using the map.

Critically, the map has to provide the means for the vehicle to localize

within it, in order to fundamentally enable automated driving. Depending

on the structure and scope of the map, localization might take on different

forms. We consider locally metric, sensor-based maps, which are widely used

in robotics research (Bosse et al., 2003, Brooks, 1985). Within such a map,

the final result of localization is an estimate for the vehicle pose—the position

and orientation—relative to a map reference frame. The vehicle pose is used

by local and global planning algorithms and for low-level vehicle control. To

enable localization, information about the environment that can be matched

to the output of the vehicle’s on-board sensors needs to be made available

through the map.

As a sensor, cameras provide rich appearance information, while being rel-

atively cheap and mechanically robust. Camera systems are already available

on the consumer market, where they are used for driver assistance systems

Page 15: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

3

such as lane-keeping assistance (Apostoloff and Zelinsky, 2003, Dickmanns

and Mysliwetz, 1992) or pedestrian protection (Franke et al., 2005). In the

robotics community, vision-based localization and mapping has been shown

to produce excellent results for large-scale map building (Sibley et al., 2010)

and long-range navigation (Furgale and Barfoot, 2010). Camera-based map-

ping and localization has also recently been successfully applied to automated

driving in urban areas (Lategahn and Stiller, 2012, Ziegler et al., 2014).

This thesis takes on one of the major unsolved challenges in vision-based

localization and mapping: long-term operation in a changing environment. We

see this as a prerequisite for the economical operation of a real-world robotic

system, such as the automated valet parking setup described earlier. The costs

for initializing a map are high, and include manual survey drives of the area,

annotation of the map with labels for lanes, parking spots, etc. Therefore, we

aim to build a map that is usable for localization over time frames of months

or years.

When using cameras for localization, we need to consider the changes in

visual appearance that an environment can undergo over the period of days,

weeks and months. Relevant phenomena include diverse weather conditions,

dynamic structure such as parked cars, seasonal variation and lighting dur-

ing different times of day. Some of the phenomena that cause visual change

are relatively abrupt—such as the exchange of parked cars at the end of a

working day. Many others occur slowly and are intermingled on different time

scales; some parts of the environment, such as buildings, appear permanent.

Environments where the number of possible visual configurations can even

be practically unbounded and change daily are problematic—consider the ar-

rangements of parked cars in a parking lot.

As robot cars navigate an environment, they will encounter all of these

phenomena. To enable automated navigation throughout the operational life

of our robots, we propose to build a lifelong visual map. This map should

Page 16: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

4 CHAPTER 1. INTRODUCTION

reflect the dynamic appearance of the world, and should be updated as infor-

mation about a changed environment becomes available. At the same time, the

map should allow for navigation by providing the basis for metric estimation.

This makes it possible to determine the vehicle pose relative to higher-level

structures—such as lanes, charging stations and parking spots—that are de-

fined in the map.

Based on these goals, we have formulated the following set of questions in

order to guide the development of the contributions presented in this thesis.

First of all, we are concerned with the limits of a state-of-the-art mapping

system, which makes no special provisions for long-term change. We wonder

when, and how, such a system will break down, as the world changes with

regard to the map. Will visual change cause localization to fail? Do we therefore

need a lifelong mapping system at all? How will visual change influence the

metric localization accuracy and precision of the system?

We demonstrate that it is necessary to handle visual change in order to

provide a reliable, lifelong map. The primary purpose of this map is to enable

localization for an automatedly navigating vehicle. For this purpose, we aim

to supply an unambiguous pose estimate, that is locally accurate, even though

the world changes. What representation for our map allows us to achieve this

goal, while still being able to incorporate information about the changed en-

vironment?

Related to this, we need to design a set of mechanisms to update and

manage our lifelong map. As we observe different appearance states of the

world (varying lighting conditions, seasons, etc.) the map will grow. If we allow

indefinite growth, the map will become prohibitively expensive to handle with

on-vehicle resources. At the same time, we want to keep as much information

about the environment as possible, in order to allow our map to grow and

remain useful even in highly changeable environments. How can we deal with

these conflicting demands?

Page 17: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

5

We aim to implement a system that works equally well in a wide variety

of environments. In order to accomplish this, we specifically avoid extracting

semantic information about the environment, e.g. the location of road mark-

ings, signs, roadside lamp posts, etc. These kind of features appear differently,

or not at all, in different locations. Rather, we use local image features that

are bound to occur in the typical environments for automated vehicles with

the expected level of structure and textures. The price we pay for abstaining

from features with specific semantics is that it is hard to gain insights into the

quality—in terms of correctness and robustness—of the mapping results. How

can we open up our system, to allow us to get a deeper understanding and

intuition about the features and the results of the mapping process?

We approach these questions through extensive real world experimentation,

as well as in-depth evaluation and analysis of recorded data.

We have collected test data from different scenarios, spanning a wide-

variety of weather, lighting and seasonal conditions. The collected data covers

all seasons, and provides a solid foundation for exploring the effects of long-

term visual change. These datasets are recorded using multiple research vehi-

cles, which are also equipped with additional reference sensors (DGPS/IMU)

that provide ground-truth data for the evaluation of the proposed algorithms.

The proposed systems are developed in the context of the research project

Autonomous Valet Charging and Parking (V-Charge) (Furgale et al., 2013),

which is concerned with the aforementioned valet charging problem. The re-

search vehicles that were built as part of this project provided a test-bed for

constantly evaluating the real-world utility of the proposed systems. Over the

course of two years, the various iterations of the localization system described

in this thesis have been extensively used as the basis for fully automated driv-

ing in the context of this project.

Finally, we implemented the proposed lifelong map on the basis of a rela-

tional database system. This foundation allows easy dissemination of mapping-

Page 18: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

6 CHAPTER 1. INTRODUCTION

results. Using relational queries, facts about the entities involved in the map

can be combined to explore their characteristics.

Contributions

The major contributions of this thesis are as follows:

1. Analysis of the limitations, both in availability and metric accuracy, of

a state-of-the-art four fisheye camera localization system (Muehlfellner

et al., 2013).

2. Proposal of a multi-session, long-term mapping and localization frame-

work (Muehlfellner et al., 2015a).

3. A general scheme for life-long vision-based navigation and mapping, that

limits the on-vehicle resource usage by requiring only compact maps

on the vehicle and deferring the majority of processing to an offline

server (Muehlfellner et al., 2015a)—reducing memory requirements to

one-tenth of the original maps, and allowing four times as many frames

to be processed per second.

4. Introduction of the concept of Summary Maps, which employ scoring

statistics and sampling policies in order to select the most pertinent in-

formation from a long-term map (Muehlfellner et al., 2015a). This pro-

vides a new abstraction of the methods used for creating summarized

maps by other authors. Based on this abstraction we are able to com-

pare our own, new methods for summarization with existing approaches

from the literature.

5. Demonstration of the feasibility of creating consistent Summary Maps

that incorporate datasets from the time-span of one year, and still pro-

vide unambiguous metric localization with high accuracy (Muehlfellner

et al., 2015a).

6. Evaluation of the trade-offs between the complexity, availability and ac-

curacy of different Summary Mapping-strategies (Muehlfellner et al.,

Page 19: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

7

2015a). We show that, with the right summary strategy, maps remain

at a constant number of landmarks while still adapting to a changing

environment as well as supporting accurate localization.

7. Proposal of a simple and efficient relational database scheme that acts

as the foundation for life-long mapping, and for the analysis and dissem-

ination of mapping results (Muehlfellner et al., 2015b).

8. Evaluation of the database-design with regards to the use-case of data-

analysis, showing examples of how relational queries can be used to im-

prove our intuition about the correctness and robustness of a long-term

map (Muehlfellner et al., 2015b). These queries allow us to interpret the

mapping results, taking in-depth looks to explore for instance during

what sessions landmarks were observed, what their appearance was and

how they were distributed in the images.

In chapter 2, a summary of the state of the art research on the long-term

visual mapping problem is provided, followed by a synopsis of our approach to

lifelong mapping. A list of the contributed publications is provided in chapter

3. These papers can be found in the appendices. We will close by providing

a discussion of these contributions, and the future work that is necessary to

further enable long-term autonomy in chapter 4.

Page 20: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved
Page 21: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Chapter 2

Synopsis

This chapter presents a summary of our work on lifelong visual localization

and mapping. We start out by presenting the problems automated driving in

unstructured environments poses to us, and what approaches we can find in

the literature in order to overcome them (Section 2.1). This is followed by a

description of the simple multi-camera localization pipeline we implemented

and evaluated (Section 2.2). As we see from this prototypical pipeline, the

long-term dynamics pose serious problems for a visual localization system. We

focus on addressing these problems through lifelong visual mapping in the next

Section (Section 2.3). The chapter finishes with a description of the relational

database framework we have implemented in order to support our lifelong

mapping system (Section 2.4).

2.1 Fundamentals of Visual Localization and

Mapping

We aim to enable automated vehicles to navigate unstructured, three dimen-

sional (3D) environments. Examples for this kind of environment include park-

9

Page 22: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

10 CHAPTER 2. SYNOPSIS

(a) (b) (c) (d)

(e) (f) (g) (h)

Figure 2.1: Snapshots from the on-vehicle camera systems in which we em-ployed our mapping and localization pipeline. The first column shows viewsof the parking lot of our research campus in Wolfsburg, Germany; the secondcolumn shows the test area and workshops at the ETH Campus Hoenggerberg,Zurich, Switzerland; the third column shows the ‘Bosch Parkhaus’ in Stuttgart,Germany; the final column shows views from a suburban area in Wolfsburg.

ing lots and garages, as well as various other private premises. In contrast to

well-maintained highways and urban roads, such places do not necessarily con-

tain features such as standardized road-markings, curbs or guide rails. These

sites are further differentiated by their three-dimensionality: we drive on roads

that go across ramps, are unsealed or contain speed-bumps, and we even aim

to explore multiple levels of car parks. Figure 2.1 shows example snapshots of

the environments that we have deployed our system in. The environments we

navigate have been constructed for human use. They are not modified through

the addition of markers, beacons, or other infrastructure that has solely been

added in order to facilitate vehicle localization.

The basic procedure of the pipeline proposed in this thesis follows a teach-

and-repeat (Furgale and Barfoot, 2010) scheme. A survey dataset is acquired,

driving the vehicle manually through the environment, along the desired paths.

Page 23: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.1. FUNDAMENTALS OF VISUAL LOCALIZATION AND MAPPING 11

This dataset is used to build a map, which can then be used for localization

when the environment is visited again.

The problems of building a map and localizing within that map are tightly

interwoven. For a long time, this has been recognized as the Simultaneous Lo-

calization and Mapping (SLAM) problem in the robotics community (Leonard

and Durrant-Whyte, 1991, Smith et al., 1987, Thrun et al., 2005). The SLAM

problem is often formulated as a belief network (Frese and Duckett, 2003, Mon-

temerlo et al., 2002, Murphy, 2000), but can also be seen as a factor graph or

a Markov Random Field (MRF) (Dellaert and Kaess, 2006). Common to the

formulations of SLAM is the optimal estimation of parameters in a map, given

a history of sensor readings and vehicle control inputs (Durrant-Whyte and

Bailey, 2006). Parameters to be estimated usually include vehicle poses and

the positions of landmarks in the environment. Note that in this thesis, we are

not explicitly concerned with finding novel solutions to the SLAM problem,

but rather this thesis deals with the practical problem of localization for au-

tomated navigation. While we feed back information from the online system

into the map, we do so through limited channels (see section 2.3) in order to

limit the on-vehicle computational complexity. Nonetheless, we still apply the

same data-association and optimization techniques used to solve SLAM.

Path planning and following in a complex environment, including obsta-

cle avoidance and off-lane maneuvers, requires metric information about the

vehicle pose. More specifically, due to the three-dimensional nature of our

environment, we aim to estimate poses with the full six degrees of freedom

(6-DOF) that the rigid body of our vehicle might move in. It is the purpose

of our localization system to provide this information. At the same time, sen-

sors that provide globally consistent information, such as GPS, are generally

unavailable, or not accurate enough, meaning that errors will accumulate over

large distances (drift).

Page 24: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

12 CHAPTER 2. SYNOPSIS

To deal with this inherent limitation of uncertain sensors, Brooks (1985)—

in a very early paper—proposed a local, relative, graph-based map representa-

tion, as did Smith et al. (1987). The Atlas framework proposed by Bosse et al.

(2003) follows in a similar vein, by defining a map as a graph of submaps.

Howard (2004) proposes a map which is embedded in a manifold, for use in

their multi-agent system. The notion of a manifold map is also central to the

continuous relative representation proposed by Sibley et al. (2009).

The common idea behind these approaches is, that the global correctness

of the map does not have to be enforced at all times. Rather, the map is built

from a set of interconnected euclidean spaces, which are related through a

graph structure. Pose estimation always takes place relative to one of these

spaces, comparing the sensor information provided by the robot about its

local surroundings with what is stored in the map. As long as the map is

locally euclidean, the resulting estimates are still useful, even if the map can-

not be expressed in a single global coordinate frame. This mixture of locally

euclidean/globally topological structure is what lends the name manifold map-

ping to this class of approaches. The teach-and-repeat scheme we employ also

benefits from this property: by estimating poses only relative to locally defined

coordinate frames, we can forego the requirement for global map correctness.

There is a wide range of sensors that have been employed for building

robotic maps. Ultrasonic sensors were popular due to their low cost and their

direct measurement of depths, but are ultimately limited to small scale envi-

ronments. Two- and three-dimensional laser scanners have been successfully

applied to the SLAM problem (Marshall et al., 2008, Newman et al., 2002).

Their utility for localization is limited in large-scale outdoor environments,

where the presence of in-range walls, or other informative structures, can not

be guaranteed. Three-dimensional systems provide richer information about

the environment, enough to allow the application of appearance based meth-

ods (McManus et al., 2013) based on this sensor. Nonetheless, even though

Page 25: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.1. FUNDAMENTALS OF VISUAL LOCALIZATION AND MAPPING 13

these sensors have become strongly associated with the public image of auto-

mated driving, they are expensive and contain moving parts. It is noteworthy

that recently several manufacturers of solid state Lidars have emerged1; they

promise to eliminate moving mechanical parts and reduce size, while still offer-

ing high resolutions. These systems might become serious contenders for the

basis of commercial localization and mapping systems in the future.

In the work presented in this thesis, we build a camera-based system for

visual localization and mapping. Compared to other sensors, vision offers an

output that is extremely rich in information, due to the high resolutions that

are available even on relatively cheap cameras. Depth can not be directly

inferred from single images, but requires multiple images taken either at the

same time (such as in a stereo rig), or in sequence. As a passive sensor, cameras

rely on the natural illumination of the environment, and are as such susceptible

to illumination changes.

When it comes to implementing vision-based localization and mapping,

there is a wide variety of approaches that are employed with different goals

and in various environments.

Purely appearance-based methods do not reconstruct the metric structure of

the environment. In this context, a large body of work deals with the problem

of place recognition (PR). PR is concerned with picking a matching image, or

place, from a map, given only a single input image. This form of localization is

useful for initializing a more fine-grained system, or for detecting loop-closures

while building a map. In order to perform automated navigation, more infor-

mation about the vehicle pose needs to be made available.

A line of work originating from Matsumoto et al. (1996) compares whole

images from the current view to a previously recorded stream of images. The

outcome of this comparison is then used to compute a steering angle that keeps

the robot on the desired path. Zhang and Kleeman (2009) present impressive

1See for example the products of Quanergy (http://www.quanergy.com/)

Page 26: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

14 CHAPTER 2. SYNOPSIS

experimental results on more than 18 km of routes using a modified version of

Matsumoto’s approach. It is in the nature of these approaches not to provide a

metric pose estimate for the vehicle. This makes performing useful tasks that

require plans that deviate from the recorded route, such as parking or obstacle

avoidance, difficult.

For road environments, matching features that are visible on the road sur-

face to similar features extracted from an overhead (aerial) image, has been

a widely used approach. For example, Pink et al. (2009) proposed to extract

lane markings from aerial images and compare them to markings visible dur-

ing vehicle localization. Similarly, Napier et al. (2010) match a local vehicle

trajectory to an overhead image in order to obtain global accuracy. Apart

from being limited to planar environments, this class of systems loses much

of its attractiveness when an overhead image is not easily obtainable—such

as when operating inside a parking garage, or under overhanging buildings or

trees. Approaches that rely on the presence of lane-markings are limited even

further, as they can only be applied in a certain set of environments (compare

Figure 2.1).

We do not wish to limit ourselves to environments with certain charac-

teristics, but rather perform estimation of the full 6-DOF vehicle pose using

generic features of the environment. One possible approach to this problem is

presented by Royer et al. (2007), who proposed a mapping system that builds

a fully metric map from local image-point features. This map is then used to

localize an automated vehicle and perform navigation on a recorded reference

path. In contrast, Šegvić et al. (2009) present a purely relative framework that

relies on local two-view reconstructions. The estimated pose is not used di-

rectly for navigation, but rather employed in order to provide a steering-angle

input for a visual servoing scheme. Furgale and Barfoot (2010) present a mixed

topological/metric system, that performs localization in multiple linked maps.

These maps are built using a Stereo Visual Odometry (VO) (Matthies, 1989,

Page 27: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.2. SIMPLE MULTI-CAMERA LOCALIZATION 15

Camera 0

Camera 2

Camera 1

Camera 3

日乳迭

Figure 2.2: Our driverless car, showing the location of the involved coordinateframes. Of the four wide-angle fisheye monocular cameras on the car, only thefront and left cameras are visible, alongside their associated coordinate framesFC0,1

. Two additional cameras are arranged symmetrically on the right andthe rear of the car.

Moravec, 1980) pipeline. They demonstrate the ability to perform navigation

on three-dimensional, multi-kilometer routes using maps that are only locally

accurate.

2.2 Simple Multi-Camera Localization

We implemented a visual mapping and localization pipeline (Muehlfellner

et al., 2013), based on the four-camera system of our research vehicle (see

Figure 2.2). Our approach is inspired by the mixed topological/metric frame-

work introduced by Furgale and Barfoot (2010); we also build a locally ac-

curate map. Instead of a stereo-camera system, we use multiple monocular

cameras with very little overlap in their fields of view. While this makes the

problem more difficult to start with, we have access to wheel-tick odometry

measurements in addition to the pure vision information. This allows us to

Page 28: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

16 CHAPTER 2. SYNOPSIS

Mapping & Fusion

Map

Database

N x Image

Features

Tracking &

Triangulation

Bundle

Adjustment

Odometry

N x Image

Features Odometry

Survey

Dataset

Localization

2D – 3D

Matching Optimization

Figure 2.3: Our simple pipeline for mapping and localization. Inputs areimage features from an N-tuple of images, as well as wheel-tick odometry.

easily recover metric structure2 from monocular images; we also make heavy

use of the odometry system to initialize and support our estimators.

In addition to the layout of the employed camera system, Figure 2.2 also

depicts the various reference frames involved in estimation:

Vehicle frames, FVj, are associated with each group of images (image-

tuple), acquired from the different cameras at time instant j. Each vehicle

frame, FVj, is attached to the rear-axle center at that time. The incremental

pose estimate provided by wheel-tick odometry is also stored alongside the

image-tuples.

Camera frames, FCkare attached to each camera. The transformation

TVjCkbetween it and the vehicle frame, FVj

is estimated during camera cali-

bration (Heng et al., 2014).

Map frames, FMi, are placed along the navigable routes and define the

map. The vehicle pose at time j is estimated as the 6-DOF transformation,

TVjMi, between the map frame, FMi

, and the vehicle frame, FVj.

The global frame, FO, is connected to each map frame, FMi, via the

transformation, TMiO. This can be used to express localization results in a

single global frame, for visualization, or to interface with other systems that

do not work in a relative representation.

2In this context, structure refers to the underlying 3D-scene geometry.

Page 29: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.2. SIMPLE MULTI-CAMERA LOCALIZATION 17

The basic buildings blocks of the first implementation of our pipeline are

depicted in Figure 2.3. During the offline mapping phase, the map frames FMi

are set up; 3d landmarks that represent the visual appearance and structure of

the environment are attached to each frame. Once we enter the online localiza-

tion phase, the vehicle poses TVjMiare estimated based on these landmarks.

Approaches for estimation from the SLAM literature are roughly split along

two lines: on the one hand, there are filtering approaches (Thrun et al., 2005,

c.f.) that take previous vehicle poses out of the problem in some way. In the

Extended Kalman Filter (EKF) (Davison et al., 2007, Smith et al., 1987),

vehicle poses are marginalized after each step, and only a map of landmarks

is kept. In Rao-Blackwellized Particle Filters (Eade and Drummond, 2006,

Montemerlo et al., 2002) a number of possible vehicle trajectories is sampled,

making landmarks conditionally independent and therefore easy to estimate.

On the other hand there are optimization approaches that work on the full

problem of estimating a history of vehicle poses at each time step. Some opti-

mization approaches only keep a graph of images around, on which estimation

is performed Konolige et al. (2010), Thrun and Montemerlo (2006). Other ap-

proaches optimize the full structure (landmarks) and motion (vehicle poses) in

what is called bundle adjustment (Hartley and Zisserman, 2003, Triggs et al.,

2000).

Triggs et al. (2000) provide convincing arguments for the flexibility, accu-

racy and efficiency of bundle adjustment. Bundle adjustment naturally fits to

the offline phase of our system, since processing is performed in batch. Within

its framework, we are able to incorporate a spherical projective model (Mei

and Rives, 2007) for the employed fisheye cameras, as well as robust M-

Estimator (Zhang, 1997) based error models. We also apply non-linear op-

timization for localization in the online phase, performing optimal pose esti-

mation based on the same projective and error models.

Page 30: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

18 CHAPTER 2. SYNOPSIS

(a) (b)

Figure 2.4: Snapshots showing extracted SURF (Leutenegger et al., 2011)(left) and BRISK (Bay et al., 2006) (right) Keypoints, as red circles, for thesame image. Circle-size indicates Keypoint-scale.

In the following, we describe the blocks involved in the pipeline in more

detail.

The proposed pipeline relies on local image point features, which are com-

posed of keypoints and descriptors. Keypoints represent interesting points that

arise from the intensity appearance of images. Descriptors capture the appear-

ance of the image patch around a keypoint in a compact fashion, providing

resistance to image noise and slight illumination changes. Typically, keypoints

have a scale and orientation associated with them. By aligning descriptors

with this scale and orientation, they can be made in/co-variant to changes

in scale and orientation of the observed feature. This provides some (limited)

invariance to viewpoint changes. We have used both the Speeded Up Robust

Features (SURF) (Bay et al., 2006), as well as the Binary Robust Invariant

Scalable Keypoints (BRISK) (Leutenegger et al., 2011) in our experiments.

Figure 2.4 shows an example of the SURF and BRISK points detected in an

image.

These keypoints are tracked through sequences of images. Tracking is ac-

complished by matching keypoints, based on their image-position and descrip-

tor across pairs of consecutive images. The tracked points are triangulated

Page 31: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.2. SIMPLE MULTI-CAMERA LOCALIZATION 19

based on the different viewpoints associated with them and stored relative to

the map frames they were seen in. A map frame is initialized for each ve-

hicle frame that satisfies a minimal baseline distance to the previous frame,

computed from wheel-tick odometry.

Once the map has been incrementally set up from vehicle frames in this

way, we identify loop closures between frames that revisit the same place in

the world. To perform this task in an automated fashion, appearance-based

place recognition (e.g. (Cummins and Newman, 2008, 2011)) is commonly

used. In fact, for later implementations of our pipeline (Muehlfellner et al.,

2015a), we used the loop closure detection described in (Lee et al., 2013). For

the pipeline described in (Muehlfellner et al., 2013) we identified candidates

for loop closures manually. Here, the user provides a set of vehicle frame-

pairs that should depict the same place. Both elements of each pair, and the

frames surrounding these, undergo feature matching and RANSAC-based pose

estimation based on an early version of the solution proposed by Kneip and

Lynen (2013). Pairs that have enough feature matches that are consistent with

the RANSAC-estimate are further processed as loop closures.

To optimize the overall structure of the map, we perform two seperate steps.

First, a pose graph of vehicle frames, based on wheel-tick odometry poses and

the loop closure pose estimates described above, is relaxed. This provides a

map skeleton that is locally geometrically accurate, across the places that were

identified as loop closures above. The relaxed pose graph is used to initialize

the second step: a full bundle adjustment (Hartley and Zisserman, 2003, Triggs

et al., 2000), which refines both the estimates for the landmarks, as well as for

the vehicle poses.

Given a map created in this way, localization is the process of estimating

the vehicle pose relative to one of the vehicle reference frames. This is done

on the basis of a set of landmarks that are visible in the map frame closest to

the currently predicted vehicle pose. These landmarks are projected into the

Page 32: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

20 CHAPTER 2. SYNOPSIS

Figure 2.5: Matches between an image in the map (left) and an imagerecorded some time later, under different environment conditions (right). Thegreen lines connect BRISK keypoints that were matched between the two im-ages.

current view and matched to the detected keypoints based on their projected

image position and the descriptor from the associated keypoint in the map. The

matching-step ties a set of keypoints in the current view to a set of keypoints

in the map view, as illustrated in Figure 2.5. Each landmark-keypoint pair

matched in this way defines a correspondence between a 2D point in the current

view and a 3D point in the map.

These 2D-3D correspondences are the basis for an optimal estimation of the

current vehicle pose using non-linear batch optimization. A similar method to

the bundle adjustment used in the mapping phase is employed here. Instead of

optimizing the whole map, only the positions of the landmarks matched to the

current vehicle frame and the associated vehicle pose are taken into account.

The parameters of the vehicle pose are estimated, the landmarks remain fixed.

In addition, an error term that punishes deviations from the initial odometry-

based pose estimate is included in the problem. This error term helps to avoid

divergent solutions if there are few, or geometrically degenerate, 2D-3D corre-

spondences.

The online pose estimation, as well as the offline bundle adjustment and

pose graph relaxation, are based on a software package developed internally at

Page 33: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.2. SIMPLE MULTI-CAMERA LOCALIZATION 21

the Autonomous Systems Lab at ETH Zurich. This package supports lineariza-

tion of parameters that involve rotations, transformations and homogeneous

points as described in Furgale (2011, Chapter 3). From the package, we also

use the support for M-Estimators (c.f. Zhang, 1997) for robust weighting of

error terms, to allow optimization in the presence of outlier correspondences.

For solving geometric problems, in (Muehlfellner et al., 2015a) we used early

versions of the software that was later published as OpenGV (Kneip and Fur-

gale, 2014). In addition, we used the Schweizer Messer library3 for representing

rigid-body kinematics and as a general toolset. On the car, our software was

run using the Automotive Data and Time-Triggered Framework4.

The proposed pipeline is a very basic approach to the problem of visual

mapping and localization.

We run the online pose estimation in a loosely coupled fashion: the op-

timizer is initialized with a pose predicted based on odometry, and provides

a (hopefully) improved pose estimate based on map information. Here, there

is room for improvement—either a filter-based or optimization-based method

(c.f. Strasdat et al. (2010) for a comparison) could be used to build a map

during online operation, perhaps in a limited window. This map could include

both landmarks from the offline map, providing registration, as well as newly

tracked features, providing richer information for pose estimation.

Furthermore, the proposed pipeline provides localization only if there is

at least a rough guess for the previous pose of the vehicle. To initialize, we

assume that the start location in the map is known beforehand. This simple

method works well in practice, allowing start-up at a drop-off area of around

1-2 meter radius. To lift this requirement, global localization based on place

recognition (Cummins and Newman, 2008) or monte-carlo techniques (Wolf

et al., 2005) could be used.

3Available online. https://github.com/ethz-asl/Schweizer-Messer.4Commercial product. https://automotive.elektrobit.com/products/eb-assist/adtf/

Page 34: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

22 CHAPTER 2. SYNOPSIS

(a) Year-Long Parking Lot Dataset

October 11, 2013

(b) Day-Night Parking Lot Dataset

December 6, 2013

December 12, 2013 July 31, 2014

Figure 2.6: Example images that illustrate the long-term dynamics of theparking-lot dataset. Each image is cropped from the rear camera view at ap-proximately the same location.

Notwithstanding the above limitations, the major drawback of this pipeline

becomes apparent during long-term operation.

We have evaluated the performance of the proposed pipeline on a number

of datasets captured during autumn and winter 2012/2013 (Muehlfellner et al.,

2013). On the one hand, the relative metric accuracy for this pipeline, calcu-

lated based on a high-precision IMU/DGPS system, is in the order of tens of

centimeters, enough for automated driving. On the other hand, the system has

limited robustness to long-term changes in the appearance of the environment.

In the paper, we show that the system fails as seasons change, when tested on

the parking lot of our research campus.

In order to deal with this problem and improve the overall performance of

the pipeline, we will look into how we can adapt to visual change in the next

section.

Page 35: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.3. LIFELONG LOCALIZATION 23

2.3 Lifelong Localization

Figure 2.6 shows how the visual appearance of a location evolves over the time

span of several months. Changes in lighting range from diffuse illumination

during overcast days to strong highlights during sunny days. Apart from af-

fecting lighting, weather also influences the visual nature of the world through

rain, snow and fog. From the figure it can also be seen that the structure of the

environment changes quite dramatically between sessions. Since the recording

was made in a parking lot, large parts of images are dominated by parked cars,

which change on a daily basis. On a longer time scale, the vegetation visible

in the images is influenced by the passage of seasons.

Related Work

Handling the effects of different lighting conditions has recently been the sub-

ject of intensive study in the literature. At the level of individual pixels—

transforming raw image data—approaches that use illumination-invariant im-

age transformations have shown very promising results (Corke et al., 2013,

McManus et al., 2014a). They employ the spectral characteristics of a sin-

gle light source in order to remove the effect of different illumination angles

and cast-shadows, at a pixel level. Lategahn et al. (2013) propose learning an

illumination invariant descriptor, using a set of basic building blocks, from

a representative set of recordings taken at different times of day. Recently,

training neural networks to extract descriptors that are robust to illumination

changes has been proposed in (Carlevaris-Bianco and Eustice, 2014).

These techniques provide an excellent way of dealing with light and shadow

variations, but are limited to scenarios with a well-known light source, e.g.

outdoors under direct sunlight. Furthermore, while these methods bring ro-

bustness to changing lighting, more general structural change in scenes is not

modelled by them. We believe that for a long-term system to be useful, it

Page 36: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

24 CHAPTER 2. SYNOPSIS

needs to cope with lighting- as well as with structural change. It should also

transition smoothly between indoor and outdoor scenes, and improve as new

information becomes available about a changed scene.

Another major branch of research is concerned with long-term place recog-

nition (compare section 2.1). This is exemplified by SeqSLAM (Milford and

Wyeth, 2012), which matches sequences of images acquired under extremely

varied conditions, matching from day to night, and rain to sunshine. Similarly,

Naseer et al. (2014) robustly select matches under severe seasonal changes us-

ing the data-association graph for full datasets. Notably, Valgren and Lilienthal

(2010) have investigated the long-term performance of visual features for place

recognition across seasons; they point out the importance of using image geom-

etry (in the form of epipolar constraints) to validate matches, and show that

the Upright SURF descriptor performs well. Krajnik et al. (2014) represent

the periodic re-occurrence of features using a spectral approach.

These systems solve the difficult problem of going from a completely un-

known pose to a rough localization (in the form of a matching place). While

this form of localization is useful for detecting loop-closures, and for helping

recover from a lost state, they do not provide enough information for plan-

ning and control of a robot on their own. Furthermore, many of the above

approaches do not incorporate new information about an environment as it

becomes available (with the exception of (Krajnik et al., 2014)).

To handle structural changes in scenes while providing metric pose esti-

mation, Biber and Duckett (2005) propose a dynamic map that represents

multiple timescales at once. Their map is composed of discretized 360 degree

range scans for each place; these local maps are updated by sampling new mea-

surements at different rates. Higher update rates allow incorporating changes

into the map quickly, representing short timescales, whereas lower update rates

favour more stable long-term structures and therefore longer timescales. Biber

and Duckett use a short-term map with a very high update rate, refreshed

Page 37: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.3. LIFELONG LOCALIZATION 25

continually during online operation. Meanwhile, all the measurements are kept

around until a later offline phase, where the various long-term timescales are

updated.

The approach of Churchill and Newman (2013) presents a successful scheme

for vision-based long-term localization and mapping. They propose a system

where the robot localizes in multiple experiences simultaneously, whenever

it re-visits a previously mapped location. If localization against the known

experiences fails, the map is augmented with a new experience built by their

Visual Odometry (VO) system, which will be available for localization on the

next visit.

Another successful visual multi-session mapping system has been presented

by Dayoub et al. (2011). Their map-representation consists of a pose graph,

where each node has an associated spherical view that stores the location of

observed image features on a 3D-sphere. Whenever a location is revisited, a

state machine modeled on human memory is used to decide which currently

observed features are to be added to the these views: features first pass through

short-term memory, which serves to discard spurious observations, before they

are admitted to long-term memory.

Similarly, Konolige and Bowman (2009) present an approach for selecting

useful views from a map. For each metric neighbourhood in their map, they aim

to identify a set of view-clusters that represent the various appearance states of

the environment. Similar view-pairs are considered connected, in a graph that

describes the local appearance dynamics of a place; from this graph clusters

are extracted based on the maximal connected subsets. In order to manage

map complexity, Konolige and Bowman propose a scheme that deletes views

from a neighbourhood, if an allowed number of views is exceeded, while still

preserving the diversity of clusters as far as possible.

The approach we take to long-term mapping is close in spirit to the method

proposed by Johns and Yang (2013, 2014). They associate image features to

Page 38: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

26 CHAPTER 2. SYNOPSIS

real-world landmarks, that can be re-observed during multiple visits to the

same place. Thus, appearance changes can be modelled at a sub-scene level,

independently updating each landmark incrementally. While Johns et al. focus

on place recognition from images only, we apply a similar approach to the

precise metric localization of an automated vehicle.

In contrast to Churchill and Newman (2013) and Biber and Duckett (2005),

we take an approach to long-term mapping which uses a single, unified repre-

sentation for the long-term information. It is unclear how semantic informa-

tion, such as road definitions for automated navigation, would be handled in

an experience-based map as only topological connections, and no metric rela-

tionship, are defined between experiences. Furthermore, by fusing information

from all experiences and timescales, we can estimate a better geometric model

of the world. Having a full set of associations, we can also make better choices

about selecting useful data.

Our approach is inspired by the systems of Konolige and Bowman (2009)

and Dayoub et al. (2011), which also handle the long-term dynamics of scenes.

We aim to provide a more general framework for handling long-term map

management. For this purpose, in the next section we present our mapping

pipeline, which combines online localization and offline map updates (com-

pare Biber and Duckett (2005)) with a summary step, that selects the most

pertinent information from the map.

Lifelong Operation based on Summary Maps

Since our goal is to enable map-based navigation for automated driving, lo-

calization is of primary importance. Map management and data fusion can be

carried out in batch, offline on a server, similar to the concept described by

Biber and Duckett (2005). If we aim to build a system that supports a large

number of robots deployed in a single area, such as automated cars in a car

park, this seems to be a natural progression. There is little benefit in having

Page 39: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.3. LIFELONG LOCALIZATION 27

Summary

Mapping /

Fusion Localization

Database

Summary

map for

localization

Localization output & raw data

Full-scale

map

(a) Overall Life Cycle

Localization

Triangulation 2D-3D

Matching

Batch

Optimization Optimization

Tracking &

RANSAC Feature

Extraction Mapping / Fusion

Database Summary

Map

Localized

Frames

Full Batch-Optimized

Map

Localized

Frames

(online or offline)

Summary

Full Map

(b) Multi-Session Pipeline

Figure 2.7: Top: The life cycle of a map, from initialization to extensionwith additional datasets, including a detailed view of the multi-session pipeline.Bottom: the blocks involved in the mapping and localization steps (comparesection 2.2).

the vehicles manage their own data, which would require expensive hardware.

Instead, automated cars could rely on centralized services that are able to pro-

duce the “best” map for a given task based on data from all robots over all

time.

The proposed lifecycle for lifelong visual maps includes four major steps.

1. Initialization populates the database with a fresh map based on data

collected with the vehicle.

Page 40: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

28 CHAPTER 2. SYNOPSIS

2. During regular online operation, vehicles localize in the environment and

(optionally) log localization and sensor data for further subsequent pro-

cessing.

3. The recorded data is transmitted to an off-vehicle computer, where the

recorded data is fused into a large full-scale map which stores all the

long-term information that is available about an environment.

4. Finally, from this full-scale map, a Summary Map can be created that is

small enough for efficient and robust online operation.

The three steps after initialization are repeated in a cycle, these are shown in

Figure 2.7.

Subsequently, we describe our multi-session mapping pipeline, giving an

overview how we perform the steps in the lifecycle. More details on this pipeline

can be found in (Muehlfellner et al., 2015a).

In this context, a session refers to a drive with the vehicle in a given

environment, where data is acquired that can subsequently be used to build

or extend a map. Sessions are recorded as discrete episodes, where the vehicle

is switched on at some location, data is logged, and the vehicle is switched off

again. This is similar to the concept of an experience used in experience-based

mapping (Churchill and Newman, 2013). Apart from automated runs, sessions

can also be recorded manually, storing raw datasets that can be localized and

fused offline.

In order to implement the proposed pipeline, we use a mixed metric-

/toplogical map representation, similar to the one described in the previous

section (2.2). The map is built from map frames, FMi, vehicle frames, FVj

, with

attached image- and keypoint information and landmarks, which represent 3D

points in the real world that arise from tracked keypoints.

The multi-session mapping performs the following basic steps:

• localize each incoming vehicle frame based on the attached image infor-

mation and create data associations to the existing map;

Page 41: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.3. LIFELONG LOCALIZATION 29

• track and triangulate points between consecutive images to initialize new

landmarks; and

• perform global batch optimization over all sessions to ensure geometric

consistency.

The components involved in these steps are depicted in Figure 2.7 (compare

also to the description of our simple pipeline in section 2.2).

The map frames, FMi, are initialized based on a state-of-the-art mapping

process (Lee, 2014) for a single dataset. As the map is extended, more and

more vehicle frames end up referring to each map frame. This turns each FMi

into an abstract place or small submap that contains contributions from all

sessions in the map. Localization is performed relative to these map frames,

via the landmarks accumulated from all sessions to each map frame. Semantic

data, such as the road and lane definitions for navigation, labels for pedestrian

crossings and traffic signs, etc. are also defined relative to these coordinate

frames.

Vehicle Frame poses and landmark positions are optimized during each it-

eration of the mapping process. After optimization, the map frames are moved

to the pose of the closest Vehicle Frame registered with them, to ensure that

they always represent a local neighbourhood of vehicle frames. Additionally,

a new Map Frame is created if a Vehicle Frame is added, with a distance to

its associated map frame that exceeds a threshold (2 meters, or more than 10

degrees in orientation). This allows the extension of the map to new places. For

the performed experiments, these relatively simple approaches for managing

the relation between Map and vehicle frames proved sufficient.

The multi-session mapping process solves the problem of adding new in-

formation to the map. In addition, we propose a framework for bounding map

size while maximizing its usefulness for on-vehicle operation. The basic idea

is to select a subset of pertinent landmarks for the map that gets downloaded

by the vehicle. The resulting Summary Map describes the content of the full

Page 42: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

30 CHAPTER 2. SYNOPSIS

map in a simplified way. This concept can be seen as an analogy to the idea

of a Summary Statistics, which concisely describe a set of observations. The

proposed framework is built from Landmark scoring functions and sampling

policies, which are applied to a multi-session map.

A scoring function provides a prediction of the usefulness of an individual

landmark for localization. Each landmark is associated with a set of observa-

tions that links the landmark to specific image features. A scoring function

assigns a real valued score to each landmark, based on the properties of its

observation set. As simple examples, the number of observations can be used

as a score, as can be the number of distinct sessions a dataset was seen in.

The state machine proposed by Dayoub et al. (2011) for modelling biological

short- and long-term memory behaviour can be turned into a scoring function

with a binary output.

Sampling policies select a subset of all landmarks based on the various

scores of individual landmarks. Furthermore, they take into account the rela-

tionship between landmarks. Multiple scoring functions can be combined in a

sampling policy. The selected subset of landmarks forms the Summary Map,

that can be used for localization on-board the vehicle. Proposed implementa-

tions of sampling policies include a simple threshold on the score, as well as a

local ranking scheme that selects a fixed number of high-scoring landmarks per

map-region. The least-recently used (LRU) view deletion algorithm proposed

by Konolige and Bowman (2009) also fits into the sampling-policy framework.

In (Muehlfellner et al., 2015a), we have investigated the performance of

the multi-session mapping pipeline in several scenarios. For this purpose, data

from our parking lot, which contains little permanent structure, and data from

a suburban area in Wolfsburg was recorded for approximately one year. Fur-

thermore, we also analyze the performance of the proposed system as lighting

conditions change from day to night.

Page 43: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.4. DATABASE DESIGN 31

In the paper, we use three criteria for evaluating the performance of a

variety of sampling schemes: availability measures the percentage of time the

robot is able to localize within the map, accuracy measures the metric error

of the robot’s six degree of freedom (6-DOF) pose estimate with respect to

ground truth, and computational complexity measures the computational cost

of localizing within the map.

We show that a full multi-session map built from a year of data provides

high availability and metric accuracy. Similarly, the proposed mapping scheme

allows building of a map that offers metrically consistent localization between

night and day scenarios. The complexity growth associated with full maps is

unbounded. We approach the complexity problem with a variety of summary

strategies that improve the complexity characteristics. The sampling scheme

adapted from Konolige and Bowman (2009) and a local-ranking based scheme

of our devising have the property of constant complexity, while offering high

availability throughout the whole year.

Designing and prototyping summary strategies is enabled by the relational

database that supports our mapping process. We describe the design of this

database in the next section.

2.4 Database Design

When performing lifelong mapping, fusing and curating data from many vis-

its is a daunting task. In (Muehlfellner et al., 2015b) we investigate how an

underlying database architecture can support this task and make extensive

long-term maps feasible. This database functions as the central repository in

the lifecycle we proposed in section 2.4.

We see three major roles associated with the use of the proposed mapping

and localization system:

Page 44: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

32 CHAPTER 2. SYNOPSIS

1. The user: typically a test engineer, who employs the localization results

in the wider context of automated driving.

2. The expert and developer: runs the mapping pipeline, is responsible for

solving problems that come up during real world operation.

3. The scientist: aims to extend the system, gain insights into its inner work-

ings through data analysis, improve results by proposing new techniques

and evaluate the performance.

While these roles might be taken on by the same person, they can also be

spread out through a larger team. This team might want to carry out tests

in geographically disparate locations, while still being able to reliable analyze

and improve system performance. These roles help us define the major use

cases for our database-system.

For the online localization use case, the database needs to make landmarks

accessible for the pose estimator. Apart from that, the online system should

also make it easy for users to report on the performance of the localization

system. They should be able to provide the developers with information that

allows them replay exactly what happened during a test run. Furthermore,

maps for the online system should be compact and access to the map should

be fast.

To enable Offline mapping, the database needs to provide a representation

for the elements of the map described in Sections 2.2 and 2.3.

The final use case, Analysis and Synthesis, is concerned with our intu-

ition about the mapping and localization processes. Since the keypoints and

landmarks stored in the map do not have any pre-determined interpretation,

working with the system can often be unintuitive. The proposed relational

database scheme gives scientists the ability to drill down into the data, asso-

ciations and parameters of the map. With this, we want to be able to answer

questions such as “What are the best, and what are the worst landmarks for

localization and what do they look like?”, “Are trees, lamp posts or repetitive

Page 45: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.4. DATABASE DESIGN 33

Scripting Interface

Boost Python Native Access

Alg

orith

ms

Pe

rsisten

ce

Ph

ysical

DB

Direct SQL

Access

RDBMS

SQLite

Localization Mapping

Map Manager

Database Abstraction

POCO Data

Serialization

Boost Serialization Atomic Storage

Figure 2.8: The basic layers involved in our database architecture. Compo-nents that rely on free- or open software and/or standards are marked in adarker color.

lane-markings useful for localization at all?”, “What is the role of ephmeral

keypoints, such as the ones arising from clouds, or from cast shadows?”.

Questions such as these require the ability to pose queries that were not

necessarily anticipated in the original design of the database. Applying the

principles of relational database design, such as normal forms, help in providing

more flexibility to such unexpected demands. On the other hand, we need to

trade off this flexbility against the implementation and runtime complexity of

the system.

Figure 2.8 shows a rough sketch of the layered architecture we use in our

system. We built these layers on top of freely/openly available software. In

choosing the libraries that these layers are built upon, we have focused on the

need to operate our system in the limited environment that is afforded to us

by the on-vehicle computers. In the figure, the need for a general library that

implements a certain concept such as ‘Serialization’ or ‘RDBMS’ (Relational

Database Management System) is indicated in the top-left corner of a block,

whereas the concrete implementation we have chose is indicated in the center.

Page 46: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

34 CHAPTER 2. SYNOPSIS

Landmark Map Frame

1 N

M N

„based in“

„seen at“

Multi Frame

M

N

„observes“

Camera System

1 N

„based in“

1 N

„observed

with“

Dataset

N 1

N 1

„from“

„from“

Figure 2.9: The conceptual Entity-Relationshop (ER) diagram for the entitiesinvolved in our pipeline.

We provide database access for the mapping, localization and other algo-

rithmic components through both a native C++ interface and a scripting lan-

guage (Python) for rapid prototyping. The regular way to access the database

is through the map manager implemented in C++. This class handles mapping

the object model used by the algorithmic components to the relational model

stored in the database. For describing and accessing the relational database,

we use Structured Query Language (SQL); queries are passed on through a

database abstraction layer, which provides independence from the specific Re-

lational Database Management System (RDBMS).

The conceptual relational model for our database is given in Figure 2.9.

With the relational model, we capture the semantics of the map, to represent

the entities that are involved in the mapping process. Landmarks and Map

Frames are directly represented in the database. Multi Frames implement the

concept of a vehicle frame with attached image and keypoint data. The Camera

System relation groups together the set of parameters that describe the cam-

Page 47: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

2.4. DATABASE DESIGN 35

eras used to capture each Multi Frame. Datasets group the stored landmarks

and Multi Frames according to the session they were observed in.

These entities are tied together by a number of relationships. In Figure 2.9,

these are indicated by diamond shapes, where the cardinality of the relation-

ship is given next to the diamond.

Both landmarks and Multi Frames are tied to a specific Map Frame. In the

case of landmarks, this is the reference frame that the estimated 3D-point for

the Landmark is expressed in. For Multi Frames, this relationship gives the

pose estimate and local coordinate frame that the corresponding vehicle frame

is tied to. Landmarks are also seen in an arbitrary number of Map Frames;

this gives no geometric relationship, but indicates that the landmark can be

used for localization at the respective place.

Multi Frames contain landmark observations in the form of keypoints.

Camera Systems, in turn observe specific Multi Frames. This chain of associa-

tions allows us to build batch optimization problems that relate the estimated

landmark 3D-points to the 2D-image keypoints via projection through a Cam-

era System. The observes relation is also important in calculating landmark

scoring functions: such as how often, or in which datasets, a landmark has been

seen. The link to a specific dataset is stored in the database for both Multi

Frames and landmarks; in the case of landmarks, this is the dataset that the

landmark has first appeared in.

In (Muehlfellner et al., 2015b), we validate the described database design

with respect to the requirements posed by the use cases. The design fulfills

the basic requirements posed by the online localization and offline mapping

use cases. Furthermore, we look into how we can gain a better intuition about

the dynamics of an environment using the database. We show that we can

pose interesting queries to the relational database, that help us get a better

intuition about the mapping process.

Page 48: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

36 CHAPTER 2. SYNOPSIS

We also identify several of the shortcomings of the implemented database

design and present lessons for future implementations of such systems. A higher

degree of normalization in the stored data structures would allow us to pose

more interesting queries. The database also offers no facilities that simplify

use by multiple agents and enable versioning of the stored maps.

Regardless, we conclude that the proposed database design has proven itself

through practical use. It has enabled us to perform the extensive experiments

described in Section 2.3, and it also has stood the test of several years of

real-world, cross-platform, operation on our research vehicles.

Page 49: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Chapter 3

List of Appended Papers

The contributions made by this thesis are contained in three papers. Of these,

the first two bridge the gap between the state-of-the-art and a lifelong mapping

and localization system. The second paper also contains the major part of the

contributions made by this thesis. The third paper rounds off our proposed

system, by describing how the relational database that underlies the long-term

map is implemented, and how it can be used to perform in-depth analysis of

mapping results.

In the following, we give the abstracts of the three appended papers.

3.1 Paper A

Evaluation of Fisheye-Camera Based Visual Multi-Session Localiza-

tion in a Real-World Scenario

Peter Muehlfellner, Paul Furgale, Wojciech Derendarz, Roland Philippsen

Intelligent Vehicles Symposium 2013, Workshop on Environment Perception

(Published)

37

Page 50: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

38 CHAPTER 3. LIST OF APPENDED PAPERS

The European V-Charge project seeks to develop fully automated valet

parking and charging of electric vehicles using only low-cost sensors. One of the

challenges is to implement robust visual localization using only cameras and

stock vehicle sensors. We integrated four monocular, wide-angle, fisheye cam-

eras on a consumer car and implemented a mapping and localization pipeline.

Visual features and odometry are combined to build and localize against a

keyframe-based three dimensional map. We report results for the first stage

of the project, based on two months worth of data acquired under varying

conditions, with the objective of localizing against a map created offline.

3.2 Paper B

Summary Maps for Lifelong Visual Localization

Peter Muehlfellner, Mathias Buerki, Mike Bosse, Wojciech Derendarz, Roland

Philippsen, Paul Furgale

Journal of Field Robotics

(Accepted for Publication)

Robots that use vision for localization need to handle environments which

are subject to seasonal and structural change, and operate under changing

lighting and weather conditions. We present a framework for lifelong localiza-

tion and mapping designed to provide robust and metrically accurate online

localization in these kinds of changing environments. Our system iterates be-

tween offline map building, map summary, and online localization. The offline

mapping fuses data from multiple visually varied datasets, thus dealing with

changing environments by incorporating new information. Before passing this

data to the online localization system, the map is summarized, selecting only

the landmarks that are deemed useful for localization. This Summary Map en-

ables online localization that is accurate and robust to the variation of visual

information in natural environments while still being computationally efficient.

Page 51: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

3.3. PAPER C 39

We present a number of summary policies for selecting useful features for

localization from the multi-session map and explore the tradeoff between local-

ization performance and computational complexity. The system is evaluated

on 77 recordings, with a total length of 30 kilometers, collected outdoors over

sixteen months. These datasets cover all seasons, various times of day, and

changing weather such as sunshine, rain, fog, and snow. We show that it is

possible to build consistent maps that span data collected over an entire year,

and cover day-to-night transitions. Simple statistics computed on landmark

observations are enough to produce a Summary Map that enables robust and

accurate localization over a wide range of seasonal, lighting, and weather con-

ditions.

3.3 Paper C

Designing a Relational Database for Long-Term Visual Mapping

Peter Muehlfellner, Paul Furgale, Wojciech Derendarz, Roland Philippsen

Journal of Software Engineering in Robotics

(Draft submitted to JOSER)

We present a map architecture based on a relational database that helps

tackle the challenge of lifelong visual localization and mapping. The proposed

design is rooted in a set of use-cases that describe the processes necessary for

creating, using and analyzing visual maps. Our database and software archi-

tecture effectively expresses the required interactions between map elements,

such as visual frames generated by multi-camera systems. One of the ma-

jor strengths of the proposed system is the ease of formulating pertinent and

novel queries. We show how these queries can help us gain a better intuition

about the map contents, taking into account complex data associations, even

as session upon session is added to the map. Furthermore, we demonstrate how

referential integrity checks, rollbacks and similar features of relational database

Page 52: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

40 CHAPTER 3. LIST OF APPENDED PAPERS

management systems are beneficial for building long-term maps. Based on our

experience with the proposed system during one year of intensive data col-

lection and analysis, we discuss key lessons learned and indicate directions

for evolving its design. These lessons show the importance of using higher

relational normal forms to make the database schema even more useful for

querying, as well as the need for a distributed, versioned system.

Page 53: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Chapter 4

Discussion

We presented a general framework for lifelong mapping and localization, based

on the concept of summary maps. In addition, we have shown the utility of the

proposed approach in an extensive set of experiments, both on recorded data

sets and through real-world usage of our system in the V-Charge project. In

the following, we look at how the work presented in this thesis has addressed

the questions posed in Chapter 1.

We started out by investigating the state-of-the-art in visual localization

and mapping (Section 2.1) and proposed our own implementation of a multi-

camera localization system (Section 2.2 and (Muehlfellner et al., 2013)). This

very simple pipeline was put to the test using datasets collected over a period

of two months, during a time of seasonal change. By evaluating the localization

outputs based on ground truth, we demonstrated that the system provides ad-

equate metric localization accuracy and precision. The evaluation also showed

that the localization pipeline breaks down dramatically as visual appearance

changes.

These results were confirmed by extensive real-world usage of the proposed

system for navigation with our experimental automated vehicle. We observed

41

Page 54: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

42 CHAPTER 4. DISCUSSION

that the maps we built of various scenarios (compare Figure 2.1, Section 2.3)

would work well over short- to medium-term time frames. Automated driv-

ing, with a path planner and controller relying on the visual localization for

making plans and correcting deviations, was quite robust when using a current

map. However, when the appearance of the world changed—most dramatically

during seasonal transitions—localization availability would get more and more

spurious, finally resulting in complete failures.

We have not observed a direct drop in the accuracy of the estimated poses

as a result of these changes. This is due to the robust nature of our optimization

system: it will either provide a reasonably accurate estimate, or completely fail.

On the other hand, the coupled odometry solution will drift as we move onward

from the last good localization that was registered to the map. In practice,

this drift makes automated driving impossible if no localization solution is

available for some time, depending on the quality of odometry and the nature

of the environment. Nonetheless, thanks to odometry, we are able to provide

a working system, even if the availability of localization is below 100%. This

goes to show the importance of incorporating odometry, as already pointed

out by Furgale (2011).

In order to improve the performance of localization over long time spans,

we proposed a system to build lifelong visual maps. Such maps incorporate

new information about a changed environment, as it becomes available. In

Section 2.3 we provide an overview of the system presented in (Muehlfellner

et al., 2015a).

The proposed mapping pipeline relies on keeping a full model of the world

which is managed on an offline platform. This offline map stores all observed

landmarks and data associations and is optimized in batch, as new datasets are

incorporated. Landmark and pose parameters are estimated based on visual

measurements, together with odometric constraints that help bridge gaps in

the registration of images.

Page 55: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

43

In order to manage the complexity of this ever growing map, we proposed

the concept of summary maps. These abridged maps are extracted from the

full scale map that is handled during offline processing. They contain land-

marks that have been selected by a sampling policy, which in turn employs

various scoring functions. In (Muehlfellner et al., 2015a) we compared several

different scoring/sampling combinations, both of our own design and based on

general strategies presented in the literature (Dayoub et al., 2011, Konolige

and Bowman, 2009). We showed that we can build summary maps of constant

size, that perform with high availability and accuracy throughout a full year.

Finally, we described the design of the underlying database system for our

lifelong map in Section 2.4 and in (Muehlfellner et al., 2015b). This database

is central to the lifecycle of the map: it is used to manage offline-, online-

and summary maps alike. Furthermore, we showed how we can exploit the

relational structure that underlies the map in order to gain a deeper insight

about the data associations and long-term processes involved in mapping.

In (Muehlfellner et al., 2015b) we identified the requirements based on use-

cases that are set up from the perspective of the three major roles involved

in the development of such a system. The users of the localization system

are concerned with the basic operation, and with the ability to submit useful

reports for analysis and debugging. The engineers work on debugging and

processing the results of these reports, as well as on mapping new areas. The

scientists aim to gain an understanding about the behaviour of the long-term

mapping process, and rely on the map to provide facilities for implementing

and testing new algorithms. We showed how these use cases are fulfilled, based

on the proposed design. In addition, we presented the lessons that have been

learned while implementing the map database, and propose improvements for

a future design.

Page 56: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

44 CHAPTER 4. DISCUSSION

4.1 Future Work

It is our long-term vision to see the proposed system implemented on a team

of field robots that is deployed continuously. This will involve significant work

automating every step of the pipeline and extending the software architecture

to meet the additional requirements. This is clearly an important next step

towards the goal of lifelong localization for mobile robots deployed in real-

world environments. In the following, we describe some extensions of proposed

mapping pipeline, as well as improvements to the database architecture, that

could be made towards this goal.

4.1.1 Extensions of the Mapping Pipeline

Different summary strategies should be explored in the context of the proposed

pipeline.

The strategies proposed in (Muehlfellner et al., 2015a) are only the very

basics of what is possible with this approach. These could be extended by

clustering landmarks both in descriptor- and appearance space, with repre-

sentative landmarks being selected in a similar scheme to the keyframe-based

approach of Konolige and Bowman (2009). With that, the diversity of visual

information available at a place could be controlled in a better way. Coobser-

vatility information—which landmarks were seen together—could also play an

important role in this clustering, enabling us to identify independent visual

experiences on a sub-scene level.

Other summary strategies would choose configurations of landmarks that

are most likely to produce the most accurate localization estimate. In a system

that is more tightly coupled with (visual) odometry, fewer landmarks would

be required for keeping the estimate well matched to the map. This could

be exploited by providing fewer, but better—in regard to their estimated 3D

location and their visual stability—landmarks. However, a more accurate lo-

Page 57: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

4.1. FUTURE WORK 45

calization system would require a different ground truth provider in order to

perform adequate quantitative evaluation. As reported in (Muehlfellner et al.,

2013), we are already close to the precision provided by the employed DGP-

S/IMU system.

The next step would be the introduction of place-specific parameters and

features generated from the summary maps. One example is the use of the

quadratic discriminant analysis described by Bosse and Zlot (2013) to create

place-specific transformations of the feature space that produce the best lo-

calization results. One could even produce place-specific features such as the

scene signatures proposed by McManus et al. (2014b). Bootstrapping the scene

descriptions via the use of point-features in our framework, combined with the

proposed offline processing approach, opens up these techniques.

The scalability of the proposed pipeline is another issue that needs to be

addressed in future work.

In order to scale this approach to larger areas, longer time scales, and

more robots, we will have to deal efficiently with torrents of incoming data.

Although we would like to produce maps using all data over all time, this

is clearly impossible for any practical case. First, we must develop cluster-

capable optimization algorithms (Agarwal et al., 2010, Ni et al., 2007, Snavely

et al., 2008) and utilize windowed or relative map representations (Sibley et al.,

2009). Second, we must also have the possibility to produce summary maps in

the offline system. It would be interesting to investigate how we can modify

summary strategies to produce a more scalable offline map, but still provide

enough information for continuous extension.

Inevitably, if the system is built to run continuously, some marginaliza-

tion of raw data and sparsification of the pose graph will be necessary. Find-

ing the best methods of summarizing data is an interesting avenue of re-

search (Carlevaris-Bianco and Eustice, 2013, Johannsson et al., 2012, Mazuran

et al., 2014). Furthermore, communication constraints may enforce that only

Page 58: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

46 CHAPTER 4. DISCUSSION

a subset of data will make it back to the server for processing. Because of this,

we would like to explore strategies for the online selection of data that is new

or surprising.

4.1.2 Improvements to the Database Architecture

The underlying database architecture could also be extended according to the

lessons learned during one-and-a-half years of continuous deployment of our

pipeline. The first step of this extension would be a redesign of the relational

model to provide a higher degree of data normalization. Currently, data is

stored in the form of serialized keyframes. Breaking these objects up into indi-

vidual keypoints, with image coordinates and descriptors that are accessible as

database columns, would open up even more possibilities for data inspection

and relational querying. It would also allow us to perform better automated

integrity checking on our databases.

Based on a higher normalization level in the databases, one could ex-

plore the possibility of employing range queries. By building on range search

trees (Guttman, 1984), the database could allow querying for spatial informa-

tion. This could prove useful for both debugging and data analysis, allowing

the retrieval of geometrically related sub-structures of the map. It is an inter-

esting question how such range searches can be made to meaningfully interact

with the manifold map representation we use. This could perhaps be achieved

by introducing a higher dimensional representation, similar to the multi-agent

system described by Howard (2004).

A more radical approach to the database design might also pay off, as we

move to bigger scales of deployment. The database could be elevated from its

current state as a repository for information about a single map, created by

a single robot. By making it accessible through a server, and providing the

means to handle concurrent reads and writes, multi-robot mapping could be

enabled. Furthermore, multiple versions of a map, referring to the same raw

Page 59: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

4.1. FUTURE WORK 47

(and expensive to store) image data could be kept in the same repository.

This would allow for easier extension and debugging of a large-scale mapping

system.

Page 60: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved
Page 61: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

Chapter 5

Conclusion

Map-based robotic navigation is a promising paradigm for the implementation

of driverless cars. The map acts as a repository of information about the street

network, it can enumerate mission goals, and it can help predict possibly dan-

gerous situations. In this framework, reliable and accurate localization with

respect to the map is the prerequisite for performing any useful automated

driving tasks.

In this thesis, we have presented our work on the problem of lifelong lo-

calization and mapping. We have built and evaluated a state-of-the art multi-

camera localization system, which was successfully deployed to our experimen-

tal vehicle. The evaluation of this localization pipeline has shown us that the

system performs well in similar visual environments, but breaks down under

visual change. In order to solve this problem, we have proposed a mapping

pipeline that can incorporate data from multiple mapping sessions, using all

information available about the environment. The generated long-term maps

are made useable in a complexity-constrained online system via the extraction

of summary maps. In addition to our work on lifelong mapping, we have also

49

Page 62: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

50 CHAPTER 5. CONCLUSION

presented the underlying relational database. The design of the database aids

us in gaining an intuition about the quality of the maps that we produce.

We were successful in building long-term maps over the time span of a

year. While we believe that the proposed approach can be scaled up to allow

lifelong, continuous mapping by a large number of vehicles, more research

is required to enable this. More elaborate summary strategies, that better

preserve visual diversity, can be designed; incorporating clustered and relative

windowed optimization can help us scale to process larger amounts of data;

loop-closure detection and pose-graph sparsification will be necessary to enable

automated operation over extremely long time scales. The proposed database

design can be improved by providing higher levels of normalization, enabling

range queries and moving to a versioned data repository.

We believe that the proposed pipeline has the potential to enable lifelong

mapping and localization on future driverless cars. The generality, robustness

and accuracy of our approach makes it well suited for a wide range of scenarios,

ranging from automated parking to driving in suburban areas.

Page 63: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

References

Sameer Agarwal, Noah Snavely, Steven M. Seitz, and Richard Szeliski. Bundle

adjustment in the large. In Computer Vision–ECCV 2010, pages 29–42.

Springer, 2010. (Cited on page 45.)

Nicholas Apostoloff and Alexander Zelinsky. Robust vision based lane tracking

using multiple cues and particle filtering. In Intelligent Vehicles Symposium,

2003. Proceedings. IEEE, pages 558–563. IEEE, 2003. (Cited on page 3.)

Herbert Bay, Tinne Tuytelaars, and Luc Van Gool. Surf: Speeded up robust

features. Computer Vision–ECCV 2006, pages 404–417, 2006. (Cited on

page 18.)

Peter Biber and Tom Duckett. Dynamic maps for long-term operation of

mobile service robots. In Robotics: science and systems, pages 17–24, 2005.

(Cited on pages 24 and 26.)

Michael Bosse and Robert Zlot. Place recognition using keypoint voting in

large 3D lidar datasets. In Robotics and Automation (ICRA), 2013 IEEE

International Conference on, pages 2677–2684, May 2013. doi: 10.1109/

ICRA.2013.6630945. (Cited on page 45.)

Michael Bosse, Paul Newman, John J. Leonard, and Seth Teller. An At-

las framework for scalable mapping. In Robotics and Automation, 2003.

51

Page 64: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

52 REFERENCES

Proceedings. ICRA’03. IEEE International Conference on, volume 2, pages

1899–1906. IEEE, 2003. (Cited on pages 2 and 12.)

Rodney Brooks. Visual map making for a mobile robot. In Robotics and Au-

tomation. Proceedings. 1985 IEEE International Conference on, volume 2,

pages 824–829. IEEE, 1985. (Cited on pages 2 and 12.)

Nicholas Carlevaris-Bianco and Ryan M. Eustice. Long-term simultaneous

localization and mapping with generic linear constraint node removal. In

Proceedings of the IEEE/RSJ International Conference on Intelligent Robots

and Systems, pages 1034–1041, Tokyo, Japan, November 2013. doi: http:

//dx.doi.org/10.1109/IROS.2013.6696478. (Cited on page 45.)

Nicholas Carlevaris-Bianco and Ryan M. Eustice. Learning visual feature de-

scriptors for dynamic lighting conditions. In IEEE/RSJ International Con-

ference on Intelligent Robots and Systems, 2014, IEEE. IEEE, 2014. (Cited

on page 23.)

Winston Churchill and Paul Newman. Experience-based navigation for long-

term localisation. The International Journal of Robotics Research, 32(14):

1645–1661, 2013. (Cited on pages 25, 26, and 28.)

Peter Corke, Rohan Paul, Winston Churchill, and Paul Newman. Dealing

with shadows: Capturing intrinsic scene appearance for image-based outdoor

localisation. In Intelligent Robots and Systems (IROS), 2013 IEEE/RSJ

International Conference on, pages 2085–2092. IEEE, 2013. (Cited on page

23.)

Mark Cummins and Paul Newman. FAB-MAP: Probabilistic localization and

mapping in the space of appearance. The International Journal of Robotics

Research, 27(6):647–665, 2008. (Cited on pages 19 and 21.)

Page 65: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

REFERENCES 53

Mark Cummins and Paul Newman. Appearance-only SLAM at large scale

with FAB-MAP 2.0. The International Journal of Robotics Research, 30(9):

1100–1123, 2011. (Cited on page 19.)

Andrew J. Davison, Ian D. Reid, Nicholas D. Molton, and Olivier Stasse.

MonoSLAM: Real-time single camera SLAM. Pattern Analysis and Machine

Intelligence, IEEE Transactions on, 29(6):1052–1067, 2007. (Cited on page

17.)

Feras Dayoub, Grzegorz Cielniak, and Tom Duckett. Long-term experiments

with an adaptive spherical view representation for navigation in changing en-

vironments. Robotics and Autonomous Systems, 59(5):285–295, 2011. (Cited

on pages 25, 26, 30, and 43.)

Frank Dellaert and Michael Kaess. Square Root SAM: Simultaneous localiza-

tion and mapping via square root information smoothing. The International

Journal of Robotics Research, 25(12):1181–1203, 2006. (Cited on page 11.)

Ernst D. Dickmanns and Birger D. Mysliwetz. Recursive 3-D road and relative

ego-state recognition. IEEE Transactions on pattern analysis and machine

intelligence, 14(2):199–213, 1992. (Cited on page 3.)

Hugh F. Durrant-Whyte and Tim Bailey. Simultaneous localization and map-

ping: part I. Robotics & Automation Magazine, IEEE, 13(2):99–110, 2006.

(Cited on page 11.)

Ethan Eade and Tom Drummond. Scalable monocular SLAM. In Computer

Vision and Pattern Recognition, 2006 IEEE Computer Society Conference

on, volume 1, pages 469–476. IEEE, 2006. (Cited on page 17.)

Uwe Franke, Clemens Rabe, Hernán Badino, and Stefan Gehrig. 6D-vision:

Fusion of stereo and motion for robust environment perception. In Pattern

Recognition, pages 216–223. Springer, 2005. (Cited on page 3.)

Page 66: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

54 REFERENCES

Udo Frese and Tom Duckett. A multigrid approach for accelerating relaxation-

based SLAM. In Autonome Mobile Systeme 2003, pages 192–202. Springer,

2003. (Cited on page 11.)

Paul T. Furgale and Timothy D. Barfoot. Visual teach and repeat for long-

range rover autonomy. Journal of Field Robotics, 27(5):534–560, 2010.

(Cited on pages 3, 10, 14, and 15.)

Paul T. Furgale, Ulrich Schwesinger, Martin Rufli, Wojciech Derendarz,

Hugo Grimmett, Peter Muehlfellner, Stefan Wonneberger, Julian Timpner,

Stephan Rottmann, Li Bo, Bastian Schmidt, Thien Nghia Nguyen, Elena

Cardarelli, Stefano Cattani, Stefan Bruning, Sven Horstmann, Martin Stell-

macher, Holger Mielenz, Kevin Koeser, Markus Beermann, Christian Hane,

Lionel Heng, Gim Hee Lee, Friedrich Fraundorfer, Rene Iser, Rudolf Triebel,

Ingmar Posner, Paul Newman, Lars Wolf, Marc Pollefeys, Stefan Brosig, Jan

Effertz, Cedric Pradalier, and Roland Siegwart. Toward automated driving

in cities using close-to-market sensors: An overview of the V-Charge project.

In Intelligent Vehicles Symposium (IV), 2013 IEEE, pages 809–816. IEEE,

2013. (Cited on page 5.)

Paul Timothy Furgale. Extensions to the visual odometry pipeline for the explo-

ration of planetary surfaces. PhD thesis, Graduate Department of Aerospace

Science and Engineering, University of Toronto, 2011. (Cited on pages 21

and 42.)

Antonin Guttman. R-trees: a dynamic index structure for spatial searching.

SIGMOD RECORD, 14(2), 1984. (Cited on page 46.)

Richard Hartley and Andrew Zisserman. Multiple view geometry in computer

vision. Cambridge university press, 2003. (Cited on pages 17 and 19.)

Page 67: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

REFERENCES 55

Lionel Heng, Paul T. Furgale, and Marc Pollefeys. Leveraging image-based lo-

calization for infrastructure-based calibration of a multi-camera rig. Journal

of Field Robotics, 2014. (Cited on page 16.)

Andrew Howard. Multi-robot mapping using manifold representations. In

Robotics and Automation, 2004. Proceedings. ICRA’04. 2004 IEEE Inter-

national Conference on, volume 4, pages 4198–4203. IEEE, 2004. (Cited on

pages 12 and 46.)

International Energy Agency IEA. Global EV outlook: understanding the

electric vehicle landscape to 2020, April 2013. (Cited on page 1.)

Hordur Johannsson, Michael Kaess, Maurice F. Fallon, and John J. Leonard.

Temporally scalable visual SLAM using a reduced pose graph. In RSS Work-

shop on Long-term Operation of Autonomous Robotic Systems in Changing

Environments, Sydney, Australia, Jul 2012. (Cited on page 45.)

Edward Johns and Guang-Zhong Yang. Feature co-occurrence maps:

Appearance-based localisation throughout the day. In Robotics and Automa-

tion (ICRA), 2013 IEEE International Conference on, pages 3212–3218.

IEEE, 2013. (Cited on page 25.)

Edward Johns and Guang-Zhong Yang. Generative methods for long-term

place recognition in dynamic scenes. International Journal of Computer

Vision, 106(3):297–314, 2014. (Cited on page 25.)

Laurent Kneip and Paul T. Furgale. OpenGV: A unified and generalized

approach to real-time calibrated geometric vision. In Proceedings of the

IEEE International Conference on Robotics and Automation (ICRA), pages

1–8, Hong Kong, China, May 3 – June 7 2014. (Cited on page 21.)

Laurent Kneip and Simon Lynen. Direct optimization of frame-to-frame ro-

tation. In Computer Vision (ICCV), 2013 IEEE International Conference

on, pages 2352–2359. IEEE, 2013. (Cited on page 19.)

Page 68: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

56 REFERENCES

Kurt Konolige and James Bowman. Towards lifelong visual maps. In Intelligent

Robots and Systems, 2009. IROS 2009. IEEE/RSJ International Conference

on, pages 1156–1163. IEEE, 2009. (Cited on pages 25, 26, 30, 31, 43, and 44.)

Kurt Konolige, James Bowman, J.D. Chen, Patrick Mihelich, Michael Calon-

der, Vincent Lepetit, and Pascal Fua. View-based maps. The International

Journal of Robotics Research, 2010. (Cited on page 17.)

Tomas Krajnik, Jaime Pulido Fentanes, Oscar Martinez Mozos, Tom Duckett,

Johan Ekekrantz, and Marc Hanheide. Long-term topological localisation for

service robots in dynamic environments using spectral maps. In IEEE/RSJ

International Conference on Intelligent Robots and Systems, 2014. (Cited

on page 24.)

Henning Lategahn and Christoph Stiller. City GPS using stereo vision. In

Vehicular Electronics and Safety (ICVES), 2012 IEEE International Con-

ference on, pages 1 –6, july 2012. doi: 10.1109/ICVES.2012.6294279. (Cited

on page 3.)

Henning Lategahn, Johannes Beck, Bernd Kitt, and Christoph Stiller. How

to learn an illumination robust image feature for place recognition. In In-

telligent Vehicles Symposium (IV), 2013 IEEE, pages 285–291. IEEE, 2013.

(Cited on page 23.)

Gim Hee Lee. Visual mapping and pose estimation for self-driving cars. PhD

thesis, Eidgenössische Technische Hochschule ETH Zürich, Nr. 21858, 2014,

2014. (Cited on page 29.)

Gim Hee Lee, Friedrich Fraundorfer, and Marc Pollefeys. Structureless pose-

graph loop-closure with a multi-camera system on a self-driving car. In

Intelligent Robots and Systems (IROS), 2013 IEEE/RSJ International Con-

ference on, pages 564–571. IEEE, 2013. (Cited on page 19.)

Page 69: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

REFERENCES 57

John J. Leonard and Hugh F. Durrant-Whyte. Simultaneous map building and

localization for an autonomous mobile robot. In IEEE/RSJ International

Workshop on Intelligent Robots and Systems, pages 1442–1447, 1991. (Cited

on page 11.)

Stefan Leutenegger, Margarita Chli, and Roland Yves Siegwart. BRISK: Bi-

nary robust invariant scalable keypoints. In Computer Vision (ICCV), 2011

IEEE International Conference on, pages 2548–2555. IEEE, 2011. (Cited

on page 18.)

Joshua Marshall, Timothy Barfoot, and Johan Larsson. Autonomous under-

ground tramming for center-articulated vehicles. Journal of Field Robotics,

25(6-7):400–421, 2008. (Cited on page 12.)

Yoshio Matsumoto, Masayuki Inaba, and Hirochika Inoue. Visual navigation

using view-sequenced route representation. In Robotics and Automation,

1996. Proceedings., 1996 IEEE International Conference on, volume 1, pages

83–88. IEEE, 1996. (Cited on page 13.)

Larry Henry Matthies. Dynamic stereo vision. PhD thesis, Carnegie Mellon

University, School of Computer Science, 1989. (Cited on page 14.)

Mladen Mazuran, Tipaldi Gian Diego, Spinello Luciano, and Wolfram Bur-

gard. Nonlinear graph sparsification for SLAM. In Proceedings of Robotics:

Science and Systems, Berkeley, USA, July 2014. (Cited on page 45.)

Colin McManus, Paul Furgale, Braden Stenning, and Timothy D Barfoot.

Lighting-invariant visual teach and repeat using appearance-based lidar.

Journal of Field Robotics, 30(2):254–287, 2013. (Cited on page 12.)

Colin McManus, Winston Churchill, Will Maddern, Alexander D. Stewart, and

Paul Newman. Shady dealings: Robust, long-term visual localisation using

illumination invariance. In Robotics and Automation (ICRA), 2014 IEEE

International Conference on, 2014a. (Cited on page 23.)

Page 70: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

58 REFERENCES

Colin McManus, Ben Upcroft, and Paul Newman. Scene signatures: Localised

and point-less features for localisation. In Proceedings of Robotics Science

and Systems (RSS), Berkley, CA, USA, July 2014b. (Cited on page 45.)

Christopher Mei and Patrick Rives. Single view point omnidirectional camera

calibration from planar grids. In Robotics and Automation, 2007 IEEE

International Conference on, pages 3945–3950. IEEE, 2007. (Cited on page

17.)

Michael J. Milford and Gordon Fraser Wyeth. SeqSLAM: Visual route-based

navigation for sunny summer days and stormy winter nights. In Robotics

and Automation (ICRA), 2012 IEEE International Conference on, pages

1643–1649. IEEE, 2012. (Cited on page 24.)

Michael Montemerlo, Sebastian Thrun, Daphne Koller, and Ben Wegbreit.

FastSLAM: A factored solution to the simultaneous localization and map-

ping problem. In AAAI/IAAI, pages 593–598, 2002. (Cited on pages 11

and 17.)

Hans Moravec. Obstacle Avoidance and Navigation in the Real World by a

Seeing Robot Rover. PhD thesis, Department of Computer Science, Stanford

University, September 1980. (Cited on page 15.)

Peter Muehlfellner, Paul T. Furgale, Wojciech Derendarz, and Roland

Philippsen. Evaluation of fisheye-camera based visual multi-session local-

ization in a real-world scenario. In Intelligent Vehicles Symposium (IV),

2013 IEEE, pages 57–62. IEEE, 2013. (Cited on pages 6, 15, 19, 22, 41,

and 45.)

Peter Muehlfellner, Mathias Buerki, Mike Bosse, Wojciech Derendarz, Roland

Philippsen, and Paul T. Furgale. Summary maps for lifelong visual local-

ization. To be published in Journal of Field Robotics, 2015a. Accepted on

Page 71: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

REFERENCES 59

March 18. 2015, Manuscript ID ROB-14-0150.R2. (Cited on pages 6, 19,

21, 28, 30, 42, 43, and 44.)

Peter Muehlfellner, Wojciech Derendarz, Paul T. Furgale, and Roland

Philippsen. A relational database design for lifelong visual mapping. Sub-

mitted to Journal of Software Engineering in Robotics, 2015b. (Cited on

pages 7, 31, 35, and 43.)

Kevin P. Murphy. Bayesian map learning in dynamic environments. In S.A.

Solla, T. K. Leen, and K. Müller, editors, Advances in Neural Information

Processing Systems 12, pages 1015–1021. MIT Press, 2000. (Cited on page

11.)

Ashley Napier, Gabe Sibley, and Paul Newman. Real-time bounded-error pose

estimation for road vehicles using vision. In Intelligent Transportation Sys-

tems (ITSC), 2010 13th International IEEE Conference on, pages 1141–

1146. IEEE, 2010. (Cited on page 14.)

Tayyab Naseer, Luciano Spinello, Wolfram Burgard, and Cyrill Stachniss. Ro-

bust visual robot localization across seasons using network flows. In AAAI,

2014. (Cited on page 24.)

Paul Newman, John J. Leonard, Juan D. Tardos, and José Neira. Explore and

return: Experimental validation of real-time concurrent mapping and local-

ization. In Robotics and Automation, 2002. Proceedings. ICRA’02. IEEE In-

ternational Conference on, volume 2, pages 1802–1809. IEEE, 2002. (Cited

on page 12.)

Kai Ni, Drew Steedly, and Frank Dellaert. Out-of-core bundle adjustment for

large-scale 3D reconstruction. In Computer Vision, 2007. ICCV 2007. IEEE

11th International Conference on, pages 1–8. IEEE, 2007. (Cited on page

45.)

Page 72: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

60 REFERENCES

Oliver Pink, Frank Moosmann, and Alexander Bachmann. Visual features

for vehicle localization and ego-motion estimation. In Intelligent Vehicles

Symposium, 2009 IEEE, pages 254–260. IEEE, 2009. (Cited on page 14.)

Eric Royer, Maxime Lhuillier, Michel Dhome, and Jean-Marc Lavest. Monoc-

ular vision for mobile robot localization and autonomous navigation. Inter-

national Journal of Computer Vision, 74(3):237–260, 2007. (Cited on page

14.)

Siniša Šegvić, Anthony Remazeilles, Albert Diosi, and François Chaumette.

A mapping and localization framework for scalable appearance-based navi-

gation. Computer Vision and Image Understanding, 113(2):172–187, 2009.

(Cited on page 14.)

Gabe Sibley, Christopher Mei, Ian Reid, and Paul Newman. Adaptive relative

bundle adjustment. In Robotics Science and Systems (RSS), Seattle, USA,

June 2009. (Cited on pages 12 and 45.)

Gabe Sibley, Christopher Mei, Ian Reid, and Paul Newman. Vast-scale outdoor

navigation using adaptive relative bundle adjustment. The International

Journal of Robotics Research, 29(8):958–980, 2010. (Cited on page 3.)

Randall Smith, Matthew Self, and Peter Cheeseman. A stochastic map for

uncertain spatial relationships. In 4th International Symposium on Robotic

Research, pages 467–474, 1987. (Cited on pages 11, 12, and 17.)

Noah Snavely, Steven M. Seitz, and Richard Szeliski. Skeletal graphs for effi-

cient structure from motion. In CVPR, volume 1, page 2, 2008. (Cited on

page 45.)

Hauke Strasdat, JosÃľ M. M. Montiel, and Andrew J. Davison. Real-time

monocular SLAM: Why filter? In Robotics and Automation (ICRA), 2010

IEEE International Conference on, pages 2657–2664. IEEE, 2010. (Cited

on page 21.)

Page 73: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

REFERENCES 61

Sebastian Thrun and Michael Montemerlo. The graph SLAM algorithm with

applications to large-scale mapping of urban structures. The International

Journal of Robotics Research, 25(5-6):403–429, 2006. (Cited on page 17.)

Sebastian Thrun, Wolfram Burgard, and Dieter Fox. Probabilistic robotics.

MIT press, 2005. (Cited on pages 11 and 17.)

Bill Triggs, Philip F McLauchlan, Richard I Hartley, and Andrew W Fitzgib-

bon. Bundle adjustmentâĂŤa modern synthesis. In Vision algorithms: the-

ory and practice, pages 298–372. Springer, 2000. (Cited on pages 17 and 19.)

Christoffer Valgren and Achim J. Lilienthal. SIFT, SURF & seasons:

Appearance-based long-term localization in outdoor environments. Robotics

and Autonomous Systems, 58(2):149–156, 2010. (Cited on page 24.)

Jürgen Wolf, Wolfram Burgard, and Hans Burkhardt. Robust vision-based

localization by combining an image-retrieval system with Monte Carlo lo-

calization. Robotics, IEEE Transactions on, 21(2):208–216, 2005. (Cited

on page 21.)

Alan M. Zhang and Lindsay Kleeman. Robust appearance based visual route

following for navigation in large-scale outdoor environments. The Interna-

tional Journal of Robotics Research, 28(3):331–356, 2009. (Cited on page

13.)

Zhengyou Zhang. Parameter estimation techniques: A tutorial with application

to conic fitting. Image and vision Computing, 15(1):59–76, 1997. (Cited on

pages 17 and 21.)

Julius Ziegler, Philipp Bender, Markus Schreiber, Henning Lategahn, To-

bias Strauss, Christoph Stiller, Thao Dang, Uwe Franke, Nils Appenrodt,

Christoph G. Keller, Eberhard Kaus, Ralf G. Herrtwich, Clemens Rabe,

David Pfeiffer, Frank Lindner, Fridtjof Stein, Friedrich Erbs, Markus En-

zweiler, Carsten Knoeppel, Jochen Hipp, Martin Haueis, Maximilian Trepte,

Page 74: Lifelong Visual Localization for Automated Vehicles - …hh.diva-portal.org/smash/get/diva2:811409/FULLTEXT01.… ·  · 2015-05-11This thesis takes on one of the major unsolved

62 REFERENCES

Carsten Brenk, Andreas Tamke, Mohammad Ghanaat, Markus Braun,

Armin Joos, Hans Fritz, Horst Mock, Martin Hein, and Eberhard Zeeb.

Making Bertha drive? an autonomous journey on a historic route. Intelli-

gent Transportation Systems Magazine, IEEE, 6(2):8–20, 2014. (Cited on

page 3.)