crowdsourcing satellite imagery (talk at giscience2012)

29
Crowdsourcing satellite imagery: study of iterative vs. parallel models Nicolas Maisonneuve, Bastien Chopard Twitter: nmaisonneuve 1 Friday, September 21, 12

Upload: nicolas-maisonneuve

Post on 24-Jan-2015

9.962 views

Category:

Technology


1 download

DESCRIPTION

Talk given at the GIScience2012 conference (http://www.giscience.org/) More info about this work on my blog http://goo.gl/giouF

TRANSCRIPT

Page 1: Crowdsourcing satellite imagery (Talk at Giscience2012)

Crowdsourcing satellite imagery: study of iterative vs. parallel models

Nicolas Maisonneuve, Bastien Chopard

Twitter: nmaisonneuve

1Friday, September 21, 12

Page 2: Crowdsourcing satellite imagery (Talk at Giscience2012)

Damage assessment after a humanitarian crisis

2Friday, September 21, 12

Page 3: Crowdsourcing satellite imagery (Talk at Giscience2012)

Port-au-prince: 300K buildings assessed in 3 months for 8 UNOSAT experts

3Friday, September 21, 12

Page 4: Crowdsourcing satellite imagery (Talk at Giscience2012)

Organizational challenges: How to organize non-trained volunteers,

especially to enforce quality?

4Friday, September 21, 12

Page 5: Crowdsourcing satellite imagery (Talk at Giscience2012)

Organizational challenges: How to organize non-trained volunteers,

especially to enforce quality?

Investigated scope:

• Qualitative + Quantitative study of 2 collaborative models inspired by Computer science: iterative vs parallel information processing

• Controlled experiment to isolate quality = F(organisation), removing other parameters e.g. training, task difficulty

• this research != studying real world collaborative practices but more extreme/symbolic cases to guide collaborative system designers

5Friday, September 21, 12

Page 6: Crowdsourcing satellite imagery (Talk at Giscience2012)

Organizational challenges: How to organize non-trained volunteers,

especially to enforce quality?

Investigated scope:

• Qualitative + Quantitative study of 2 collaborative models inspired by Computer science: iterative vs parallel information processing

• Controlled experiment to isolate quality = F(organisation), removing other parameters e.g. training, task difficulty

• this research != studying real world collaborative practices but more extreme/symbolic cases to guide collaborative system designers

6Friday, September 21, 12

Page 7: Crowdsourcing satellite imagery (Talk at Giscience2012)

Organizational challenges: How to organize non-trained volunteers,

especially to enforce quality?

Investigated scope:

• Qualitative + Quantitative study of 2 collaborative models inspired by Computer science: iterative vs parallel information processing

• Controlled experiment to isolate quality = F(organisation), removing other parameters e.g. training, task difficulty

• this research != studying real world collaborative practices but more extreme/symbolic cases to guide collaborative system designers

7Friday, September 21, 12

Page 8: Crowdsourcing satellite imagery (Talk at Giscience2012)

Tested Collaborative Models (1/2) iterative model

e.g. wikipedia, open street map, assembly lines

8Friday, September 21, 12

Page 9: Crowdsourcing satellite imagery (Talk at Giscience2012)

Tested Collaborative Models (2/2) parallel model

aggregation

e.g. voting systems in society, distributed computing9Friday, September 21, 12

Page 10: Crowdsourcing satellite imagery (Talk at Giscience2012)

Tested Collaborative Models (2/2) parallel model

old version (17th to mid 20th century): when computers were human/women (Mathematical Table project - (1938 -1948)

10Friday, September 21, 12

Page 11: Crowdsourcing satellite imagery (Talk at Giscience2012)

Qualitative comparisonIterative Parallel

problem divisibility

No need to divide complex problem

Complex problem need to be divided in easier pieces

11Friday, September 21, 12

Page 12: Crowdsourcing satellite imagery (Talk at Giscience2012)

Qualitative comparisonIterative Parallel

problem divisibility

No need to divide complex problem

Complex problem need to be divided in easier pieces

optimization tradeoff

copy emphasizing exploitation

isolation emphasizing exploration

12Friday, September 21, 12

Page 13: Crowdsourcing satellite imagery (Talk at Giscience2012)

Qualitative comparisonIterative Parallel

problem divisibility

No need to divide complex problem

Complex problem need to be divided in easier pieces

optimization tradeoff

copy emphasizing exploitation

isolation emphasizing exploration

quality mechanism sequential improvement redundancy + diversity of

opinions

13Friday, September 21, 12

Page 14: Crowdsourcing satellite imagery (Talk at Giscience2012)

Qualitative comparisonIterative Parallel

problem divisibility

No need to divide complex problem

Complex problem need to be divided in easier pieces

optimization tradeoff

copy emphasizing exploitation

isolation emphasizing exploration

quality mechanism sequential improvement redundancy + diversity of

opinions

side effect path dependency effect + sensitivity to vandalism

useless redundancy for obvious decisions + pb of aggregation

14Friday, September 21, 12

Page 15: Crowdsourcing satellite imagery (Talk at Giscience2012)

Controlled Experiment: web platform

Interface/instruction for the Parallel model

15Friday, September 21, 12

Page 16: Crowdsourcing satellite imagery (Talk at Giscience2012)

on 3 maps with different topologies (annotated by 1 UNITAR expert)

16Friday, September 21, 12

Page 17: Crowdsourcing satellite imagery (Talk at Giscience2012)

Participants used for the experiments: Mechanical Turk as simulator

17Friday, September 21, 12

Page 18: Crowdsourcing satellite imagery (Talk at Giscience2012)

Quality of the collective output• type I errors = p(wrong annotation)• type II errors = p(missing a building)• Consistency

Data Quality Metrics

Analogy with the information retrieval field:• Precision = p(an annotation is a building)• Recall = p(a building is annotated)• F-measure = score mixing recall + precision• (metrics adjusted with tolerance distance)

18Friday, September 21, 12

Page 19: Crowdsourcing satellite imagery (Talk at Giscience2012)

Step 1 - collecting independent contribution: N for (map1, map2, map3) = (121,120,113)

Methodology for parallel model

19Friday, September 21, 12

Page 20: Crowdsourcing satellite imagery (Talk at Giscience2012)

Step 2 - for each map,generating the set of groups of m=[1 to N] participants

Methodology for parallel model

m = 1

m = 2

m = 3

20Friday, September 21, 12

Page 21: Crowdsourcing satellite imagery (Talk at Giscience2012)

Step 3 - for each group: aggregating + computing quality

Methodology for parallel model

groups of m = 2

Spatial Clustering of points + quorum

Compute Data Quality with Gold Standard

Precision Recall F-measure

21Friday, September 21, 12

Page 22: Crowdsourcing satellite imagery (Talk at Giscience2012)

The more = the better? (parallel model)

yes but until some points.. • (Adding more people wont change the consensus panel) • Limitation of Linus’ law (compared to iterative model e.g. openstreetmap)

• Wisdom != skill: we can’t replace training by more people

avg.

F-m

easu

re

22Friday, September 21, 12

Page 23: Crowdsourcing satellite imagery (Talk at Giscience2012)

Methodology for Iterative model

sample of an iterative process for map3

23Friday, September 21, 12

Page 24: Crowdsourcing satellite imagery (Talk at Giscience2012)

Methodology for Iterative model

Collected data for map1, map2, map3 = 13, 21,25 instances of about 10 iterations

n instances of about m iterations

24Friday, September 21, 12

Page 25: Crowdsourcing satellite imagery (Talk at Giscience2012)

Methodology for Iterative model

Step 2- for each iteration, we compute the precision, recall, f-measure of all the instances

Precision Recall F-measure

25Friday, September 21, 12

Page 26: Crowdsourcing satellite imagery (Talk at Giscience2012)

Intrepretation of results / Comparison on data quality

(*) but parallel < iterative in difficult cases (map 2) (lack of consensus)

Parallel Iterative

Accuracy - wrong annotations

consensual results (*) error propagation

Accuracy - missing buildings

useless redundancy on obvious buildings

accumulation of knowledge driving attention on uncovered area

Consistency redundancy naive last = best

26Friday, September 21, 12

Page 27: Crowdsourcing satellite imagery (Talk at Giscience2012)

Side-objective: Measuring how the crowd spatially agrees

Method: taking randomly 2 participants and measure their spatial inter-agreement (e.g. ratio of points matching) and repeat

the process N time

27Friday, September 21, 12

Page 28: Crowdsourcing satellite imagery (Talk at Giscience2012)

Side-objective: Measuring how the crowd spatially agrees

Method: taking randomly 2 participants and measure their spatial inter-agreement (e.g. ratio of points matching) and repeat

the process N time

way to measure the intrinsic difficulty of a task (map 1 = easy , map 2 = quite hard)

28Friday, September 21, 12

Page 29: Crowdsourcing satellite imagery (Talk at Giscience2012)

Impact of the organization beyond data quality• Energy / Footprint to collectively solve a problem, • Participation sustainability, • On Individual behavior (skill Learning & Enjoyment)

Skill complementarity: Is the best group of 3 people the best 3 people at the individual level? data says no!

Other symbolic organisations / mechanism: • human cellular automata (cell = 1 person, resubmit a task at time t, because influenced by peers results generated at time t-1)• Integration of Game design / Gamification

future tracks

29Friday, September 21, 12