an online learning approach to improving the …youngliu/pub/sigmetrics15_slides.pdfto make the most...

38
An Online Learning Approach to Improving the Quality of Crowd-Sourcing Yang Liu (Joint work with Mingyan Liu) Department of Electrical Engineering and Computer Science University of Michigan, Ann Arbor, MI June 2015

Upload: others

Post on 13-Jul-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

An Online Learning Approach to Improving theQuality of Crowd-Sourcing

Yang Liu(Joint work with Mingyan Liu)

Department of Electrical Engineering and Computer ScienceUniversity of Michigan, Ann Arbor, MI

June 2015

Page 2: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

The power of crowdsourcing

Data collection: participatory sensing, user-generated map :

Liu (Michigan) Online Labeler Selection June 2015 2 / 36

Page 3: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Recommendation: rating of movies, news, restaurants, services:

Liu (Michigan) Online Labeler Selection June 2015 3 / 36

Page 4: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Social studies: opinion survey, the science of opinion survey:

Liu (Michigan) Online Labeler Selection June 2015 4 / 36

Page 5: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Crowd-sourcing market

Data processing: image labeling, annotation

Assignment

Labelers

Labeling 0 1 1

I Paid workers perform computational tasks.

I Hard to measure and evaluate quality objective: competence,bias, irresponsible behavior, etc.

Liu (Michigan) Online Labeler Selection June 2015 5 / 36

Page 6: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

One step further: Use the wisdom of crowd

Labeling

Aggregation

Assignment

Labelers

0,0,1 1,0,1 0,1,0

0 1 0

I Redundant assignment.

I Label aggregation.

Liu (Michigan) Online Labeler Selection June 2015 6 / 36

Page 7: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Our vision:

Labeling Time

Aggregation

Assignment

Labelers

0,0,1 1,0,1 0,1,0

0 1 0

I Labeler selection. (first step)

I Adaptive learning. (second step)

I A weighted version of selection. (one more step)

Liu (Michigan) Online Labeler Selection June 2015 7 / 36

Page 8: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Our objective

To make the most effective use of the crowdsourcing system

I Cost in having large amount of data labeled is non-trivial

I Time constrained machine learning tasks.

A sequential/online learning framework

I Over time learn which labelers are more competent, or whosereviews/opinion should be valued more.

I Quality control rather than random assignment

I Closed-loop, causal.

Liu (Michigan) Online Labeler Selection June 2015 8 / 36

Page 9: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Multiarmed bandit (MAB) framework

A sequential decision making and learning framework:

I Objective: select the best of a set of choices (“arms”)

I Principle: repeated sampling of different choices (“exploration”),while controlling how often each choice is used based on theirempirical quality (“exploitation”).

I Performance measure: “regret” – difference between an algorithmand a benchmark.

Liu (Michigan) Online Labeler Selection June 2015 9 / 36

Page 10: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Challenges and key ideas

Main challenge in crowdsourcing: ground truth

I True label of data remains unknown

I If view each labeler as a choice/arm: unknown quality of outcome(“reward”).

Key features:

I Mild assumption on the collective quality of the crowd; quality ofan individual is estimated against the crowd.

I Online learning: Learning occurs as data/labeling tasks arrive.

I Comparing against optimal static selections.

Liu (Michigan) Online Labeler Selection June 2015 10 / 36

Page 11: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Introduction Introduction

Outline of the talk

Problem formulation

Online solution

I Simple/weighted majority voting

Extensions and discussions

Experiment results

I Numerical results & Results on AMT data

Conclusion and on-going works

Liu (Michigan) Online Labeler Selection June 2015 11 / 36

Page 12: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Labeler selection

M labelers; labelers i has accuracy pi (can be task-dependent).

I No two exactly the same: pi 6= pj for i 6= j , and 0 < pi < 1, ∀i .I Collective quality: p̄ :=

∑i pi/M > 1/2.

I Probability that a simple majority vote over all M labelers iscorrect: amin := P(

∑i Xi/M > 1/2).

I If p̄ > 1/2 and M > log 2p̄−1/2

, then amin > 1/2.

Unlabeled tasks arrive at t = 1, 2, · · · .I User selects a subset St of labelers for task at t.

I Labeling payment of ci for each task performed by labeler i .

Liu (Michigan) Online Labeler Selection June 2015 12 / 36

Page 13: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Labeler selection

M labelers; labelers i has accuracy pi (can be task-dependent).

I No two exactly the same: pi 6= pj for i 6= j , and 0 < pi < 1, ∀i .I Collective quality: p̄ :=

∑i pi/M > 1/2.

I Probability that a simple majority vote over all M labelers iscorrect: amin := P(

∑i Xi/M > 1/2).

I If p̄ > 1/2 and M > log 2p̄−1/2

, then amin > 1/2.

Unlabeled tasks arrive at t = 1, 2, · · · .I User selects a subset St of labelers for task at t.

I Labeling payment of ci for each task performed by labeler i .

Liu (Michigan) Online Labeler Selection June 2015 12 / 36

Page 14: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Labeling outcome/Information aggregation

Labeling Time

Aggregation

Assignment

Labelers

0,0,1 1,0,1 0,1,0

0 1 0

Aggregating results from multiple labelers:

I A task receives a set of labels: {Li (t)}i∈St .

I Use simple majority voting to compute the label output: L∗(t).(extensible to weighted majority voting)

Liu (Michigan) Online Labeler Selection June 2015 13 / 36

Page 15: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Probability of correct labeling outcome: π(St).

I Optimal set of labelers: S∗ that maximizes π(S).

π(St) =∑

S :S⊆St ,|S|≥d |St |+12 e

∏i∈S

pi ·∏

j∈St\S

(1− pj)

︸ ︷︷ ︸Majority wins

+

∑S:S⊆St ,|S|= |St |2

∏i∈S pi ·

∏j∈St\S(1− pj)

2︸ ︷︷ ︸Ties broken equally likely

.

Liu (Michigan) Online Labeler Selection June 2015 14 / 36

Page 16: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Assuming known {pi}, S∗ can be obtained using a linear search

Theorem

Under the simple majority voting rule, |S∗| is an odd number.Furthermore, S∗ is monotonic: if i ∈ S∗ and j 6∈ S∗, then we musthave pi > pj .

Liu (Michigan) Online Labeler Selection June 2015 15 / 36

Page 17: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Assuming known {pi}, S∗ can be obtained using a linear search

Theorem

Under the simple majority voting rule, |S∗| is an odd number.Furthermore, S∗ is monotonic: if i ∈ S∗ and j 6∈ S∗, then we musthave pi > pj .

Liu (Michigan) Online Labeler Selection June 2015 15 / 36

Page 18: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Labeler selection Model

Objective of our online solution

Labeler expertise pi s being unknown a priori

I Goal: Gradually learn labelers’ quality and make selectionsadaptively

Performance measure:

I Comparing with the optimal selection (static):

R(T ) = Tπ(S∗)− E [T∑t=1

π(St)]

Liu (Michigan) Online Labeler Selection June 2015 16 / 36

Page 19: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution

Outline of the talk

Problem formulation

Online solution

I Simple/weighted majority voting

Extensions and discussions

Experiment results

I Numerical results & Results on AMT data

Conclusion and on-going works

Liu (Michigan) Online Labeler Selection June 2015 17 / 36

Page 20: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Algorithm

An online learning algorithm

Exploration Exploitation

Task arrivals Time

Tasks used for testing

x

Interlveaing explorations of MAB and Crowd-sourcing: Doubleexploration

I There is a set of tasks E (t) (∼ log t) used for testing purposes.

I These or their independent and identical variants are repeatedlyassigned to the labelers (∼ log t). 1

1More discussions follow later on independence.Liu (Michigan) Online Labeler Selection June 2015 18 / 36

Page 21: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Algorithm

Exploration Exploitation

Task arrivals Time

Tasks used for testing

x

Two types of time steps:

I Double exploration: all M labelers are used. Exploration isentered if (1) the number of testers falls below a threshold(∼ log t), or if (2) the number of times a tester has been testedfalls below a threshold (∼ log t).

I Exploitation: the estimated S̃∗ is used to label the arriving taskbased on the current estimated {p̃i}.

Liu (Michigan) Online Labeler Selection June 2015 19 / 36

Page 22: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Algorithm

Exploration Exploitation

Task arrivals Time

Tasks used for testing

x

Three types of tasks:

I Testers: those arriving to find (1) true and (2) false. These areadded to E (t) and are repeatedly used to collect independentlabels whenever (2) is true subsequently.

I Throw-aways: those arriving to find (2) true. These are given arandom label.

I Keepers: those arriving to find both (1) and (2) false. These aregiven a label outcome using the best estimated set of labelers.

Liu (Michigan) Online Labeler Selection June 2015 20 / 36

Page 23: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Algorithm

Exploration Exploitation

Task arrivals Time

Tasks used for testing

x

Accuracy update

I Estimated label on tester k at time t: majority label over all testoutcomes up to time t.

I p̃i at time t: the % of times i ’s label matches the majority voteknown at t out of all tests on all testers.

Liu (Michigan) Online Labeler Selection June 2015 21 / 36

Page 24: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Algorithm

Regret

Main result:

R(T ) ≤ Const(S∗,∆max,∆min, δmax, δmin, amin) log2(T ) + Const

I ∆max = maxS 6=S∗ π(S∗)− π(S), δmax = maxi 6=j |pi − pj |.I First term due to exploration; second due to exploitation.

I Can obtain similar result on the cost C (T ).

Liu (Michigan) Online Labeler Selection June 2015 22 / 36

Page 25: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Weighted majority voting

Weighted majority voting

Labeling Time

Aggregation

Assignment

Labelers

0,0,1 1,0,1 0,1,0

log𝑝

1 − 𝑝

I Each labeler i ’s decision is weighed by log pi1−pi .

Theorem

Under the weighted majority vote and assuming pi ≥ 0.5,∀i , theoptimal set S∗ is monotonic, i.e., if we have i ∈ S∗ and j 6∈ S∗ thenwe must have pi > pj .

Liu (Michigan) Online Labeler Selection June 2015 23 / 36

Page 26: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Online solution Weighted majority voting

Main results on weighted majority voting:

I R(T ) ≤ O(log2 T ), but with strictly larger constants.

I Have to account for additional error in estimating the weightswhen determining label outcome.

I A larger constant: slower convergence to a better target.

Liu (Michigan) Online Labeler Selection June 2015 24 / 36

Page 27: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Discussions

Outline of the talk

Problem formulation

Online solution

I Simple/weighted majority voting

Extensions and discussions

Experiment results

I Numerical results & Results on AMT data

Conclusion and on-going works

Liu (Michigan) Online Labeler Selection June 2015 25 / 36

Page 28: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Discussions Discussion

Re-assignment of the testers

IID noise insertion

Random delay

I Commonly adopted in survey methodology to ensure validresponses.

I Bounded delay τmax leads to τmax-fold regret.

Liu (Michigan) Online Labeler Selection June 2015 26 / 36

Page 29: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Discussions Discussion

Other extensions

Prior knowledge on several constants

I Exploration length depends on several system parameters.

I Sub-logarithmic remedy without knowing prior knowledge.

Improve the bound by improving amin: weed out bad labelers.

I Ranking based on counting of disagreement.

I Start the weeding out from the end of list.

I Requires only O(logT ) samples to achieve a bounded regret.

Liu (Michigan) Online Labeler Selection June 2015 27 / 36

Page 30: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Discussions Discussion

Other extensions cont.

Labelers with different type of tasks

I Finite number of types: Similar results.

I Infinite number of types.

I Continuous MAB.

I Sub-linear regret bound.

With delayed arrival of ground-truth

I O(logT ) regret time uniformly.

Liu (Michigan) Online Labeler Selection June 2015 28 / 36

Page 31: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Experiment results

Outline of the talk

Problem formulation

Online solution

I Simple/weighted majority voting

Extensions and discussions

Experiment results

I Numerical results & Results on AMT data

Conclusion and on-going works

Liu (Michigan) Online Labeler Selection June 2015 29 / 36

Page 32: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Experiment results Synthetic data

Experiment I: simulation with M = 5

Accumulative regret & Average regret R(T )/T

0 500 1000 1500 20000

1

2

3

4

5

6

Time steps

Acc

umat

ed re

gret

s

1 1000 2,0000

0.005

0.01

0.015

0.02

0.025

0.03

0.035

Time steps

R(T

)/T

Effect of amin: higher amin leads to much better performance.

500 1000 1500 20000

0.5

1

Time steps

Ave

rag

e e

rro

r ra

te

a_min = 0.6

a_min = 0.7

a_min = 0.8

Liu (Michigan) Online Labeler Selection June 2015 30 / 36

Page 33: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Experiment results Synthetic data

Performance comparison

labeler selection v.s. full crowd-sourcing (simple majority vote)

0 200 400 600 800 10000.65

0.7

0.75

0.8

Time

Av

era

ge

re

wa

rd

w/ LS_OL

Full crowd-sourcing

Liu (Michigan) Online Labeler Selection June 2015 31 / 36

Page 34: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Experiment results Synthetic data

Comparing weighted and simple majority vote

0 20 40 60 80 1000

50

100

Time steps

Accu

mu

late

d r

ew

ard

Simple majority votingWeighted majority voting

M 5 10 15 20

Full crowd-sourcing (majority vote) 0.5154 0.5686 0.7000 0.7997Majority vote w/ LS OL 0.8320 0.9186 0.9434 0.9820

Weighted majority vote w/ LS OL 0.8726 0.9393 0.9641 0.9890

Table: Average reward per labeler: there is a clear gap between with andwithout using LS OL.

Liu (Michigan) Online Labeler Selection June 2015 32 / 36

Page 35: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Experiment results AMT datasets

Experiment II: on a real AMT dataset

The dataset

I Contains 1,000 images each labeled by the same set of 5 AMTs.

I Labels are on a scale from 0 to 5, indicating how many scenes areseen from each image.

I A second dataset summarizing keywords for scenes of each image:use this count as the ground truth.

Counting number of disagreement (online):

AMT1 AMT2 AMT3 AMT4 AMT5

# of disagree 348 353 376 338 441

Table: Total number of disagreement each AMT has

Liu (Michigan) Online Labeler Selection June 2015 33 / 36

Page 36: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Experiment results AMT datasets

Performance comparison

0 200 400 600 800 10000.8

1

1.2

1.4

1.6

1.8

Ordered image number

Ave

rag

e e

rro

r ra

te

Full crowdw/ LS_OL

AMT5 ruled out

0 1 2 3 4 50

0.2

0.4

0.6

0.8

1

Error in labeling

CD

F

Full Crowd : avg=1.63w/ LS_OL: avg=1.36

(L) AMT 5 was quickly weeded out; eventually settled on the optimalset of AMTs 1, 2, and 4 for most of the time.

(R) CDF of all images’ labeling error at the end of this process.

Liu (Michigan) Online Labeler Selection June 2015 34 / 36

Page 37: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Conclusion Conclusion

Conclusion

We discussed a quality control problem in labeler market

I How to select the best set of labelers over a sequence of tasks.

I An algorithm that estimates labeler’s quality by comparing against(weighted) majority vote; new regret bound.

Currently under investigation

I Lower bound on the regret in the labeler selection problem.

I Hypothesis testing & coupling argument.

Liu (Michigan) Online Labeler Selection June 2015 35 / 36

Page 38: An Online Learning Approach to Improving the …youngliu/pub/sigmetrics15_slides.pdfTo make the most e ective use of the crowdsourcing system I Cost in having large amount of data

Conclusion Conclusion

Q & A

Thank you. Any question?

http://www.umich.edu/~youngliu

Liu (Michigan) Online Labeler Selection June 2015 36 / 36