pick a crowd

23
Pick-A-Crowd: Tell Me What You Like, and I’ll Tell You What to Do A Crowdsourcing Platform for Personalized Human Intelligence Task Assignment Based on Social Networks Djellel E. Difallah , Gianluca Demartini, Philippe Cudré-Mauroux eXascale Infolab University of Fribourg, Switzerland 1 15 th May 2013, WWW 2013 - Rio De Janeiro, Brazil

Upload: exascale-infolab

Post on 10-May-2015

183 views

Category:

Business


0 download

TRANSCRIPT

Page 1: Pick a Crowd

1

Pick-A-Crowd: Tell Me What You Like,and I’ll Tell You What to Do

A Crowdsourcing Platform for PersonalizedHuman Intelligence Task Assignment Based on Social

Networks

Djellel E. Difallah, Gianluca Demartini, Philippe Cudré-MaurouxeXascale Infolab

University of Fribourg, Switzerland15th May 2013, WWW 2013 - Rio De Janeiro, Brazil

Page 2: Pick a Crowd

2

Crowdsourcing

• Exploit human intelligence to solve tasks that are simple for Humans and complex for machines

• Examples: – Wikipedia, reCaptcha, Duolingo

• Incentives– Financial, fun, visibility

Page 3: Pick a Crowd

3

Motivation

• The Pull Methodology is suboptimal

Effective workers

Actual workers

Max Overlap

Page 4: Pick a Crowd

4

Motivation

• The Push Methodology is a Task-to-Worker Recommender System.

Page 5: Pick a Crowd

5

Contribution and Claim

• Pick-A-Crowd: A system architecture that uses Task-to-Worker matching:– The worker’s social profile – The task context

• Workers can provide higher quality answers on tasks they relate to

Page 6: Pick a Crowd

7

Worker Social Profiling

“YouAreWhatYouLike”

Page 7: Pick a Crowd

8

Image Tagging

Problem Definition (1)-The Human Intelligence Task (HIT)

Data CollectionSurveyCategorization

Batch of Tasks:TitleBatch InstructionSpecific task instruction*Task data:

- Text.- Options.- Additional data (image, Url)

List of categories*

Page 8: Pick a Crowd

9

Problem Definition (2)-The Worker

Completed HITs: 256Approval Rate: 96%Qualification TypesGeneric Qualifications

Page:- Title- Category- Description- Feed, etc.

Page:- Title- Category- Description- Feed, etc.

Page:- Title- Category- Description- Feed, etc.

Page 9: Pick a Crowd

10

Problem Definition (3) –Task-to-Worker Matching

Batch of Tasks:TitleBatch InstructionSpecific task instruction*Task data:

- Text.- Options.- Additional data (image,

Url)List of categories*

Page:- Title- Category- Description- Feed, etc.

Page:- Title- Category- Description- Feed, etc.

Page:- Title- Category- Description- Feed, etc.

1- Task-to-Page Matching Function- Category- Expert finding- Semantic

2- Worker Ranking

Page 10: Pick a Crowd

11

Matching Models (1/3)–Category Based

• The requester provides a list of categories related to the batch• We create a subset of pages whose category is in the category

list of the batch• Rank the workers by the number of liked pages in the subset

Page 11: Pick a Crowd

12

Matching Models (2/3) –Expert Finding

• Build an inverted index on the pages’ titles and description• Use the title/description of the tasks as a key word query on

the inverted index and get a subset of pages• Rank the workers by the number of liked pages in the subset

Page 12: Pick a Crowd

Matching Models (3/3) –Semantic Based

• Link the context to an external knowledge base (e.g., DBPedia)• Exploit the underlying graph structure to determine the Hits and Pages similarity

– Assumption that a worker who likes a page is able to answer questions about related entities– Worker who likes a page is able to answer questions about entities of the same type

• Rank the workers by the number of liked pages in the subset

13

HIT FB Pages

Similarity

Relatedness

Type-Similarity

Page 13: Pick a Crowd

15

Pick-A-Crowd Architecture

Page 14: Pick a Crowd

16

Experimental Evaluation

• The Facebook app OpenTurk implements part of the Pick-A-Crowd architecture:– More than 170 registered workers participated– Over 12k pages crawled

• Covered both multiple answer questions as well as open-ended questions– 50 images with multiple choice question and 5 candidate answers

(Soccer, Actors, Music, Authors,Movies, Animes)– Answer 20 open-ended questions related to the topic (Cricket)

Page 15: Pick a Crowd

18

OpenTurk app

Page 16: Pick a Crowd

19

Evaluation -Correlation between the crowd accuracy and the number of relevant likes (Category Based)

WO

RKER

PR

ECIS

ION

NUMBER OF RELEVANT LIKES

Page 17: Pick a Crowd

20

Evaluation (Baseline) –Amazon Mechanical Turk (AMT)

AMT 3 = Majority vote of 3 workersAMT 5 = Majority vote of 5 workers

Page 18: Pick a Crowd

21

Evaluation – HIT Assignment Models

CATEGORY APPROACH

Page 19: Pick a Crowd

22

Evaluation – HIT Assignment Models

EXPERT FINDING BASED

TITLE/INSTRUCTION CONTENT

Page 20: Pick a Crowd

23

Evaluation – HIT Assignment Models

SEMANTIC BASED

TYPE RELATEDNESS

Page 21: Pick a Crowd

24

Evaluation -Comparison With Mechanical Turk

AMT

PICK

-A-C

ROW

D

Page 22: Pick a Crowd

25

Conclusions and Future Work

• Pull vs. Push methodologies in Crowdsourcing • Pick-A-Crowd system architecture with Task-

to-Worker recommendation• Experimental comparison with AMT shows a

consistent quality improvement“Workers Know what they Like”

• Exploit more of the social activity, and handle content-less tasks

Page 23: Pick a Crowd

26

Next Step

• We are building a Crowdsourcing platform for the research community

• Pre-register on:

www.openturk.com

Thank You!