dominantly truthful multi-task peer prediction mechanism ......we de ne alice’s strategy for each...

45
Dominantly truthful multi-task peer prediction mechanism with a constant number of tasks Yuqing Kong CFCS, Peking University

Upload: others

Post on 16-Oct-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Dominantly truthful multi-task peer prediction

mechanism with a constant number of tasks

Yuqing Kong

CFCS, Peking University

Page 2: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Restaurant reviews

• Do you like McDonald?

• Do you like Subway?

• Do you like KFC?

1

Page 3: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Central question

How to incentivize honest reviews?

designing proper reward systems

2

Page 4: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Central question

How to incentivize honest reviews? designing proper reward systems

2

Page 5: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Problems of traditional reward systems

Spot-checking: check the answer with some probability.

• The elicited information is unverifiable!

Agreements based reward: pay Alice one dollar if she is in the majority

(Majority vote). When there are only two participants, pay Alice and Bob

based on the number of agreements. However, it

• discourages minority

• encourages uninformative feedbacks

3

Page 6: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

This work

A novel reward system that is

• strategy-proof: truth-telling pays Alice higher than any other

strategy in expectation, regardless of other people’s behaviors; and

strictly higher than less “informative” strategies in expectation, with

mild conditions.

• prior-independent: does not need any prior knowledge;

• practical: only needs 2C∗ tasks.

To the best of our knowledge, this is the first strategy-proof reward

system that only needs finite number of tasks, not to say a small

constant number of tasks

∗Each task is a multi-choice question and C is the number of choices. For example,

C = 2 in our running example.

4

Page 7: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Model and assumptions

Page 8: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Model

Multi-task information elicitation

5

Page 9: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Model

T restaurants with a review space C (e.g. for binary reviews C = 0, 1)

Alice’s honest reviews (unknown): x1A, x

2A, · · · , xTA

Alice’s reports: x̂1A, x̂

2A, · · · , x̂TA

Bob’s honest reviews (unknown): x1B , x

2B , · · · , xTB

Bob’s reports: x̂1B , x̂

2B , · · · , x̂TB

6

Page 10: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

“Peer” prediction

The reward function only depends on their reports

{x̂ tA, x̂ tB}t

7

Page 11: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Strategy

Definition: Strategy

We define Alice’s strategy for each task t as a C × C transition matrix

StA where St

A(x tA, x̂tA) is the probability that Alice reports x̂ tA given that

she receives private signal x tA.

The profile of Alice and Bob’s strategies is called strategy profile.

8

Page 12: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Strategy

Truth-telling[1 0

0 1

]Random guessing[.7 .3

.7 .3

]Mixed strategy[.7 .3

.5 .5

]

Permutation[0 1

1 0

]Always say 1[

0 1

0 1

]Mixed strategy[

0 1

.5 .5

]

9

Page 13: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Strategy

Truth-telling[1 0

0 1

]Fully informative strategies

Random guessing[.7 .3

.7 .3

]Uninformative strategies

Mixed strategy[.7 .3

.5 .5

]Partially informative strategies

Permutation[0 1

1 0

]

Always say 1[0 1

0 1

]

Mixed strategy[0 1

.5 .5

]

10

Page 14: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Assumptions

Assumption 1.1: A priori similar tasks

All tasks have the same prior, i.e., the same joint distribution over Alice

and Bob’s honest reviews.

This implies that {x tA, x tB}t are i.i.d. samples.

Assumption 1.2: Consistent strategies

Alice (Bob) play the same strategy for all tasks.

∃SA s.t. S tA = SA,∀t

This implies that {x̂ tA, x̂ tB}t are i.i.d. samples.

Assumption 1.3: Informative prior

Alice and Bob’s honest reviews are “strictly correlated”.

11

Page 15: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Assumptions

Assumption 1.1: A priori similar tasks

All tasks have the same prior, i.e., the same joint distribution over Alice

and Bob’s honest reviews.

This implies that {x tA, x tB}t are i.i.d. samples.

Assumption 1.2: Consistent strategies

Alice (Bob) play the same strategy for all tasks.

∃SA s.t. S tA = SA,∀t

This implies that {x̂ tA, x̂ tB}t are i.i.d. samples.

Assumption 1.3: Informative prior

Alice and Bob’s honest reviews are “strictly correlated”.

11

Page 16: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Assumptions

Assumption 1.1: A priori similar tasks

All tasks have the same prior, i.e., the same joint distribution over Alice

and Bob’s honest reviews.

This implies that {x tA, x tB}t are i.i.d. samples.

Assumption 1.2: Consistent strategies

Alice (Bob) play the same strategy for all tasks.

∃SA s.t. S tA = SA,∀t

This implies that {x̂ tA, x̂ tB}t are i.i.d. samples.

Assumption 1.3: Informative prior

Alice and Bob’s honest reviews are “strictly correlated”.

11

Page 17: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

“Strictly Correlated”

Random variables XA,XB : Alice and Bob’s honest reviews for a random

task.

Definition: Joint distribution matrix

UX ,Y (x , y) = Pr[X = x ,Y = y ].

Alice and Bob’s honest reviews are “strictly correlated” means

det(UXA,XB) 6= 0

12

Page 18: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Mechanism design goals

Dominant truthfulness (strategy-proof)

Truth-telling is

• a dominant strategy

• strictly dominant “non-permutation” strategies with mild conditions;

Approximated version: (ε, δ)-Dominant truthfulness:

with probability 1− δ, truth-telling > other strategy −ε

13

Page 19: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Previous works

Page 20: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Previous works

Dasgupta and Ghosh 2013 [1]: starts multi-task peer prediction and

proposes a mechanism where truth-telling is a strict equilibrium and the

best† equilibrium. However,

• binary-choice

• limited prior: only works for positively correlated participants

• not strategy-proof

†the highest amount agent welfare

14

Page 21: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Previous works

(ε, δ)-Dominantly Truthful Dominantly Truthful

CA Mechanism(N/A) Infinite

Shnayder, Agarwal, Frongillo, Parkes 2016 [4]

f -Mutual Information(N/A) Infinite

Mechanism Kong and Schoenebeck 2016 [2]

Dominant Truthful SerumO(− log δ

ε2 ) InfiniteLiu and Chen 2018 [3]

Table 1: A Task Sample Complexity Comparison of Multi-task Peer Prediction

15

Page 22: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Previous works

Dominantly truthful, but

• require infinite number of tasks

Finite number of tasks, but

• only has approximated truthfulness

16

Page 23: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Gap

The existence of dominantly truthful mechanism that works for finite

number of tasks.

17

Page 24: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Why need an infinite number

The mechanism needs an infinite number of tasks to know that whether

Alice and Bob have

• similar tastes and should be rewarded for agreements

• opposite tastes and should be rewarded for disagreements

• complicated relation and should be rewarded for complicated

agreements

18

Page 25: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Why need an infinite number

The mechanism needs an infinite number of tasks to know that whether

Alice and Bob have

• similar tastes and should be rewarded for agreements

• opposite tastes and should be rewarded for disagreements

• complicated relation and should be rewarded for complicated

agreements

18

Page 26: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Why need an infinite number

The mechanism needs an infinite number of tasks to know that whether

Alice and Bob have

• similar tastes and should be rewarded for agreements

• opposite tastes and should be rewarded for disagreements

• complicated relation and should be rewarded for complicated

agreements

18

Page 27: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Why need an infinite number

The mechanism needs an infinite number of tasks to learn Alice and

Bob’s information structure.

19

Page 28: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

This work

(ε, δ)-Dominantly Truthful Dominantly Truthful

CA Mechanism(N/A) Infinite

[Shnayder, Agarwal, Frongillo, Parkes 2016]

f -Mutual Information(N/A) Infinite

Mechanism [Kong and Schoenebeck 2016]

Dominant Truthful SerumO(− log δ

ε2 ) Infinite[Liu and Chen 2018]

DMI-Mechanism 2C 2C

Table 2: A Task Sample Complexity Comparison of Multi-task Peer Prediction

20

Page 29: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Solution

Page 30: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Solution

DMI-Mechanism (Determinant based Mutual Information Mechanism)

21

Page 31: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Main theorem

Main theorem

When T1,T2 ≥ C , DMI-mechanism is dominantly truthful,

prior-independent and works for ≥ 2C tasks and ≥ 2 participants.

22

Page 32: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

The core of DMI-Mechanism

For two random variables X ,Y which have the same support C,

Joint distribution matrix

UX ,Y (x , y) = Pr[X = x ,Y = y ].

Determinant based Mutual Information (DMI)

DMI(X ;Y ) = | det(UX ,Y )|.

Like Shannon’s mutual information, DMI is information-monotone.

Ideally, to get strategy-proof, we want the expected payment to be the

DMI between Alice’s report and Bob’s report.

However, the “absolute” symbol is a problem. Thus, we design the

system such that the expected payment is DMI’s square to remove the

“absolute” symbol. That’s why we divide the questions into two sets.

23

Page 33: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Proof sketch

• Step 1: DMI is information-monotone

• DMI(X ′;Y ) ≤ DMI(X ;Y )‡

• DMI is also symmetric, non-negative, and bounded.

• Step 2: The expected payment is the square of the DMI between

Alice’s report and Bob’s report:

EpA = EpB ∝ DMI(X̂A; X̂B)2§

• E detM1,E detM2 ∝ det(UX̂A;X̂B)

• EpA = EpB = E detM1 detM2 = ¶E detM1E detM2 =

DMI(X̂A; X̂B)2

‡X ′ is independent of Y conditioning X .§X̂A, X̂B : Alice and Bob’s actual reviews for a random task.¶This equality is valid since M1 and M2 are independent.

24

Page 34: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

DMI is information-monotone

Proof.When X ′ is independent of Y conditioning X ,

UX ′,Y (x ′, y) = Pr[X ′ = x ′,Y = y ] =∑x

Pr[X ′ = x ′|X = x ] Pr[X = x ,Y = y ].

Thus, UX ′,Y = UX ′|XUX ,Y .

DMI(X ′;Y ) =| det(UX ′,Y )|=| det(UX ′|XUX ,Y )| (UX ′,Y = UX ′|XUX ,Y )

=DMI(X ;Y )| det(UX ′|X )| (det(AB) = det(A) det(B))

≤DMI(X ;Y )

25

Page 35: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

EpA = EpB ∝ DMI2(X̂A; X̂B)

EM`(0, 0) = T`UX̂A;X̂B(0, 0)

E det(M`) = (T`)C det(UX̂A;X̂B

)?

26

Page 36: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

EpA = EpB ∝ DMI2(X̂A; X̂B)

EM`(0, 0) = T`UX̂A;X̂B(0, 0)

E det(M`) = (T`)C det(UX̂A;X̂B

)?

No, the matrix’s entries are not independent.

27

Page 37: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

EpA = EpB ∝ DMI2(X̂A; X̂B)

It(c , c′) := 1 ((x̂ tA, x̂

tB) = (c , c ′))

28

Page 38: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

EpA = EpB ∝ DMI2(X̂A; X̂B)

It(c , c′) := 1 ((x̂ tA, x̂

tB) = (c , c ′))

t` = (t`(1), t`(2), · · · , t`(C )): ordered C distinct tasks in T`

It`(π) := Πc It`(c)(c , π(c))

We can show

det(M`) =∑

t`

∑π

sgn(π)It`(π)

29

Page 39: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

EpA = EpB ∝ DMI2(X̂A; X̂B)

Moreover since distinct tasks are independent,

EX̂A,X̂BIt`(π) =EX̂A,X̂B

Πc It`(c)(c , π(c))

=∏c

UX̂A,X̂B(c , π(c))

Then

E det(M`) = (#t`s) ∗ det(UX̂A;X̂B)

Thus EpA = EpB ∝ DMI2(X̂A; X̂B).

30

Page 40: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Information evaluation without verification

Additionally, this work also proposes a novel scoring rule that

• identifies high-quality information without verification;

• only needs ≥ 3 participants

Score every participant via the sum of the DMI between her report and

other participants’ reports

See the full paper for more properties of this evaluation rule.

31

Page 41: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Conclusion and future work

Page 42: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Conclusion and future work

A novel reward system that is

• dominantly truthful

• prior-independent

• practical

Future work

• real-world experiments

• full characterization of all strategy-proof mechanisms

• relaxing i.i.d. assumption

32

Page 43: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

References i

A. Dasgupta and A. Ghosh.

Crowdsourced judgement elicitation with endogenous

proficiency.

In Proceedings of the 22nd international conference on World Wide

Web, pages 319–330. International World Wide Web Conferences

Steering Committee, 2013.

Y. Kong and G. Schoenebeck.

An information theoretic framework for designing information

elicitation mechanisms that reward truth-telling.

ACM Trans. Econ. Comput., 7(1):2:1–2:33, Jan. 2019.

Y. Liu and Y. Chen.

Surrogate scoring rules and a dominant truth serum for

information elicitation.

CoRR, abs/1802.09158, 2018.

33

Page 44: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

References ii

V. Shnayder, A. Agarwal, R. Frongillo, and D. C. Parkes.

Informed truthfulness in multi-task peer prediction.

In Proceedings of the 2016 ACM Conference on Economics and

Computation, pages 179–196. ACM, 2016.

34

Page 45: Dominantly truthful multi-task peer prediction mechanism ......We de ne Alice’s strategy for each task t as a C C transition matrix S t A where S A (x t A;x^ A) is the probability

Thanks for listening!

34