composite marginal likelihood methods for …plackett-luce model theorems 1 & 2: strict...

1
RESEARCH POSTER PRESENTATION DESIGN © 2015 www.PosterPresentations.com Rank Aggregation Introduction Random Utility Models High-level message: Breaking: symmetric RUMs: uniform breaking Plackett-Luce: weighted union of position-k breakings CML weights: connected and symmetric Theoretical Results Gaussian Random Utility Model Experiments Summary and Future Work RBCML: fast and accurate due to strict concavity and asymptotic normality. Future Work: to extend RBCML to partial orders. References Hossein Azari Soufiani, David C. Parkes, and Lirong Xia, “Computing Parametric Ranking Models via Rank-Breaking”, In proceedings of the 31 st International Conference on Machine Learning, 2014. Lucas Maystre and Matthias Grossglauser , “Fast and Accurate Inference of Plackett-Luce Models”, in Advances in Neural Information Processing Systems, 2015. Ashish Khetan and Sewoong Oh, “Data-Driven Rank Breaking for Efficient Rank Aggregation”, in Journal of Machine Learning Research, 2016. Parameter space: Θ ={ Ԧ | 1 ,… } Sample space: full rankings Rensselaer Polytechnic Institute Zhibing Zhao and Lirong Xia Composite Marginal Likelihood Methods for Random Utility Models Alternatives: Agents: Plackett-Luce Model Parameter space: Θ PL ={ Ԧ | 1 ,… } Sample space: full rankings 0.3 0.2 0.5 Pr = 0.5 0.5 + 0.3 + 0.2 Rank-Breaking Dunkin’ Donuts Haagen- Dazs Veggiegrill Dunkin’ Donuts 2 2 Haagen- Dazs 1 2 Veggiegrill 1 1 Composite Marginal Likelihood Ԧ = arg max log Pr | Ԧ + strictly concave + asymptotically normal + fast + easy to implement - restricted to full rankings (future direction) Plackett-Luce Model Theorems 1 & 2: Strict log-concavity is preserved under convolution and marginalization. Theorems 3 & 4: (strongly connected ⊗ () is desired.) For Plackett-Luce model or RUMs where CDF of each utility distribution is strictly log-concave, the composite likelihood function (objective function of RBCML) is strictly log-concave if and only if ⊗ () is weakly connected. RBCML is bounded if and only if ⊗ () is strongly connected. Theorem 5: RBCML is consistent and asymptotic normal. Theorems 6 & 7: When CML weight is uniform, RBCML is consistent if and only if (i) for Plackett-Luce model, the breaking is weighted union of position-k breakings; (ii) for symmetric RUMs, the breaking is uniform. Theorems 8 & 9: RBCML is consistent if and only if W is connected and symmetric and (i) for Plackett-Luce model, the breaking is weighted union of position-k breakings; (ii) for symmetric RUMs, the breaking is uniform. Adaptive RBCML Input: Profile P of n rankings, number of iterations T, the heuristic of breaking ( Ԧ ) and weights ( Ԧ ). Output: estimated parameter Ԧ Initialize Ԧ (0) = 0 For =1 to T do Compute ( Ԧ (−1) ) and W( Ԧ (−1) ) Estimate Ԧ () using ( Ԧ (−1) ) and W( Ԧ (−1) ) using RBCML End for Ԧ () RBCML Ԧ (+1) ( Ԧ () ), ( Ԧ () ) Ranking Models Parameter Data (distribution) Model Parameter Estimation [email protected] http://homepages.rpi.edu/~zhaoz6 /

Upload: others

Post on 16-Aug-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Composite Marginal Likelihood Methods for …Plackett-Luce Model Theorems 1 & 2: Strict log-concavity is preserved under convolution and marginalization. Theorems 3 & 4: (strongly

RESEARCH POSTER PRESENTATION DESIGN © 2015

www.PosterPresentations.com

Rank Aggregation

Introduction

Random Utility Models

High-level message:

Breaking: symmetric RUMs: uniform breaking

Plackett-Luce: weighted union of position-k breakings

CML weights: connected and symmetric

Theoretical Results

Gaussian Random Utility Model

Experiments

Summary and Future Work

RBCML: fast and accurate due to strict concavity

and asymptotic normality.

Future Work: to extend RBCML to partial orders.

References

Hossein Azari Soufiani, David C. Parkes, and Lirong Xia,

“Computing Parametric Ranking Models via Rank-Breaking”, In

proceedings of the 31st International Conference on Machine

Learning, 2014.

Lucas Maystre and Matthias Grossglauser, “Fast and Accurate

Inference of Plackett-Luce Models”, in Advances in Neural

Information Processing Systems, 2015.

Ashish Khetan and Sewoong Oh, “Data-Driven Rank Breaking for

Efficient Rank Aggregation”, in Journal of Machine Learning

Research, 2016.

Parameter space: Θ𝑅𝑈𝑀 = { Ԧ𝜃|𝜃1, … 𝜃𝑚}

Sample space: full rankings

Rensselaer Polytechnic Institute

Zhibing Zhao and Lirong Xia

Composite Marginal Likelihood Methods for Random Utility Models

≻ ≻

≻≻

≻ ≻

Alternatives:

Agents:

Plackett-Luce Model

Parameter space: ΘPL = { Ԧ𝜃|𝜃1, … 𝜃𝑚}

Sample space: full rankings

0.3 0.2

0.5

Pr 𝑡𝑜𝑝 =0.5

0.5 + 0.3 + 0.2

Rank-Breaking

Dunkin’ Donuts

Haagen-Dazs

Veggiegrill

Dunkin’ Donuts

2 2

Haagen-Dazs

1 2

Veggiegrill 1 1

Composite Marginal Likelihood

Ԧ𝜃∗ = argmax

𝑖≠𝑗

log Pr 𝑎𝑖 ≻ 𝑎𝑗| Ԧ𝜃𝜅𝑖𝑗𝑤𝑖𝑗

+ strictly concave

+ asymptotically normal

+ fast

+ easy to implement

- restricted to full rankings

(future direction)

Plackett-Luce Model

Theorems 1 & 2: Strict log-concavity is preserved under

convolution and marginalization.

Theorems 3 & 4: (strongly connected 𝑊⊗𝐺(𝑃) is desired.) For

Plackett-Luce model or RUMs where CDF of each utility

distribution is strictly log-concave, the composite likelihood

function (objective function of RBCML) is strictly log-concave if

and only if 𝑊⊗𝐺(𝑃) is weakly connected. RBCML is bounded if

and only if 𝑊⊗𝐺(𝑃) is strongly connected.

Theorem 5: RBCML is consistent and asymptotic normal.

Theorems 6 & 7: When CML weight is uniform, RBCML is

consistent if and only if (i) for Plackett-Luce model, the breaking is

weighted union of position-k breakings; (ii) for symmetric RUMs,

the breaking is uniform.

Theorems 8 & 9: RBCML is consistent if and only if W is

connected and symmetric and (i) for Plackett-Luce model, the

breaking is weighted union of position-k breakings; (ii) for

symmetric RUMs, the breaking is uniform.

Adaptive RBCML

Input: Profile P of n rankings, number of iterations T, the heuristic

of breaking 𝐺( Ԧ𝜃) and weights 𝑊( Ԧ𝜃).

Output: estimated parameter Ԧ𝜃∗

Initialize Ԧ𝜃(0) = 0

For 𝑡 = 1 to T do

Compute 𝐺( Ԧ𝜃(𝑡−1)) and W( Ԧ𝜃(𝑡−1))

Estimate Ԧ𝜃(𝑡) using 𝐺( Ԧ𝜃(𝑡−1)) and W( Ԧ𝜃(𝑡−1)) using RBCML

End for

𝐺 𝑊

Ԧ𝜃(𝑡) RBCML Ԧ𝜃(𝑡+1)

𝐺( Ԧ𝜃(𝑡)),𝑊( Ԧ𝜃(𝑡))

Ranking Models

ParameterData

(distribution)

Model

Parameter Estimation

[email protected]

http://homepages.rpi.edu/~zhaoz6/