recommendation independence - kamishima · model-based approaches for independence-enhanced...

41
Recommendation Independence Toshihiro Kamishima*, Shotaro Akaho*, Hideki Asoh*, and Jun Sakuma** www.kamishima.net *National Institute of Advanced Industrial Science and Technology (AIST), Japan **University of Tsukuba; and RIKEN Center for Advanced Intelligence Project, Japan 1st Conference on Fairness, Accountability, and Transparency (FAT* 2018) @ New York, USA, Feb. 24, 2018 1 START

Upload: others

Post on 17-Aug-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Recommendation Independence

Toshihiro Kamishima*, Shotaro Akaho*, Hideki Asoh*, and Jun Sakuma**www.kamishima.net

*National Institute of Advanced Industrial Science and Technology (AIST), Japan **University of Tsukuba; and RIKEN Center for Advanced Intelligence Project, Japan

1st Conference on Fairness, Accountability, and Transparency (FAT* 2018)@ New York, USA, Feb. 24, 2018

1START

Page 2: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Outline

2

Basics of Recommendation recommendation tasks, collaborative filtering

Recommendation Independence sensitive feature, recommendation independence

Applications of Recommendation Independence adherence to laws and regulations, fair treatment of content providers, exclusion of Unwanted Information

Independence-Enhanced Recommendation regularization approach, model-based approach

Conclusions & Future work

We overview our series of researches on fairness in recommendationrather then focusing on this paper

Page 3: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

3

Basics of Recommendation

Page 4: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Recommender System

4

Recommenders: Tools to help identify worthwhile stuff[Konstan+ 03]

Find Good Items Predicting Ratings[Herlocker+ 04, Gunawardana+ 09]

✽ Screen-shots are acquired from Amazon.co.jp and Movielens.org on 2007-07-26

Ranking items according to users' preference, to help for finding at least one target item

Presenting items with predicted ratings for a user, to help for exploring items

Page 5: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Collaborative Filtering

5

✽ There are other approaches: content-based filtering or knowledge-based filtering

Collaborative filtering is a major approach for predicting users' preference in a word-of-mouth manner

recommending items liked by those who having similar preferences

Any good sushi

restaurant?

I'll go tothe “Taro”

The “Taro”is awesome

They like the “Taro” restaurant

people who like similar tastes of sushi

I like the “Taro”

Page 6: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

6

Recommendation Independence

Page 7: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Sensitive Feature

7

S : sensitive feature This represents information that should be ignored in a recommendation processIts values are determined depending on a user and/or an item

As in a case of standard recommendation, we use random variablesX: a user, Y: an item, and R: a recommendation outcome

A sensitive feature is restricted to a binary type

We adopt a variable required for recommendation independence

Ex. Sensitive feature = movie’s popularity / user’s gender

Page 8: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Recommendation Independence

8

No information about a sensitive feature influences the outcomeThe status of the sensitive feature is explicitly excluded from the inference of the recommendation outcome

Recommendation Independence statistical independence

between a recommendation outcome, R, and a sensitive feature, S

Independence-Enhanced Recommendation Preferred items are predicted

so as to satisfy a constraint of recommendation independence

Pr[R | S] = Pr[R] ≡ R ⫫ S

[Kamishima+ 12, Kamishima+ 13, Kamishima+ 16, Kamishima+18]

Page 9: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

dislike likedislike like

Effect of Independence Enhancement

9

Standard Independence-enhanced

two distributions are largely diverged

two distributions become closer

The bias that older movies were rated higher could be successfully canceled by enhancing independence

✽ each bin of histograms of predicted scores for older and newer moviesa sensitive feature = whether a movie is newer or older

Page 10: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

10

Applications of Recommendation Independence

Page 11: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Adherence to Laws and Regulations

11

[Sweeney 13]

A recommendation service must be managed while adhering to laws and regulations

suspicious placement in keyword-matching advertisements

Advertisements indicating arrest records were more frequently displayed for names that are more popular among individuals of African descent than those of European descent

↓Socially discriminative treatments must be avoided

sensitive feature = users’ demographic information ↓

Legally or socially sensitive informationcan be excluded from the inference process of recommendation

Page 12: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Fair Treatment of Content Providers

12

System managers should fairly treat their content providers

The US FTC has investigated Google to determine whether the search engine ranks its own services higher than those of competitors

Fair treatment in search engines

sensitive feature = a content provider of a candidate item ↓

Information about who provides a candidate item can be ignored,and providers are treated fairly

Fair treatment in recommendation

A hotel booking site should not abuse their position to recommend hotels of its group company

[Bloomberg]

Page 13: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Exclusion of Unwanted Information

13

[TED Talk by Eli Pariser, http://www.filterbubble.com/]

sensitive feature = a political conviction of a friend candidate ↓

Information about whether a candidate is conservative or progressivecan be ignored in a recommendation process

Filter Bubble: To fit for Pariser’s preference, conservative people are eliminated from his friend recommendation list in Facebook

Information unwanted by a user is excluded from recommendation

Page 14: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

14

Independence-Enhanced Recommendation

Page 15: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Independence-Enhanced Recommendation

15

[Kamishima+ 12, Kamishima+ 13, Kamishima+ 16, Kamishima+18]

Predicting Ratings: a task to predict a rating value that a user would provide to an item

Dataset

Random variables: user X, item Y, rating R, sensitive feature S

Prediction Function

Dataset

Prediction Function

Standard Recommendation Independence-Enhanced Recommendation

D = {(xi, yi, ri)} D = {(xi, yi, ri, si)}

Çr(x, y) Çr(x, y, s)

Page 16: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Regularization Approach

16

[Kamishima+ 12, Kamishima+ 13, Kamishima+18]

Objective Function independence parameter: control the balance between independence and accuracy

independence term: a regularizer to constrain independenceThe larger value indicates that recommendation outcomes and sensitive values are more independent

Regularization Approach: Adopting a regularizer imposing a constraint of independence while training a recommendation model

L2 regularizerregularization

parameterempirical loss

≥D loss�ri, Çr(xi, yi, si)

�* ⌘ indep(R,S) + 1

2�Ò⇥Ò2<latexit sha1_base64="8wP7gPw9zmMwC1CZuBT8jcc/W50=">AAADh3ichVJLbxMxEJ50oZTwSIBLEReLKCgRIXJygNJTeBzg1m4SGqkui3fjJFb3pfUmSrrsiRuXHjlwASQOiJ/BhT/AoT+h4likXkBidpPwqlps2R5/nm/m89imb0sVUrqXWdBOnV48s3Q2e+78hYu5/KXLT5Q3DCzRtjzbCzomV8KWrmiHMrRFxw8Ed0xbbJjbD5LzjZEIlPTcVjjxxZbD+67sSYuHCBn5F0wNHYNZ3H5ImMPDQeBEtqdUzEzZL5HAkBXCBjyMgrg0TjaTZFKGLJPEo0xuESZCziq/2NLtCj8u6RXSLJObhPUCbkW1OKrHzEZdXU7Yc+T2WgPkof20buQLtErTRo4atZlRaJR3d54BwJqXPwQGXfDAgiE4IMCFEG0bOCjsm1ADCj5iWxAhFqAl03MBMWSRO0QvgR4c0W2c+7jbnKEu7pOYKmVbmMXGESCTQJF+oR/oAf1MP9J9+v3YWFEaI9EywdWccoVv5F4uNw//y3JwDWHwm3Wi5hB6sJJqlajdT5HkFtaUP9p5ddBc1YvRDfqOfkX9b+ke/YQ3cEffrPfrQn+N0YtAdn/ob/TFEzK5WIUk8nhWV5VWdTzNgmOumqAeL32RVbSb0ILH0PkDPf728wjzqiWVV8mL4Qep/fsdjhrtevVula7XCo37MG1LcA2uQwm/wx1owCNYgzYm3M/kMsuZq1pWo9ptbWXqupCZca7AX0279xPLQNen</latexit><latexit sha1_base64="Wl/8P0HKUViixFbqSLJCtPOamUk=">AAADh3ichVJLbxMxEJ50oYTwSIBLEReLKCgRIXJygNJTeBzg1uZBI3XLyrtxEqv70tqJki574salR5C4ABJIiJ/BhT/AoT+h4likXkBidpPwqtrasj3+PN/M57FN3xZSUbqbWtBOnV48kz6bOXf+wsVs7tLlJ9IbBhZvW57tBR2TSW4Ll7eVUDbv+AFnjmnzdXPrQXy+PuKBFJ7bUhOfbzqs74qesJhCyMg91+XQMXSL2Q+J7jA1CJzQ9qSMdFP0iyQwRJnoA6bCICqO480knqQhSiT2KJFbROeK6eXfbOF2uR8VG2XSLJGbRO8FzAqrUViLdBt1dRnRnyG31xogD+2nNSOXpxWaNHLYqM6MfL20s+2nP7xa9XIHoEMXPLBgCA5wcEGhbQMDiX0DqkDBR2wTQsQCtERyziGCDHKH6MXRgyG6hXMfdxsz1MV9HFMmbAuz2DgCZBIo0K/0I92nX+gnukd/HBkrTGLEWia4mlMu943si6XmwYksB1cFgz+sYzUr6MFyolWgdj9B4ltYU/5o++V+c6VRCG/Qd/Qb6n9Ld+lnvIE7+m69X+ON1xi9AGTnZ+NNY/GYTC5WIY48ntVVJlUdT7PgmKsmqMdLXmQF7Sa04DF0/kKPvv08wrxqceVl/GL4Qar/f4fDRrtWuVuha9V8/T5MWxquwXUo4ne4A3V4BKvQxoR7qWxqKXVVy2hUu60tT10XUjPOFfinafd+AQ4C2P0=</latexit><latexit sha1_base64="Wl/8P0HKUViixFbqSLJCtPOamUk=">AAADh3ichVJLbxMxEJ50oYTwSIBLEReLKCgRIXJygNJTeBzg1uZBI3XLyrtxEqv70tqJki574salR5C4ABJIiJ/BhT/AoT+h4likXkBidpPwqtrasj3+PN/M57FN3xZSUbqbWtBOnV48kz6bOXf+wsVs7tLlJ9IbBhZvW57tBR2TSW4Ll7eVUDbv+AFnjmnzdXPrQXy+PuKBFJ7bUhOfbzqs74qesJhCyMg91+XQMXSL2Q+J7jA1CJzQ9qSMdFP0iyQwRJnoA6bCICqO480knqQhSiT2KJFbROeK6eXfbOF2uR8VG2XSLJGbRO8FzAqrUViLdBt1dRnRnyG31xogD+2nNSOXpxWaNHLYqM6MfL20s+2nP7xa9XIHoEMXPLBgCA5wcEGhbQMDiX0DqkDBR2wTQsQCtERyziGCDHKH6MXRgyG6hXMfdxsz1MV9HFMmbAuz2DgCZBIo0K/0I92nX+gnukd/HBkrTGLEWia4mlMu943si6XmwYksB1cFgz+sYzUr6MFyolWgdj9B4ltYU/5o++V+c6VRCG/Qd/Qb6n9Ld+lnvIE7+m69X+ON1xi9AGTnZ+NNY/GYTC5WIY48ntVVJlUdT7PgmKsmqMdLXmQF7Sa04DF0/kKPvv08wrxqceVl/GL4Qar/f4fDRrtWuVuha9V8/T5MWxquwXUo4ne4A3V4BKvQxoR7qWxqKXVVy2hUu60tT10XUjPOFfinafd+AQ4C2P0=</latexit>

Page 17: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Independence Terms

17

Mutual Information with Histogram Models [Kamishima+ 12]computationally inefficient

Mean Matching [Kamishima+ 13]

matching means of predicted ratings for distinct sensitive groupsimproved computational efficiency, but considering only means

Mutual Information with Normal Distributions [Kamishima+ 18]

Distribution Matching with Bhattacharyya Distance [Kamishima+ 18]

These two terms can take both means and variances into account, and are computationally efficient

*�mean

�D(0)� *mean�D(1)��2

*⇠* ln î ˘

Pr[rS=0] Pr[rS=1]dr⇡

*

⇠H (R) *

≥s Pr[s] H (Rs)

Page 18: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Model-based Approach

18

Model-based Approach: a sensitive variable is added to a recommendation model so that it satisfies an independence constraint

[Kamishima+ 16]

z

y

x r

s

z

y

x rstandard model independence-enhanced model

user

itemrating

latent variable sensitivevariable

independent

A model-based approach is inferior to a regularization approach↑

Generative process of ratings are assumed to be probabilistic in this model, but it is actually deterministic [Kamishima+ 18b]

[Hofmann 99]

Page 19: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Experiments: Accuracy vs Fairness

19

✽ Details of experimental conditions are shown in Sec. 4.2.1 and Fig. 2

We apply a regularization method with mutual information with normal distributions to Movielens 1M with the Year sensitive featureThe changes of accuracy and independence measures according as the enhancement of Independence

η0

0.05

0.10

0.15

0.20

10−2 10−1 1 101 102η0.64

0.66

0.68

0.70

0.72

0.74

10−2 10−1 1 101 102fa

ir

accu

rate independence-enhanced

standard independence-enhanced

standard

Recommendation independence could be successfully enhanced by slightly sacrificing prediction accuracy

Accuracy mean absolute error

Independence Kolmogorov-Smirnov test statistics

Page 20: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Experiments: Means & Variances

20

η3.0

3.2

3.4

3.6

3.8

4.0

0.4

0.6

0.8

1.0

1.2

1.4

10−2 10−1 1 101 102η

3.0

3.2

3.4

3.6

3.8

4.0

0.4

0.6

0.8

1.0

1.2

1.4

10−2 10−1 1 101 102

✽ Details of experimental conditions are shown in Sec. 4.2.1 and Fig. 3

Comparison between our previous method, mean match, with our latest method, mutual information with normal distributionsThe changes of means and variances of predicted ratings according as the enhancement of independence

differences ofstandard deviations

differences ofmeans

mean matching mutual information with Gaussian models

Our previous method cannot control the variances of predicted ratings, but our new method can

matchnot match

Page 21: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Conclusions

21

Contributions We proposed a notion of recommendation independenceWe have been developed methods for independence-enhanced recommendationEnhancement of recommendation independence have been empirically examinedOur new independence terms could take variances of outcomes into account

Future work Recommendation independence for a find-good-items taskSensitive features other than a binary type, such as a continuous typeOther types of independence, such as equalized odds [Hardt+ 17, Yao+ 18]

In cases that are not point-estimation, such as Bayesian inferenceIntroduce conditional fairness or confounding variables

Page 22: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Additional Information

22

Program Codes (plan to open in March) http://www.kamishima.net/iers/

Our Survey Slide of Fairness-Aware Data Mining http://www.kamishima.net/archive/fadm.pdf

Acknowledgment We gratefully acknowledge the valuable comments and suggestions of Dr. Paul ResnickWe would like to thank the Grouplens research lab and Dr. Mohsen Jamali for providing datasetsThis work is supported by MEXT/JSPS KAKENHI Grant Number JP24500194, JP15K00327, and JP16H02864

Page 23: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Bibliography I

Ò. Celma and P. Cano.

From hits to niches?: or how popular artists can bias music recommendation and

discovery.

In Proc. of the 2nd KDD Workshop on Large-Scale Recommender Systems and the Netflix

Prize Competition, 2008.

S. Forden.

Google said to face ultimatum from FTC in antitrust talks.

Bloomberg, Nov. 13 2012.

hhttp://bloom.bg/PPNEaSi.

A. Gunawardana and G. Shani.

A survey of accuracy evaluation metrics of recommendation tasks.

Journal of Machine Learning Research, 10:2935–2962, 2009.

M. Hardt, E. Price, and N. Srebro.

Equality of opportunity in supervised learning.

In Advances in Neural Information Processing Systems 29, 2016.

J. L. Herlocker, J. A. Konstan, L. G. Terveen, and J. T. Riedl.

Evaluating collaborative filtering recommender systems.

ACM Trans. on Information Systems, 22(1):5–53, 2004.

23

Page 24: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Bibliography II

T. Hofmann and J. Puzicha.

Latent class models for collaborative filtering.

In Proc. of the 16th Int’l Joint Conf. on Artificial Intelligence, pages 688–693, 1999.

M. Jamali and M. Ester.

A matrix factorization technique with trust propagation for recommendation in social

networks.

In Proc. of the 4th ACM Conf. on Recommender Systems, pages 135–142, 2010.

T. Kamishima and S. Akaho.

Considerations on recommendation independence for a find-good-items task.

In Workshop on Responsible Recommendation, 2017.

T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma.

Enhancement of the neutrality in recommendation.

In The 2nd Workshop on Human Decision Making in Recommender Systems, 2012.

T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma.

E�ciency improvement of neutrality-enhanced recommendation.

In The 3rd Workshop on Human Decision Making in Recommender Systems, 2013.

T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma.

Model-based and actual independence for fairness-aware classification.

Data Mining and Knowledge Discovery, 32:258–286, 2018.

24

Page 25: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Bibliography III

T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma.

Recommendation independence.

In The 1st Conf. on Fairness, Accountability and Transparency, volume 81 of PMLR,

pages 187–201, 2018.

T. Kamishima, S. Akaho, H. Asoh, and I. Sato.

Model-based approaches for independence-enhanced recommendation.

In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages 860–867, 2016.

J. A. Konstan and J. Riedl.

Recommender systems: Collaborating in commerce and communities.

In Proc. of the SIGCHI Conf. on Human Factors in Computing Systems, Tutorial, 2003.

Y. Koren.

Factorization meets the neighborhood: A multifaceted collaborative filtering model.

In Proc. of the 14th ACM SIGKDD Int’l Conf. on Knowledge Discovery and Data Mining,

pages 426–434, 2008.

E. Pariser.

The Filter Bubble: What The Internet Is Hiding From You.

Viking, 2011.

R. Salakhutdinov and A. Mnih.

Probabilistic matrix factorization.

In Advances in Neural Information Processing Systems 20, pages 1257–1264, 2008.

25

Page 26: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Bibliography IV

L. Sweeney.

Discrimination in online ad delivery.

Communications of the ACM, 56(5):44–54, 2013.

S. Yao and B. Huang.

Beyond parity: Fairness objectives for collaborative filtering.

In Advances in Neural Information Processing Systems 30, 2017.

C. N. Ziegler, S. M. McNee, J. A. Konstan, and G. Lausen.

Improving recommendation lists through topic diversification.

In Proc. of the 14th Int’l Conf. on World Wide Web, pages 22–32, 2005.

26

Page 27: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

27

Extra Slides

Page 28: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Probabilistic Matrix Factorization

28

[Salakhutdinov 08, Koren 08]

Çr(x, y) = � + bx + cy + pxqÒy

Probabilistic Matrix Factorization Model predict a preference rating of an item y rated by a user x

well-performed and widely used

For a given training dataset, model parameters are learned by minimizing the squared loss function with an L2 regularizer.

cross effect ofusers and itemsglobal bias

user-dependent bias item-dependent bias

≥D (ri * Çr(xi, yi))2 + � Ò⇥Ò2

Prediction Function

Objective Function

L2 regularizer

regularization parameter

squared loss function

Page 29: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Regularization Approach

29

[Kamishima+ 12, Kamishima+ 13, Kamishima+18]

≥D�ri * Çr(xi, yi, si)

�2 * ⌘ indep(R,S) + �Ò⇥Ò2

a prediction function is selected according to a sensitive value

sensitive feature

Çr(x, y, s) = �(s) + b(s)x + c(s)y + p(s)x q(s)yÒ

Prediction Function

Objective Function independence parameter: control the balance between the independence and accuracy

independence term: a regularizer to constrain independenceThe larger value indicates that ratings and sensitive values are more independent Matching means of predicted ratings for two sensitive values

Page 30: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Latent Class Model

30

[Hofmann 99]

Prediction:

z

y

x

r

Latent Class Model: A probabilistic model for collaborative filtering

A basic topic model, pLSA

extended so as to be able to deal with ratings r given by users x to items y

Çr(x, y) = EPr[rx,y][level(r)]= ≥

r Pr[rx, y] level(r)the r-th rating value

A rating value can be predicted by the expectation of ratings

Model parameters can be learned by an EM algorithm

latent topic variable

Page 31: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Independence-Enhanced LCM

31

[Kamishima+ 16]

z

y

x

r

sz

y

x

r

s

Independence-Enhancement by a Model-based Approach A sensitive variable is embedded into the original LCM

A rating and a sensitive variable are mutually independentA user, an item, and a rating are conditionally independent given Z

A type 2 model can more strictly enhance recommendation independence, because in addition to X and Y, Z depends on a sensitive variable

Type 1 model Type 2 model

Page 32: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Ugly Duckling Theorem

32

[Watanabe 69]

Ugly Duckling Theorem In classification, one must emphasize some features of objects and must ignore the other features

It is impossible to make recommendation that is independent from any sensitive features

a sensitive feature must be specified by a user and other features are ignored

In a case of a Facebook example, A recommender system enhances independent from a political conviction, but it is allowed to make biased recommendations in terms of other features, for example, the birthplace or age of friend candidates

Independence-enhanced Recommendation

Page 33: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

For a Find-Good-Items Task

33

cross-entropy loss

Find Good Items: a task to find some items preferred by a user↓

making a preference score independent, instead of a predicted rating

≥D CE�ri * Çr(xi, yi, si)

�* ⌘ indep(R,S) + �Ò⇥Ò2

Çr(x, y, s) = sig��(s) + b(s)x + c(s)y + p(s)x q(s)y

Ò�

Preference Score: How strongly a user prefers an item

enhancing the independence between a preference score and a sensitive feature

sigmoid function

[Kamishima+ 17]

Page 34: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Kolmogorov-Smirnov Statistic

34

The statistic of the two-sample Kolmogorov-Smirnov test a nonparametric test for the equality of two distribution

↓Evaluating the degree of independence

by measuring the equality between Pr[R | S=0] and Pr[R | S=1]

Kolmogorov-Smirnov statistic the area between two empirical

cumulative distributions

Wikipedia

Page 35: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Preference Score vs Sensitive Feature

35

Observation 1: A preference score could be successfully made independent from a sensitive feature

Observation 2: A ranking accuracy (AUC) did not worsen so much by enhancement of the recommendation independenceThis is was contrasted with the increase of a prediction error (MAE) in a predicting ratings task

AUC

η0.70

0.75

0.80

0.85

0.90

10−2 10−1 1 101 102

KS

η0

0.05

0.10

0.15

0.20

10−2 10−1 1 101 102

Kolmogolov-Smirnov statisticbetween Pr[R | S=0] and Pr[R | S=1]

ranking accuracy (AUC)

[Kamishima+ 17]

Page 36: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Relevance and Sensitive Feature

36

1.0 0.9 0.7 0.10.30.51.0

Recommending top-k items whose preference scores are the largest

sort according to their preference scoresselect top-k items

irrelevant itemsrelevant items

check the independence from a relevance, not from a preference score

Observation 3: The relevance of items was not independent from a sensitive feature for some values of k, in particular, small k

↓A need for a new method that fits for a ranked item list

[Kamishima+ 17]

Page 37: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Popularity Bias

37

Popularity Biasthe tendency for popular items to be recommended more frequently

Flixster data The degree popularity of an item is measured

by the number of users who rated the item

Short-head items are frequently and highly rated

[Jamali+ 10]

long-tail (bottom 99%) share in ratings: 52.8%

mean rating: 3.53

short-head (top 1%) share in ratings: 47.2%

mean rating: 3.71

sensitive feature = popularity of items ↓

Popularity bias can be corrected

[Celma 08]

Page 38: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Recommendation Diversity

38

[Ziegler+ 05]

Recommendation Diversity Similar items are not recommended in a single list, to a single user,

to all users, or in a temporally successive listsrecommendation list

similar itemsexcluded

Diversity Items that are similar in a specified metric are excluded from recommendation results

The mutual relations among results

Independence Information about a sensitive feature is excluded from recommendation results

The relations between results and sensitive values

Page 39: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Diversity vs Independence

39

Diversity Depending on the definition of

similarity measures

Independence Depending on the specification of

sensitive feature

Similarity A function of two items

Sensitive Feature A function of a user-item pair

Because a sensitive feature depends on a user, neutrality can be applicable for coping with users’ factor, such as, users’ gender or age, which cannot be straightforwardly dealt by using diversity

Page 40: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Diversity vs Independence

40

short-head

long-tail

short-head

long-tail

standard recommendation diversified recommendation

Because a set of recommendations are diversified by abandoning short-head items, predicted ratings are still biased

Prediction ratings themselves are unbiased by enhancing recommendation independence

Page 41: Recommendation Independence - Kamishima · Model-based approaches for independence-enhanced recommendation. In Proc. of the IEEE 16th Int’l Conf. on Data Mining Workshops, pages

Privacy-Preserving Data Mining

41

recommendation results, R, and sensitive features, S,are statistically independent

In a context of privacy-preservation Even if the information about R is disclosed,

the information about S will not exposed

mutual information between a recommendation result, R,and a sensitive feature, S, is zero

I(R; S) = 0

In particular, a notion of the t-closeness has strong connection