presented by: john paisley duke university, ece

10
The Phylogenetic Indian Buffet Process: A Non-Exchangeable Nonparametric Prior for Latent Features By: Kurt T. Miller, Thomas L. Griffiths and Michael I. Jordan ICML 2008 Presented by: John Paisley Duke University, ECE

Upload: may

Post on 13-Jan-2016

30 views

Category:

Documents


2 download

DESCRIPTION

The Phylogenetic Indian Buffet Process : A Non-Exchangeable Nonparametric Prior for Latent Features By: Kurt T. Miller, Thomas L. Griffiths and Michael I. Jordan ICML 2008. Presented by: John Paisley Duke University, ECE. Motivation. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Presented by: John Paisley Duke University, ECE

The Phylogenetic Indian Buffet Process: A Non-Exchangeable Nonparametric Prior for Latent Features

By: Kurt T. Miller, Thomas L. Griffiths and Michael I. JordanICML 2008

Presented by: John Paisley

Duke University, ECE

Page 2: Presented by: John Paisley Duke University, ECE

Motivation

• Nonparametric models are often used with the assumption of exchangeability. – The Indian Buffet Process is an example

• Sometimes, non-exchangeable models might be more appropriate.– The Phylogenetic Indian Buffet Process– Similar to the IBF, but uses additional information of

how related diners are with each other.– These relationships are captured in a tree structure.

Page 3: Presented by: John Paisley Duke University, ECE

Indian Buffet Process

Page 4: Presented by: John Paisley Duke University, ECE

Phylogenetic Indian Buffet Process

• Uses a tree to model columns zk

• This is done as follows:– Assign the root node to be zero

– Along an edge of distance t, let this change to a 1 with probability , where

. The distance from every leaf to the root is 1.

– If a 0 is changed to a 1 along a path to a node, all subsequent nodes are 1 and therefore so are the leaves.

Page 5: Presented by: John Paisley Duke University, ECE
Page 6: Presented by: John Paisley Duke University, ECE

Sampling Issues

• For (1), use the sum-product algorithm (Pearl, 1988).

• For (2), use the chain rule of probability.

• An MCMC inference algorithm is given in detail.

Page 7: Presented by: John Paisley Duke University, ECE

Experimental Results• Elimination by Aspects (EBA) model

– A Choice Model

• Let there be i objects and zik indicate the ith object has the kth feature. Let each feature have a weight, wk. The EBA model defines the probability of choosing object I over j as

• The likelihood of an observation matrix, X, is

• This has been modeled using the IBP.

Page 8: Presented by: John Paisley Duke University, ECE

Experimental Results• Consider now an underlying tree

structure to this model.

• Preference trees: Out of 9 personalities, 3 movie stars, 3 athletes and 3 politicians, people made the 36 pairwise choices of whom they would rather spend time with. Here, L is the length of the edge of each general category to a leaf.

• A soft version of this tree is modeled with the pIBP using data generated from this model with L = 0.1

Page 9: Presented by: John Paisley Duke University, ECE

Experimental Results

• Example results: As the number of samples decreases, the pIBP is able to infer the structure better than the IBP because of the prior.

Page 10: Presented by: John Paisley Duke University, ECE

Experimental Results

• As can be seen, the additional structure in the model produces better results.