undirected probabilistic graphical models (markov nets)

64
Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis Lecture)

Upload: pancho

Post on 22-Feb-2016

47 views

Category:

Documents


0 download

DESCRIPTION

Undirected Probabilistic Graphical Models (Markov Nets). (Slides from Sam Roweis Lecture). Connection to MCMC: MCMC requires sampling a node given its markov blanket Need to use P( x|MB (x)). For Bayes nets MB(x) contains more - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Undirected Probabilistic Graphical Models (Markov Nets)

Undirected Probabilistic Graphical Models(Markov Nets)

(Slides from Sam Roweis Lecture)

Page 2: Undirected Probabilistic Graphical Models (Markov Nets)
Page 3: Undirected Probabilistic Graphical Models (Markov Nets)
Page 4: Undirected Probabilistic Graphical Models (Markov Nets)
Page 5: Undirected Probabilistic Graphical Models (Markov Nets)
Page 6: Undirected Probabilistic Graphical Models (Markov Nets)

Connection to MCMC: MCMC requires sampling a node given its markov blanket Need to use P(x|MB(x)). For Bayes nets MB(x) contains more nodes than are mentioned in the local distribution CPT(x) For Markov nets,

Page 7: Undirected Probabilistic Graphical Models (Markov Nets)
Page 8: Undirected Probabilistic Graphical Models (Markov Nets)
Page 9: Undirected Probabilistic Graphical Models (Markov Nets)

Because neighbor relation is symmetric nodes xi and xj are both neighbors of each other..

In contrast, note that in Bayes Nets, CPTs can be filled with any real numbers between 0 and 1, and we can be sure the ensuing product will define a valid joint distribution!

Page 10: Undirected Probabilistic Graphical Models (Markov Nets)
Page 11: Undirected Probabilistic Graphical Models (Markov Nets)
Page 12: Undirected Probabilistic Graphical Models (Markov Nets)
Page 13: Undirected Probabilistic Graphical Models (Markov Nets)

12/2

All project presentations on 12/14 (10min each)

All project reports due on 12/14On 12/7, we will read and discuss MLN paper

Today: Complete discussion of Markov Nets; Start towards MLN

Page 14: Undirected Probabilistic Graphical Models (Markov Nets)

A

B

C

D Qn: What is the most likely configuration of A&B?

Fact

or sa

ys a=

b=0

But, marginal says

a=0;b=1!

Moral: Factors are not marginals!

Although A,B wouldLike to agree, B&CNeed to agree, C&D need to disagreeAnd D&A need to agree.and the latter three haveHigher weights! Mr. & Mrs. Smith example

Okay, you convinced methat given any potentialswe will have a consistentJoint. But given any joint,will there be a potentials I can provide?

Hammersley-Clifford theorem…

We can have potentials on any cliques—not just the maximal ones. So, for example we can have a potential on A in addition to the other four pairwise potentials

Page 15: Undirected Probabilistic Graphical Models (Markov Nets)
Page 16: Undirected Probabilistic Graphical Models (Markov Nets)
Page 17: Undirected Probabilistic Graphical Models (Markov Nets)

Markov Networks• Undirected graphical models

Cancer

CoughAsthma

Smoking

Potential functions defined over cliquesSmoking Cancer Ф(S,C)

False False 4.5

False True 4.5

True False 2.7

True True 4.5

c

cc xZxP )(1)(

x c

cc xZ )(

Page 18: Undirected Probabilistic Graphical Models (Markov Nets)
Page 19: Undirected Probabilistic Graphical Models (Markov Nets)

Log-Linear models for Markov NetsA

B

C

D

Factors are “functions” over their domainsLog linear model consists of Features fi (Di ) (functions over domains) Weights wi for features s.t.

Without loss of generality!

Page 20: Undirected Probabilistic Graphical Models (Markov Nets)

Markov Networks• Undirected graphical models

Log-linear model:

Weight of Feature i Feature i

otherwise0

CancerSmokingif1)CancerSmoking,(1f

5.11 w

Cancer

CoughAsthma

Smoking

iii xfw

ZxP )(exp1)(

Page 21: Undirected Probabilistic Graphical Models (Markov Nets)
Page 22: Undirected Probabilistic Graphical Models (Markov Nets)
Page 23: Undirected Probabilistic Graphical Models (Markov Nets)

Markov Nets vs. Bayes Nets

Property Markov Nets Bayes NetsForm Prod. potentials Prod. potentials

Potentials Arbitrary Cond. probabilities

Cycles Allowed Forbidden

Partition func. Z = ? global Z = 1 localIndep. check Graph separation D-separation

Indep. props. Some Some

Inference MCMC, BP, etc. Convert to Markov

Page 24: Undirected Probabilistic Graphical Models (Markov Nets)

Inference in Markov Networks• Goal: Compute marginals & conditionals of

• Exact inference is #P-complete• Most BN inference approaches work for MNs too

– Variable Elimination used factor multiplication—and should work without change..

• Conditioning on Markov blanket is easy:

• Gibbs sampling exploits this

exp ( )( | ( ))

exp ( 0) exp ( 1)i ii

i i i ii i

w f xP x MB x

w f x w f x

1( ) exp ( )i ii

P X w f XZ

exp ( )i i

X i

Z w f X

Page 25: Undirected Probabilistic Graphical Models (Markov Nets)

MCMC: Gibbs Sampling

state ← random truth assignmentfor i ← 1 to num-samples do for each variable x sample x according to P(x|neighbors(x)) state ← state with new value of xP(F) ← fraction of states in which F is true

Page 26: Undirected Probabilistic Graphical Models (Markov Nets)

Other Inference Methods

• Many variations of MCMC• Belief propagation (sum-product)• Variational approximation• Exact methods

Page 27: Undirected Probabilistic Graphical Models (Markov Nets)

Learning Markov Networks

• Learning parameters (weights)– Generatively– Discriminatively

• Learning structure (features)• Easy Case: Assume complete data

(If not: EM versions of algorithms)

Page 28: Undirected Probabilistic Graphical Models (Markov Nets)

Entanglement in log likelihood…a b c

Page 29: Undirected Probabilistic Graphical Models (Markov Nets)

Learning for log-linear formulation

Use gradient ascent

Unimodal, because Hessian is Co-variance matrix over features

What is the expectedValue of the feature given the current parameterizationof the network?

Requires inference to answer(inference at every iteration— sort of like EM )

Page 30: Undirected Probabilistic Graphical Models (Markov Nets)

Why should we spend so much time computing gradient?

• Given that gradient is being used only in doing the gradient ascent iteration, it might look as if we should just be able to approximate it in any which way– Afterall, we are going to take a step with some arbitrary

step size anyway..• ..But the thing to keep in mind is that the gradient is a

vector. We are talking not just of magnitude but direction. A mistake in magnitude can change the direction of the vector and push the search into a completely wrong direction…

Page 31: Undirected Probabilistic Graphical Models (Markov Nets)

Generative Weight Learning• Maximize likelihood or posterior probability• Numerical optimization (gradient or 2nd order) • No local maxima

• Requires inference at each step (slow!)

No. of times feature i is true in data

Expected no. times feature i is true according to model

)()()(log xnExnxPw iwiwi

1( ) exp ( )i ii

P X w f XZ

exp ( )i i

X i

Z w f X

Page 32: Undirected Probabilistic Graphical Models (Markov Nets)

Alternative Objectives to maximize..

• Since log-likelihood requires network inference to compute the derivative, we might want to focus on other objectives whose gradients are easier to compute (and which also –hopefully—have optima at the same parameter values).

• Two options:– Pseudo Likelihood– Contrastive Divergence

Given a single data instance x log-likelihood is

Log prob of data Log prob of all other possible data instances (w.r.t. current q

Maximize the distance (“increase the divergence”)

Pick a sample of typical other instances(need to sample from Pq Run MCMC initializing withthe data..)

Compute likelihood ofeach possible data instancejust using markov blanket (approximate chain rule)

Page 33: Undirected Probabilistic Graphical Models (Markov Nets)

Pseudo-Likelihood

• Likelihood of each variable given its neighbors in the data

• Does not require inference at each step• Consistent estimator• Widely used in vision, spatial statistics, etc.• But PL parameters may not work well for

long inference chains

i

ii xneighborsxPxPL ))(|()(

[Which can lead to disasterous results]

Page 34: Undirected Probabilistic Graphical Models (Markov Nets)

Discriminative Weight Learning

• Maximize conditional likelihood of query (y) given evidence (x)

• Approximate expected counts by counts in MAP state of y given x

No. of true groundings of clause i in data

Expected no. true groundings according to model

),(),()|(log yxnEyxnxyPw iwiwi

Page 35: Undirected Probabilistic Graphical Models (Markov Nets)

Structure Learning

• How to learn the structure of a Markov network?– … not too different from learning structure for a

Bayes network: discrete search through space of possible graphs, trying to maximize data probability….

Page 36: Undirected Probabilistic Graphical Models (Markov Nets)

MLNs: Points to ponder• Compared to ground

representations, MLNs have easier learning but equal harder inference– MLNs need to learn significantly fewer

parameters than a ground network of similar size

– MLNs may be compelled to exploit the “relational” structure and thus may spend time inventing lifted inference methods

• Inference approaches• Learning

– Parameter• Why Pseudo Likelihood?

– Structure—implies learning clauses.. (what ILP does)

• Connection to Dynamic Bayes Nets?

• Relational

Page 37: Undirected Probabilistic Graphical Models (Markov Nets)

Markov Logic: Intuition

• A logical KB is a set of hard constraintson the set of possible worlds

• Let’s make them soft constraints:When a world violates a formula,It becomes less probable, not impossible

• Give each formula a weight(Higher weight Stronger constraint)

satisfiesit formulas of weightsexpP(world)

Page 38: Undirected Probabilistic Graphical Models (Markov Nets)

Markov Logic: Definition

• A Markov Logic Network (MLN) is a set of pairs (F, w) where– F is a formula in first-order logic– w is a real number

• Together with a set of constants,it defines a Markov network with– One node for each grounding of each predicate in

the MLN– One feature for each grounding of each formula F

in the MLN, with the corresponding weight w

Page 39: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

habits. smoking similar have Friendscancer. causes Smoking

Page 40: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

Page 41: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

1.15.1

Page 42: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

1.15.1

Two constants: Anna (A) and Bob (B)

Page 43: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

1.15.1

Cancer(A)

Smokes(A) Smokes(B)

Cancer(B)

Two constants: Anna (A) and Bob (B)

Page 44: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

1.15.1

Cancer(A)

Smokes(A)Friends(A,A)

Friends(B,A)

Smokes(B)

Friends(A,B)

Cancer(B)

Friends(B,B)

Two constants: Anna (A) and Bob (B)

Page 45: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

1.15.1

Cancer(A)

Smokes(A)Friends(A,A)

Friends(B,A)

Smokes(B)

Friends(A,B)

Cancer(B)

Friends(B,B)

Two constants: Anna (A) and Bob (B)

Page 46: Undirected Probabilistic Graphical Models (Markov Nets)

Example: Friends & Smokers

)()(),(,)()(

ySmokesxSmokesyxFriendsyxxCancerxSmokesx

1.15.1

Cancer(A)

Smokes(A)Friends(A,A)

Friends(B,A)

Smokes(B)

Friends(A,B)

Cancer(B)

Friends(B,B)

Two constants: Anna (A) and Bob (B)

Page 47: Undirected Probabilistic Graphical Models (Markov Nets)

Markov Logic Networks• MLN is template for ground Markov nets• Probability of a world x:

• Typed variables and constants greatly reduce size of ground Markov net

• Functions, existential quantifiers, etc.• Infinite and continuous domains

Weight of formula i No. of true groundings of formula i in x

iii xnw

ZxP )(exp1)(

Page 48: Undirected Probabilistic Graphical Models (Markov Nets)

Relation to Statistical Models

• Special cases:– Markov networks– Markov random fields– Bayesian networks– Log-linear models– Exponential models– Max. entropy models– Gibbs distributions– Boltzmann machines– Logistic regression– Hidden Markov models– Conditional random fields

• Obtained by making all predicates zero-arity

• Markov logic allows objects to be interdependent (non-i.i.d.)

Page 49: Undirected Probabilistic Graphical Models (Markov Nets)

Relation to First-Order Logic

• Infinite weights First-order logic• Satisfiable KB, positive weights

Satisfying assignments = Modes of distribution• Markov logic allows contradictions between

formulas

Page 50: Undirected Probabilistic Graphical Models (Markov Nets)

MAP/MPE Inference

• Problem: Find most likely state of world given evidence

)|(maxarg xyPy

Query Evidence

Page 51: Undirected Probabilistic Graphical Models (Markov Nets)

MAP/MPE Inference

• Problem: Find most likely state of world given evidence

i

iixy

yxnwZ

),(exp1maxarg

Page 52: Undirected Probabilistic Graphical Models (Markov Nets)

MAP/MPE Inference

• Problem: Find most likely state of world given evidence

i

iiy

yxnw ),(maxarg

Page 53: Undirected Probabilistic Graphical Models (Markov Nets)

MAP/MPE Inference

• Problem: Find most likely state of world given evidence

• This is just the weighted MaxSAT problem• Use weighted SAT solver

(e.g., MaxWalkSAT [Kautz et al., 1997] )• Potentially faster than logical inference (!)

i

iiy

yxnw ),(maxarg

Page 54: Undirected Probabilistic Graphical Models (Markov Nets)

The MaxWalkSAT Algorithm

for i ← 1 to max-tries do solution = random truth assignment for j ← 1 to max-flips do if ∑ weights(sat. clauses) > threshold then return solution c ← random unsatisfied clause with probability p flip a random variable in c else flip variable in c that maximizes ∑ weights(sat. clauses) return failure, best solution found

Page 55: Undirected Probabilistic Graphical Models (Markov Nets)

But … Memory Explosion

• Problem: If there are n constantsand the highest clause arity is c,the ground network requires O(n ) memory

• Solution:Exploit sparseness; ground clauses lazily→ LazySAT algorithm [Singla & Domingos, 2006]

c

Page 56: Undirected Probabilistic Graphical Models (Markov Nets)

Computing Probabilities

• P(Formula|MLN,C) = ?• MCMC: Sample worlds, check formula holds• P(Formula1|Formula2,MLN,C) = ?• If Formula2 = Conjunction of ground atoms

– First construct min subset of network necessary to answer query (generalization of KBMC)

– Then apply MCMC (or other)• Can also do lifted inference [Braz et al, 2005]

Page 57: Undirected Probabilistic Graphical Models (Markov Nets)

Ground Network Construction

network ← Øqueue ← query nodesrepeat node ← front(queue) remove node from queue add node to network if node not in evidence then add neighbors(node) to queue until queue = Ø

Page 58: Undirected Probabilistic Graphical Models (Markov Nets)

But … Insufficient for Logic

• Problem:Deterministic dependencies break MCMCNear-deterministic ones make it very slow

• Solution:Combine MCMC and WalkSAT→ MC-SAT algorithm [Poon & Domingos, 2006]

Page 59: Undirected Probabilistic Graphical Models (Markov Nets)

Learning

• Data is a relational database• Closed world assumption (if not: EM)• Learning parameters (weights)• Learning structure (formulas)

Page 60: Undirected Probabilistic Graphical Models (Markov Nets)

• Parameter tying: Groundings of same clause

• Generative learning: Pseudo-likelihood• Discriminative learning: Cond. likelihood,

use MC-SAT or MaxWalkSAT for inference

Weight Learning

No. of times clause i is true in data

Expected no. times clause i is true according to MLN

)()()(log xnExnxPw iwiwi

Page 61: Undirected Probabilistic Graphical Models (Markov Nets)

Structure Learning

• Generalizes feature induction in Markov nets• Any inductive logic programming approach can be used,

but . . .• Goal is to induce any clauses, not just Horn• Evaluation function should be likelihood• Requires learning weights for each candidate• Turns out not to be bottleneck• Bottleneck is counting clause groundings• Solution: Subsampling

Page 62: Undirected Probabilistic Graphical Models (Markov Nets)

Structure Learning

• Initial state: Unit clauses or hand-coded KB• Operators: Add/remove literal, flip sign• Evaluation function:

Pseudo-likelihood + Structure prior• Search: Beam, shortest-first, bottom-up

[Kok & Domingos, 2005; Mihalkova & Mooney, 2007]

Page 63: Undirected Probabilistic Graphical Models (Markov Nets)

Alchemy

Open-source software including:• Full first-order logic syntax• Generative & discriminative weight learning• Structure learning• Weighted satisfiability and MCMC• Programming language features

alchemy.cs.washington.edu

Page 64: Undirected Probabilistic Graphical Models (Markov Nets)

Alchemy Prolog BUGS

Represent-ation

F.O. Logic + Markov nets

Horn clauses

Bayes nets

Inference Model check- ing, MC-SAT

Theorem proving

Gibbs sampling

Learning Parameters& structure

No Params.

Uncertainty Yes No Yes

Relational Yes Yes No