DOC PREVIEW
CMU CS 10708 - Lecture

This preview shows page 1-2-3-20-21-40-41-42 out of 42 pages.

Save
View full document
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience
Premium Document
Do you want full access? Go Premium and unlock all 42 pages.
Access to all documents
Download any document
Ad free experience

Unformatted text preview:

The Elimination AlgorithmProbabilistic InferenceQuery 1: LikelihoodQuery 2: Conditional ProbabilityApplications of a posteriori BeliefQuery 3: Most Probable AssignmentApplications of MPAComplexity of InferenceApproaches to inferenceMarginalization and EliminationElimination on ChainsSlide 12Elimination in ChainsSlide 14Undirected ChainsThe Sum-Product OperationOutcome of eliminationDealing with evidenceInference on General GM via Variable EliminationThe elimination algorithmSlide 21Slide 22A more complex networkExample: Variable EliminationSlide 25Slide 26Slide 27Slide 28Slide 29Slide 30Slide 31Complexity of variable eliminationUnderstanding Variable EliminationGraph eliminationSlide 35Graph elimination and marginalizationComplexityExamplesLimitation of Procedure EliminationFrom Elimination to Message PassingSlide 41Summary1School of Computer ScienceThe Elimination AlgorithmProbabilistic Graphical Models (10-708)Probabilistic Graphical Models (10-708)Lecture 4, Sep 26, 2007Eric XingEric XingReading: J-Chap 3, KF-Chap. 8, 9Eric Xing 2Probabilistic InferenceWe now have compact representations of probability distributions: Graphical ModelsA GM G describes a unique probability distribution PHow do we answer queries about P?We use inference as a name for the process of computing answers to such queriesEric Xing 3 11x xkk,,x,xPP )()( ee Query 1: LikelihoodMost of the queries one may ask involve evidenceEvidence e is an assignment of values to a set E variables in the domainWithout loss of generality E = { Xk+1, …, Xn }Simplest query: compute probability of evidencethis is often referred to as computing the likelihood of eEric Xing 4xx,XPX,PPX,PXP)()()()()|(eeeeezezZYY )|()|( ,PePQuery 2: Conditional ProbabilityOften we are interested in the conditional probability distribution of a variable given the evidencethis is the a posteriori belief in X, given evidence eWe usually query a subset Y of all domain variables X={Y,Z} and "don't care" about the remaining, Z:the process of summing out the "don't care" variables z is called marginalization, and the resulting P(y|e) is called a marginal prob.Eric Xing 5A CBA CB??Applications of a posteriori BeliefPrediction: what is the probability of an outcome given the starting conditionthe query node is a descendent of the evidenceDiagnosis: what is the probability of disease/fault given symptomsthe query node an ancestor of the evidenceLearning under partial observationfill in the unobserved values under an "EM" setting (more later)The directionality of information flow between variables is not restricted by the directionality of the edges in a GMprobabilistic inference can combine evidence form all parts of the networkEric Xing 6In this query we want to find the most probable joint assignment (MPA) for some variables of interestSuch reasoning is usually performed under some given evidence e, and ignoring (the values of) other variables z :this is the maximum a posteriori configuration of y.zyyezyeyeY )|,(maxarg)|(maxarg)|(MPA PPYYQuery 3: Most Probable AssignmentEric Xing 7Applications of MPAClassification find most likely label, given the evidenceExplanation what is the most likely scenario, given the evidenceCautionary note:The MPA of a variable depends on its "context"---the set of variables been jointly queriedExample:MPA of X ?MPA of (X, Y) ?Eric Xing 8Thm:Computing P(X = x | e) in a GM is NP-hardHardness does not mean we cannot solve inferenceIt implies that we cannot find a general procedure that works efficiently for arbitrary GMsFor particular families of GMs, we can have provably efficient proceduresComplexity of InferenceEric Xing 9Approaches to inferenceExact inference algorithmsThe elimination algorithmMessage-passing algorithm (sum-product, belief propagation)The junction tree algorithms Approximate inference techniquesStochastic simulation / sampling methodsMarkov chain Monte Carlo methodsVariational algorithmsEric Xing 10Query: P(e) By chain decomposition, we getA B CEDd c b ad c b adePcdPbcPabPaPedcbaPeP)|()|()|()|()(),,,,()(a naïve summation needs to enumerate over an exponential number of termsA signal transduction pathway:What is the likelihood that protein E is active?Marginalization and EliminationEric Xing 11Rearranging terms ...A B CED d c b ad c b aabPaPdePcdPbcPdePcdPbcPabPaPeP)|()()|()|()|()|()|()|()|()()(Elimination on ChainsEric Xing 12Now we can perform innermost summationThis summation "eliminates" one variable from our summation argument at a "local cost".A B CEDX d c bd c b abpdePcdPbcPabPaPdePcdPbcPeP)()|()|()|()|()()|()|()|()(Elimination on ChainsEric Xing 13A B CEDRearranging and then summing again, we get d cd c bd c bcpdePcdPbpbcPdePcdPbpdePcdPbcPeP)()|()|()()|()|()|()()|()|()|()(XXElimination in ChainsEric Xing 14A B CEDEliminate nodes one by one all the way to the end, we getddpdePeP )()|()(XX X XComplexity:•Each step costs O(|Val(Xi)|*|Val(Xi+1)|) operations: O(kn2)•Compare to naïve evaluation that sums over joint values of n-1 variables O(nk)Elimination in ChainsEric Xing 15Rearranging terms ...A B CEDUndirected Chains d c b ad c b aabdecdbcZdecdbcabZeP),(),(),(),(),(),(),(),()(11Eric Xing 16The Sum-Product OperationIn general, we can view the task at hand as that of computing the value of an expression of the form:where F is a set of factorsWe call this task the sum-product inference task.zFEric Xing 17Outcome of eliminationLet X be some set of variables, let F be a set of factors such that for each  ∈ F , Scope[ ] ⊆ X, let Y ⊂ X be a set of query variables, and let Z = X−Y be the variable to be eliminated The result of eliminating the variable Z is a factorThis factor does not necessarily correspond to any probability or conditional probability in this network. (example forthcoming)zYF)(Eric Xing 18Dealing with evidenceConditioning as a Sum-Product OperationThe evidence potential:Total evidence potential:Introducing evidence --- restricted factors:iiiiiieEeEeE if 0 if


View Full Document

CMU CS 10708 - Lecture

Documents in this Course
Lecture

Lecture

15 pages

Lecture

Lecture

25 pages

Lecture

Lecture

24 pages

causality

causality

53 pages

lecture11

lecture11

16 pages

Exam

Exam

15 pages

Notes

Notes

12 pages

lecture

lecture

18 pages

lecture

lecture

16 pages

Lecture

Lecture

17 pages

Lecture

Lecture

15 pages

Lecture

Lecture

17 pages

Lecture

Lecture

19 pages

Lecture

Lecture

16 pages

r6

r6

22 pages

lecture

lecture

20 pages

lecture

lecture

35 pages

Lecture

Lecture

19 pages

Lecture

Lecture

21 pages

lecture

lecture

21 pages

lecture

lecture

13 pages

review

review

50 pages

Semantics

Semantics

30 pages

lecture21

lecture21

26 pages

MN-crf

MN-crf

20 pages

hw4

hw4

5 pages

lecture

lecture

12 pages

Lecture

Lecture

25 pages

Lecture

Lecture

25 pages

Lecture

Lecture

14 pages

Lecture

Lecture

15 pages

Load more
Download Lecture
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view Lecture and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view Lecture 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?