Slide 1Structure and Semantics of BNSlide 3Slide 4Slide 5Slide 6Slide 7Slide 8Slide 9Slide 10Lumiere – Office AssistantInference TasksSlide 13Slide 14Slide 15•Probabilistic (Bayesian) representations of knowledge have had a major impact on AI–contrast with symbolic/logical knowledge bases–necessity to handle uncertainty in real world apps–recent advances allow scaling up to larger networks•Example applications of Bayesian networks–HCI: inferring intent in conversation/action, plan recognition, intelligent tutoring–vision – image interpretation, de-noising–control – variables that influence flight–medicine–economicsStructure and Semantics of BN•draw causal nodes first•draw directed edges to effects (“direct causes”)•links encode conditional probability tables (CPT over parents)•fewer parameters than full joint PDF•absence of link is related to independence•types of independence–A is indep of non-descendants given parents–Markov blanket–d-separation – all paths between A and B are “blocked”–useful for determining if obtaining knowledge of B would change belief about A•child is cond.dep. on parent: P(B|A)•parent is cond.dep. on child:–P(A|B)=P(B|A)P(A)/P(B)•what about when one node is not an ancestor of the other? e.g. siblingsABA and B are only conditionally independent given Csimple treespoly-trees (singly connected, one path between any pair of nodes)“cyclic” (using undirected edges) – much harder to do computationsexplaining away: P(sprinkler | wetGrass) = 0.43P(sprinkler | wetGrass,rain) = 0.19•Compact representations of CPT–Noisy-Or–prob. version of: cold flu malaria fever–only have to represent 3 numbers (“strengths”) instead of 8Network Engineering for Complex Belief Networks, Mahoney and LaskeyA Bayesian network approach to threat valuation with application to an air defense scenario, Johansson and FalkmanLumiere – Office AssistantInference Tasks•posterior: P(Xi|{Zi})–Zi observed vars, with unobserved variables Yi, marginalized out–prediction vs. diagnosis–evidence combination is crucial–handling unobserved variables is crucial•all marginals: P(Ai) – like priors, but for interior nodes too•subjoint: P(A,B)•boolean queries•most-probable explanation: –argmax{Yi} P(Yi U Zi) – state with highest joint probability(see slides 4-10 in http://aima.eecs.berkeley.edu/slides-pdf/chapter14b.pdffor discussion of Enumeration and VariableElimination)Inference in Bayesian Networks, D’AmbrosioBelief Propagation (this figure happens to come from http://www.pr-owl.org/basics/bn.php)see also: wiki, Ch. 8 in Bishop
View Full Document