Published in: Tenth Tbilisi Symposium on Language, Logic and Computation: Gudauri, Georgia, September 2013

Similar documents
Citation for published version (APA): Harinck, S. (2001). Conflict issues matter : how conflict issues influence negotiation

Recent revisions of phosphate rock reserves and resources: a critique Edixhoven, J.D.; Gupta, J.; Savenije, H.H.G.

NMDA receptor dependent functions of hippocampal networks in spatial navigation and memory formation de Oliveira Cabral, H.

UvA-DARE (Digital Academic Repository)

Rotational symmetry breaking in the topological superconductor SrxBi2Se3 probed by uppercritical

UvA-DARE (Digital Academic Repository) Converting lignin to aromatics: step by step Strassberger, Z.I. Link to publication

Mass loss and evolution of asymptotic giant branch stars in the Magellanic Clouds van Loon, J.T.

UvA-DARE (Digital Academic Repository) Phenotypic variation in plants Lauss, K. Link to publication

On a unified description of non-abelian charges, monopoles and dyons Kampmeijer, L.

Citation for published version (APA): Weber, B. A. (2017). Sliding friction: From microscopic contacts to Amontons law

Data-driven methods in application to flood defence systems monitoring and analysis Pyayt, A.

UvA-DARE (Digital Academic Repository)

Physiological and genetic studies towards biofuel production in cyanobacteria Schuurmans, R.M.

NMDA receptor dependent functions of hippocampal networks in spatial navigation and memory formation de Oliveira Cabral, H.

Citation for published version (APA): Hin, V. (2017). Ontogenesis: Eco-evolutionary perspective on life history complexity.

Coherent X-ray scattering of charge order dynamics and phase separation in titanates Shi, B.

UvA-DARE (Digital Academic Repository) The syntax of relativization de Vries, M. Link to publication

UvA-DARE (Digital Academic Repository) Electrokinetics in porous media Luong, D.T. Link to publication

Combining radar systems to get a 3D - picture of the bird migration Liechti, F.; Dokter, A.; Shamoun-Baranes, J.Z.; van Gasteren, J.R.; Holleman, I.

Monitoring and modelling hydrological fluxes in support of nutrient cycling studies in Amazonian rain forest ecosystems Tobon-Marin, C.

Stochastic Sampling and Search in Belief Updating Algorithms for Very Large Bayesian Networks

Belief Update in CLG Bayesian Networks With Lazy Propagation

Integrating Correlated Bayesian Networks Using Maximum Entropy

TDT70: Uncertainty in Artificial Intelligence. Chapter 1 and 2

Bayesian networks: approximate inference

Citation for published version (APA): Jak, S. (2013). Cluster bias: Testing measurement invariance in multilevel data

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =

Published in: ICCSA 2014: Fourth International Conference on Complex Systems and Applications: June 23-26, 2014, Le Havre, Normandie, France

Expectation Propagation Algorithm

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Bayesian Networks BY: MOHAMAD ALSABBAGH

Decisiveness in Loopy Propagation

A Stochastic l-calculus

Citation for published version (APA): Adhyaksa, G. W. P. (2018). Understanding losses in halide perovskite thin films

Citation for published version (APA): Ochea, M. I. (2010). Essays on nonlinear evolutionary game dynamics Amsterdam: Thela Thesis

CUADERNOS DE TRABAJO ESCUELA UNIVERSITARIA DE ESTADÍSTICA

Probability Propagation in Singly Connected Networks

Citation for published version (APA): Nguyen, X. C. (2017). Different nanocrystal systems for carrier multiplication

Statistical Approaches to Learning and Discovery

Measurement of the charged particle density with the ATLAS detector: First data at vs = 0.9, 2.36 and 7 TeV Kayl, M.S.

Bayesian networks in Mastermind

The Necessity of Bounded Treewidth for Efficient Inference in Bayesian Networks

Complexity Results for Enhanced Qualitative Probabilistic Networks

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Uncertainty and Bayesian Networks

Undirected Graphical Models

Introduction to Artificial Intelligence. Unit # 11

6.867 Machine learning, lecture 23 (Jaakkola)

p L yi z n m x N n xi

The Parameterized Complexity of Approximate Inference in Bayesian Networks

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Machine Learning Lecture 14

Inference in Hybrid Bayesian Networks Using Mixtures of Gaussians

Optimization and approximation on systems of geometric objects van Leeuwen, E.J.

Bayesian Networks: Belief Propagation in Singly Connected Networks

Applying Bayesian networks in the game of Minesweeper

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Abductive Inference in Bayesian Belief Networks Using Swarm Intelligence

Young intermediate-mass stars: from a HIFI spectral survey to the physics of the disk inner rim Kama, M.

An Empirical Study of Probability Elicitation under Noisy-OR Assumption

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

A Decision Theoretic View on Choosing Heuristics for Discovery of Graphical Models

Bayes-Ball: The Rational Pastime (for Determining Irrelevance and Requisite Information in Belief Networks and Influence Diagrams)

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

UvA-DARE (Digital Academic Repository) Charge carrier dynamics in photovoltaic materials Jensen, S.A. Link to publication

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Artificial Intelligence Bayes Nets: Independence

Bayesian networks. Chapter Chapter

Probabilistic Reasoning Systems

Part I Qualitative Probabilistic Networks

Bayesian network modeling. 1

Bayesian Methods in Artificial Intelligence

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Improving Search-Based Inference in Bayesian Networks

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Machine Learning 3. week

Climate change and topography as drivers of Latin American biome dynamics Flantua, S.G.A.

Variational algorithms for marginal MAP

SIMULATION APPROACHES TO GENERAL PROBABILISTIC INFERENCE ON BELIEF NETWORKS

Lecture 4: Probability Propagation in Singly Connected Bayesian Networks

Love and fear of water: Water dynamics around charged and apolar solutes van der Post, S.T.

The Minimum Message Length Principle for Inductive Inference

Introduction to Probabilistic Graphical Models

Abstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables

Lower Bound Bayesian Networks An Efficient Inference of Lower Bounds on Probability Distributions in Bayesian Networks

Methods of Exact Inference. Cutset Conditioning. Cutset Conditioning Example 1. Cutset Conditioning Example 2. Cutset Conditioning Example 3

13: Variational inference II

Dynamic importance sampling in Bayesian networks using factorisation of probability trees

What Causality Is (stats for mathematicians)

Lecture 18 Generalized Belief Propagation and Free Energy Approximations

Probabilistic Graphical Networks: Definitions and Basic Results

Generalized Loopy 2U: A New Algorithm for Approximate Inference in Credal Networks

Measuring more or less: Estimating product period penetrations from incomplete panel data Hoogendoorn, A.W.

Bayesian networks. Chapter 14, Sections 1 4

A polarized view on DNA under tension van Mameren, Joost; Vermeulen, K.; Wuite, G.J.L.; Peterman, E.J.G.

Improving Bound Propagation

Probabilistic and Bayesian Machine Learning

Separable and Transitive Graphoids

Sensitivity Analysis for Threshold Decision Making with Dynamic Networks

UvA-DARE (Digital Academic Repository) Matrix perturbations: bounding and computing eigenvalues Reis da Silva, R.J. Link to publication

Transcription:

UvA-DARE (Digital Academic Repository) Estimating the Impact of Variables in Bayesian Belief Networks van Gosliga, S.P.; Groen, F.C.A. Published in: Tenth Tbilisi Symposium on Language, Logic and Computation: Gudauri, Georgia, 23-27 September 2013 Link to publication Citation for published version (APA): van Gosliga, S. P., & Groen, F. (2013). Estimating the Impact of Variables in Bayesian Belief Networks. In Tenth Tbilisi Symposium on Language, Logic and Computation: Gudauri, Georgia, 23-27 September 2013 (pp. 87-90). Tbilisi: Centre for Language, Logic and Speech, and Razmadze Mathematical Institute, Tbilisi State University. General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: http://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (http://dare.uva.nl) Download date: 26 Feb 2019

Tenth Tbilisi Symposium on Language, Logic and Computation Gudauri, Georgia 23 27 September 2013 Centre for Language, Logic and Speech and Razmadze Mathematical Institute at the Tbilisi State University Institute for Logic, Language and Computation

Estimating the Impact of Variables in Bayesian Belief Networks Sicco Pier van Gosliga and Frans Groen Introduction Bayesian belief networks (BBNs) are often designed to aid decision making. A BBN models uncertainty and enables to compute posterior probabilities given prior information and current observations [8]. In this paper we focus on a solution to two practical problems that arise with the application of BBNs: First, in real world applications observations are associated with costs. To keep these costs within acceptable limits we would like to prioritize observations most relevant to our decision. Second, models can grow too large for feasible computations [2]. Rather than restricting the design of a BBN for decision making, we pursue an ad-hoc sub-model tailored to its relevance for the decision maker. For these reasons, we propose an efficient approximation algorithm to compute the maximum impact of observing a variable in respect to the posteriors of other variables in a BBN. The algorithm is guaranteed to never underestimate the real impact. First the impact of the variables within the markov blanket of a variable are calculated, followed by a message passing algorithm to include other variables in the network. The method is closely related the field of sensitivity analysis [3][5], which mostly focuses on aiding the design of a BBN. Methodology Distance measure. To quantify belief changes in posterior probability distributions we will use the maximum absolute distance as defined by the Chebyshev distance function D Ch pp, Qq which takes the largest difference between pairs of probability values in two discrete probability distributions P and Q: D Ch pp, Qq max ˇˇpi q i iˇˇ (1) Distance functions to quantify differences in probability distributions are often based on entropy [1][6][7]. Entropy based distance functions are relative distance measures. As a result, small absolute differences can be valued equally important as a large difference when both span an equal order of magnitude. Also, entropy based distance functions evaluate the general difference between two discrete probability distributions, while the Chebyshev distance focuses on the maximum difference. Since decisions are based upon the absolute posterior probability of a specific outcome, the Chebyshev distance is the measure that directly relates to the decision making. Van Engelen [9] introduced a method to compute an absolute upper bound for the maximum absolute error based on the K-L divergence. However, its reliance on computing prior marginals in advance limits its applicability for pruning BBNs. We base our method directly on the Chebyshev distance and local prior conditional distributions rather than prior marginals. 87

Aim. We aim to get a safe estimate of the real maximum impact. Given a BBN G containing variables V 1 and we first define, in Eq. 2, the real maximum impact of on V 1 in the context of evidence e. We then define, in Eq. 3, the generalized real maximum impact of on V 1 as the maximum absolute difference that two different instantiations of can cause in the posterior probability for any state of V 1 given any possible combination of evidence e for other variables in G, where set E holds all evidence configurations for G. δ e pv 1 q max P pv 1 v 2i, eq, P pv 1 v 2j, eq (2) Algorithm δpv 1 q max epe v 2iP;v 2jP D Ch δ e pv 1 q First Phase. We first calculate the maximum impact a variable may potentially have on other variables within its markov blanket, if the BBN beyond the markov blanket could take any form. Figure 1 shows V 1 s markov blanket, a subgraph of G that contains V 1 s parents, children and parents of children. Suppose e is a set of evidence, and we want to compute the posteriors for V 1 given this evidence set: P pv 1 eq. Each edge to V 1 can be considered to partition G in subgraphs: be the subset of e that concerns the variables in G. Likewise, edges Ñ V 1, V 1 Ñ and V 1 Ñ create the following subgraphs end evidence sets: G with e, G` with e` and G` with e`. edge Ñ V 1 divides G in an upper subgraph G and a lower subgraph G`. Let e Applying Bayes rule, the posterior probability distribution for V 1 given e can then be computed as follows: P pv 1 eq P pv 1 e, e, e`, e` q (4) η P pv 1 e, e q P pe` V 1 q P pe` V 1 q ÿ η` P pv 1, qp p e qp p e q λpv 1 q (5) In Eq. 4 and 5, η is a normalizing constant and λpv 1 q P pe` V 1 qp pe` V 1 q. The posteriors of V 1 are computed by combining prior information P pv 1, q and current observations in evidence set e. Each parent and child variable of V 1 contributes a parameter conditioned by a subset of e. The maximizing causal parameters for V 1 s parents can be derived from the conditional probability table P pv 1, q. For computing the local potential maximum impact of on V 1, V1 pv 1 q, we set these as p P pv 1k v 2i, v 3c q and q P pv 1k v 2j, v 3c q. For the diagnostic parameters of V 1 s children, we set λ λpv 1k q. The λ value that maximizes V1 pv 1 q can be calculated in closed form with Eq. 7. The value V1 pv 1 q can then be computed with Eq. 6. V1 pv 1 q max 1 i,j,k,c ˇ2 λ p λ p ` 1 λ p 1 2 λ q λ q ` 1 λ q ˇ unless p or q is 0, or p or q is 1, then V1 pv 1 q 1 (6) b`p 1 p`q2 q ` p` q ` p ` q 1 (3) λ p ` q 1 unless p ` q 1 0, then: λ 1 2 (7) 88

Second Phase. The second phase assesses the maximum potential impact between nodes further apart. A propagation algorithm is used to investigate all d-connecting paths between each pair of nodes. Multiple d-connecting paths may exist between a pair of variables, and are likely to partly overlap. For a single path segment, we take the product of all local potential impact values along intermediate edges as a safe overestimate of the real maximum impact. When parallel path segments converge we take their sum to approximate their joint influence. Experiments To assess the quality of our approximation of the real maximum impact values, the algorithm was tested on randomly generated BBNs. The BBNs were generated by using an algorithm designed and implemented by [4]. The pv a V b q values were compared to the corresponding δpv a V b q values, based on simulated evidence. In each BBN, all possible constellations of evidence were taken into account to discover the real maximum impact. In total 48000 pairs of variables were evaluated in singly connected graphs, and 90000 pairs of variables in multiply connected graphs. The experiments show that accuracy of the pv a V b q values decreases with the number of states a variable may have, the maximum degree of the graph, and the number of edges between V a and V b. The algorithm never returned an underestimate of the real maximum impact. Figure 2 gives the estimated impact as function of the real impact in these experiments. Discussion The method was successfully used to estimate the impact of variables in a BBN. Instead of getting an optimally accurate assessment of the real maximum impact, we have chosen an approach that is guaranteed never to return an underestimate of the real impact. Not underestimating a variable s impact in the context of decision making can be crucial. To improve the method s accuracy and applicability, it could be extended to respect existing evidence and to get the joint impact of multiple variables to a single goal variable. References [1] H. Chan and A. Darwiche, A distance measure for bounding probabilistic belief change, Int. Journal of Approximate Reasoning, vol. 38, no. 2, 149 174 (2005) [2] G. F. Cooper, Probabilistic Inference Using Belief Networks is NP-hard, Technical Report KSL-87-27, Medical Computer Science Group, Stanford University (1987) 89

[3] V. M. H. Coupé and L. C. van der Gaag, Properties of sensitivity analysis of bayesian belief networks, Annals of Mathematics and A.I., vol. 36, no. 4, 323 356 (2002) [4] J. S. Ide and F. G. Cozman, Random generation of bayesian networks, Proc. of the 16th Brazilian Symp. on Artificial Intelligence, Springer-Verlag, 366 375 (2002) [5] F.V. Jensen, S.H. Aldenryd, and K.B. Jensen, Sensitivity analysis in bayesian networks, Symbolic and Quantitative Approaches to Reasoning and Uncertainty, Lecture Notes in Computer Science, vol. 946, Springer, 243 250 (1995) [6] S. Kullback and R. A. Leibler, On Information and Sufficiency, The Annals of Mathematical Statistics, vol. 22, no. 1, 79 86 (1951) [7] J. Lin, Divergence measures based on the shannon entropy, Information Theory, IEEE Transactions on Information Theory, vol. 37, no. 1, 145 151 (1991) [8] J. Pearl, Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, Morgan Kaufmann (1988). [9] R. A. van Engelen, Approximating Bayesian belief networks by arc removal, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, 916 920 (1997)