Deductive and Inductive Probabilistic Programming

Size: px
Start display at page:

Download "Deductive and Inductive Probabilistic Programming"

Transcription

1 Deductive and Inductive Probabilistic Programming Fabrizio Riguzzi F. Riguzzi Deductive and Inductive PP 1 / 43

2 Outline Probabilistic programming Probabilistic logic programming Inference Learning Applications F. Riguzzi Deductive and Inductive PP 2 / 43

3 Probabilistic Programming Users specify a probabilistic model in its entirety (e.g., by writing code that generates a sample from the joint distribution) and inference follows automatically given the specification. PP languages provide the full power of modern programming languages for describing complex distributions Reuse of libraries of models Interactive modeling Abstraction F. Riguzzi Deductive and Inductive PP 3 / 43

4 Probabilistic Programming F. Riguzzi Deductive and Inductive PP 4 / 43

5 Probabilistic Programming Languages Name Extends from Host language Venture Scheme C++ Probabilistic-C C C Anglican Scheme Clojure IBAL OCaml PRISM B-Prolog Infer.NET.NET Framework.NET Framework dimple MATLAB, Java chimple MATLAB, Java BLOG Java PSQL SQL BUGS FACTORIE Scala PMTK MATLAB MATLAB Alchemy C++ Dyna Prolog Name Extends from Host language Figaro Scala Church Scheme JavaScript, Scheme ProbLog Prolog Python, Jython ProBT C++, Python Stan (software) C++ Hakaru Haskell Haskell BAli-Phy (software) Haskell C++ ProbCog Java, Python Gamble Racket Tuffy Java PyMC Python Python Lea Python Python WebPPL JavaScript JavaScript Picture Julia Julia Turing.jl Julia Julia Source programming_language F. Riguzzi Deductive and Inductive PP 5 / 43

6 Probabilistic Programming Languages Only three are logic, the others imperative/functional/object oriented DARPA released in 2013 the funding call Probabilistic Programming for Advancing Machine Learning (PPAML) Aim: develop probabilistic programming languages and accompanying tools to facilitate the construction of new machine learning applications across a wide range of domains. Focus: functional PP F. Riguzzi Deductive and Inductive PP 6 / 43

7 Probabilistic Logic Programming What are we missing? Is logic programming to blame? F. Riguzzi Deductive and Inductive PP 7 / 43

8 Thesis Probabilistic logic programming is alive and kicking! F. Riguzzi Deductive and Inductive PP 8 / 43

9 Strengths Relationships are first class citizens Conceptually easier to lift Strong semantics Inductive systems F. Riguzzi Deductive and Inductive PP 9 / 43

10 Weaknesses Handling non-termination Continuous variables F. Riguzzi Deductive and Inductive PP 10 / 43

11 Non-termination Possible when the number of explanations for the query is infinite F. Riguzzi Deductive and Inductive PP 11 / 43

12 Non-termination: Inducing Arithmetic Functions Church code (define (random-arithmetic-fn) (if (flip 0.3) (random-combination (random-arithmetic-fn) (random-arithmetic-fn)) (if (flip) (lambda (x) x) (random-constant-fn)))) (define (random-combination f g) (define op (uniform-draw (list + -))) (lambda (x) (op (f x) (g x)))) (define (random-constant-fn) (define i (sample-integer 10)) (lambda (x) i)) F. Riguzzi Deductive and Inductive PP 12 / 43

13 Non-termination: Inducing Arithmetic Functions LPAD (cplint) code example/inference/arithm.pl eval(x,y):- random_fn(x,0,f), Y is F. op(l,+):0.5;op(l,-):0.5. random_fn(x,l,f):- comb(l), random_fn(x,l(l),f1), random_fn(x,r(l),f2), op(l,op), F=..[Op,F1,F2]. random_fn(x,l,f):- \+ comb(l), base_random_fn(x,l,f). comb(_):0.3. base_random_fn(x,l,x):- identity(l). base_random_fn(_x,l,c):- \+ identity(l), random_const(l,c). identity(_):0.5. random_const(_,c):discrete(c,[0:0.1,1:0.1,2:0.1,3:0.1,4:0.1, 5:0.1,6:0.1,7:0.1,8:0.1,9:0.1]). F. Riguzzi Deductive and Inductive PP 13 / 43

14 Non-termination: Inducing Arithmetic Functions Aim: given observations of couples input-output for the random function, predict the output for a new input Arbitrarily complex functions have a non-zero probability of being selected The program has a non-terminating execution. Exact inference: infinite number of explanations F. Riguzzi Deductive and Inductive PP 14 / 43

15 Non-termination: Inducing Arithmetic Functions (define (sample) (rejection-query (define my-proc (random-arithmetic-fn)) (my-proc 2) (= (my-proc 1) 3))) (hist (repeat 100 sample)) F. Riguzzi Deductive and Inductive PP 15 / 43

16 Solution Use (T. Sato, P. Meyer, Infinite probability computation by cyclic explanation graphs, Theor. Pract. Log. Prog ) or (A. Gorlin, C. R. Ramakrishnan, S. A. Smolka, Model checking with proba- bilistic tabled logic programming, Theor. Pract. Log. Prog. 12 (4-5) 2012) or resort to sampling: with the increase of complexity, the probability of functions tend to 0 and the probability of the infinite trace is 0 Metropolis Hastings: (Nampally, A., Ramakrishnan, C.: Adaptive MCMC-based inference in probabilistic logic programs. arxiv preprint arxiv: ) Monte Carlo sampling is attractive for the simplicity of its implementation and because you can improve the estimate as more time is available. F. Riguzzi Deductive and Inductive PP 16 / 43

17 Monte Carlo The disjunctive clause C r = H 1 : α 1... H n : α n L 1,..., L m. is transformed into the set of clauses MC(C r ) MC(C r, 1) = H 1 L 1,..., L m, sample_head(n, r, VC, NH), NH = MC(C r, n) = H n L 1,..., L m, sample_head(n, r, VC, NH), NH = n. Sample truth value of query Q:... (call(q)->... NT1 is NT+1 ; NT1 =NT), F. Riguzzi Deductive and Inductive PP 17 / 43

18 Metropolis-Hastings MCMC A Markov chain is built by taking an initial sample and by generating successor samples. The initial sample is built by randomly sampling choices so that the evidence is true. A successor sample is obtained by deleting a fixed number of sampled probabilistic choices. Then the evidence is queried If the query succeeds, the goal is queried. The sample is accepted with a probability of min{1, N 0 N 1 } where N 0 (N 1 ) is the number of choices sampled in the previous (current) sample. F. Riguzzi Deductive and Inductive PP 18 / 43

19 Solution In cplint:?- mc_mh_sample(eval(2,4),eval(1,3),100,100,3,t,f,p). Probability of eval(2,4) given that eval(1,3) is true F = 90, T = 10, P = 0.1 You can also try rejection sampling (usually slower)?- mc_rejection_sample(eval(2,4),eval(1,3),100, T,F,P). F. Riguzzi Deductive and Inductive PP 19 / 43

20 Solution You may be interested in the distribution of the output In cplint:?- mc_mh_sample_arg_bar(eval(2,y),eval(1,3),100, 100,3,Y,V). F. Riguzzi Deductive and Inductive PP 20 / 43

21 Solution You may be interested in the expected value of the output In cplint:?- mc_mh_expectation(eval(2,y),eval(1,3), 100,100,3,Y,E). E = 3.21 F. Riguzzi Deductive and Inductive PP 21 / 43

22 Continuous Random Variables Distributional clauses (B. Gutmann, I. Thon, A. Kimmig, M. Bruynooghe, and L. De Raedt, The magic of logical inference in probabilistic programming, Theory and Practice of Logic Programming, 2011) Gaussian mixture model in cplint: heads:0.6;tails:0.4. g(x): gaussian(x,0, 1). h(x): gaussian(x,5, 2). mix(x) :- heads, g(x). mix(x) :- tails, h(x). F. Riguzzi Deductive and Inductive PP 22 / 43

23 Continuous Random Variables Inference by sampling Without evidence or evidence on discrete random variables, you can reuse the same methods Sampling arguments of goals for building a probability density of the arguments. F. Riguzzi Deductive and Inductive PP 23 / 43

24 Gaussian Mixture Model heads:0.6;tails:0.4. g(x): gaussian(x,0, 1). h(x): gaussian(x,5, 2). mix(x) :- heads, g(x). mix(x) :- tails, h(x).?- mc_sample_arg(mix(x),10000,x,l0), histogram(l0,40,chart). F. Riguzzi Deductive and Inductive PP 24 / 43

25 Evidence on Continuous Random Variables You cannot use rejection sampling or Metropolis-Hastings, as the probability of the evidence is 0 You can use likelihood weighting to obtain samples of continuous arguments of a goal. (Nitti, D., De Laet, T., De Raedt, L.: Probabilistic logic programming for hybrid relational domains. Mach. Learn. 103(3), ) F. Riguzzi Deductive and Inductive PP 25 / 43

26 Likelihood Weighting For each sample to be taken, likelihood weighting samples the query and then assigns a weight to the sample on the basis of evidence. The weight is computed by deriving the evidence backward in the same sample of the query starting with a weight of one Each time a choice should be taken or a continuous variable sampled, if the choice/variable has already been taken, the current weight is multiplied by probability of the choice/by the density value of the continuous value. F. Riguzzi Deductive and Inductive PP 26 / 43

27 Bayesian Estimation Problem from examples/viewer/?worksheet=gaussian-posteriors Estimate the true value of a Gaussian distributed random variable, given some observed data. The variance is known and we suppose that the mean has itself a Gaussian distribution with mean 1 and variance 5 (prior on the parameter) We take different measurement (e.g. at different times), indexed with an integer. F. Riguzzi Deductive and Inductive PP 27 / 43

28 Bayesian Estimation Anglican code (def dataset [9 8]) (defquery gaussian-model [data] (let [mu (sample (normal 1 (sqrt 5))) sigma (sqrt 2)] (doall (map (fn [x] (observe (normal mu sigma) x)) data)) mu)) (def posterior ((conditional gaussian-model :smc :number-of-particles 10) dataset)) (def posterior-samples (repeatedly #(sample* posterior))) F. Riguzzi Deductive and Inductive PP 28 / 43

29 Bayesian Estimation cplint code example/inference/gauss_mean_est.pl value(i,x) :- mean(m), value(i,m,x). mean(m): gaussian(m,1.0, 5.0). value(_,m,x): gaussian(x,m, 2.0).?- mc_sample_arg(value(0,y),10000,y,l0), mc_lw_sample_arg(value(0,x),(value(1,9),value(2,8)),10000,x,l), densities(l0,l,40,chart). F. Riguzzi Deductive and Inductive PP 29 / 43

30 Learning Parameter learning Structure learning more developed for PLP, but (Perov, Yura N., and Frank D. Wood. Learning Probabilistic Programs. arxiv preprint arxiv: ). (Lake, Brenden M., Ruslan Salakhutdinov, and Joshua B. Tenenbaum. Human-level concept learning through probabilistic program induction. Science ). (Gaunt, Alexander L., et al. TerpreT: A Probabilistic Programming Language for Program Induction. arxiv preprint arxiv: ). F. Riguzzi Deductive and Inductive PP 30 / 43

31 Parameter Learning Problem: given a set of interpretations, a program, find the parameters maximizing the likelihood of the interpretations (or of instances of a target predicate) Exploit the equivalence with BN to use BN learning algorithms The interpretations record the truth value of ground atoms, not of the choice variables Unseen data: relative frequency can t be used F. Riguzzi Deductive and Inductive PP 31 / 43

32 Parameter Learning (Thon et al. ECML 2008) proposed an adaptation of EM for CPT-L, a simplified version of LPADs The algorithm computes the counts efficiently by repeatedly traversing the BDDs representing the explanations (Ishihata et al. ILP 2008) independently proposed a similar algorithm LFI-PROBLOG (Gutamnn et al. ECML 2011) is the adaptation of EM to ProbLog EMBLEM (Riguzzi & Bellodi IDAJ 2013) adapts (Ishihata et al. ILP 2008) to LPADs F. Riguzzi Deductive and Inductive PP 32 / 43

33 Structure Learning Given a trivial LPAD or an empty one, a set of interpretations (data) Find the model and the parameters that maximize the probability of the data (log-likelihood) SLIPCOVER: Structure LearnIng of Probabilistic logic program by searching OVER the clause space 1 Beam search in the space of clauses to find the promising ones 2 Greedy search in the space of probabilistic programs guided by the LL of the data. Parameter learning by means of EMBLEM F. Riguzzi Deductive and Inductive PP 33 / 43

34 Applications Link prediction: given a (social) network, compute the probability of the existence of a link between two entities (UWCSE) advisedby(x, Y) :0.3 :- publication(p, X), publication(p, Y), student(x). F. Riguzzi Deductive and Inductive PP 34 / 43

35 Applications Classify web pages on the basis of the link structure (WebKB) coursepage(page1): 0.3 :- linkto(page2,page1),coursepage(page2). coursepage(page1): 0.3 :- linkto(page2,page1),facultypage(page2).... coursepage(page): 0.3 :- has( abstract,page).... F. Riguzzi Deductive and Inductive PP 35 / 43

36 Applications Entity resolution: identify identical entities in text or databases samebib(a,b):0.3 :- samebib(a,c), samebib(c,b). sameauthor(a,b):0.3 :- sameauthor(a,c), sameauthor(c,b). sametitle(a,b):0.3 :- sametitle(a,c), sametitle(c,b). samevenue(a,b):0.3 :- samevenue(a,c), samevenue(c,b). samebib(b,c):0.3 :- author(b,d),author(c,e),sameauthor(d,e). samebib(b,c):0.3 :- title(b,d),title(c,e),sametitle(d,e). samebib(b,c):0.3 :- venue(b,d),venue(c,e),samevenue(d,e). samevenue(b,c):0.3 :- haswordvenue(b,word_06), haswordvenue(c,word_06).... F. Riguzzi Deductive and Inductive PP 36 / 43

37 Applications Chemistry: given the chemical composition of a substance, predict its mutagenicity or its carcenogenicity active(a):0.5 :- atm(a,b,c,29,c), gteq(c,-0.003), ring_size_5(a,d). active(a):0.5 :- lumo(a,b), lteq(b,-2.072). active(a):0.5 :- bond(a,b,c,2), bond(a,c,d,1), ring_size_5(a,e). active(a):0.5 :- carbon_6_ring(a,b). active(a):0.5 :- anthracene(a,b).... F. Riguzzi Deductive and Inductive PP 37 / 43

38 Applications Medicine: diagnose diseases on the basis of patient information (Hepatitis), influence of genes on HIV, risk of falling of elderly people (FFRAT) F. Riguzzi Deductive and Inductive PP 38 / 43

39 Experiments - Area Under the PR Curve System HIV UW-CSE Mondial SLIPCOVER 0.82 ± ± ± 0.07 SLIPCASE 0.78 ± ± ± 0.06 LSM 0.37 ± ± ALEPH ± ± 0.07 RDN-B 0.28 ± ± ± 0.07 MLN-BT 0.29 ± ± ± 0.10 MLN-BC 0.51 ± ± ± 0.09 BUSL 0.38 ± ± F. Riguzzi Deductive and Inductive PP 39 / 43

40 Experiments - Area Under the PR Curve System Carcinogenesis Mutagenesis Hepatitis SLIPCOVER ± ± 0.01 SLIPCASE ± ± 0.05 LSM ± 0.04 ALEPH ± RDN-B ± ± 0.01 MLN-BT ± ± 0.02 MLN-BC ± ± 0.02 BUSL ± 0.03 F. Riguzzi Deductive and Inductive PP 40 / 43

41 PLP Online Inference (knwoledge compilation, Monte Carlo) Parameter learning (EMBLEM) Structure learning (SLIPCOVER) Inference (knwoledge compilation, Monte Carlo) Parameter learning (LFI-ProbLog) F. Riguzzi Deductive and Inductive PP 41 / 43

42 Conclusions PLP is still a fertile field but......we must look at other communities and build bridges and......join forces! Much is left to do: Tractable sublanguages (see following talk) Lifted inference Structure/Parameter learning (also for programs with continuous variables) F. Riguzzi Deductive and Inductive PP 42 / 43

43 F. Riguzzi Deductive and Inductive PP 43 / 43

Probabilistic Logical Inference On the Web

Probabilistic Logical Inference On the Web Probabilistic Logical Inference On the Web Marco Alberti 1, Giuseppe Cota 2, Fabrizio Riguzzi 1, and Riccardo Zese 2 1 Dipartimento di Matematica e Informatica University of Ferrara Via Saragat 1, I-44122,

More information

The PITA System for Logical-Probabilistic Inference

The PITA System for Logical-Probabilistic Inference The System for Logical-Probabilistic Inference Fabrizio Riguzzi 1 and Terrance Swift 2 1 EDIF University of Ferrara, Via Saragat 1, I-44122, Ferrara, Italy fabrizio.riguzzi@unife.it 2 CETRIA Universidade

More information

Structure Learning of Probabilistic Logic Programs by MapReduce

Structure Learning of Probabilistic Logic Programs by MapReduce Structure Learning of Probabilistic Logic Programs by MapReduce Fabrizio Riguzzi 1, Elena Bellodi 2, Riccardo Zese 2, Giuseppe Cota 2, and Evelina Lamma 2 1 Dipartimento di Matematica e Informatica University

More information

Lifted Discriminative Learning of Probabilistic Logic Programs

Lifted Discriminative Learning of Probabilistic Logic Programs Lifted Discriminative Learning of Probabilistic Logic Programs Arnaud Nguembang Fadja 1 and Fabrizio Riguzzi 2 1 Dipartimento di Ingegneria University of Ferrara Via Saragat 1, I-44122, Ferrara, Italy

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Introduction to Probabilistic Programming Languages (PPL)

Introduction to Probabilistic Programming Languages (PPL) Introduction to Probabilistic Programming Languages (PPL) Anton Andreev Ingénieur d'études CNRS andreev.anton@.grenoble-inp.fr Something simple int function add(int a, int b) return a + b add(3, 2) 5 Deterministic

More information

Logic Programming Techniques for Reasoning with Probabilistic Ontologies

Logic Programming Techniques for Reasoning with Probabilistic Ontologies Logic Programming Techniques for Reasoning with Probabilistic Ontologies Riccardo Zese, Elena Bellodi, Evelina Lamma and Fabrizio Riguzzi University of Ferrara, Italy riccardo.zese@unife.it Zese, Bellodi,

More information

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

MCINTYRE: A Monte Carlo Algorithm for Probabilistic Logic Programming

MCINTYRE: A Monte Carlo Algorithm for Probabilistic Logic Programming : A Monte Carlo Algorithm for Probabilistic Logic Programming Fabrizio Riguzzi ENDIF Università di Ferrara Via Saragat, 1 44122 Ferrara, Italy. {fabrizio.riguzzi}@unife.it Abstract. Probabilistic Logic

More information

Partial inconsistency and vector semantics: sampling, animation, and program learning

Partial inconsistency and vector semantics: sampling, animation, and program learning Partial inconsistency and vector semantics: sampling, animation, and program learning Michael Bukatin Nokia Corporation, Cambridge, MA Joint work with Ralph Kopperman and Steve Matthews - - - 29th Summer

More information

Probabilistic Programming with Infer.NET. John Bronskill University of Cambridge 19 th October 2017

Probabilistic Programming with Infer.NET. John Bronskill University of Cambridge 19 th October 2017 Probabilistic Programming with Infer.NET John Bronskill University of Cambridge 19 th October 2017 The Promise of Probabilistic Programming We want to do for machine learning what the advent of high-level

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

Introduction to Probabilistic Programming Language (with Church as an example) Presenter: Enrique Rosales, Xing Zeng

Introduction to Probabilistic Programming Language (with Church as an example) Presenter: Enrique Rosales, Xing Zeng Introduction to Probabilistic Programming Language (with Church as an example) Presenter: Enrique Rosales, Xing Zeng 1 Knowledge How can we infer knowledge from observations? 2 Ron s box Bob has a box

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Inference in Probabilistic Logic Programs using Weighted CNF s

Inference in Probabilistic Logic Programs using Weighted CNF s Inference in Probabilistic Logic Programs using Weighted CNF s Daan Fierens, Guy Van den Broeck, Ingo Thon, Bernd Gutmann, Luc De Raedt Department of Computer Science Katholieke Universiteit Leuven Celestijnenlaan

More information

Bayes Nets: Sampling

Bayes Nets: Sampling Bayes Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Approximate Inference:

More information

Machine Learning. Probabilistic KNN.

Machine Learning. Probabilistic KNN. Machine Learning. Mark Girolami girolami@dcs.gla.ac.uk Department of Computing Science University of Glasgow June 21, 2007 p. 1/3 KNN is a remarkably simple algorithm with proven error-rates June 21, 2007

More information

Probabilistic Inductive Constraint Logic

Probabilistic Inductive Constraint Logic Probabilistic Inductive Constraint Logic Fabrizio Riguzzi 1, Elena Bellodi 2, Riccardo Zese 2, Giuseppe Cota 2, and Evelina Lamma 2 1 Dipartimento di Matematica e Informatica University of Ferrara Via

More information

Tutorial on Probabilistic Programming with PyMC3

Tutorial on Probabilistic Programming with PyMC3 185.A83 Machine Learning for Health Informatics 2017S, VU, 2.0 h, 3.0 ECTS Tutorial 02-04.04.2017 Tutorial on Probabilistic Programming with PyMC3 florian.endel@tuwien.ac.at http://hci-kdd.org/machine-learning-for-health-informatics-course

More information

Inference and Learning in Probabilistic Logic Programs using Weighted Boolean Formulas

Inference and Learning in Probabilistic Logic Programs using Weighted Boolean Formulas Under consideration for publication in Theory and Practice of Logic Programming 1 Inference and Learning in Probabilistic Logic Programs using Weighted Boolean Formulas DAAN FIERENS, GUY VAN DEN BROECK,

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science Inference in Probabilistic Logic Programs using Weighted CNF s Daan Fierens Guy Van den Broeck Ingo Thon Bernd Gutmann Luc De Raedt Report CW 607, June 2011 Katholieke Universiteit Leuven Department of

More information

MCMC notes by Mark Holder

MCMC notes by Mark Holder MCMC notes by Mark Holder Bayesian inference Ultimately, we want to make probability statements about true values of parameters, given our data. For example P(α 0 < α 1 X). According to Bayes theorem:

More information

Probabilistic Logic Languages

Probabilistic Logic Languages Probabilistic Logic Languages Fabrizio Riguzzi ENDIF University of Ferrara, Italy fabrizio.riguzzi@unife.it Fabrizio Riguzzi (University of Ferrara) Probabilistic Logic Languages 1 / 65 Outline 1 Probabilistic

More information

Probabilistic Programming; Ways Forward. Frank Wood

Probabilistic Programming; Ways Forward. Frank Wood Probabilistic Programming; Ways Forward Frank Wood Outline What is probabilistic programming? What are the goals of the field? What are some challenges? Where are we now? Ways forward What is probabilistic

More information

Semantics and Inference for Probabilistic Ontologies

Semantics and Inference for Probabilistic Ontologies Semantics and Inference for Probabilistic Ontologies Fabrizio Riguzzi, Elena Bellodi, Evelina Lamma, and Riccardo Zese ENDIF University of Ferrara, Italy, email: {fabrizio.riguzzi, elena.bellodi, evelina.lamma}@unife.it,

More information

This is an Open Access document downloaded from ORCA, Cardiff University's institutional repository:

This is an Open Access document downloaded from ORCA, Cardiff University's institutional repository: This is an Open Access document downloaded from ORCA, Cardiff University's institutional repository: http://orca.cf.ac.uk/106737/ This is the author s version of a work that was submitted to / accepted

More information

Toward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming

Toward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming Toward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming Arjen Hommersom 1,2 and Marcos L.P. Bueno 2 1 Open University of the Netherlands 2 Radboud University Nijmegen, The Netherlands

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Advanced Statistical Modelling

Advanced Statistical Modelling Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Probabilistic Machine Learning

Probabilistic Machine Learning Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Human-level concept learning through probabilistic program induction

Human-level concept learning through probabilistic program induction B.M Lake, R. Salakhutdinov, J.B. Tenenbaum Human-level concept learning through probabilistic program induction journal club at two aspects in which machine learning spectacularly lags behind human learning

More information

Closed-form Gibbs Sampling for Graphical Models with Algebraic constraints. Hadi Mohasel Afshar Scott Sanner Christfried Webers

Closed-form Gibbs Sampling for Graphical Models with Algebraic constraints. Hadi Mohasel Afshar Scott Sanner Christfried Webers Closed-form Gibbs Sampling for Graphical Models with Algebraic constraints Hadi Mohasel Afshar Scott Sanner Christfried Webers Inference in Hybrid Graphical Models / Probabilistic Programs Limitations

More information

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling 12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference

More information

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22. Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides

More information

Epistemic and Statistical Probabilistic Ontologies

Epistemic and Statistical Probabilistic Ontologies Epistemic and Statistical Probabilistic Ontologies Fabrizio Riguzzi, Elena Bellodi, Evelina Lamma, Riccardo Zese ENDIF University of Ferrara, Italy fabrizio.riguzzi@unife.it, elena.bellodi@unife.it, evelina.lamma@unife.it,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 10-708 Probabilistic Graphical Models Homework 3 (v1.1.0) Due Apr 14, 7:00 PM Rules: 1. Homework is due on the due date at 7:00 PM. The homework should be submitted via Gradescope. Solution to each problem

More information

Fast Likelihood-Free Inference via Bayesian Optimization

Fast Likelihood-Free Inference via Bayesian Optimization Fast Likelihood-Free Inference via Bayesian Optimization Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology

More information

KU Leuven Department of Computer Science

KU Leuven Department of Computer Science Implementation and Performance of Probabilistic Inference Pipelines: Data and Results Dimitar Shterionov Gerda Janssens Report CW 679, March 2015 KU Leuven Department of Computer Science Celestijnenlaan

More information

Sum-Product Networks: A New Deep Architecture

Sum-Product Networks: A New Deep Architecture Sum-Product Networks: A New Deep Architecture Pedro Domingos Dept. Computer Science & Eng. University of Washington Joint work with Hoifung Poon 1 Graphical Models: Challenges Bayesian Network Markov Network

More information

Expert Systems! Knowledge Based Systems!

Expert Systems! Knowledge Based Systems! Expert Systems Knowledge Based Systems ES-1 Medical diagnosis» Disease identification Example Areas of Use ES-2 Example Areas of Use 2 Medical diagnosis» Disease identification Natural resource exploration»

More information

Expert Systems! Knowledge Based Systems!

Expert Systems! Knowledge Based Systems! Expert Systems Knowledge Based Systems ES-1 Medical diagnosis» Disease identification Example Areas of Use ES-2 Example Areas of Use 2 Medical diagnosis» Disease identification Natural resource exploration»

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate

More information

Probabilistic programming and Stan. mc-stan.org

Probabilistic programming and Stan. mc-stan.org Probabilistic programming and Stan mc-stan.org Outline What is probabilistic programming Stan now Stan in the future A bit about other software Probabilistic programming Probabilistic programming framework

More information

Artificial Intelligence

Artificial Intelligence ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning

More information

Automatic Sampler Discovery via Probabilistic Programming and Approximate Bayesian Computation

Automatic Sampler Discovery via Probabilistic Programming and Approximate Bayesian Computation Automatic Sampler Discovery via Probabilistic Programming and Approximate Bayesian Computation Yura Perov and Frank Wood Department of Engineering Science, University of Oxford, United Kingdom, {perov,fwood}@robots.ox.ac.uk

More information

Part IV: Monte Carlo and nonparametric Bayes

Part IV: Monte Carlo and nonparametric Bayes Part IV: Monte Carlo and nonparametric Bayes Outline Monte Carlo methods Nonparametric Bayesian models Outline Monte Carlo methods Nonparametric Bayesian models The Monte Carlo principle The expectation

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Recent Advances in Bayesian Inference Techniques

Recent Advances in Bayesian Inference Techniques Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian

More information

an introduction to bayesian inference

an introduction to bayesian inference with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Learning Bayesian Networks for Biomedical Data

Learning Bayesian Networks for Biomedical Data Learning Bayesian Networks for Biomedical Data Faming Liang (Texas A&M University ) Liang, F. and Zhang, J. (2009) Learning Bayesian Networks for Discrete Data. Computational Statistics and Data Analysis,

More information

Introduction to Gaussian Processes

Introduction to Gaussian Processes Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of

More information

StarAI Full, 6+1 pages Short, 2 page position paper or abstract

StarAI Full, 6+1 pages Short, 2 page position paper or abstract StarAI 2015 Fifth International Workshop on Statistical Relational AI At the 31st Conference on Uncertainty in Artificial Intelligence (UAI) (right after ICML) In Amsterdam, The Netherlands, on July 16.

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

A Bayesian Approach to Phylogenetics

A Bayesian Approach to Phylogenetics A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte

More information

Brief Introduction of Machine Learning Techniques for Content Analysis

Brief Introduction of Machine Learning Techniques for Content Analysis 1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview

More information

A Probabilistic Extension of the Stable Model Semantics

A Probabilistic Extension of the Stable Model Semantics A Probabilistic Extension of the Stable Model Semantics Joohyung Lee and Yi Wang School of Computing, Informatics, and Decision Systems Engineering Arizona State University, Tempe, USA {joolee, ywang485}@asu.edu

More information

Learning With Bayesian Networks. Markus Kalisch ETH Zürich

Learning With Bayesian Networks. Markus Kalisch ETH Zürich Learning With Bayesian Networks Markus Kalisch ETH Zürich Inference in BNs - Review P(Burglary JohnCalls=TRUE, MaryCalls=TRUE) Exact Inference: P(b j,m) = c Sum e Sum a P(b)P(e)P(a b,e)p(j a)p(m a) Deal

More information

Approximate inference in Energy-Based Models

Approximate inference in Energy-Based Models CSC 2535: 2013 Lecture 3b Approximate inference in Energy-Based Models Geoffrey Hinton Two types of density model Stochastic generative model using directed acyclic graph (e.g. Bayes Net) Energy-based

More information

Forward Problems and their Inverse Solutions

Forward Problems and their Inverse Solutions Forward Problems and their Inverse Solutions Sarah Zedler 1,2 1 King Abdullah University of Science and Technology 2 University of Texas at Austin February, 2013 Outline 1 Forward Problem Example Weather

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

Bayesian Learning. CSL603 - Fall 2017 Narayanan C Krishnan

Bayesian Learning. CSL603 - Fall 2017 Narayanan C Krishnan Bayesian Learning CSL603 - Fall 2017 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Bayes Theorem MAP Learners Bayes optimal classifier Naïve Bayes classifier Example text classification Bayesian networks

More information

A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace

A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I kevin small & byron wallace today a review of probability random variables, maximum likelihood, etc. crucial for clinical

More information

Probabilistic Logic CNF for Reasoning

Probabilistic Logic CNF for Reasoning Probabilistic Logic CNF for Reasoning Song- Chun Zhu, Sinisa Todorovic, and Ales Leonardis At CVPR, Providence, Rhode Island June 16, 2012 Goal input tracking parsing reasoning activity recognition in

More information

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return

More information

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks

More information

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Approximate Bayesian computation for the parameters of PRISM programs

Approximate Bayesian computation for the parameters of PRISM programs Approximate Bayesian computation for the parameters of PRISM programs James Cussens Department of Computer Science & York Centre for Complex Systems Analysis University of York Heslington, York, YO10 5DD,

More information

Learning Effect Axioms via Probabilistic Logic Programming

Learning Effect Axioms via Probabilistic Logic Programming Learning Effect Axioms via Probabilistic Logic Programming Rolf Schwitter Department of Computing, Macquarie University, Sydney NSW 2109, Australia Rolf.Schwitter@mq.edu.au Abstract Events have effects

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

Probabilistic Time Series Classification

Probabilistic Time Series Classification Probabilistic Time Series Classification Y. Cem Sübakan Boğaziçi University 25.06.2013 Y. Cem Sübakan (Boğaziçi University) M.Sc. Thesis Defense 25.06.2013 1 / 54 Problem Statement The goal is to assign

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

CSCE 478/878 Lecture 6: Bayesian Learning

CSCE 478/878 Lecture 6: Bayesian Learning Bayesian Methods Not all hypotheses are created equal (even if they are all consistent with the training data) Outline CSCE 478/878 Lecture 6: Bayesian Learning Stephen D. Scott (Adapted from Tom Mitchell

More information

Lecture 8: Bayesian Networks

Lecture 8: Bayesian Networks Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1

More information

1 Inference for binomial proportion (Matlab/Python)

1 Inference for binomial proportion (Matlab/Python) Bayesian data analysis exercises from 2015 1 Inference for binomial proportion (Matlab/Python) Algae status is monitored in 274 sites at Finnish lakes and rivers. The observations for the 2008 algae status

More information

Bayesian leave-one-out cross-validation for large data sets

Bayesian leave-one-out cross-validation for large data sets 1st Symposium on Advances in Approximate Bayesian Inference, 2018 1 5 Bayesian leave-one-out cross-validation for large data sets Måns Magnusson Michael Riis Andersen Aki Vehtari Aalto University mans.magnusson@aalto.fi

More information

Bayesian Learning (II)

Bayesian Learning (II) Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Lecture 16 Deep Neural Generative Models

Lecture 16 Deep Neural Generative Models Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods. Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman

15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods. Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman 15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman Admin Reminder: midterm March 29 Reminder: project milestone

More information

Approximate Inference using MCMC

Approximate Inference using MCMC Approximate Inference using MCMC 9.520 Class 22 Ruslan Salakhutdinov BCS and CSAIL, MIT 1 Plan 1. Introduction/Notation. 2. Examples of successful Bayesian models. 3. Basic Sampling Algorithms. 4. Markov

More information

Lifted MAP Inference for Markov Logic

Lifted MAP Inference for Markov Logic Lifted MAP Inference for Markov Logic - Somdeb Sarkhel - Deepak Venugopal - Happy Mittal - Parag Singla - Vibhav Gogate Outline Introduction Lifted MAP Inference for Markov Logic Networks S. Sarkhel, D.

More information

BN Semantics 3 Now it s personal! Parameter Learning 1

BN Semantics 3 Now it s personal! Parameter Learning 1 Readings: K&F: 3.4, 14.1, 14.2 BN Semantics 3 Now it s personal! Parameter Learning 1 Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 22 nd, 2006 1 Building BNs from independence

More information

PROBABILISTIC REASONING SYSTEMS

PROBABILISTIC REASONING SYSTEMS PROBABILISTIC REASONING SYSTEMS In which we explain how to build reasoning systems that use network models to reason with uncertainty according to the laws of probability theory. Outline Knowledge in uncertain

More information