Deductive and Inductive Probabilistic Programming
|
|
- Agatha Cain
- 5 years ago
- Views:
Transcription
1 Deductive and Inductive Probabilistic Programming Fabrizio Riguzzi F. Riguzzi Deductive and Inductive PP 1 / 43
2 Outline Probabilistic programming Probabilistic logic programming Inference Learning Applications F. Riguzzi Deductive and Inductive PP 2 / 43
3 Probabilistic Programming Users specify a probabilistic model in its entirety (e.g., by writing code that generates a sample from the joint distribution) and inference follows automatically given the specification. PP languages provide the full power of modern programming languages for describing complex distributions Reuse of libraries of models Interactive modeling Abstraction F. Riguzzi Deductive and Inductive PP 3 / 43
4 Probabilistic Programming F. Riguzzi Deductive and Inductive PP 4 / 43
5 Probabilistic Programming Languages Name Extends from Host language Venture Scheme C++ Probabilistic-C C C Anglican Scheme Clojure IBAL OCaml PRISM B-Prolog Infer.NET.NET Framework.NET Framework dimple MATLAB, Java chimple MATLAB, Java BLOG Java PSQL SQL BUGS FACTORIE Scala PMTK MATLAB MATLAB Alchemy C++ Dyna Prolog Name Extends from Host language Figaro Scala Church Scheme JavaScript, Scheme ProbLog Prolog Python, Jython ProBT C++, Python Stan (software) C++ Hakaru Haskell Haskell BAli-Phy (software) Haskell C++ ProbCog Java, Python Gamble Racket Tuffy Java PyMC Python Python Lea Python Python WebPPL JavaScript JavaScript Picture Julia Julia Turing.jl Julia Julia Source programming_language F. Riguzzi Deductive and Inductive PP 5 / 43
6 Probabilistic Programming Languages Only three are logic, the others imperative/functional/object oriented DARPA released in 2013 the funding call Probabilistic Programming for Advancing Machine Learning (PPAML) Aim: develop probabilistic programming languages and accompanying tools to facilitate the construction of new machine learning applications across a wide range of domains. Focus: functional PP F. Riguzzi Deductive and Inductive PP 6 / 43
7 Probabilistic Logic Programming What are we missing? Is logic programming to blame? F. Riguzzi Deductive and Inductive PP 7 / 43
8 Thesis Probabilistic logic programming is alive and kicking! F. Riguzzi Deductive and Inductive PP 8 / 43
9 Strengths Relationships are first class citizens Conceptually easier to lift Strong semantics Inductive systems F. Riguzzi Deductive and Inductive PP 9 / 43
10 Weaknesses Handling non-termination Continuous variables F. Riguzzi Deductive and Inductive PP 10 / 43
11 Non-termination Possible when the number of explanations for the query is infinite F. Riguzzi Deductive and Inductive PP 11 / 43
12 Non-termination: Inducing Arithmetic Functions Church code (define (random-arithmetic-fn) (if (flip 0.3) (random-combination (random-arithmetic-fn) (random-arithmetic-fn)) (if (flip) (lambda (x) x) (random-constant-fn)))) (define (random-combination f g) (define op (uniform-draw (list + -))) (lambda (x) (op (f x) (g x)))) (define (random-constant-fn) (define i (sample-integer 10)) (lambda (x) i)) F. Riguzzi Deductive and Inductive PP 12 / 43
13 Non-termination: Inducing Arithmetic Functions LPAD (cplint) code example/inference/arithm.pl eval(x,y):- random_fn(x,0,f), Y is F. op(l,+):0.5;op(l,-):0.5. random_fn(x,l,f):- comb(l), random_fn(x,l(l),f1), random_fn(x,r(l),f2), op(l,op), F=..[Op,F1,F2]. random_fn(x,l,f):- \+ comb(l), base_random_fn(x,l,f). comb(_):0.3. base_random_fn(x,l,x):- identity(l). base_random_fn(_x,l,c):- \+ identity(l), random_const(l,c). identity(_):0.5. random_const(_,c):discrete(c,[0:0.1,1:0.1,2:0.1,3:0.1,4:0.1, 5:0.1,6:0.1,7:0.1,8:0.1,9:0.1]). F. Riguzzi Deductive and Inductive PP 13 / 43
14 Non-termination: Inducing Arithmetic Functions Aim: given observations of couples input-output for the random function, predict the output for a new input Arbitrarily complex functions have a non-zero probability of being selected The program has a non-terminating execution. Exact inference: infinite number of explanations F. Riguzzi Deductive and Inductive PP 14 / 43
15 Non-termination: Inducing Arithmetic Functions (define (sample) (rejection-query (define my-proc (random-arithmetic-fn)) (my-proc 2) (= (my-proc 1) 3))) (hist (repeat 100 sample)) F. Riguzzi Deductive and Inductive PP 15 / 43
16 Solution Use (T. Sato, P. Meyer, Infinite probability computation by cyclic explanation graphs, Theor. Pract. Log. Prog ) or (A. Gorlin, C. R. Ramakrishnan, S. A. Smolka, Model checking with proba- bilistic tabled logic programming, Theor. Pract. Log. Prog. 12 (4-5) 2012) or resort to sampling: with the increase of complexity, the probability of functions tend to 0 and the probability of the infinite trace is 0 Metropolis Hastings: (Nampally, A., Ramakrishnan, C.: Adaptive MCMC-based inference in probabilistic logic programs. arxiv preprint arxiv: ) Monte Carlo sampling is attractive for the simplicity of its implementation and because you can improve the estimate as more time is available. F. Riguzzi Deductive and Inductive PP 16 / 43
17 Monte Carlo The disjunctive clause C r = H 1 : α 1... H n : α n L 1,..., L m. is transformed into the set of clauses MC(C r ) MC(C r, 1) = H 1 L 1,..., L m, sample_head(n, r, VC, NH), NH = MC(C r, n) = H n L 1,..., L m, sample_head(n, r, VC, NH), NH = n. Sample truth value of query Q:... (call(q)->... NT1 is NT+1 ; NT1 =NT), F. Riguzzi Deductive and Inductive PP 17 / 43
18 Metropolis-Hastings MCMC A Markov chain is built by taking an initial sample and by generating successor samples. The initial sample is built by randomly sampling choices so that the evidence is true. A successor sample is obtained by deleting a fixed number of sampled probabilistic choices. Then the evidence is queried If the query succeeds, the goal is queried. The sample is accepted with a probability of min{1, N 0 N 1 } where N 0 (N 1 ) is the number of choices sampled in the previous (current) sample. F. Riguzzi Deductive and Inductive PP 18 / 43
19 Solution In cplint:?- mc_mh_sample(eval(2,4),eval(1,3),100,100,3,t,f,p). Probability of eval(2,4) given that eval(1,3) is true F = 90, T = 10, P = 0.1 You can also try rejection sampling (usually slower)?- mc_rejection_sample(eval(2,4),eval(1,3),100, T,F,P). F. Riguzzi Deductive and Inductive PP 19 / 43
20 Solution You may be interested in the distribution of the output In cplint:?- mc_mh_sample_arg_bar(eval(2,y),eval(1,3),100, 100,3,Y,V). F. Riguzzi Deductive and Inductive PP 20 / 43
21 Solution You may be interested in the expected value of the output In cplint:?- mc_mh_expectation(eval(2,y),eval(1,3), 100,100,3,Y,E). E = 3.21 F. Riguzzi Deductive and Inductive PP 21 / 43
22 Continuous Random Variables Distributional clauses (B. Gutmann, I. Thon, A. Kimmig, M. Bruynooghe, and L. De Raedt, The magic of logical inference in probabilistic programming, Theory and Practice of Logic Programming, 2011) Gaussian mixture model in cplint: heads:0.6;tails:0.4. g(x): gaussian(x,0, 1). h(x): gaussian(x,5, 2). mix(x) :- heads, g(x). mix(x) :- tails, h(x). F. Riguzzi Deductive and Inductive PP 22 / 43
23 Continuous Random Variables Inference by sampling Without evidence or evidence on discrete random variables, you can reuse the same methods Sampling arguments of goals for building a probability density of the arguments. F. Riguzzi Deductive and Inductive PP 23 / 43
24 Gaussian Mixture Model heads:0.6;tails:0.4. g(x): gaussian(x,0, 1). h(x): gaussian(x,5, 2). mix(x) :- heads, g(x). mix(x) :- tails, h(x).?- mc_sample_arg(mix(x),10000,x,l0), histogram(l0,40,chart). F. Riguzzi Deductive and Inductive PP 24 / 43
25 Evidence on Continuous Random Variables You cannot use rejection sampling or Metropolis-Hastings, as the probability of the evidence is 0 You can use likelihood weighting to obtain samples of continuous arguments of a goal. (Nitti, D., De Laet, T., De Raedt, L.: Probabilistic logic programming for hybrid relational domains. Mach. Learn. 103(3), ) F. Riguzzi Deductive and Inductive PP 25 / 43
26 Likelihood Weighting For each sample to be taken, likelihood weighting samples the query and then assigns a weight to the sample on the basis of evidence. The weight is computed by deriving the evidence backward in the same sample of the query starting with a weight of one Each time a choice should be taken or a continuous variable sampled, if the choice/variable has already been taken, the current weight is multiplied by probability of the choice/by the density value of the continuous value. F. Riguzzi Deductive and Inductive PP 26 / 43
27 Bayesian Estimation Problem from examples/viewer/?worksheet=gaussian-posteriors Estimate the true value of a Gaussian distributed random variable, given some observed data. The variance is known and we suppose that the mean has itself a Gaussian distribution with mean 1 and variance 5 (prior on the parameter) We take different measurement (e.g. at different times), indexed with an integer. F. Riguzzi Deductive and Inductive PP 27 / 43
28 Bayesian Estimation Anglican code (def dataset [9 8]) (defquery gaussian-model [data] (let [mu (sample (normal 1 (sqrt 5))) sigma (sqrt 2)] (doall (map (fn [x] (observe (normal mu sigma) x)) data)) mu)) (def posterior ((conditional gaussian-model :smc :number-of-particles 10) dataset)) (def posterior-samples (repeatedly #(sample* posterior))) F. Riguzzi Deductive and Inductive PP 28 / 43
29 Bayesian Estimation cplint code example/inference/gauss_mean_est.pl value(i,x) :- mean(m), value(i,m,x). mean(m): gaussian(m,1.0, 5.0). value(_,m,x): gaussian(x,m, 2.0).?- mc_sample_arg(value(0,y),10000,y,l0), mc_lw_sample_arg(value(0,x),(value(1,9),value(2,8)),10000,x,l), densities(l0,l,40,chart). F. Riguzzi Deductive and Inductive PP 29 / 43
30 Learning Parameter learning Structure learning more developed for PLP, but (Perov, Yura N., and Frank D. Wood. Learning Probabilistic Programs. arxiv preprint arxiv: ). (Lake, Brenden M., Ruslan Salakhutdinov, and Joshua B. Tenenbaum. Human-level concept learning through probabilistic program induction. Science ). (Gaunt, Alexander L., et al. TerpreT: A Probabilistic Programming Language for Program Induction. arxiv preprint arxiv: ). F. Riguzzi Deductive and Inductive PP 30 / 43
31 Parameter Learning Problem: given a set of interpretations, a program, find the parameters maximizing the likelihood of the interpretations (or of instances of a target predicate) Exploit the equivalence with BN to use BN learning algorithms The interpretations record the truth value of ground atoms, not of the choice variables Unseen data: relative frequency can t be used F. Riguzzi Deductive and Inductive PP 31 / 43
32 Parameter Learning (Thon et al. ECML 2008) proposed an adaptation of EM for CPT-L, a simplified version of LPADs The algorithm computes the counts efficiently by repeatedly traversing the BDDs representing the explanations (Ishihata et al. ILP 2008) independently proposed a similar algorithm LFI-PROBLOG (Gutamnn et al. ECML 2011) is the adaptation of EM to ProbLog EMBLEM (Riguzzi & Bellodi IDAJ 2013) adapts (Ishihata et al. ILP 2008) to LPADs F. Riguzzi Deductive and Inductive PP 32 / 43
33 Structure Learning Given a trivial LPAD or an empty one, a set of interpretations (data) Find the model and the parameters that maximize the probability of the data (log-likelihood) SLIPCOVER: Structure LearnIng of Probabilistic logic program by searching OVER the clause space 1 Beam search in the space of clauses to find the promising ones 2 Greedy search in the space of probabilistic programs guided by the LL of the data. Parameter learning by means of EMBLEM F. Riguzzi Deductive and Inductive PP 33 / 43
34 Applications Link prediction: given a (social) network, compute the probability of the existence of a link between two entities (UWCSE) advisedby(x, Y) :0.3 :- publication(p, X), publication(p, Y), student(x). F. Riguzzi Deductive and Inductive PP 34 / 43
35 Applications Classify web pages on the basis of the link structure (WebKB) coursepage(page1): 0.3 :- linkto(page2,page1),coursepage(page2). coursepage(page1): 0.3 :- linkto(page2,page1),facultypage(page2).... coursepage(page): 0.3 :- has( abstract,page).... F. Riguzzi Deductive and Inductive PP 35 / 43
36 Applications Entity resolution: identify identical entities in text or databases samebib(a,b):0.3 :- samebib(a,c), samebib(c,b). sameauthor(a,b):0.3 :- sameauthor(a,c), sameauthor(c,b). sametitle(a,b):0.3 :- sametitle(a,c), sametitle(c,b). samevenue(a,b):0.3 :- samevenue(a,c), samevenue(c,b). samebib(b,c):0.3 :- author(b,d),author(c,e),sameauthor(d,e). samebib(b,c):0.3 :- title(b,d),title(c,e),sametitle(d,e). samebib(b,c):0.3 :- venue(b,d),venue(c,e),samevenue(d,e). samevenue(b,c):0.3 :- haswordvenue(b,word_06), haswordvenue(c,word_06).... F. Riguzzi Deductive and Inductive PP 36 / 43
37 Applications Chemistry: given the chemical composition of a substance, predict its mutagenicity or its carcenogenicity active(a):0.5 :- atm(a,b,c,29,c), gteq(c,-0.003), ring_size_5(a,d). active(a):0.5 :- lumo(a,b), lteq(b,-2.072). active(a):0.5 :- bond(a,b,c,2), bond(a,c,d,1), ring_size_5(a,e). active(a):0.5 :- carbon_6_ring(a,b). active(a):0.5 :- anthracene(a,b).... F. Riguzzi Deductive and Inductive PP 37 / 43
38 Applications Medicine: diagnose diseases on the basis of patient information (Hepatitis), influence of genes on HIV, risk of falling of elderly people (FFRAT) F. Riguzzi Deductive and Inductive PP 38 / 43
39 Experiments - Area Under the PR Curve System HIV UW-CSE Mondial SLIPCOVER 0.82 ± ± ± 0.07 SLIPCASE 0.78 ± ± ± 0.06 LSM 0.37 ± ± ALEPH ± ± 0.07 RDN-B 0.28 ± ± ± 0.07 MLN-BT 0.29 ± ± ± 0.10 MLN-BC 0.51 ± ± ± 0.09 BUSL 0.38 ± ± F. Riguzzi Deductive and Inductive PP 39 / 43
40 Experiments - Area Under the PR Curve System Carcinogenesis Mutagenesis Hepatitis SLIPCOVER ± ± 0.01 SLIPCASE ± ± 0.05 LSM ± 0.04 ALEPH ± RDN-B ± ± 0.01 MLN-BT ± ± 0.02 MLN-BC ± ± 0.02 BUSL ± 0.03 F. Riguzzi Deductive and Inductive PP 40 / 43
41 PLP Online Inference (knwoledge compilation, Monte Carlo) Parameter learning (EMBLEM) Structure learning (SLIPCOVER) Inference (knwoledge compilation, Monte Carlo) Parameter learning (LFI-ProbLog) F. Riguzzi Deductive and Inductive PP 41 / 43
42 Conclusions PLP is still a fertile field but......we must look at other communities and build bridges and......join forces! Much is left to do: Tractable sublanguages (see following talk) Lifted inference Structure/Parameter learning (also for programs with continuous variables) F. Riguzzi Deductive and Inductive PP 42 / 43
43 F. Riguzzi Deductive and Inductive PP 43 / 43
Probabilistic Logical Inference On the Web
Probabilistic Logical Inference On the Web Marco Alberti 1, Giuseppe Cota 2, Fabrizio Riguzzi 1, and Riccardo Zese 2 1 Dipartimento di Matematica e Informatica University of Ferrara Via Saragat 1, I-44122,
More informationThe PITA System for Logical-Probabilistic Inference
The System for Logical-Probabilistic Inference Fabrizio Riguzzi 1 and Terrance Swift 2 1 EDIF University of Ferrara, Via Saragat 1, I-44122, Ferrara, Italy fabrizio.riguzzi@unife.it 2 CETRIA Universidade
More informationStructure Learning of Probabilistic Logic Programs by MapReduce
Structure Learning of Probabilistic Logic Programs by MapReduce Fabrizio Riguzzi 1, Elena Bellodi 2, Riccardo Zese 2, Giuseppe Cota 2, and Evelina Lamma 2 1 Dipartimento di Matematica e Informatica University
More informationLifted Discriminative Learning of Probabilistic Logic Programs
Lifted Discriminative Learning of Probabilistic Logic Programs Arnaud Nguembang Fadja 1 and Fabrizio Riguzzi 2 1 Dipartimento di Ingegneria University of Ferrara Via Saragat 1, I-44122, Ferrara, Italy
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationMarkov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationIntroduction to Probabilistic Programming Languages (PPL)
Introduction to Probabilistic Programming Languages (PPL) Anton Andreev Ingénieur d'études CNRS andreev.anton@.grenoble-inp.fr Something simple int function add(int a, int b) return a + b add(3, 2) 5 Deterministic
More informationLogic Programming Techniques for Reasoning with Probabilistic Ontologies
Logic Programming Techniques for Reasoning with Probabilistic Ontologies Riccardo Zese, Elena Bellodi, Evelina Lamma and Fabrizio Riguzzi University of Ferrara, Italy riccardo.zese@unife.it Zese, Bellodi,
More informationMarkov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationMCINTYRE: A Monte Carlo Algorithm for Probabilistic Logic Programming
: A Monte Carlo Algorithm for Probabilistic Logic Programming Fabrizio Riguzzi ENDIF Università di Ferrara Via Saragat, 1 44122 Ferrara, Italy. {fabrizio.riguzzi}@unife.it Abstract. Probabilistic Logic
More informationPartial inconsistency and vector semantics: sampling, animation, and program learning
Partial inconsistency and vector semantics: sampling, animation, and program learning Michael Bukatin Nokia Corporation, Cambridge, MA Joint work with Ralph Kopperman and Steve Matthews - - - 29th Summer
More informationProbabilistic Programming with Infer.NET. John Bronskill University of Cambridge 19 th October 2017
Probabilistic Programming with Infer.NET John Bronskill University of Cambridge 19 th October 2017 The Promise of Probabilistic Programming We want to do for machine learning what the advent of high-level
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationIntroduction to Probabilistic Programming Language (with Church as an example) Presenter: Enrique Rosales, Xing Zeng
Introduction to Probabilistic Programming Language (with Church as an example) Presenter: Enrique Rosales, Xing Zeng 1 Knowledge How can we infer knowledge from observations? 2 Ron s box Bob has a box
More informationCS 188: Artificial Intelligence. Bayes Nets
CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationInference in Probabilistic Logic Programs using Weighted CNF s
Inference in Probabilistic Logic Programs using Weighted CNF s Daan Fierens, Guy Van den Broeck, Ingo Thon, Bernd Gutmann, Luc De Raedt Department of Computer Science Katholieke Universiteit Leuven Celestijnenlaan
More informationBayes Nets: Sampling
Bayes Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Approximate Inference:
More informationMachine Learning. Probabilistic KNN.
Machine Learning. Mark Girolami girolami@dcs.gla.ac.uk Department of Computing Science University of Glasgow June 21, 2007 p. 1/3 KNN is a remarkably simple algorithm with proven error-rates June 21, 2007
More informationProbabilistic Inductive Constraint Logic
Probabilistic Inductive Constraint Logic Fabrizio Riguzzi 1, Elena Bellodi 2, Riccardo Zese 2, Giuseppe Cota 2, and Evelina Lamma 2 1 Dipartimento di Matematica e Informatica University of Ferrara Via
More informationTutorial on Probabilistic Programming with PyMC3
185.A83 Machine Learning for Health Informatics 2017S, VU, 2.0 h, 3.0 ECTS Tutorial 02-04.04.2017 Tutorial on Probabilistic Programming with PyMC3 florian.endel@tuwien.ac.at http://hci-kdd.org/machine-learning-for-health-informatics-course
More informationInference and Learning in Probabilistic Logic Programs using Weighted Boolean Formulas
Under consideration for publication in Theory and Practice of Logic Programming 1 Inference and Learning in Probabilistic Logic Programs using Weighted Boolean Formulas DAAN FIERENS, GUY VAN DEN BROECK,
More informationKatholieke Universiteit Leuven Department of Computer Science
Inference in Probabilistic Logic Programs using Weighted CNF s Daan Fierens Guy Van den Broeck Ingo Thon Bernd Gutmann Luc De Raedt Report CW 607, June 2011 Katholieke Universiteit Leuven Department of
More informationMCMC notes by Mark Holder
MCMC notes by Mark Holder Bayesian inference Ultimately, we want to make probability statements about true values of parameters, given our data. For example P(α 0 < α 1 X). According to Bayes theorem:
More informationProbabilistic Logic Languages
Probabilistic Logic Languages Fabrizio Riguzzi ENDIF University of Ferrara, Italy fabrizio.riguzzi@unife.it Fabrizio Riguzzi (University of Ferrara) Probabilistic Logic Languages 1 / 65 Outline 1 Probabilistic
More informationProbabilistic Programming; Ways Forward. Frank Wood
Probabilistic Programming; Ways Forward Frank Wood Outline What is probabilistic programming? What are the goals of the field? What are some challenges? Where are we now? Ways forward What is probabilistic
More informationSemantics and Inference for Probabilistic Ontologies
Semantics and Inference for Probabilistic Ontologies Fabrizio Riguzzi, Elena Bellodi, Evelina Lamma, and Riccardo Zese ENDIF University of Ferrara, Italy, email: {fabrizio.riguzzi, elena.bellodi, evelina.lamma}@unife.it,
More informationThis is an Open Access document downloaded from ORCA, Cardiff University's institutional repository:
This is an Open Access document downloaded from ORCA, Cardiff University's institutional repository: http://orca.cf.ac.uk/106737/ This is the author s version of a work that was submitted to / accepted
More informationToward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming
Toward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming Arjen Hommersom 1,2 and Marcos L.P. Bueno 2 1 Open University of the Netherlands 2 Radboud University Nijmegen, The Netherlands
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationAdvanced Statistical Modelling
Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1
More informationIntroduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016
Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An
More informationProbabilistic Machine Learning
Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationHuman-level concept learning through probabilistic program induction
B.M Lake, R. Salakhutdinov, J.B. Tenenbaum Human-level concept learning through probabilistic program induction journal club at two aspects in which machine learning spectacularly lags behind human learning
More informationClosed-form Gibbs Sampling for Graphical Models with Algebraic constraints. Hadi Mohasel Afshar Scott Sanner Christfried Webers
Closed-form Gibbs Sampling for Graphical Models with Algebraic constraints Hadi Mohasel Afshar Scott Sanner Christfried Webers Inference in Hybrid Graphical Models / Probabilistic Programs Limitations
More informationBayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling
12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference
More informationAnnouncements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.
Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides
More informationEpistemic and Statistical Probabilistic Ontologies
Epistemic and Statistical Probabilistic Ontologies Fabrizio Riguzzi, Elena Bellodi, Evelina Lamma, Riccardo Zese ENDIF University of Ferrara, Italy fabrizio.riguzzi@unife.it, elena.bellodi@unife.it, evelina.lamma@unife.it,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationProbabilistic Graphical Models
10-708 Probabilistic Graphical Models Homework 3 (v1.1.0) Due Apr 14, 7:00 PM Rules: 1. Homework is due on the due date at 7:00 PM. The homework should be submitted via Gradescope. Solution to each problem
More informationFast Likelihood-Free Inference via Bayesian Optimization
Fast Likelihood-Free Inference via Bayesian Optimization Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology
More informationKU Leuven Department of Computer Science
Implementation and Performance of Probabilistic Inference Pipelines: Data and Results Dimitar Shterionov Gerda Janssens Report CW 679, March 2015 KU Leuven Department of Computer Science Celestijnenlaan
More informationSum-Product Networks: A New Deep Architecture
Sum-Product Networks: A New Deep Architecture Pedro Domingos Dept. Computer Science & Eng. University of Washington Joint work with Hoifung Poon 1 Graphical Models: Challenges Bayesian Network Markov Network
More informationExpert Systems! Knowledge Based Systems!
Expert Systems Knowledge Based Systems ES-1 Medical diagnosis» Disease identification Example Areas of Use ES-2 Example Areas of Use 2 Medical diagnosis» Disease identification Natural resource exploration»
More informationExpert Systems! Knowledge Based Systems!
Expert Systems Knowledge Based Systems ES-1 Medical diagnosis» Disease identification Example Areas of Use ES-2 Example Areas of Use 2 Medical diagnosis» Disease identification Natural resource exploration»
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate
More informationProbabilistic programming and Stan. mc-stan.org
Probabilistic programming and Stan mc-stan.org Outline What is probabilistic programming Stan now Stan in the future A bit about other software Probabilistic programming Probabilistic programming framework
More informationArtificial Intelligence
ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationMCMC: Markov Chain Monte Carlo
I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov
More informationSampling Algorithms for Probabilistic Graphical models
Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning
More informationAutomatic Sampler Discovery via Probabilistic Programming and Approximate Bayesian Computation
Automatic Sampler Discovery via Probabilistic Programming and Approximate Bayesian Computation Yura Perov and Frank Wood Department of Engineering Science, University of Oxford, United Kingdom, {perov,fwood}@robots.ox.ac.uk
More informationPart IV: Monte Carlo and nonparametric Bayes
Part IV: Monte Carlo and nonparametric Bayes Outline Monte Carlo methods Nonparametric Bayesian models Outline Monte Carlo methods Nonparametric Bayesian models The Monte Carlo principle The expectation
More informationReminder of some Markov Chain properties:
Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationan introduction to bayesian inference
with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationLearning Bayesian Networks for Biomedical Data
Learning Bayesian Networks for Biomedical Data Faming Liang (Texas A&M University ) Liang, F. and Zhang, J. (2009) Learning Bayesian Networks for Discrete Data. Computational Statistics and Data Analysis,
More informationIntroduction to Gaussian Processes
Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of
More informationStarAI Full, 6+1 pages Short, 2 page position paper or abstract
StarAI 2015 Fifth International Workshop on Statistical Relational AI At the 31st Conference on Uncertainty in Artificial Intelligence (UAI) (right after ICML) In Amsterdam, The Netherlands, on July 16.
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationA Bayesian Approach to Phylogenetics
A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte
More informationBrief Introduction of Machine Learning Techniques for Content Analysis
1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview
More informationA Probabilistic Extension of the Stable Model Semantics
A Probabilistic Extension of the Stable Model Semantics Joohyung Lee and Yi Wang School of Computing, Informatics, and Decision Systems Engineering Arizona State University, Tempe, USA {joolee, ywang485}@asu.edu
More informationLearning With Bayesian Networks. Markus Kalisch ETH Zürich
Learning With Bayesian Networks Markus Kalisch ETH Zürich Inference in BNs - Review P(Burglary JohnCalls=TRUE, MaryCalls=TRUE) Exact Inference: P(b j,m) = c Sum e Sum a P(b)P(e)P(a b,e)p(j a)p(m a) Deal
More informationApproximate inference in Energy-Based Models
CSC 2535: 2013 Lecture 3b Approximate inference in Energy-Based Models Geoffrey Hinton Two types of density model Stochastic generative model using directed acyclic graph (e.g. Bayes Net) Energy-based
More informationForward Problems and their Inverse Solutions
Forward Problems and their Inverse Solutions Sarah Zedler 1,2 1 King Abdullah University of Science and Technology 2 University of Texas at Austin February, 2013 Outline 1 Forward Problem Example Weather
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationBayesian Learning. CSL603 - Fall 2017 Narayanan C Krishnan
Bayesian Learning CSL603 - Fall 2017 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Bayes Theorem MAP Learners Bayes optimal classifier Naïve Bayes classifier Example text classification Bayesian networks
More informationA.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace
A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I kevin small & byron wallace today a review of probability random variables, maximum likelihood, etc. crucial for clinical
More informationProbabilistic Logic CNF for Reasoning
Probabilistic Logic CNF for Reasoning Song- Chun Zhu, Sinisa Todorovic, and Ales Leonardis At CVPR, Providence, Rhode Island June 16, 2012 Goal input tracking parsing reasoning activity recognition in
More informationAnnouncements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic
CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return
More informationEE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks
More informationProbabilistic Graphical Models Lecture 17: Markov chain Monte Carlo
Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationApproximate Bayesian computation for the parameters of PRISM programs
Approximate Bayesian computation for the parameters of PRISM programs James Cussens Department of Computer Science & York Centre for Complex Systems Analysis University of York Heslington, York, YO10 5DD,
More informationLearning Effect Axioms via Probabilistic Logic Programming
Learning Effect Axioms via Probabilistic Logic Programming Rolf Schwitter Department of Computing, Macquarie University, Sydney NSW 2109, Australia Rolf.Schwitter@mq.edu.au Abstract Events have effects
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationProbabilistic Time Series Classification
Probabilistic Time Series Classification Y. Cem Sübakan Boğaziçi University 25.06.2013 Y. Cem Sübakan (Boğaziçi University) M.Sc. Thesis Defense 25.06.2013 1 / 54 Problem Statement The goal is to assign
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationCSCE 478/878 Lecture 6: Bayesian Learning
Bayesian Methods Not all hypotheses are created equal (even if they are all consistent with the training data) Outline CSCE 478/878 Lecture 6: Bayesian Learning Stephen D. Scott (Adapted from Tom Mitchell
More informationLecture 8: Bayesian Networks
Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1
More information1 Inference for binomial proportion (Matlab/Python)
Bayesian data analysis exercises from 2015 1 Inference for binomial proportion (Matlab/Python) Algae status is monitored in 274 sites at Finnish lakes and rivers. The observations for the 2008 algae status
More informationBayesian leave-one-out cross-validation for large data sets
1st Symposium on Advances in Approximate Bayesian Inference, 2018 1 5 Bayesian leave-one-out cross-validation for large data sets Måns Magnusson Michael Riis Andersen Aki Vehtari Aalto University mans.magnusson@aalto.fi
More informationBayesian Learning (II)
Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationLecture 16 Deep Neural Generative Models
Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed
More informationIntroduction to Probability and Statistics (Continued)
Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:
More information15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods. Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman
15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman Admin Reminder: midterm March 29 Reminder: project milestone
More informationApproximate Inference using MCMC
Approximate Inference using MCMC 9.520 Class 22 Ruslan Salakhutdinov BCS and CSAIL, MIT 1 Plan 1. Introduction/Notation. 2. Examples of successful Bayesian models. 3. Basic Sampling Algorithms. 4. Markov
More informationLifted MAP Inference for Markov Logic
Lifted MAP Inference for Markov Logic - Somdeb Sarkhel - Deepak Venugopal - Happy Mittal - Parag Singla - Vibhav Gogate Outline Introduction Lifted MAP Inference for Markov Logic Networks S. Sarkhel, D.
More informationBN Semantics 3 Now it s personal! Parameter Learning 1
Readings: K&F: 3.4, 14.1, 14.2 BN Semantics 3 Now it s personal! Parameter Learning 1 Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 22 nd, 2006 1 Building BNs from independence
More informationPROBABILISTIC REASONING SYSTEMS
PROBABILISTIC REASONING SYSTEMS In which we explain how to build reasoning systems that use network models to reason with uncertainty according to the laws of probability theory. Outline Knowledge in uncertain
More information