Probabilistic Abductive Logic Programming using Possible Worlds
|
|
- Elmer Collins
- 6 years ago
- Views:
Transcription
1 Probabilistic Abductive Logic Programming using Possible Worlds Fulvio Rotella 1 and Stefano Ferilli 1,2 {fulvio.rotella, stefano.ferilli}@uniba.it 1 DIB Dipartimento di Informatica Università di Bari 2 CILA Centro Interdipartimentale per la Logica e sue Applicazioni Università di Bari XXVIII Convegno Italiano di Logica Computazionale - CILC September 2013 Fulvio Rotella, and Stefano Ferilli
2 Motivation Artificial Intelligence: two approaches Numerical/statistical Relational Strengths and weaknesses Numerical/statistical Relational + handle amount of data + handle incompleteness and uncertainty - flat representations - no relationships between objects/attributes + complex representations of data + comprehensibility - no incompleteness - no noise and uncertainty
3 Motivation Problem: Real World data multi-relational, heterogeneous and semi-structured noisy and uncertain Solution: Relational Representations + Probability Logic Programming representation language and reasoning strategies Probabilistic Reasoning robustness Solutions Statistical Relational Learning (SRL) [Getoor, 2002] Probabilistic Inductive Logic Programming (PILP) [Raedt and Kersting, 2004]
4 Problems : High degree of complexity lack and incompleteness of observations deductive reasoning not enough Solution: Exploit Abduction! Abductive statement: given an observation that can not be derived in the theory, make assumptions that explain it All the beans from this bag are white.(bk) These beans (oddly) are white. (observation) These beans are from this bag.(diagnosis) Logic-based approaches multiple sets of assumptions integrity constraints Probabilistic-based approaches multiple explanations with probability (uncertainty)
5 Problems Logic-based too many logical explanations Probabilistic-based independent variables and unstructured data Some solutions Probabilistic Horn Abduction and Bayesian Networks (PHA) [Poole, 1993] Bayesian Abductive Logic Programs: A Probabilistic Logic for Abductive Reasoning (BALP) [Raghavan, 2011] Probabilistic Abduction using Markov Logic Networks (MLN) [Kate and Mooney, 2009] Abduction with stochastic logic programs based on a possible worlds semantics et al., 2006] [Arvanitis Implementing Probabilistic Abductive Logic Programming with Constraint Handling Rules [Christiansen, 2008]
6 Preliminaries: Abductive Logic Programming (ALP) Abductive Logic Program T = P, A, I [Kakas and Mancarella, 1990] P is a standard logic program A (Abducibles) is a set of predicate names IC (Integrity Constraints or domain-specific properties) Problem formulation Given an observation O and a theory T = P, A, I Find an abductive explanation s.t. P = O ( explains O) and P = IC ( is consistent). T abductively entails G (T = A O). Abductive Logic Programming [Kakas and Mancarella, 1990] extends Logic Programming: some predicates (abducibles) incompletely defined deriving hypotheses on these abducible predicates (abductive hypotheses) Goal: observations to be explained
7 Preliminaries: Abductive Logic Programming (ALP) Abductive Logic proof procedure [Kakas and Riguzzi, 2000] Two phases abductive (A) and consistency derivations (B) (A) is the standard Logic derivation extended in order to consider abducibles when an atom δ has to be proved, it is added to the current set of assumptions the addition of δ must not violate any integrity constraint (B) starts to check that all integrity constraints containing δ fails (B) calls (A) to solve each goal Considerations there are constraints that prevent an abduction? constraints verification involves: facts deductively verified true hypotheses evaluating all possible explanations constraints: classical vs typed and crisp vs soft?
8 Probabilistic Abductive Logic Programming (PALP) A new approach using Possible Worlds each time one assumes something he hypothesizes that situation in a specific world each abductive explanation can be seen as a possible world likelihood assessed considering what we have seen and what we should expect to see typed probabilistic constraints: personal belief in the likelihood of whole constraint {nand, or, xor}-constraints Classical vs Probabilistic ALP ALP PALP looks for the minimal explanation handles crisp nand-constraint looks for the most probable explanation handles probabilistic typed constraint Prob, Literals, Type : Prob = [0, 1], Type = {nand, or, xor}, Literals = l 1,..., l n
9 Probabilistic Abductive Logic Programming (PALP) New probabilistic proof procedure Two perspectives: Logical exploits ALP to generate many logical explanations extends ALP to handle typed constraints Probabilistic rank all explanations according to their chance of being true
10 Logical perspective New Logical Proof Procedure extends Abductive and Consistency Derivation: Classical: when an atom δ has to be proved, it is added to the current set of assumption New: when an atom δ has to be proved, two sets of assumptions are considered: one where it holds and another where it does not. extends Consistency Derivation: integrity checking on constraints NAND,OR,XOR NAND satisfied when: at least one condition is false OR satisfied when: at least one condition is true XOR satisfied when: only one condition is true each conclusion is a possible consistent world New Approach Classical + (new rules and backtracking on each choice point)
11 Logical perspective Example (Observation o 1, Query and Possible Explanations) P : {printable(x) a4(x), text(x)} a4(o 1 ) A = {image, text, black_white, printable, table, a4, a5, a3} I = {ic 2, ic 3, ic 4 } ic 2 = 0.9,[table(X), text(x), image(x)], or ic 3 = 0.3,[text(X), color(x)], nand ic 4 = 0.3,[table(X), color(x)], nand table(o 1 ) text(o 1 ) table(o 1 )?- printable(o 1 ) printable(o 1 ) a4(o 1 ), text(o 1 ). image(o 1 ) 1 = {text(o 1 ), table(o 1 )} 2 = {text(o 1 ), table(o 1 ), image(o 1 )}.
12 Probabilistic perspective The chance of being true of a ground literal δ j (1). The unnormalized probability of the abductive explanation (2). P(δ j ) = n(δ j ) J K (1) P n(cons)! ( i,ic i ) = P(δ j ) P(ic k ) (2) (n(cons) a(δ j ))! j=1 k=1 The probability of δ j is equal to 1 P(δ j ). = {P 1 : ( 1, Ic 1 ),..., P T : ( T, Ic T )}, T consistent possible worlds for goal G i = {δ 1,...,δ J }, the ground literals δ j abduced in an abductive proof Ic i = {ic 1,..., ic K } is the set of the constraints involved in i n(δ j ) true groundings of the predicate used in literal δ j n(cons) is total number of constants encountered in the world a(δ j ) is the arity of literal δ j P(ic k ) is the probability of the kth-constraint.
13 Probabilistic perspective Example (Compute explanations probability ) P ( 1,Ic 1 ) = P(text(o 1)) P(table(o 1 )) P(ic 2 ) P(ic 3 ) P(ic 4 ) P ( 1,Ic 1 ) = = Example (Probability assessment of the Abductive Explanations) A = {0.2:image, 0.4:text, 0.1:black_white, 0.6:printable, 0.1:table, 0.9:a4, 0.1:a5, 0.1:a3} P ( 1, Ic 1 ) = P ( 2, Ic 2 ) = P (printable(o 1 )) = max 1 i T P i : ( i, Ic i ) = P ( 2, Ic 2 ) =
14 Improving Classification Exploiting Probabilistic Abductive Reasoning Exploiting our probabilistic abductive logic proof procedure learns the model (i.e. the Abductive Logic Program < P, A, IC >) and the parameters (i.e. literals probabilities) classify never-seen instances Solution: A new system for classification tasks given a Training set and a abducibles set A (possibly empty), it learns: the corresponding theory T by INTHELEX [Esposito et al., 2000] the integrity constraints nand, xor by [Ferilli et al., 2005] given a Test set, tries to cover the example considering both as positive and as negative for the class c < P_max(c, e), p > probabilistic_abductive_proof(problit i, c, e) < P_max( c, e), n > probabilistic_abductive_proof(problit i, c, e) compute the higher between them selects the best classification between all concepts
15 Experimental Settings Goal: assessing the quality of the results in presence of incomplete and noisy data comparing with deductive-reasoning with increasing levels of data corruption Methodology: 10-fold split to obtain < Train, Test > replace each test-set by corrupted versions: removed at random K% of each example (K varying from 10% to 70% with step 10) 5 runs to randomize (35 test-sets for each fold) assume learned constraints true with probability 1.0 (no prev. knowledge) Dataset: Breast-Cancer Congressional Voting Records Tic-Tac-Toe
16 Results and Discussion Breast-Cancer (#Pos = 201; #Neg: 85) Each instance: 9 literals Theory: 30 clauses; 6 lits/clause Learned IC: 1784 nand-constraints (55% -> 4, 35% -> 3 and 10% -> 2); 9 type-domain Congressional Voting Records (#Republicans = 267; #Democrats: 168) Each instance: 16 literals Theory: 35 clauses; 4.5 lits/clause Learned IC: 4173 nand-constraints (16% -> 4, 37% -> 3 and 47% -> 2); 16 type-domain Tic-Tac-Toe (#Pos = 626; #Neg: 332) Each instance: 8 literals Theory: 18 clauses; 4 lits/clause Learned IC: 1863 nand-constraints (99% -> 4, 1% -> 3); 16 type-domain Accuracy Breast Cancer Congress Tic Tac Toe Corruption
17 Results and Discussion Dataset Breast Congress TikTakToe Corr. Abductive Reas. Deductive Reas. Prec. Rec. F 1 Prec. Rec. F 1 0% % % % % % % % % % % % % % % % % % % % % % % %
18 Probabilistic Abductive Logic Approach Reasoning in complex contexts deduction is not enough. Abduction might help it should be logical + probabilistic. Our approach: Abductive Logic Programming generates multiple explanations; Probabilistic assessment of each explanation. Our strategy to classification works correctly in presence of noisy and corruption. Current and Future works Learning the probabilistic constraints. Enriching the probabilistic model of literal distribution. Test our procedure on other tasks such as: NLU and plan recognition.
19 Thanks for attention Questions?
20 References I A. Arvanitis, S. H. Muggleton, J. Chen, and H. Watanabe. Abduction with stochastic logic programs based on a possible worlds semantics. In In Short Paper Proc. of 16th ILP, H. Christiansen. Implementing probabilistic abductive logic programming with constraint handling rules. In T. Schrijvers and T. Frà 1 hwirth, editors, Constraint 4 Handling Rules, volume 5388 of Lecture Notes in Computer Science, pages Springer Berlin Heidelberg, ISBN doi: / _5. URL F. Esposito, G. Semeraro, N. Fanizzi, and S. Ferilli. Multistrategy theory revision: Induction and abduction in inthelex. Machine Learning, 38: , ISSN doi: /A: URL S. Ferilli, T. M. A. Basile, N. Di Mauro, and F. Esposito. Automatic induction of abduction and abstraction theories from observations. In Proc. of the 15th ILP, ILP 05, pages , Berlin, Heidelberg, Springer-Verlag. ISBN , doi: / _7. URL
21 References II L. C. Getoor. Learning statistical models from relational data. PhD thesis, Stanford, CA, USA, AAI A. C. Kakas and P. Mancarella. Generalized stable models: A semantics for abduction. In ECAI, pages , A. C. Kakas and F. Riguzzi. Abductive concept learning. New Generation Comput., 18 (3): , R. J. Kate and R. J. Mooney. Probabilistic abduction using markov logic networks. In Proceedings of the IJCAI-09 Workshop on Plan, Activity, and Intent Recognition (PAIR-09), Pasadena, CA, July URL D. Poole. Probabilistic horn abduction and bayesian networks. Artif. Intell., 64(1): , L. D. Raedt and K. Kersting. Probabilistic inductive logic programming. In ALT, pages 19 36, S. V. Raghavan. Bayesian abductive logic programs: A probabilistic logic for abductive reasoning. In T. Walsh, editor, IJCAI, pages IJCAI/AAAI, ISBN URL
Probabilistic Abductive Logic Programming: a Joint Approach of Logic and Probability
Probabilistic Abductive Logic Programming: a Joint Approach of Logic and Probability F. Rotella 1 and S. Ferilli 1,2 1 Dipartimento di Informatica Università di Bari {fulvio.rotella, stefano.ferilli}@uniba.it
More informationProbabilistic Abductive Logic Programming using Possible Worlds
Probabilistic Abductive Logic Programming using Possible Worlds F. Rotella 1 and S. Ferilli 1,2 1 Dipartimento di Informatica Università di Bari {fulvio.rotella, stefano.ferilli}@uniba.it 2 Centro Interdipartimentale
More informationProbabilistic Abduction using Markov Logic Networks
In Proceedings of the IJCAI-09 Workshop on Plan, Activity, and Intent Recognition (PAIR-09), Pasadena, CA, July 2009. Probabilistic Abduction using Markov Logic Networks Rohit J. Kate Raymond J. Mooney
More informationThe PITA System for Logical-Probabilistic Inference
The System for Logical-Probabilistic Inference Fabrizio Riguzzi 1 and Terrance Swift 2 1 EDIF University of Ferrara, Via Saragat 1, I-44122, Ferrara, Italy fabrizio.riguzzi@unife.it 2 CETRIA Universidade
More informationAbduction in Classification Tasks
Abduction in Classification Tasks Maurizio Atzori, Paolo Mancarella, and Franco Turini Dipartimento di Informatica University of Pisa, Italy {atzori,paolo,turini}@di.unipi.it Abstract. The aim of this
More informationProbabilistic Inductive Constraint Logic
Probabilistic Inductive Constraint Logic Fabrizio Riguzzi 1, Elena Bellodi 2, Riccardo Zese 2, Giuseppe Cota 2, and Evelina Lamma 2 1 Dipartimento di Matematica e Informatica University of Ferrara Via
More informationToward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming
Toward Computing Conflict-Based Diagnoses in Probabilistic Logic Programming Arjen Hommersom 1,2 and Marcos L.P. Bueno 2 1 Open University of the Netherlands 2 Radboud University Nijmegen, The Netherlands
More informationDisjunctive Bottom Set and Its Computation
Disjunctive Bottom Set and Its Computation Wenjin Lu and Ross King Department of Computer Science, University of Wales, Aberystwyth Ceredigion, SY23 3DB, Wales, UK e-mail:{wwl, rdk}@aber.ac.uk Abstract
More informationSemantics and Inference for Probabilistic Ontologies
Semantics and Inference for Probabilistic Ontologies Fabrizio Riguzzi, Elena Bellodi, Evelina Lamma, and Riccardo Zese ENDIF University of Ferrara, Italy, email: {fabrizio.riguzzi, elena.bellodi, evelina.lamma}@unife.it,
More informationSymmetry Breaking for Relational Weighted Model Finding
Symmetry Breaking for Relational Weighted Model Finding Tim Kopp Parag Singla Henry Kautz WL4AI July 27, 2015 Outline Weighted logics. Symmetry-breaking in SAT. Symmetry-breaking for weighted logics. Weighted
More informationAn Algorithm for Learning with Probabilistic Description Logics
An Algorithm for Learning with Probabilistic Description Logics José Eduardo Ochoa Luna and Fabio Gagliardi Cozman Escola Politécnica, Universidade de São Paulo, Av. Prof. Mello Morais 2231, São Paulo
More informationLogical Aspects of Several Bottom-Up Fittings
Logical Aspects of Several Bottom-Up Fittings Akihiro Yamamoto Division of Electronics and Information Engineering and Meme Media Laboratory Hokkaido University N 13 W 8, Kita-ku Sapporo 060-8628 JAPAN
More informationA First-Order Stochastic Modeling Language for Diagnosis
A First-Order Stochastic Modeling Language for Diagnosis Chayan Chakrabarti, Roshan Rammohan and George F. Luger Department of Computer Science University of New Mexico Albuquerque, NM 87131 {cc, roshan,
More informationReasoning by Analogy Using Past Experiences
Reasoning by Analogy Using Past Experiences F. Leuzzi 1 and S. Ferilli 1,2 1 Dipartimento di Informatica Università di Bari {fabio.leuzzi, stefano.ferilli}@uniba.it 2 Centro Interdipartimentale per la
More informationA Simple Model for Sequences of Relational State Descriptions
A Simple Model for Sequences of Relational State Descriptions Ingo Thon, Niels Landwehr, and Luc De Raedt Department of Computer Science, Katholieke Universiteit Leuven, Celestijnenlaan 200A, 3001 Heverlee,
More informationFinite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier
Finite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier Kaizhu Huang, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More informationMAI0203 Lecture 7: Inference and Predicate Calculus
MAI0203 Lecture 7: Inference and Predicate Calculus Methods of Artificial Intelligence WS 2002/2003 Part II: Inference and Knowledge Representation II.7 Inference and Predicate Calculus MAI0203 Lecture
More informationAbductive Disjunctive Logic Programming. Ullrich Hustadt
Abductive Disjunctive Logic Programming Ullrich Hustadt Max-Planck-Institut fur Informatik, Im Stadtwald, W-6600 Saarbrucken, Germany Phone: (+9 681) 302-531, Fax: (+9 681) 302-530, E-mail: Ullrich.Hustadt@mpi-sb.mpg.de
More informationIncremental Construction of Complex Aggregates: Counting over a Secondary Table
Incremental Construction of Complex Aggregates: Counting over a Secondary Table Clément Charnay 1, Nicolas Lachiche 1, and Agnès Braud 1 ICube, Université de Strasbourg, CNRS 300 Bd Sébastien Brant - CS
More informationComputing the acceptability semantics. London SW7 2BZ, UK, Nicosia P.O. Box 537, Cyprus,
Computing the acceptability semantics Francesca Toni 1 and Antonios C. Kakas 2 1 Department of Computing, Imperial College, 180 Queen's Gate, London SW7 2BZ, UK, ft@doc.ic.ac.uk 2 Department of Computer
More informationIntelligent Agents. First Order Logic. Ute Schmid. Cognitive Systems, Applied Computer Science, Bamberg University. last change: 19.
Intelligent Agents First Order Logic Ute Schmid Cognitive Systems, Applied Computer Science, Bamberg University last change: 19. Mai 2015 U. Schmid (CogSys) Intelligent Agents last change: 19. Mai 2015
More informationA Partition-Based First-Order Probabilistic Logic to Represent Interactive Beliefs
A Partition-Based First-Order Probabilistic Logic to Represent Interactive Beliefs Alessandro Panella and Piotr Gmytrasiewicz Fifth International Conference on Scalable Uncertainty Management Dayton, OH
More informationIntroduction to Artificial Intelligence Propositional Logic & SAT Solving. UIUC CS 440 / ECE 448 Professor: Eyal Amir Spring Semester 2010
Introduction to Artificial Intelligence Propositional Logic & SAT Solving UIUC CS 440 / ECE 448 Professor: Eyal Amir Spring Semester 2010 Today Representation in Propositional Logic Semantics & Deduction
More informationBayesian Identity Clustering
Bayesian Identity Clustering Simon JD Prince Department of Computer Science University College London James Elder Centre for Vision Research York University http://pvlcsuclacuk sprince@csuclacuk The problem
More informationMarkov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationConcept Learning.
. Machine Learning Concept Learning Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg Martin.Riedmiller@uos.de
More informationParts 3-6 are EXAMPLES for cse634
1 Parts 3-6 are EXAMPLES for cse634 FINAL TEST CSE 352 ARTIFICIAL INTELLIGENCE Fall 2008 There are 6 pages in this exam. Please make sure you have all of them INTRODUCTION Philosophical AI Questions Q1.
More informationA Revised Comparison of Bayesian Logic Programs and Stochastic Logic Programs
A Revised Comparison of Bayesian Logic Programs and Stochastic Logic Programs Jianzhong Chen and Stephen Muggleton Department of Computing, Imperial College London, 180 Queen s Gate, South Kensington,
More informationFoundations of Artificial Intelligence
Foundations of Artificial Intelligence 31. Propositional Logic: DPLL Algorithm Malte Helmert and Gabriele Röger University of Basel April 24, 2017 Propositional Logic: Overview Chapter overview: propositional
More informationPropositions. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 5.1, Page 1
Propositions An interpretation is an assignment of values to all variables. A model is an interpretation that satisfies the constraints. Often we don t want to just find a model, but want to know what
More informationRelational Information Gain
Relational Information Gain Marco Lippi 1, Manfred Jaeger 2, Paolo Frasconi 1, and Andrea Passerini 3 1 Dipartimento di Sistemi e Informatica, Università degli Studi di Firenze, Italy 2 Department for
More information1. Courses are either tough or boring. 2. Not all courses are boring. 3. Therefore there are tough courses. (Cx, Tx, Bx, )
Logic FOL Syntax FOL Rules (Copi) 1. Courses are either tough or boring. 2. Not all courses are boring. 3. Therefore there are tough courses. (Cx, Tx, Bx, ) Dealing with Time Translate into first-order
More informationKarl-Rudolf Koch Introduction to Bayesian Statistics Second Edition
Karl-Rudolf Koch Introduction to Bayesian Statistics Second Edition Karl-Rudolf Koch Introduction to Bayesian Statistics Second, updated and enlarged Edition With 17 Figures Professor Dr.-Ing., Dr.-Ing.
More informationMCINTYRE: A Monte Carlo Algorithm for Probabilistic Logic Programming
: A Monte Carlo Algorithm for Probabilistic Logic Programming Fabrizio Riguzzi ENDIF Università di Ferrara Via Saragat, 1 44122 Ferrara, Italy. {fabrizio.riguzzi}@unife.it Abstract. Probabilistic Logic
More informationSkeptical Abduction: A Connectionist Network
Skeptical Abduction: A Connectionist Network Technische Universität Dresden Luis Palacios Medinacelli Supervisors: Prof. Steffen Hölldobler Emmanuelle-Anna Dietz 1 To my family. Abstract A key research
More informationReduction of abductive logic programs. In this paper we study a form of abductive logic programming which combines
Reduction of abductive logic programs to normal logic programs Francesca Toni, Robert A. Kowalski Department of Computing, Imperial College 18 Queen's Gate, London SW7 2BZ, UK fft, rakg@doc.ic.ac.uk Abstract
More informationarxiv: v1 [cs.lo] 18 Sep 2018
On a Convex Logic Fragment for Learning and Reasoning Francesco Giannini, Michelangelo Diligenti, Marco Gori, and Marco Maggini arxiv:1809.06778v1 [cs.lo] 18 Sep 2018 Department of Information Engineering
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationLearning Probabilistic Logic Models from Probabilistic Examples
Learning Probabilistic Logic Models from Probabilistic Examples Jianzhong Chen 1, Stephen Muggleton 1, and José Santos 2 1 Department of Computing, Imperial College London, London SW7 2AZ, UK {cjz, shm}@doc.ic.ac.uk
More informationPropositional Logic: Logical Agents (Part I)
Propositional Logic: Logical Agents (Part I) First Lecture Today (Tue 21 Jun) Read Chapters 1 and 2 Second Lecture Today (Tue 21 Jun) Read Chapter 7.1-7.4 Next Lecture (Thu 23 Jun) Read Chapters 7.5 (optional:
More informationUnifying Logical and Statistical AI
Unifying Logical and Statistical AI Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with Stanley Kok, Hoifung Poon, Matt Richardson and Parag Singla Overview Motivation
More informationTopLog: ILP Using a Logic Program Declarative Bias
TopLog: ILP Using a Logic Program Declarative Bias Stephen H. Muggleton, José C. A. Santos, and Alireza Tamaddoni-Nezhad Department of Computing, Imperial College, London {shm,jcs06,atn}@doc.ic.ac.uk Abstract.
More informationFuzzy Systems. Introduction
Fuzzy Systems Introduction Prof. Dr. Rudolf Kruse Christoph Doell {kruse,doell}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge Processing
More informationBayesian Reasoning. Adapted from slides by Tim Finin and Marie desjardins.
Bayesian Reasoning Adapted from slides by Tim Finin and Marie desjardins. 1 Outline Probability theory Bayesian inference From the joint distribution Using independence/factoring From sources of evidence
More informationLogic Programming Techniques for Reasoning with Probabilistic Ontologies
Logic Programming Techniques for Reasoning with Probabilistic Ontologies Riccardo Zese, Elena Bellodi, Evelina Lamma and Fabrizio Riguzzi University of Ferrara, Italy riccardo.zese@unife.it Zese, Bellodi,
More informationMarkov Logic Networks for Spoken Language Interpretation
Intelligent Information Systems 9999 ISBN 666-666-666, pages 1 10 Markov Logic Networks for Spoken Language Interpretation Marie-Jean Meurs, Frédéric Duvert, Fabrice Lefèvre, and Renato De Mori Université
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationEncoding formulas with partially constrained weights in a possibilistic-like many-sorted propositional logic
Encoding formulas with partially constrained weights in a possibilistic-like many-sorted propositional logic Salem Benferhat CRIL-CNRS, Université d Artois rue Jean Souvraz 62307 Lens Cedex France benferhat@criluniv-artoisfr
More informationUnifying Version Space Representations: Part II
Unifying Version Space Representations: Part II E.N. Smirnov, I.G. Sprinkhuizen-Kuyper, and H.J. van den Herik IKAT, Department of Computer Science, Maastricht University, P.O.Box 616, 6200 MD Maastricht,
More informationWarm-Up Problem. Is the following true or false? 1/35
Warm-Up Problem Is the following true or false? 1/35 Propositional Logic: Resolution Carmen Bruni Lecture 6 Based on work by J Buss, A Gao, L Kari, A Lubiw, B Bonakdarpour, D Maftuleac, C Roberts, R Trefler,
More informationIntroduction to Bayesian Learning
Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline
More informationVersion Spaces.
. Machine Learning Version Spaces Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg riedmiller@informatik.uni-freiburg.de
More informationCombining logic with probability Motivation
Combining logic with probability Motivation is a powerful language to represent complex relational information Probability is the standard way to represent uncertainty in knowledge Combining the two would
More informationDeep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści
Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, 2017 Spis treści Website Acknowledgments Notation xiii xv xix 1 Introduction 1 1.1 Who Should Read This Book?
More informationPropositional Logic Language
Propositional Logic Language A logic consists of: an alphabet A, a language L, i.e., a set of formulas, and a binary relation = between a set of formulas and a formula. An alphabet A consists of a finite
More informationBayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction
15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive
More informationPei Wang( 王培 ) Temple University, Philadelphia, USA
Pei Wang( 王培 ) Temple University, Philadelphia, USA Artificial General Intelligence (AGI): a small research community in AI that believes Intelligence is a general-purpose capability Intelligence should
More informationRelationship between Loss Functions and Confirmation Measures
Relationship between Loss Functions and Confirmation Measures Krzysztof Dembczyński 1 and Salvatore Greco 2 and Wojciech Kotłowski 1 and Roman Słowiński 1,3 1 Institute of Computing Science, Poznań University
More informationConnectionist Artificial Neural Networks
Imperial College London Department of Computing Connectionist Artificial Neural Networks Master s Thesis by Mathieu Guillame-Bert Supervised by Dr. Krysia Broda Submitted in partial fulfillment of the
More informationMarkov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationDesign of Logic-based Intelligent Systems Lecture 1: Introduction Part 1. K. Truemper University of Texas at Dallas
Design of Logic-based Intelligent Systems Lecture 1: Introduction Part 1 K. Truemper University of Texas at Dallas Acknowledgements 2 Acknowledgements (Names in Alphabetical Order) Organizations: Alexander
More informationA Neural Network Approach for First-Order Abductive Inference
A Neural Network Approach for First-Order Abductive Inference Oliver Ray and Bruno Golénia Department of Computer Science University of Bristol Bristol, BS8 1UB, UK {oray,goleniab}@cs.bris.ac.uk Abstract
More informationCMPT 310 Artificial Intelligence Survey. Simon Fraser University Summer Instructor: Oliver Schulte
CMPT 310 Artificial Intelligence Survey Simon Fraser University Summer 2017 Instructor: Oliver Schulte Assignment 3: Chapters 13, 14, 18, 20. Probabilistic Reasoning and Learning Instructions: The university
More informationModeling User s Cognitive Dynamics in Information Access and Retrieval using Quantum Probability (ESR-6)
Modeling User s Cognitive Dynamics in Information Access and Retrieval using Quantum Probability (ESR-6) SAGAR UPRETY THE OPEN UNIVERSITY, UK QUARTZ Mid-term Review Meeting, Padova, Italy, 07/12/2018 Self-Introduction
More informationInducing Probabilistic Relational Rules from Probabilistic Examples
Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 25) Inducing Probabilistic Relational Rules from Probabilistic Examples Luc De Raedt, Anton Dries, Ingo
More informationTopics in Model-Based Reasoning
Towards Integration of Proving and Solving Dipartimento di Informatica Università degli Studi di Verona Verona, Italy March, 2014 Automated reasoning Artificial Intelligence Automated Reasoning Computational
More informationProbabilistic Explanation Based Learning
Probabilistic Explanation Based Learning Angelika Kimmig 1, Luc De Raedt 1, and Hannu Toivonen 2 1 Department of Computer Science, Katholieke Universiteit Leuven 2 Department of Computer Science, University
More informationSplitting a Default Theory. Hudson Turner. University of Texas at Austin.
Splitting a Default Theory Hudson Turner Department of Computer Sciences University of Texas at Austin Austin, TX 7872-88, USA hudson@cs.utexas.edu Abstract This paper presents mathematical results that
More informationOctober 23, :49 WSPC/INSTRUCTION FILE ijait06 revised THE DESIGN AND TESTING OF A FIRST-ORDER LOGIC-BASED STOCHASTIC MODELING LANGUAGE
International Journal on Artificial Intelligence Tools c World Scientific Publishing Company THE DESIGN AND TESTING OF A FIRST-ORDER LOGIC-BASED STOCHASTIC MODELING LANGUAGE DANIEL J. PLESS Sandia National
More informationIndex. Cambridge University Press Relational Knowledge Discovery M E Müller. Index. More information
s/r. See quotient, 93 R, 122 [x] R. See class, equivalence [[P Q]]s, 142 =, 173, 164 A α, 162, 178, 179 =, 163, 193 σ RES, 166, 22, 174 Ɣ, 178, 179, 175, 176, 179 i, 191, 172, 21, 26, 29 χ R. See rough
More informationProbabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning
Probabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning Guy Van den Broeck KULeuven Symposium Dec 12, 2018 Outline Learning Adding knowledge to deep learning Logistic circuits
More informationSemantics and Complexity of Abduction from Default Theories
Semantics and Complexity of Abduction from Default Theories (Extended abstract*) Thomas Eiter, Georg Gottlob Christian Doppler Lab for Expert Systems Technical University of Vienna Paniglgasse 16, A-1040
More informationGaussian Mixture Model Uncertainty Learning (GMMUL) Version 1.0 User Guide
Gaussian Mixture Model Uncertainty Learning (GMMUL) Version 1. User Guide Alexey Ozerov 1, Mathieu Lagrange and Emmanuel Vincent 1 1 INRIA, Centre de Rennes - Bretagne Atlantique Campus de Beaulieu, 3
More informationLogic in AI Chapter 7. Mausam (Based on slides of Dan Weld, Stuart Russell, Subbarao Kambhampati, Dieter Fox, Henry Kautz )
Logic in AI Chapter 7 Mausam (Based on slides of Dan Weld, Stuart Russell, Subbarao Kambhampati, Dieter Fox, Henry Kautz ) 2 Knowledge Representation represent knowledge about the world in a manner that
More informationReasoning Under Uncertainty: Introduction to Probability
Reasoning Under Uncertainty: Introduction to Probability CPSC 322 Uncertainty 1 Textbook 6.1 Reasoning Under Uncertainty: Introduction to Probability CPSC 322 Uncertainty 1, Slide 1 Lecture Overview 1
More informationReasoning in Uncertain Situations
9 Reasoning in Uncertain Situations 9.0 Introduction 9.1 Logic-Based Abductive Inference 9.2 Abduction: Alternatives to Logic 9.3 The Stochastic Approach to Uncertainty 9.4 Epilogue and References 9.5
More information1. True/False (40 points) total
Name: 1 2 3 4 5 6 7 total 40 20 45 45 20 15 15 200 UMBC CMSC 671 Final Exam December 20, 2009 Please write all of your answers on this exam. The exam is closed book and has seven problems that add up to
More informationLecture Notes in Machine Learning Chapter 4: Version space learning
Lecture Notes in Machine Learning Chapter 4: Version space learning Zdravko Markov February 17, 2004 Let us consider an example. We shall use an attribute-value language for both the examples and the hypotheses
More informationReasoning in the Description Logic BEL using Bayesian Networks
Reasoning in the Description Logic BEL using Bayesian Networks İsmail İlkan Ceylan Theoretical Computer Science TU Dresden, Germany ceylan@tcs.inf.tu-dresden.de Rafael Peñaloza Theoretical Computer Science,
More informationBayesian network modeling. 1
Bayesian network modeling http://springuniversity.bc3research.org/ 1 Probabilistic vs. deterministic modeling approaches Probabilistic Explanatory power (e.g., r 2 ) Explanation why Based on inductive
More informationLecture 4: Perceptrons and Multilayer Perceptrons
Lecture 4: Perceptrons and Multilayer Perceptrons Cognitive Systems II - Machine Learning SS 2005 Part I: Basic Approaches of Concept Learning Perceptrons, Artificial Neuronal Networks Lecture 4: Perceptrons
More informationLifted Discriminative Learning of Probabilistic Logic Programs
Lifted Discriminative Learning of Probabilistic Logic Programs Arnaud Nguembang Fadja 1 and Fabrizio Riguzzi 2 1 Dipartimento di Ingegneria University of Ferrara Via Saragat 1, I-44122, Ferrara, Italy
More informationProf. Dr. Ralf Möller Dr. Özgür L. Özçep Universität zu Lübeck Institut für Informationssysteme. Tanya Braun (Exercises)
Prof. Dr. Ralf Möller Dr. Özgür L. Özçep Universität zu Lübeck Institut für Informationssysteme Tanya Braun (Exercises) Slides taken from the presentation (subset only) Learning Statistical Models From
More informationLearning Gaussian Process Models from Uncertain Data
Learning Gaussian Process Models from Uncertain Data Patrick Dallaire, Camille Besse, and Brahim Chaib-draa DAMAS Laboratory, Computer Science & Software Engineering Department, Laval University, Canada
More informationA Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems
A Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems Daniel Meyer-Delius 1, Christian Plagemann 1, Georg von Wichert 2, Wendelin Feiten 2, Gisbert Lawitzky 2, and
More informationKnowledge Extraction from DBNs for Images
Knowledge Extraction from DBNs for Images Son N. Tran and Artur d Avila Garcez Department of Computer Science City University London Contents 1 Introduction 2 Knowledge Extraction from DBNs 3 Experimental
More informationArtificial Intelligence
Artificial Intelligence Propositional Logic Marc Toussaint University of Stuttgart Winter 2016/17 (slides based on Stuart Russell s AI course) Motivation: Most students will have learnt about propositional
More informationData Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts
Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification
More informationClassification Based on Logical Concept Analysis
Classification Based on Logical Concept Analysis Yan Zhao and Yiyu Yao Department of Computer Science, University of Regina, Regina, Saskatchewan, Canada S4S 0A2 E-mail: {yanzhao, yyao}@cs.uregina.ca Abstract.
More informationInverse Resolution as Belief Change
Inverse Resolution as Belief Change Maurice Pagnucco ARC Centre of Excel. for Autonomous Sys. School of Comp. Science and Engineering The University of New South Wales Sydney, NSW, 2052, Australia. Email:
More informationTowards Maximum Geometric Margin Minimum Error Classification
THE SCIENCE AND ENGINEERING REVIEW OF DOSHISHA UNIVERSITY, VOL. 50, NO. 3 October 2009 Towards Maximum Geometric Margin Minimum Error Classification Kouta YAMADA*, Shigeru KATAGIRI*, Erik MCDERMOTT**,
More informationThe Perceptron. Volker Tresp Summer 2014
The Perceptron Volker Tresp Summer 2014 1 Introduction One of the first serious learning machines Most important elements in learning tasks Collection and preprocessing of training data Definition of a
More informationHYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH
HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi
More informationExtracting Provably Correct Rules from Artificial Neural Networks
Extracting Provably Correct Rules from Artificial Neural Networks Sebastian B. Thrun University of Bonn Dept. of Computer Science III Römerstr. 64, D-53 Bonn, Germany E-mail: thrun@cs.uni-bonn.de thrun@cmu.edu
More informationarxiv: v2 [cs.ai] 1 May 2018
Under consideration for publication in Theory and Practice of Logic Programming 1 Incremental and Iterative Learning of Answer Set Programs from Mutually Distinct Examples arxiv:1802.07966v2 [cs.ai] 1
More informationDeclarative Merging of and Reasoning about Decision Diagrams
Declarative Merging of and Reasoning about Decision Diagrams Thomas Eiter Thomas Krennwallner Christoph Redl {eiter,tkren,redl}@kr.tuwien.ac.at September 12, 2011 Eiter T., Krennwallner T., Redl C. (TU
More informationData-Driven Logical Reasoning
Data-Driven Logical Reasoning Claudia d Amato Volha Bryl, Luciano Serafini November 11, 2012 8 th International Workshop on Uncertainty Reasoning for the Semantic Web 11 th ISWC, Boston (MA), USA. Heterogeneous
More informationFrom Constructibility and Absoluteness to Computability and Domain Independence
From Constructibility and Absoluteness to Computability and Domain Independence Arnon Avron School of Computer Science Tel Aviv University, Tel Aviv 69978, Israel aa@math.tau.ac.il Abstract. Gödel s main
More informationProbabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning
Probabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning Guy Van den Broeck HRL/ACTIONS @ KR Oct 28, 2018 Foundation: Logical Circuit Languages Negation Normal Form Circuits Δ
More informationA Schema for Generating Update Semantics
A Schema for Generating Update Semantics José Luis Carballido Carranza Benemérita Universidad Autónoma de Puebla jlcarballido7@gmail.com Abstract In this paper, we present a general schema for defining
More information