ANALYSIS OF BEHAVIOR OF CHEMICAL MOLECULES: RULE FORMATION ON NON-HOMOGENEOUS CLASSES OF OBJECTS

Size: px
Start display at page:

Download "ANALYSIS OF BEHAVIOR OF CHEMICAL MOLECULES: RULE FORMATION ON NON-HOMOGENEOUS CLASSES OF OBJECTS"

Transcription

1 ANALYSIS OF BEHAVIOR OF CHEMICAL MOLECULES: RULE FORMATION ON NON-HOMOGENEOUS CLASSES OF OBJECTS Session 5 Applications and Implications of Artificial Intelligence Bruce G. Buchanan N.S. Sridharan Stanford University Stanford, California ABSTRACT An information processing model of some important aspects of inductive reasoning is presented within the context of one scientific discipline. Given a collection of experimental (mass spectrometry) data from several chemical molecules the computer program described here separates the molecules into "wellbehaved" subclasses and selects from the space of all explanatory processes the "characteristic" processes for each subclass. The definitions of "well-behaved" and "characteristic" embody several heuristics which are discussed. Some results of the program are discussed which have been useful to chemists and which lend credibility to this approach. INTRODUCTION Induction in science has been understood to encompass many different levels of tasks, from theory construction as performed by Einstein to everyday non-deductive inferences as made by scientists looking for explanations of routine data. For the most part, it is not well defined however one understands it (a notable exception being statistical inference). Although general statements can be made about non-deductive inference, it is unlikely that there exists one general "inductive method" that guides scientific inference at all levels. Nor does it seem likely that a method of scientific inference at any one level can succeed without recourse to task-specific information, that is, information specific to the particular science. Within these assumptions we are exploring an information processing model of scientific inference in one discipline- A unifying theme in our explorations is that induction is efficient selection from the domain of all possible answers. Previous papers on the Heuristic DENDRAL Program have advanced this theme with respect to hypothesis formation in routine scientific work. Recently, we have been exploring this theme with respect to the higher-order task of finding general rules to explain large collections of data. This paper extends the previous work to the task of finding" rules for subclasses of objects, given empirical data for the objects but without prior knowledge of the number of subclasses or the features that characterize them. THE TASK AREA 2 For reasons discussed previously, the task area is mass spectrometry, a branch of organic chemistry. The rule formation task is to find rules that characterize the behavior of classes of molecules in the mass spectrometer, given the mass spectrometric data from several known molecules. The chemical structure of each molecule is known. The data for each molecule are a) the masses of various molecular fragments produced from the electron bombardment of the molecule in the instrument and b) the relative abundances of fragments at each mass. The data for each molecule are arranged in a fragmentmass table {FMT), or mass spectrum. Typically, there are data points in one FMT. The task is to characterize the experimental behavior of the whole class of molecules. Rules which characterize the behavior of the molecules are represented as conditional sentences in our system. The antecedent of a simple conditional rule is a predicate which is true or false of a molecule (or class of molecules); the consequent is a description of a mass spectrometric action (henceforth "process") which is thought to occur when that molecule is in the experimental context. We have termed these rules "situation-action rules" (or "S-A rules"). The rule syntax has been described previously 3 and is not critical to an understanding of the present paper. An example of a rule, rewritten in English, is: "IF the graph of the molecule contains the estrogen skeleton, THEN break the bonds between nodes labeled and " This process (the consequent of this rule) is named BRK10L in Table T. The graph of the estrogen skeleton mentioned in the antecedent is shown with the conventional node numbering in Figure 3. The rules will be used in the Heuristic DENDRAL performance program to determine the structure of compounds, reasoning from the mass spectrometric data of each. They are also of use to chemists interested in extending the theory of mass spectrometry. OVERVIEW OF METHOD The rule formation program contains three major sub-programs, which are described below under the headings Data Interpretation, Process Selection, and Molecule Selection. The control structure for the overall program is described after the discussions of the three major sub-programs. A brief overview of the whole program will be given first, however, in order to set the context. The purpose of the program is to find the characteristic processes which determine separable subclasses of molecules given the experimental data and molecular structure of each molecule. The overall flow of the program, as described below, is shown in Figure 1. The three major steps are to reinterpret the experimental data as molecular processes, find the characteristic processes for the given molecules, and select the set of molecules that are "well behaved" with regard to the characteristic processes. The reinterpretation of the data is done once for each molecule in the whole set, and the results are summarized once. The second and third sub-programs are called successively until they isolate a well-behaved subclass of molecules and determine the processes which characterize their behavior. The monitor then subtracts the well-behaved subclass from the starting class of molecules, and repeats the successive calls to the second and third subprograms. The whole program stops when there are N or fewer molecules not yet in some well-behaved subclass. (For now, N=3.) The data interpretation program has been described previously with some aspects of the process selection program 3. The molecule selection program and class refinement loop in the control sequence are new additions. 67

2 DATA INTERPRETATION As mentioned above, the purpose of the data interpretation and summary program (INTSUM) is to reinterpret the experimentally determined data, the FMT, for each molecule and summarize the results. Because the program has been described previously details will be omitted here. It should be noted that the successful apolication of this program to a subclass of estrogens has already been reported in the chemical literature. The INTSUM program is general in that it will work on FMT's for any class of molecules with a common skeletal graph and it is flexible in that the knowledge used by the program is easily changed and there are numerous optionr, controlling the operation of the program. The INTSUM program is called with the initial set of molecules and their FMT's, It is also Riven the graph structure of the skeleton common to all molecules in the initial set. The first step is to search the space of all possible processes which could explain data points in the FMT of any molecule with the given skeleton. The space of explanatory processes is combinatorial; simple processes that cut the graph into two fragments are generated first, followed by pairs of simple processes, triples, and so on. The heuristics listed below constrain the search: Simplicity (Occam's Razor), If two or more processes explain the same data point, prefer the simpler one, i.e., the process involving fewer simple steps. Chemical Constraints. (a) Break no more than NB bonds in any process, whether simple or multi-step (NB=5 in our current version); (b) Do not allow any process to break two bonds to the same carbon atom; (c) Do not allow a fragment to contain fewer than NA atoms (NA=5 currently); (d) Do not allow any process to contain more than NP simple processes (NP=2 currently); (e) Break only single bonds (no double or triple bonds). The heuristic search produces a list of plausible processes without reference to the data. The second step of the INTSUM program is to determine for each process and each FMT whether there is evidence for the process in the FMT. If so, then that process can explain the data point and the strength of the evidence is saved. The final step is to summarize for each process and all molecules the freauency, total strength of evidence and number of alternative explanations. (Freauency for a given process is the percentage of all molecules that have evidence for the process.) These statistics are passed to the process selection program. PROCESS SELECTION The process selection program chooses the most characteristic processes for the given class of molecules from the list of a. priori plausible processes that are output by the INTSUM program. It assumes that the molecules given to it are all in one well-behaved class. Thus, it can merely filter the list of processes to find those which satisfy the criteria for characteristic processes. A process mentioned in a rule statement must satisfy several criteria in order to be counted as a characteristic process for the molecules under consideration. The INTSUM program provides a summary of statistics for the plausible processes it has chosen from the space of all processes. The process selection program applies heuristic criteria to sort out the most likely processes and to distinguish among alternative explanations, when alternatives remain. It uses the information from the data for filtering, in contrast to the a priori filtering in the INTSUM program. For example, an a priori simplicity criterion filters out processes that break too many bonds. The criteria for "most likely processes" frequency, strength of evidence, and degree of uniqueness are discussed below. To a large extent the choice of these criteria and particularly the choice of parameter settings are arbitrary. However, the following discussion provides some rationale for our choices. Frequency. If nature presented clear and unambiguous data to us we could expect all and only characteristic processes for a class of molecules to occur 100% of the time. This is what we would like to mean by 'characteristic' process. However, the data contain noise and, more importantly, we are forced to interpret the data in terms of processes that we construct. Thus, in the literature one finds discussions of exceptions to rules together with presentation of the rules. A low frequency threshold (60%) is used as a criterion for plausible process instead of a high one because the marginal processes which are included at one step can be excluded at a later refinement step if they prove to be uncharacteristic of a class of molecules. Strength of Evidence. The program considers the strength of evidence found for each process, besides the frequency of molecules that show the process. Associated with each fragment mass in the experimental data is a measure of the percent of total Jons (or ion current) contributed by fragments of that mass. (The evidence from mass spectrometry is not merely binary, i.e., yes/no, although we have considered it that way in the past.) The total ion current for any molecule can be visualized as the sum of all y-values in a bar graph in which the x-values represent fragment masses. The strength of evidence for a process, then, is the percent of the total of all ion currents (for all molecules) that can be explained by the process. The present value of this parameter is 0.005, i.e., 0.5% of the data must be explained by any process that will be said to be characteristic of the given molecules. There may be much information in the weaker data points, but until we can interpret the strong signals, we do not want to start looking critically at the weak ones. This is why we have a strength of evidence threshold (although in our trials we have kept it fairly low). Degree of Uniqueness. The program will discard processes that cannot uniquely explain at least n data points for each molecule. The rationale behind this criterion is that processes that are always (or often) redundant with other processes have no explanatory power of their own. In spite of the intuitive appeal of this criterion, it was not used for the trials reported here in which molecule selection is coupled with process selection. For process selection alone, it is a useful filter. These three criteria filter the processes to provide the characteristic processes for the molecules given to the program. However, the processes may still overlap in the data points that they explain. If two (or more) processes are ambiguous, i.e., they explain most of the same data points, the program tries to resolve the ambiguity in favor of a single explanation. This is not easy, for the competing explanations have all passed the tests for "most likely processes" Just discussed. Thus, they all appear good enough to be 68

3 rules on their own. The resolution of ambiguities among processes is made according to relative values of the criteria used to Judge them likely in the first place. That is, the values of frequency, strength of evidence and degree of uniqueness are compared - in any order - to determine which process is preferred, if any. MOLECULE SELECTION Molecule selection, by itself, is a simple program whose purpose is to find a subclass of molecules that are "well-behaved" with respect to a set of processes. Its inputs are (a) a class of molecules and (b) a set of processes that are characteristic of those molecules {output of the process selection program just described). The processes that are chosen as roughly characteristic of a class of molecules are used by the molecule selection program to refine the extension of the class. Several processes will each have a few exceptions - the number permitted depending on the frequency threshold used by the program. But if the same molecules appear as exceptions over and over again (for several processes) then they probably do not belong in the same subclass with the molecules whose behavior is characterized by those processes. A molecule is said to be well-behaved with respect to a set of processes (or well-behaved) if it shows evidence for at least MP of the processes. The current value of MP is 85% of the number of processes in the set. Currently this is the only criterion used to identify members of the subclass, although other features of the molecules could also be used for clustering. For example, the structural features of chemical molecules could also help classify molecules which "belong" together. The reason descriptive features such as these are not used during molecule selection is that they constitute a good check (by chemists) on the adequacy of the results of the molecule separation procedure. CONTROL STRUCTURE OF THE RULE FORMATION MONITOR The overall flow of control has been briefly described and diagrammed in Figure 1, and the three major components of the whole program have been discussed. The interaction between process selection and molecule selection is the last important detail in the description of the program. It is shown schematically in Figure 2 and selected portions of intermediate output are shown in Table II. After the INTSUM program interprets and summarizes the data for a set of molecules, the process selection program is asked to find a set of processes that characterize those molecules. However, process selection starts with the assumption that the molecules should be characterized all together, i.e., that the molecules are homogeneous, or belong in one class with respect to mass spectrometry. The purpose of the rule formation monitor, and the molecule selection program in particular, is to remove the necessity of working within this assumption. Because a class of molecules has a common skeleton, there is reason to believe that they are homogeneous (with respect to mass spectrometry processes). But this is not necessarily true. Many of the molecules whose structures contain the graph common to estrogens (e.g., the equilenins discussed with Table II in the Results section) fail to exhibit behavior that is characteristic of most estrogens in the mass spectrometer. The monitor begins with the Null Hypothesis that the initial set M of molecules is homogeneous with respect to all the relevant processes given as input. With the process selection program it finds plausible processes that roughly characterize the whole class of molecules. It attempts to confirm the hypothesis by finding the subclass S of molecules that are wellbehaved for those processes. If this subclass S is the same as the initial set M, then the assumption of homogeneity is taken to be true. In that case, there is no proper subset to be separated. When the subclass S is different from the starting class M, however, the program loops back to process selection as shown in Figure 2. This figure shows the procedure for producing one homogeneous subclass of molecules (and the characteristic processes for the subclass); this procedure, rule formation, is itself used repeatedly in the main program as shown in Figure 1. The inputs to the rule formation procedure are (a) the set RP of relevant processes and statistics for them, viz., the output of INTSUM, and (b) a class M' of molecules, where M' is initially the same as the entire class of molecules, M, given to INTSUM. M' is used to keep track of the best refinement of M so far. The process selection program selects a set of processes P from RP in the manner described above. P characterizes the class M' s insofar as M' can be characterized at all. The criteria for characteristic process can be made more restrictive if the class is known to be homogeneous (e.g., frequency >95%). In this case, however, the loose criteria listed above are used (e.g., frequency >60%) in order to allow many exceptions to the "characteristic" processes. The molecule selection program selects a subclass of molecules S, from M', that ere best characterized by the processes in P. The subclass S includes molecules that show evidence for most (85% or more) of the processes in P, and excludes molecules that are exceptions to many. Thus S is at least as well behaved as M' with respect to P. And since the two measures of selection are not perfectly complementary, S is likely to be better behaved than M' with respect to P, (If molecule selection uses less restrictive measures than process selection, then S will be less well behaved than M' and the procedure will fail except when the initial set of molecules is homogeneous.) One interesting part of the procedure is that after processes are selected, ALL of the molecules are reclassified with regard to the number of times they appear as exceptions to the processes. This is shown in Figure 2: at step 2 of each level all molecules in the initial set, H (not M' or S), are tested against the processes. Thus, a molecule can be excluded at one level (because it is an exception to too many of the processes at that level), but be included again at another level for a slightly different set of processes. The condition under which we want the program to stop is that the subclass S of molecules after an iteration is the same as the class M' from which the iteration started (condition 1 in Figure 2). In other words, under this condition the program has found an S and a P such that P characterizes S (S=M') and S is vell-behaved with respect to P. The subclass S is taken to be homogeneous, and the processes in P can be taken to be mass spectrometry rules for molecules in S. The refinement level in Figure 2 is the number of times the procedure has been invoked in trying to find 69

4 one homogeneous subclass of molecules. The second of the stopping conditions tests whether the refinement level is equal to an arbitrary maximum, which is currently 3. This condition is necessary to avoid an infinite loop in the case where the program can find no subclass S that is homogeneous with respect to P. The level 3 has been observed to produce fairly acceptable results: after three iterations through this loop, the subclass S is about as refined as it will get. After more iterations the procedure appears to oscillate in that molecules added to S in one iteration are subtracted, from S in a later iteration. Our experience is very limited. Because there is no guarantee that the procedure converges, however, some stopping condition like the maximum refinement level is necessary. The last stopping condition shown in Figure 2 tests whether there are enough molecules in the subclass to warrant further refinement. If there are fewer than an arbitrary minimum number (=3) of molecules in S, then further refinements will be unreliable. This minimum is not completely arbitrary, since it depends to some extent on the frequency measures used in process and molecule selection. But, intuitively, when the number of molecules in G is small there is little value in breaking S up into subclasses anyway. As shown in the overall flow diagram, Figure 1, after the first major subclass (S) has been defined, all molecules in S are removed from any further consideration by subtracting them from M. The entire procedure is then repeated with the new M. It stops only when there are so few molecules left in M (3 or fewer) that process selection is unreliable and molecule selection appears pointless. The output of the whole program now is merely the collected set of outputs from all iterations, viz., the collected S,P pairs, as shown in Figure 2. Future work will focus on automatically generalizing the descriptions of the molecules. This is now done by hand, except when the initial class M is homogeneous - then the generalized description is the common graph structure. RESULTS The INTSUM program alone has already provided useful new results for chemists, as reported in the chemical literature. The process selection program, working with output from INTSUM (but without molecule selection), has successfully found sets of characteristic processes for a well-understood class of molecules (estrogens, Figure 3) and for classes whose behavior is still under investigation (e.g., equilenins, progesterones, amino acids). For 47 estrogens, which were assumed by both an expert and the program to be in one class, rules found by the program agree closely with rules formed by the expert from the same data. (This result is not shown in a table, but the comparison with the expert's rules looks much like that shown in Table I.) Expert chemists have made suggestions for improvements, but were generally in agreement with the processes selected by the program. The rule formation program with molecule selection has been tested on several sets of molecules. The results of running the program on a set of 15 estrogens (a subset of the 47 mentioned above) are shown in Table I. The program separated two of the 15 compounds into a second class because they were not as well behaved as the rest - they were exceptions to about 20% of the characteristic processes. However, the chemist thought the separation was reasonable. The processes selected by the program are shown with indications of the discrepancies between the program's choices and the chemist's. The discrepancies mostly arose from the program's applying different criteria to select one process from viable alternatives. Table II shows the success of the molecule separation part of the program when rule formation was done on data from 19 non-homogeneous estrogenic steroids. The major subclass of chemical interest is the set of 5 equilenins which are identified by common modifications to the skeleton shown in Figure 3. The structural properties were not used by the program although the chemist did classify the compounds by such features. By selecting well-behaved subclasses of molecules the program grouped four or five "equilenins" (molecules #4, 8, 10, 19) and all three "3-acetates" {#3, 11, l8) in the first subclass. The fifth equilenin (#2) was removed from that subclass on the last refinement because it was an exception to 3 of 9 characteristic processes used to determine the subclass. In the third iteration shown in Table II, the program grouped three of the chemist's four "3-benzoates" together (molecules #12, 13, 14). In the fourth iteration it grouped together the chemist's two "diacetates" and one "triacetate" (molecules #9, 15, 16). Two iterations produced subclasses with only two members - when put together they encompass two "17-acetates" (#1, 17), one "17-benzoate", and one "gamma-lactone" (#5). The two molecules remaining unclassified at the end of the procedure were the last "equilenin" (molecule 2) and the last "3-benzoate" (#6). CONCLUSIONS Building an information processing model of scientific reasoning in mass spectrometry, although not completed, has already led to interesting and useful results. The model incorporates heuristic search in process selection. The procedure for selecting molecules can be thought of as a planning procedure insofar as it reduces the problem of formulating rules for a class of diverse molecules to a number of smaller subproblems, viz., formulating rules for smaller classes of well-behaved molecules. However, the molecule selection procedure is highly dependent on process selection, as described in detail. The incompleteness of the program as a model of the entire rule formation procedure should be readily apparent. We have not described anything that approximates confrontation of rules with new data, for example. But as the results section indicates, the program can separate subclasses of well-behaved molecules and can find characteristic processes for the subclasses with enough accuracy (on a few examples) to gain preliminary acceptance by an expert in the field. ACKNOWLEDGEMENTS We have received invaluable assistance from colleagues. In particular, Mssrs. William White and Stephen Reiss have done almost all of the programming; Dr. Dennis Smith has provided chemical expertise; and Professors E.A. Feigenbaum and J. Lederberg have provided much of the intelligence in the design of the system. Financial support is gratefully acknowledged from the Advanced Research Projects Agency (SD-183) and from the National Institutes of Health (RR ). 70

5 REFERENCES (1) E.G. Buchanan and J. Lederberg, "The Heuristic DENDRAL Program for Explaining Empirical Data". In proceedings of the IFIP Congress 71, Ljubljana, Yugoslavia (1971). (Also Stanford Artificial Intelligence Project Memo No. 141} (2) B.G. Buchanan, E.A. Feigenbaum, and J. Lederberg, "A Heuristic Programming Study of Theory Formation in Science." In proceedings of the Second International Joint Conference on Artificial Intelligence, Imperial College, London (September, 1971). (Also Stanford Artificial Intelligence Project Memo No. 145) (3) B.G. Buchanan, E.A. Feigenbaum, and U.S. Sridharan, "Heuristic Theory Formation: Data Interpretation and Rule Formation". In Machine Intelligence 7, Edinburgh University Press (1972). (4) D.H. Smith, B.G. Buchanan, V.C. White, E.A. Feigenbaum, C. DJerassi and J. Lederberg, "Applications of Artificial Intelligence for Chemical Inference X. INTSUM. A Data Interpretation Program as Applied to the Collected Mass Spectra of Estrogenic Steroids". Tetrahedron (in press). * Details in Figure 2. 71

6 72

7 TABLE I. PROCESSES SELECTED FOR 15 ESTROGENS BELIEVED TO BE IN ONE WELL-BEHAVED CLASS * The underlined processes are those selected by an expert chemist on the basis of data from 47 well-behaved estrogens, including these 15-73

8 PROCESS LABEL* 11. BRK11L TABLE I, Page 2 PICTORIAL DESCRIPTION % 07 ALL DATA POINTS EXPLAINED 12. BRK2L/11L (preferred BRK20L) over 13. BRK5H/10L 14. BRK5H/12L 15. BRK12L/15H or BRK12L/14H * The underlined processes are those selected by an expert chemist on the basis of data from 47 well-behaved estrogens, including these

9 TABLE II SUMMARY OF STEPS IN THE RULE FORMATION PROCEDURE WITH 19 ESTROGENIC STEROIDS ITERATION #1 Initial Set: First Refinement: Second Refinement: Third Refinement: = Subclass 1 ITERATION #2 Initial Set [- Subclass 1, Third Refinement - Subclass 2 ITERATION #3 Third Refinement = Subclass 3 ITERATION #4 Last Refinement: = Subclass 4 75

10 TABLE II* Page 2 Molecules 76

Coh. cjtankxd Univeraity Libraries Dept. of Sc&aai Cotedtons Trt*

Coh. cjtankxd Univeraity Libraries Dept. of Sc&aai Cotedtons Trt* Coh cjtankxd Univeraity Libraries Dept. of Sc&aai Cotedtons Trt* STANFORD ARTIFICIAL INTELLIGENCE MEMO AIM- 10^ PROJECT NOVEMBER 1969 A HEURISTIC PROGRAM FOR SOLVING A SCIENTIFIC INFERENCE PROBLEM: SUMMARY

More information

Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples

Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples Published in: Journal of Global Optimization, 5, pp. 69-9, 199. Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples Evangelos Triantaphyllou Assistant

More information

PHENOMENOLOGICAL LAWS AND THEIR APPLICATION TO SCIENTIFIC EPISTEMIC EXPLANATION PROBLEMS. Erik WEBER(*)

PHENOMENOLOGICAL LAWS AND THEIR APPLICATION TO SCIENTIFIC EPISTEMIC EXPLANATION PROBLEMS. Erik WEBER(*) Logique & Analyse 129-130 (1990), 175-189 PHENOMENOLOGICAL LAWS AND THEIR APPLICATION TO SCIENTIFIC EPISTEMIC EXPLANATION PROBLEMS Erik WEBER(*) 1. Introduction Scientific knowledge has both a practical

More information

CHAPTER 2 INTRODUCTION TO CLASSICAL PROPOSITIONAL LOGIC

CHAPTER 2 INTRODUCTION TO CLASSICAL PROPOSITIONAL LOGIC CHAPTER 2 INTRODUCTION TO CLASSICAL PROPOSITIONAL LOGIC 1 Motivation and History The origins of the classical propositional logic, classical propositional calculus, as it was, and still often is called,

More information

CS 6375 Machine Learning

CS 6375 Machine Learning CS 6375 Machine Learning Decision Trees Instructor: Yang Liu 1 Supervised Classifier X 1 X 2. X M Ref class label 2 1 Three variables: Attribute 1: Hair = {blond, dark} Attribute 2: Height = {tall, short}

More information

Lecture Notes on From Rules to Propositions

Lecture Notes on From Rules to Propositions Lecture Notes on From Rules to Propositions 15-816: Substructural Logics Frank Pfenning Lecture 2 September 1, 2016 We review the ideas of ephemeral truth and linear inference with another example from

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 14, 2015 Today: The Big Picture Overfitting Review: probability Readings: Decision trees, overfiting

More information

CS340 Machine learning Lecture 5 Learning theory cont'd. Some slides are borrowed from Stuart Russell and Thorsten Joachims

CS340 Machine learning Lecture 5 Learning theory cont'd. Some slides are borrowed from Stuart Russell and Thorsten Joachims CS340 Machine learning Lecture 5 Learning theory cont'd Some slides are borrowed from Stuart Russell and Thorsten Joachims Inductive learning Simplest form: learn a function from examples f is the target

More information

DENDRALANDMETA-DENDRAL: THEIR APPLICATIONS DIMENSION. Bruce G. Buchanan and Edward A. Feigenbaum

DENDRALANDMETA-DENDRAL: THEIR APPLICATIONS DIMENSION. Bruce G. Buchanan and Edward A. Feigenbaum Stanford Heuristic Programming Project Memo HPP-78-I February1978 Computer Science Department Report No. STAN-CS-78-649 DENDRALANDMETA-DENDRAL: THEIR APPLICATIONS DIMENSION by Bruce G. Buchanan and Edward

More information

Lecture Presentation. Chapter 1. Matter, Measurement, and Problem Solving. Christian Madu, Ph.D. Collin College Pearson Education, Inc.

Lecture Presentation. Chapter 1. Matter, Measurement, and Problem Solving. Christian Madu, Ph.D. Collin College Pearson Education, Inc. Lecture Presentation Chapter 1 Matter, Measurement, and Problem Solving Christian Madu, Ph.D. Collin College What Do You Think? What do you think is the most important idea in all of human knowledge? If

More information

Analysis of a Large Structure/Biological Activity. Data Set Using Recursive Partitioning and. Simulated Annealing

Analysis of a Large Structure/Biological Activity. Data Set Using Recursive Partitioning and. Simulated Annealing Analysis of a Large Structure/Biological Activity Data Set Using Recursive Partitioning and Simulated Annealing Student: Ke Zhang MBMA Committee: Dr. Charles E. Smith (Chair) Dr. Jacqueline M. Hughes-Oliver

More information

Efficiently merging symbolic rules into integrated rules

Efficiently merging symbolic rules into integrated rules Efficiently merging symbolic rules into integrated rules Jim Prentzas a, Ioannis Hatzilygeroudis b a Democritus University of Thrace, School of Education Sciences Department of Education Sciences in Pre-School

More information

Supplementary Logic Notes CSE 321 Winter 2009

Supplementary Logic Notes CSE 321 Winter 2009 1 Propositional Logic Supplementary Logic Notes CSE 321 Winter 2009 1.1 More efficient truth table methods The method of using truth tables to prove facts about propositional formulas can be a very tedious

More information

Continuous symmetry and shape measures, or how to measure the distance between polyhedra representing molecules

Continuous symmetry and shape measures, or how to measure the distance between polyhedra representing molecules Continuous symmetry and shape measures, or how to measure the distance between polyhedra representing molecules Pere Alemany Institut de Química Teòrica i Computacional de la Universitat de Barcelona (IQTCUB)

More information

A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997

A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997 A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997 Vasant Honavar Artificial Intelligence Research Laboratory Department of Computer Science

More information

Natural Language Processing Prof. Pawan Goyal Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

Natural Language Processing Prof. Pawan Goyal Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Natural Language Processing Prof. Pawan Goyal Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Lecture - 18 Maximum Entropy Models I Welcome back for the 3rd module

More information

Computational Tasks and Models

Computational Tasks and Models 1 Computational Tasks and Models Overview: We assume that the reader is familiar with computing devices but may associate the notion of computation with specific incarnations of it. Our first goal is to

More information

Explanation and Argument in Mathematical Practice

Explanation and Argument in Mathematical Practice Explanation and Argument in Mathematical Practice Andrew Aberdein Humanities and Communication, Florida Institute of Technology, 50 West University Blvd, Melbourne, Florida 3290-6975, U.S.A. my.fit.edu/

More information

Introduction to Metalogic

Introduction to Metalogic Philosophy 135 Spring 2008 Tony Martin Introduction to Metalogic 1 The semantics of sentential logic. The language L of sentential logic. Symbols of L: Remarks: (i) sentence letters p 0, p 1, p 2,... (ii)

More information

PAC Learning. prof. dr Arno Siebes. Algorithmic Data Analysis Group Department of Information and Computing Sciences Universiteit Utrecht

PAC Learning. prof. dr Arno Siebes. Algorithmic Data Analysis Group Department of Information and Computing Sciences Universiteit Utrecht PAC Learning prof. dr Arno Siebes Algorithmic Data Analysis Group Department of Information and Computing Sciences Universiteit Utrecht Recall: PAC Learning (Version 1) A hypothesis class H is PAC learnable

More information

Computational Learning Theory

Computational Learning Theory CS 446 Machine Learning Fall 2016 OCT 11, 2016 Computational Learning Theory Professor: Dan Roth Scribe: Ben Zhou, C. Cervantes 1 PAC Learning We want to develop a theory to relate the probability of successful

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University References: 1. Machine Learning, Chapter 3 2. Data Mining: Concepts, Models,

More information

Lecture 3: Decision Trees

Lecture 3: Decision Trees Lecture 3: Decision Trees Cognitive Systems II - Machine Learning SS 2005 Part I: Basic Approaches of Concept Learning ID3, Information Gain, Overfitting, Pruning Lecture 3: Decision Trees p. Decision

More information

Stochastic Histories. Chapter Introduction

Stochastic Histories. Chapter Introduction Chapter 8 Stochastic Histories 8.1 Introduction Despite the fact that classical mechanics employs deterministic dynamical laws, random dynamical processes often arise in classical physics, as well as in

More information

HS AP Physics 1 Science

HS AP Physics 1 Science Scope And Sequence Timeframe Unit Instructional Topics 5 Day(s) 20 Day(s) 5 Day(s) Kinematics Course AP Physics 1 is an introductory first-year, algebra-based, college level course for the student interested

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Goals for the lecture you should understand the following concepts the decision tree representation the standard top-down approach to learning a tree Occam s razor entropy and information

More information

CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE

CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE 3.1 Model Violations If a set of items does not form a perfect Guttman scale but contains a few wrong responses, we do not necessarily need to discard it. A wrong

More information

A Necessary Condition for Learning from Positive Examples

A Necessary Condition for Learning from Positive Examples Machine Learning, 5, 101-113 (1990) 1990 Kluwer Academic Publishers. Manufactured in The Netherlands. A Necessary Condition for Learning from Positive Examples HAIM SHVAYTSER* (HAIM%SARNOFF@PRINCETON.EDU)

More information

C3020 Molecular Evolution. Exercises #3: Phylogenetics

C3020 Molecular Evolution. Exercises #3: Phylogenetics C3020 Molecular Evolution Exercises #3: Phylogenetics Consider the following sequences for five taxa 1-5 and the known outgroup O, which has the ancestral states (note that sequence 3 has changed from

More information

A DEFINITION-DRIVEN THEOREM PROVER

A DEFINITION-DRIVEN THEOREM PROVER Session 3 Logic; I Theorem Proving and A DEFINITION-DRIVEN THEOREM PROVER George W. Ernst Department of Computing and Information Sciences Case Western Reserve University Cleveland, Ohio 44106 Abstract

More information

Preferred Mental Models in Qualitative Spatial Reasoning: A Cognitive Assessment of Allen s Calculus

Preferred Mental Models in Qualitative Spatial Reasoning: A Cognitive Assessment of Allen s Calculus Knauff, M., Rauh, R., & Schlieder, C. (1995). Preferred mental models in qualitative spatial reasoning: A cognitive assessment of Allen's calculus. In Proceedings of the Seventeenth Annual Conference of

More information

Logic, Sets, and Proofs

Logic, Sets, and Proofs Logic, Sets, and Proofs David A. Cox and Catherine C. McGeoch Amherst College 1 Logic Logical Operators. A logical statement is a mathematical statement that can be assigned a value either true or false.

More information

CS6375: Machine Learning Gautam Kunapuli. Decision Trees

CS6375: Machine Learning Gautam Kunapuli. Decision Trees Gautam Kunapuli Example: Restaurant Recommendation Example: Develop a model to recommend restaurants to users depending on their past dining experiences. Here, the features are cost (x ) and the user s

More information

X X (2) X Pr(X = x θ) (3)

X X (2) X Pr(X = x θ) (3) Notes for 848 lecture 6: A ML basis for compatibility and parsimony Notation θ Θ (1) Θ is the space of all possible trees (and model parameters) θ is a point in the parameter space = a particular tree

More information

Supplementary material to Structure Learning of Linear Gaussian Structural Equation Models with Weak Edges

Supplementary material to Structure Learning of Linear Gaussian Structural Equation Models with Weak Edges Supplementary material to Structure Learning of Linear Gaussian Structural Equation Models with Weak Edges 1 PRELIMINARIES Two vertices X i and X j are adjacent if there is an edge between them. A path

More information

Solutions to Quiz 1. Problem 1 (20 points). Suppose S(n) is a predicate on natural numbers, n, and suppose. k N S(k) S(k + 2). (1)

Solutions to Quiz 1. Problem 1 (20 points). Suppose S(n) is a predicate on natural numbers, n, and suppose. k N S(k) S(k + 2). (1) Massachusetts Institute of Technology 6.042J/18.062J, Fall 02: Mathematics for Computer Science Professor Albert Meyer and Dr. Radhika Nagpal Solutions to Quiz 1 Problem 1 (20 points). Suppose S(n) is

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Fall 2018 Introduction to Artificial Intelligence Practice Final You have approximately 2 hours 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Mathematics 114L Spring 2018 D.A. Martin. Mathematical Logic

Mathematics 114L Spring 2018 D.A. Martin. Mathematical Logic Mathematics 114L Spring 2018 D.A. Martin Mathematical Logic 1 First-Order Languages. Symbols. All first-order languages we consider will have the following symbols: (i) variables v 1, v 2, v 3,... ; (ii)

More information

A Generalized Algorithm for Computing Matching Polynomials using Determinants. Asa Scherer

A Generalized Algorithm for Computing Matching Polynomials using Determinants. Asa Scherer A Generalized Algorithm for Computing Matching Polynomials using Determinants Asa Scherer April 17, 2007 Abstract We discuss the connection between determinants of modified matching matrices and the matching

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

Final Exam, Machine Learning, Spring 2009

Final Exam, Machine Learning, Spring 2009 Name: Andrew ID: Final Exam, 10701 Machine Learning, Spring 2009 - The exam is open-book, open-notes, no electronics other than calculators. - The maximum possible score on this exam is 100. You have 3

More information

CS 6901 (Applied Algorithms) Lecture 3

CS 6901 (Applied Algorithms) Lecture 3 CS 6901 (Applied Algorithms) Lecture 3 Antonina Kolokolova September 16, 2014 1 Representative problems: brief overview In this lecture we will look at several problems which, although look somewhat similar

More information

Min/Max-Poly Weighting Schemes and the NL vs UL Problem

Min/Max-Poly Weighting Schemes and the NL vs UL Problem Min/Max-Poly Weighting Schemes and the NL vs UL Problem Anant Dhayal Jayalal Sarma Saurabh Sawlani May 3, 2016 Abstract For a graph G(V, E) ( V = n) and a vertex s V, a weighting scheme (w : E N) is called

More information

Propositions and Proofs

Propositions and Proofs Chapter 2 Propositions and Proofs The goal of this chapter is to develop the two principal notions of logic, namely propositions and proofs There is no universal agreement about the proper foundations

More information

Boosting & Deep Learning

Boosting & Deep Learning Boosting & Deep Learning Ensemble Learning n So far learning methods that learn a single hypothesis, chosen form a hypothesis space that is used to make predictions n Ensemble learning à select a collection

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

WADA Technical Document TD2003IDCR

WADA Technical Document TD2003IDCR IDENTIFICATION CRITERIA FOR QUALITATIVE ASSAYS INCORPORATING CHROMATOGRAPHY AND MASS SPECTROMETRY The appropriate analytical characteristics must be documented for a particular assay. The Laboratory must

More information

THe use of mixed data-type expressions, although not

THe use of mixed data-type expressions, although not ARTICLE PREPRINT: AUGUST 24 1 Mixed Data-Type Circuit Modeling: The Ideal Diode John Rugis Abstract This article presents a mixed numeric plus Boolean data-type algebraic representation that can be used

More information

On improving matchings in trees, via bounded-length augmentations 1

On improving matchings in trees, via bounded-length augmentations 1 On improving matchings in trees, via bounded-length augmentations 1 Julien Bensmail a, Valentin Garnero a, Nicolas Nisse a a Université Côte d Azur, CNRS, Inria, I3S, France Abstract Due to a classical

More information

CS 6783 (Applied Algorithms) Lecture 3

CS 6783 (Applied Algorithms) Lecture 3 CS 6783 (Applied Algorithms) Lecture 3 Antonina Kolokolova January 14, 2013 1 Representative problems: brief overview of the course In this lecture we will look at several problems which, although look

More information

Clocking the Effect of Molarity on Speed of Reaction. reaction. While most people do assume that the temperature of the solution is often the most

Clocking the Effect of Molarity on Speed of Reaction. reaction. While most people do assume that the temperature of the solution is often the most Ding 1 Chunyang Ding Mr. Rierson AP/IB Chemistry SL 28 January 2013 Clocking the Effect of Molarity on Speed of Reaction In basic levels of chemistry, most of the experimenter s attention is on the reaction

More information

WADA Technical Document TD2015IDCR

WADA Technical Document TD2015IDCR MINIMUM CRITERIA FOR CHROMATOGRAPHIC-MASS SPECTROMETRIC CONFIRMATION OF THE IDENTITY OF ANALYTES FOR DOPING CONTROL PURPOSES. The ability of a method to identify an analyte is a function of the entire

More information

Optimization of Quadratic Forms: NP Hard Problems : Neural Networks

Optimization of Quadratic Forms: NP Hard Problems : Neural Networks 1 Optimization of Quadratic Forms: NP Hard Problems : Neural Networks Garimella Rama Murthy, Associate Professor, International Institute of Information Technology, Gachibowli, HYDERABAD, AP, INDIA ABSTRACT

More information

1 Computational problems

1 Computational problems 80240233: Computational Complexity Lecture 1 ITCS, Tsinghua Univesity, Fall 2007 9 October 2007 Instructor: Andrej Bogdanov Notes by: Andrej Bogdanov The aim of computational complexity theory is to study

More information

Supervised Learning! Algorithm Implementations! Inferring Rudimentary Rules and Decision Trees!

Supervised Learning! Algorithm Implementations! Inferring Rudimentary Rules and Decision Trees! Supervised Learning! Algorithm Implementations! Inferring Rudimentary Rules and Decision Trees! Summary! Input Knowledge representation! Preparing data for learning! Input: Concept, Instances, Attributes"

More information

10. Alternative case influence statistics

10. Alternative case influence statistics 10. Alternative case influence statistics a. Alternative to D i : dffits i (and others) b. Alternative to studres i : externally-studentized residual c. Suggestion: use whatever is convenient with the

More information

Notes for course EE1.1 Circuit Analysis TOPIC 4 NODAL ANALYSIS

Notes for course EE1.1 Circuit Analysis TOPIC 4 NODAL ANALYSIS Notes for course EE1.1 Circuit Analysis 2004-05 TOPIC 4 NODAL ANALYSIS OBJECTIVES 1) To develop Nodal Analysis of Circuits without Voltage Sources 2) To develop Nodal Analysis of Circuits with Voltage

More information

cse541 LOGIC FOR COMPUTER SCIENCE

cse541 LOGIC FOR COMPUTER SCIENCE cse541 LOGIC FOR COMPUTER SCIENCE Professor Anita Wasilewska Spring 2015 LECTURE 2 Chapter 2 Introduction to Classical Propositional Logic PART 1: Classical Propositional Model Assumptions PART 2: Syntax

More information

A Note on Backward Dual Resolution and Its Application to Proving Completeness of Rule-Based Systems*

A Note on Backward Dual Resolution and Its Application to Proving Completeness of Rule-Based Systems* A Note on Backward Dual Resolution and Its Application to Proving Completeness of Rule-Based Systems* Antoni Ligeza Institute of Automatics AGH al. Mickiewicza 30, 30-059 Krakow, Poland tel. &; fax: (48

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

Machine Learning

Machine Learning Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 13, 2011 Today: The Big Picture Overfitting Review: probability Readings: Decision trees, overfiting

More information

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric?

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/

More information

Quantitative Understanding in Biology Module II: Model Parameter Estimation Lecture I: Linear Correlation and Regression

Quantitative Understanding in Biology Module II: Model Parameter Estimation Lecture I: Linear Correlation and Regression Quantitative Understanding in Biology Module II: Model Parameter Estimation Lecture I: Linear Correlation and Regression Correlation Linear correlation and linear regression are often confused, mostly

More information

Decision Tree Learning - ID3

Decision Tree Learning - ID3 Decision Tree Learning - ID3 n Decision tree examples n ID3 algorithm n Occam Razor n Top-Down Induction in Decision Trees n Information Theory n gain from property 1 Training Examples Day Outlook Temp.

More information

I. Induction, Probability and Confirmation: Introduction

I. Induction, Probability and Confirmation: Introduction I. Induction, Probability and Confirmation: Introduction 1. Basic Definitions and Distinctions Singular statements vs. universal statements Observational terms vs. theoretical terms Observational statement

More information

Module 7-2 Decomposition Approach

Module 7-2 Decomposition Approach Module 7-2 Decomposition Approach Chanan Singh Texas A&M University Decomposition Approach l Now we will describe a method of decomposing the state space into subsets for the purpose of calculating the

More information

CMSC 451: Lecture 7 Greedy Algorithms for Scheduling Tuesday, Sep 19, 2017

CMSC 451: Lecture 7 Greedy Algorithms for Scheduling Tuesday, Sep 19, 2017 CMSC CMSC : Lecture Greedy Algorithms for Scheduling Tuesday, Sep 9, 0 Reading: Sects.. and. of KT. (Not covered in DPV.) Interval Scheduling: We continue our discussion of greedy algorithms with a number

More information

Name (NetID): (1 Point)

Name (NetID): (1 Point) CS446: Machine Learning (D) Spring 2017 March 16 th, 2017 This is a closed book exam. Everything you need in order to solve the problems is supplied in the body of this exam. This exam booklet contains

More information

Statistical and Computational Learning Theory

Statistical and Computational Learning Theory Statistical and Computational Learning Theory Fundamental Question: Predict Error Rates Given: Find: The space H of hypotheses The number and distribution of the training examples S The complexity of the

More information

A Little Deductive Logic

A Little Deductive Logic A Little Deductive Logic In propositional or sentential deductive logic, we begin by specifying that we will use capital letters (like A, B, C, D, and so on) to stand in for sentences, and we assume that

More information

Algorithms and Complexity theory

Algorithms and Complexity theory Algorithms and Complexity theory Thibaut Barthelemy Some slides kindly provided by Fabien Tricoire University of Vienna WS 2014 Outline 1 Algorithms Overview How to write an algorithm 2 Complexity theory

More information

Learning Decision Trees

Learning Decision Trees Learning Decision Trees Machine Learning Fall 2018 Some slides from Tom Mitchell, Dan Roth and others 1 Key issues in machine learning Modeling How to formulate your problem as a machine learning problem?

More information

Approximating the Partition Function by Deleting and then Correcting for Model Edges (Extended Abstract)

Approximating the Partition Function by Deleting and then Correcting for Model Edges (Extended Abstract) Approximating the Partition Function by Deleting and then Correcting for Model Edges (Extended Abstract) Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles

More information

Lecture Notes on From Rules to Propositions

Lecture Notes on From Rules to Propositions Lecture Notes on From Rules to Propositions 15-816: Linear Logic Frank Pfenning Lecture 2 January 18, 2012 We review the ideas of ephemeral truth and linear inference with another example from graph theory:

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Goals for the lecture you should understand the following concepts the decision tree representation the standard top-down approach to learning a tree Occam s razor entropy and information

More information

Logical Abduction and its Applications in Program Verification. Isil Dillig MSR Cambridge

Logical Abduction and its Applications in Program Verification. Isil Dillig MSR Cambridge Logical Abduction and its Applications in Program Verification? Isil Dillig MSR Cambridge What is Abduction? Abduction: Opposite of deduction 2 / 21 What is Abduction? Abduction: Opposite of deduction

More information

Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4

Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4 Inference: Exploiting Local Structure aphne Koller Stanford University CS228 Handout #4 We have seen that N inference exploits the network structure, in particular the conditional independence and the

More information

The Beginning of Graph Theory. Theory and Applications of Complex Networks. Eulerian paths. Graph Theory. Class Three. College of the Atlantic

The Beginning of Graph Theory. Theory and Applications of Complex Networks. Eulerian paths. Graph Theory. Class Three. College of the Atlantic Theory and Applications of Complex Networs 1 Theory and Applications of Complex Networs 2 Theory and Applications of Complex Networs Class Three The Beginning of Graph Theory Leonhard Euler wonders, can

More information

Conjunctive Normal Form and SAT

Conjunctive Normal Form and SAT Notes on Satisfiability-Based Problem Solving Conjunctive Normal Form and SAT David Mitchell mitchell@cs.sfu.ca September 19, 2013 This is a preliminary draft of these notes. Please do not distribute without

More information

Chapter 2. Mathematical Reasoning. 2.1 Mathematical Models

Chapter 2. Mathematical Reasoning. 2.1 Mathematical Models Contents Mathematical Reasoning 3.1 Mathematical Models........................... 3. Mathematical Proof............................ 4..1 Structure of Proofs........................ 4.. Direct Method..........................

More information

The Importance of Being Formal. Martin Henz. February 5, Propositional Logic

The Importance of Being Formal. Martin Henz. February 5, Propositional Logic The Importance of Being Formal Martin Henz February 5, 2014 Propositional Logic 1 Motivation In traditional logic, terms represent sets, and therefore, propositions are limited to stating facts on sets

More information

CHAPTER 4 CLASSICAL PROPOSITIONAL SEMANTICS

CHAPTER 4 CLASSICAL PROPOSITIONAL SEMANTICS CHAPTER 4 CLASSICAL PROPOSITIONAL SEMANTICS 1 Language There are several propositional languages that are routinely called classical propositional logic languages. It is due to the functional dependency

More information

Top-down Causality the missing link in our physical theories

Top-down Causality the missing link in our physical theories Top-down Causality the missing link in our physical theories Jose P Koshy josepkoshy@gmail.com Abstract: Confining is a top-down effect. Particles have to be confined in a region if these are to bond together.

More information

KE/Tableaux. What is it for?

KE/Tableaux. What is it for? CS3UR: utomated Reasoning 2002 The term Tableaux refers to a family of deduction methods for different logics. We start by introducing one of them: non-free-variable KE for classical FOL What is it for?

More information

Reading 5 : Induction

Reading 5 : Induction CS/Math 40: Introduction to Discrete Mathematics Fall 015 Instructors: Beck Hasti and Gautam Prakriya Reading 5 : Induction In the last reading we began discussing proofs. We mentioned some proof paradigms

More information

Strong AI vs. Weak AI Automated Reasoning

Strong AI vs. Weak AI Automated Reasoning Strong AI vs. Weak AI Automated Reasoning George F Luger ARTIFICIAL INTELLIGENCE 6th edition Structures and Strategies for Complex Problem Solving Artificial intelligence can be classified into two categories:

More information

Supplementary Notes on Inductive Definitions

Supplementary Notes on Inductive Definitions Supplementary Notes on Inductive Definitions 15-312: Foundations of Programming Languages Frank Pfenning Lecture 2 August 29, 2002 These supplementary notes review the notion of an inductive definition

More information

A SYSTEM OF AXIOMATIC SET THEORY PART VI 62

A SYSTEM OF AXIOMATIC SET THEORY PART VI 62 THE JOOBNAL OF SYMBOLIC LOGIC Volume 13, Number 2, June 1948 A SYSTEM OF AXIOMATIC SET THEORY PART VI 62 PAUL BEKNAYS 16. The r61e of the restrictive axiom. Comparability of classes. Till now we tried

More information

Machine Learning Alternatives to Manual Knowledge Acquisition

Machine Learning Alternatives to Manual Knowledge Acquisition Machine Learning Alternatives to Manual Knowledge Acquisition Interactive programs which elicit knowledge from the expert during the course of a conversation at the terminal. Programs which learn by scanning

More information

Wooldridge, Introductory Econometrics, 3d ed. Chapter 9: More on specification and data problems

Wooldridge, Introductory Econometrics, 3d ed. Chapter 9: More on specification and data problems Wooldridge, Introductory Econometrics, 3d ed. Chapter 9: More on specification and data problems Functional form misspecification We may have a model that is correctly specified, in terms of including

More information

Computational Methods for Mass Spectrometry Proteomics

Computational Methods for Mass Spectrometry Proteomics Computational Methods for Mass Spectrometry Proteomics Eidhammer, Ingvar ISBN-13: 9780470512975 Table of Contents Preface. Acknowledgements. 1 Protein, Proteome, and Proteomics. 1.1 Primary goals for studying

More information

CS 5522: Artificial Intelligence II

CS 5522: Artificial Intelligence II CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]

More information

MA103 Introduction to Abstract Mathematics Second part, Analysis and Algebra

MA103 Introduction to Abstract Mathematics Second part, Analysis and Algebra 206/7 MA03 Introduction to Abstract Mathematics Second part, Analysis and Algebra Amol Sasane Revised by Jozef Skokan, Konrad Swanepoel, and Graham Brightwell Copyright c London School of Economics 206

More information

Lesson 7: Classification of Solutions

Lesson 7: Classification of Solutions Student Outcomes Students know the conditions for which a linear equation will have a unique solution, no solution, or infinitely many solutions. Lesson Notes Part of the discussion on the second page

More information

MATHEMATICS GRADE 7. THE EWING PUBLIC SCHOOLS 2099 Pennington Road Ewing, NJ 08618

MATHEMATICS GRADE 7. THE EWING PUBLIC SCHOOLS 2099 Pennington Road Ewing, NJ 08618 MATHEMATICS GRADE 7 THE EWING PUBLIC SCHOOLS 2099 Pennington Road Ewing, NJ 08618 Board Approval Date: August 29, 2016 Michael Nitti Revised by: District Math Staff Superintendent In accordance with The

More information

Propositional Logic. Fall () Propositional Logic Fall / 30

Propositional Logic. Fall () Propositional Logic Fall / 30 Propositional Logic Fall 2013 () Propositional Logic Fall 2013 1 / 30 1 Introduction Learning Outcomes for this Presentation 2 Definitions Statements Logical connectives Interpretations, contexts,... Logically

More information

Classification and Regression Trees

Classification and Regression Trees Classification and Regression Trees Ryan P Adams So far, we have primarily examined linear classifiers and regressors, and considered several different ways to train them When we ve found the linearity

More information

A Little Deductive Logic

A Little Deductive Logic A Little Deductive Logic In propositional or sentential deductive logic, we begin by specifying that we will use capital letters (like A, B, C, D, and so on) to stand in for sentences, and we assume that

More information

Mathematics Standards for High School Advanced Quantitative Reasoning

Mathematics Standards for High School Advanced Quantitative Reasoning Mathematics Standards for High School Advanced Quantitative Reasoning The Advanced Quantitative Reasoning (AQR) course is a fourth-year launch course designed to be an alternative to Precalculus that prepares

More information