Data Fusion with Entropic Priors

Size: px
Start display at page:

Download "Data Fusion with Entropic Priors"

Transcription

1 Data Fusion with Entropic Priors Francesco PALMIERI, Domenico CIUONZO Dipartimento di Ingegneria dell Informazione, Seconda Università di Napoli, Italy Abstract. In classification problems, lack of knowledge of the prior distribution may make the application of Bayes rule inadequate. Uniform or arbitrary priors may often provide classification answers that, even in simple examples, may end up contradicting our common sense about the problem. Entropic priors, via application of the maximum entropy principle, seem to provide a much better answer and can be easily derived and applied to classification tasks when no more than the likelihood funtions are available. In this paper we present an application example in which the use of the entropic priors is compared to the results of the application of Dempster-Shafer theory. Keywords. Bayesian Networks, Artificial Intelligence, Propagation of Belief, Data Fusion Introduction In providing solutions to data fusion problems we often have to deal with incomplete data. More specifically, even if the likelihood functions may be available, lack of knowledge of the priors may strongly affect the results of our classifications. Classical use of uniform priors, that may be traced back to the work of Laplace [], may result inappropriate in many applications [7] [2]. The basic question is: what is the information on the priors that is carried only by the likelihoods? Or better: what is the prior distribution to be used in Bayes fusion formula that ensures that no other information beyond that included in likelihoods is injected into the problem? Various attempts have appeared in the literature in trying to obtain the least informative solution as an alternative to the inadequacy of using uniform, or arbitrary priors [7][8][9][0]. Evidence theory [] and alternative theories to probability [7], have been developed in the attempt of providing answers when Bayes theorem seems to lead to solutions that are inadequate, or contradictory with respect to our common sense. Dempster-Shafer approach to data fusion [3][4][5] is becoming very popular in some domain areas [2][7] and to many it seems to be a promising useful alternative to standard Bayesian approaches. We do not necessarily make classical probability theory our flagship, but try fillin our lack of knowledge about the model within the context of classical information theory [3]. More specifically, we show that, when prior probabilities are not available, by global maximization of the entropy we obtain a solution for the priors that is easy to Corresponding Author: Dipartimento di Ingegneria dell Informazione, Seconda Università di Napoli, via Roma 29, 803 Aversa (CE), Italy; francesco.palmieri@unina2.it.

2 compute and that, when applied to application scenarios, seems to fix the inadequacies of Bayes rule based on uniform or arbitrary priors. Entropy maximization has a long history [4][5][6] and entropic priors [7][8][9][0] have been advocated mostly within the context of theoretical physics. To our knowledge however, their full potential has not been exploited yet within the data fusion context. In this paper we derive the entropic priors and apply them to an example taken from [2] and [7] where Dempster-Shafer (DS) theory has been applied to a target classification problem. Our application of the maximum entropy priors shows strikingly similar results in providing good common sense answers to a problem where Bayes rule with uniform priors is clearly inaquate.. Bayes Data Fusion In classification, or pattern recognition problems, the typical scenario is one with available a sequence of N vectors of random attributes X = {X[], X[2],..., X[N]}, X[n] X n, n =,..., N () directly linked to a set of discrete labels S = {S[], S[2],..., S[M]}, S[m] S m, m =,..., M. (2) Each vector X[n] is a D-dimensional random vector that contains continuous or discrete attibutes. They may represent the results of measurements or discrete information coming from other subsystems. The classification consists in associating the measurements to the labels, that may be representative of different classes, or other discrete attributes. More compactly (X, S) X N S M, X N = X X N, S M = S S M. (3) The classification problem consists in estimating S from a realization X = x = {x[], x[2],..., x[n]}. (4) Detailed inference about S can be obtained if we were able to compute for each s S the a posteriori probabilities via application of Bayes theorem P r {S = s X = x} = f(x S = s)p r{s = s}, (5) f(x) where f(x S = s) are the likelihood functions and P r {S = s} are the a priori probabilities (priors). In many practical cases the system designer has available the likelihood funtions, that represent the data generative model, but he is not sure about the priors. Incorrect use of the priors may strongly affect the results of the classification and this has often put doubts on the effectiveness of using of Bayes theorem [7]. Assuming uniform priors is very common in statistical methods and it appears to be reasonable when such information is not available. After all the well-known maximum likelyhood method is equivalent to using uniform priors. Unfortunately, many cases have been shown in the literature in which the use of uniform priors shows contradictory results if compared to our intuitive understanding of the problem [2][5][7]. The assumption of uniform priors may be sometimes too strong, or completely arbitrary.

3 2. Entropic Priors We approach the problem here searching for the prior distribution that, given the likelihood functions, maximizes the joint entropy [3] H(X, S) = f(x S = s)p r{s = s} log dx. (6) s S x X f(x S = s)p r{s = s} The rationale behind this choice is that entropy effectively represents the degree of uncertainty about the model [4][5][6] and appears to be the natural choice when no further hypotheses can be injected into the system. Maximum entropy methods have been used extensively in the literature showing great success in many application areas. Wellknown is the exponential solution to entropy maximization [3][4] when constraints are imposed in the form of moments. We focus here on the classification problem and point to a solution in which no more than the likelihood funtions are needed. Proposition (Entropic Priors): The prior distribution that maximizes the joint entropy H(X, S) for given likelihoods f(x S = s), s S, is P re {S = s} = y S e H(X S=s) where H(X S = y) is the conditional entropy H(X S = s) = x X eh(x S=y), (7) f(x S = s) log [f(x S = s)] dx. (8) Entropic priors have been proposed in different forms [7][8][9][0][6] mostly within the context of theoretical physics. In [7][9][0] constraints on Fisher s information matrix are also imposed. Our derivation is different from Neumann and Skilling s [8] because we maximize joint entropy, while they maximize prior entropy with a contraint on the average likelihood entropy. Their result contains a scale parameter that needs to be resolved by other means, while in our case we obtain a closed-form solution that is totally self-contained. In anycase, to our knowledge, the real power of using entropy priors has not found full application yet in data fusion problems. For completeness we report here our proof. Proof: For maximization of H(X, S), the Lagrange function is L = H(X, S) λ ( s P (s) ) = H(X S) + H(S) λ ( s P (s) ) = s H(X s)p (s) s P (s) log P (s) λ ( s P (s) ) (9) where P (s) = P r{s = s} and H(X s) = H(X S = s) and the only constraint is P (s) =. Taking the derivative with respect to P (s) we get s L = H(X s) log P (s) λ. (0) P (s) Setting the derivative to zero we get p(s) = e H(X s) λ. () By imposing the constraint we get

4 e λ = y eh(x y), (2) which gives the solution (7). Joint, conditional and marginal entropies with entropy priors are respectively H E (X, S) = log C, H E (X S) = C s H(X s)eh(x s), H E (S) = log C C s H(X s)eh(x s), (3) where C = s eh(x s). Note that the entropies are expressed in nats for easier notation, but the base of the logarithm is totally irrelevant. Entropy priors can be computed numerically when we have empirical or dicrete distributions, or analytically for some canonical densities. In the example that follows we compute the entropy priors for Gaussian likelihoods for scalar variables. A more complete account for other densities will be reported elsewhere. The final fusion formula with entropy priors is then P re {S = s X = x} = 2.. A geometric interpretation f(x S = s)e H(X S=s). (4) f(x S = y)eh(x S=y) y S Entropy priors have a nice geometric interpretation. Recall that the volume of the typical set for an independent sequence of i.i.d. random variables X, X 2,..., X n is A n (X) = e nh(x) where H(X) is the entropy of all the X i. In our case we have that the entropy priors are just the relative volumes of the conditional typical sets, i.e. A n (X S = s) = A n (X S = s) y S A n(x S = y). (5) Likelihoods with larger volumes have the largest entropy priors. 3. Application Example We consider here an application with only one scalar attribute (D = ) presented N times independently and the class to estimate s {c, c 2, c 3 } is constant over the whole set (M = ). The example is extracted from [2] and [7] where the attribute x is an acceleration parameter and the classes are naval targets ({c, c 2, c 3 } = {slowship, f astship, f astraf t}. The various crafts are assumed to have accelerations in the ranges {[ 0.5g, 0.5g], [ g, g], [.5g,.5g]}, respectively. Following the approach taken in [2] and [7], the accelerations in the three cases are assumed to be zero-mean gaussian with σ = g/6, σ 2 = g/3 and σ 3 = g/2. The likelihoods are then N N f(x[],..., x[n] c j ) = f(x[i] c j ) = N (x[i]; 0, σ j ) j =, 2, 3. (6) i= i=

5 We compare the results obtained with the classical Bayesian approach based on uniform priors, with the results obtained from application of Dempster-Shafer theory and with those resulting from the entropic priors. 3.. Classical Bayesian Approach If the priors are assumed to be uniform (Laplace s principle []), the posterior probabilities are N i= P r {S = c j x[],..., x[n]} = N (x[i]; 0, σ j) 3 N l= i= N (x[i]; 0, σ l). (7) 3.2. Dempster-Shafer Approach The DS approach for posterior calculation can be summarized as follows by a four step procedure. A complete descrition of the DS technique is beyond the scope of this paper. We report here only the essential steps in support of the comparison outlined in the following section.. For every likelihood f(x[i] c j ) compute the respective plausibility function P l X[i] (x[i] c j ) corresponding to the q LC isopignistic belief density function [2][5], which in the case of a gaussian density N (x[i]; 0, σ j ) is defined as: P l X[i] (x[i] c j ) = 2y j 2π e y2 j /2 + erfc( y j 2 ) y j = x[i] σ j. 2. For every plausibility function P l X[i] (x[i] c j ) derive the conditional Basic Belief Assignment (BBA) m S (A x[i]) of each possible subset A S (the union of all these subsets is called the Power-set of S) through the Generalized Bayes theorem [3]: m S (A x[i]) = K c P j A lx[i] (x[i] c j ) [ c j Ā P l X[i] (x[i] c j ) ], where K is a normalization constant which is the sum of the BBAs over the whole Power-set. 3. Compute the total BBA m S (A x[],..., x[n]) from the all the evidences m S (A x[i]) given by the Dempster Combination Rule (denoted by ) []: m S (A x[],..., x[n]) = m S (A x[])... m S (A x[n]). 4. From the total BBA m S (A x[],..., x[n]) compute the posterior probabilities P rds {S = c j x[],..., x[n]}, through the Pignistic transformation [4]: P rds {S = c j x[],..., x[n]} = c j A S A ms (A x[],..., x[n]) Bayesian Approach with Entropic Priors For gaussian likelihoods [3] N N H(X S = c j ) = H(X[i] S = c j ) = 2 log 2πeσ2 j, (8) i= and the entropic priors are P re {S = c j } = σ j N 3 l= σn l i=. (9)

6 0 8 6 Acceleration Profile Acceleration Profile of a specific target Posterior Probabilities Profile Classic Bayesian Approach Slow Ship Fast Ship Fast Raft acceleration Time Slot posterior probability Time Slot Posterior Probabilities Profile DS Approach Slow Ship Fast Ship Fast Raft Posterior Probabilities Profile Entropy Prior Bayesian Approach Slow Ship Fast Ship Fast Raft posterior probability posterior probability Time slot Time Slot Figure. Acceleration profile (top left); Posterior probabilities with vaying N up to 34 estimated with: (a) Bayesian approach with uniform priors (top right); (b) Dempster-Shafer theory (bottom left); (c) Bayesian approach with entropic priors (bottom right). Bayes rule becomes P r {S = c j x[],..., x[n]} = σj N 3 l= σn l N i= N (x[i]; 0, σ j) N i= N (x[i]; 0, σ l). (20) 3.4. Simulations To visualize the results of the three approaches, we have simulated the identification process for a target of type c 3 (fast raft) that shows a specific acceleration profile over 34 time slots x = {x[], x[2],..., x[34]}. In particular the target moves at constant speed for 5 time slots, then it accelerates at g for 2 time slots, it decelerates at -g for 2 time slots and finally moves again at constant speed for the remaining 5 time slots. The acceleration profile is shown in Figure (top left). The posterior probabilities are computed with the three approaches for varying N up to 34 and shown in the rest of Figure. As we can see from the figure, the classical Bayesian approach with the uniform priors, in the first 5 slots favours the target with smaller variance, i.e. the target class c. After the 4 slots of manoeuvre the result changes, because the target c cannot reach acceleration magnitude g and its probability rapidly falls to zero. Unfortunately, in the following steps, even though initially target class c 3 is favoured, in the final 5 slots more confidence is definitely given to target class c 2, which is the wrong result.

7 With the DS approach the results are quite different. In the first 5 slots of constant speed the posterior probabilities remain equal, correctly reflecting the ambiguity of the observation. During the 4 slots of manoeuvre, target class c 3 becomes the favorite, for the same reason expressed by the classic Bayesian approach. In the final 5 slots the constant speed doesn t make the probabilities evolve any further correctly reflecting the lack of new information, giving rise then to a correct result. The Bayesian approach, simply by using different priors, reflects all the common sense that seems to be underlying the (complicated) DS approach. The result is strikingly similar and gives also the correct answer. We have carried out other simulations in which our results seem to confirm consistently the validity of the maximum entropy idea. 4. Conclusions In this paper we have shown that in classification problems, our lack of knowledge about the prior distribution, can be effectively substituted with a maximum entropy solution (entropy priors). The entropy method, very popular in many application areas, is yet to show its potential in classification where we seem to find strikingly effective results in many scenarios. In this paper we have derived the entropic prior distribution for a multiclass problem and have applied it to a simple example of identification based on a set of measurements from a scalar variable with gaussian likelihoods. The results are preliminary, but they are very promising, also because the method seems to provide answers that are consistent with our common sense and that have motivated the development of alternatives to probability theory. References [] P. S. Laplace, Théorie analytique des Probabilités, Veuve Courcier, Paris, 82. [2] J. Pearl, Probabilistic Reasoning in Intelligent Systems, 2nd ed. San Francisco: Morgan Kaufmann, 988. [3] T. M. Cover and J. A. Thomas, Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing), 2nd ed.: Wiley-Interscience, [4] E. T. Jaynes, Information Theory and Statistical Mechanics, Phys. Rev., Vol. 06, NO. 4, pp , American Physical Society, May 957. [5] E. T. Jaynes, Information Theory and Statistical Mechanics. II, Phys. Rev., Vol. 08, NO. 2, pp. 7-90, American Physical Society, Oct 957. [6] E. T. Jaynes, Probability Theory : The Logic of Science, Cambridge University Press, [7] A. Caticha and R. Preuss, Maximum Entropy and Bayesian Data Analysis: Entropic Priors, Physical Review, E 70, pp :-2, [8] T. Neumann, Bayesian Inference Featuring Entropic Priors, Bayesian Inference and Maximum Entropy Methods in Science and Engineering: 27th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Vol. 954, NO., pp , AIP, [9] A. Caticha, "Maximum entropy, fluctuations and priors," Bayesian Inference and Maximum Entropy Methods in Science and Engineering: 20th International Workshop. AIP Conference Proceedings, Vol. 568, pp , 200. [0] C. C. Rodriguez, Entropic priors for discrete probabilistic networks and for mixtures of Gaussians models, Bayesian Inference and Maximu Entropy Methods in Science and Engineering. AIP Conference Proceedings, Vol. 67, pp , 2002.

8 [] G. Shafer, A mathematical theory of evidence, Princeton University Press, 976. [2] B. Ristic and P. Smets, Target classification approach based on the belief function theory, IEEE Transactions on Aerospace and Electronic Systems, Vol. 4, NO. 2, pp , April [3] P. Smets, Belief functions: The disjunctive rule of combination and the generalized Bayesian theorem, International Journal of Approximate Reasoning, Vol. 9, NO., pp. - 35, August 993. [4] P. Smets, Decision making in the TBM: the necessity of the pignistic transformation, International Journal of Approximate Reasoning, Vol. 38, NO. 2, pp , February [5] B. Ristic and P. Smets, Belief function theory on the continuous space with an application to model based classification, Proceedings of 0th International Conference IPMU 2004, pp. 9-26, July [6] A. Giffin and A. Caticha, Updating Probabilities with Data and Moments," Bayesian Inference and Maximum Entropy Methods in Science and Engineering: 27th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Vol. 954, NO., pp , AIP, [7] A, Benavoli, L. Chisci, B. Ristic, A. Farina and A. Graziano, Reasoning under uncertainty: from Bayes to Valuation Based Systems. Applications to target classification and threat evaluation, Dipartimento di Sistemi e Informatica, Universitá degli Studi di Firenze, Italy, Selex Sistemi Integrati, Sept. 2007, ISBN:

Tracking and Identification of Multiple targets

Tracking and Identification of Multiple targets Tracking and Identification of Multiple targets Samir Hachour, François Delmotte, Eric Lefèvre, David Mercier Laboratoire de Génie Informatique et d'automatique de l'artois, EA 3926 LGI2A first name.last

More information

Approximation of Belief Functions by Minimizing Euclidean Distances

Approximation of Belief Functions by Minimizing Euclidean Distances Approximation of Belief Functions by Minimizing Euclidean Distances Thomas Weiler and Ulrich Bodenhofer Software Competence Center Hagenberg A-4232 Hagenberg, Austria e-mail: {thomas.weiler,ulrich.bodenhofer}@scch.at

More information

Distributed Classification of Multiple Moving Targets with Binary Wireless Sensor Networks

Distributed Classification of Multiple Moving Targets with Binary Wireless Sensor Networks 14th International Conference on Information Fusion Chicago, Illinois, USA, July 5-8, 2011 Distributed Classification of Multiple Moving argets with Binary Wireless Sensor Networks Domenico Ciuonzo, Aniello

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012 Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

in Evidence Modelling Uncertain Implication Rul~es Theory

in Evidence Modelling Uncertain Implication Rul~es Theory Modelling Uncertain Implication Rul~es Theory in Evidence Alessio Benavoli Engineering Division SELEX Sistemi Integrati Italy benavoli@gmail.com Luigi Chisci DSI Universita di Firenze Italy chisci@dslunifi.it

More information

Integrating Correlated Bayesian Networks Using Maximum Entropy

Integrating Correlated Bayesian Networks Using Maximum Entropy Applied Mathematical Sciences, Vol. 5, 2011, no. 48, 2361-2371 Integrating Correlated Bayesian Networks Using Maximum Entropy Kenneth D. Jarman Pacific Northwest National Laboratory PO Box 999, MSIN K7-90

More information

Bayesian Learning (II)

Bayesian Learning (II) Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

Data Fusion Using a Factor Graph for Ship Tracking in Harbour Scenarios

Data Fusion Using a Factor Graph for Ship Tracking in Harbour Scenarios Data Fusion Using a Factor Graph for Ship Tracing in Harbour Scenarios Francesco Castaldo Francesco A.N. Palmieri Dipartimento di Ingegneria Industriale e dell Informazione Seconda Universitá di Napoli

More information

Probability, Entropy, and Inference / More About Inference

Probability, Entropy, and Inference / More About Inference Probability, Entropy, and Inference / More About Inference Mário S. Alvim (msalvim@dcc.ufmg.br) Information Theory DCC-UFMG (2018/02) Mário S. Alvim (msalvim@dcc.ufmg.br) Probability, Entropy, and Inference

More information

Probabilistic Graphical Models for Image Analysis - Lecture 1

Probabilistic Graphical Models for Image Analysis - Lecture 1 Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.

More information

12 : Variational Inference I

12 : Variational Inference I 10-708: Probabilistic Graphical Models, Spring 2015 12 : Variational Inference I Lecturer: Eric P. Xing Scribes: Fattaneh Jabbari, Eric Lei, Evan Shapiro 1 Introduction Probabilistic inference is one of

More information

8 Basics of Hypothesis Testing

8 Basics of Hypothesis Testing 8 Basics of Hypothesis Testing 4 Problems Problem : The stochastic signal S is either 0 or E with equal probability, for a known value E > 0. Consider an observation X = x of the stochastic variable X

More information

Contradiction Measures and Specificity Degrees of Basic Belief Assignments

Contradiction Measures and Specificity Degrees of Basic Belief Assignments Contradiction Measures and Specificity Degrees of Basic Belief Assignments Florentin Smarandache Arnaud Martin Christophe Osswald Originally published as: Smarandache F., Martin A., Osswald C - Contradiction

More information

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,

More information

Handling imprecise and uncertain class labels in classification and clustering

Handling imprecise and uncertain class labels in classification and clustering Handling imprecise and uncertain class labels in classification and clustering Thierry Denœux 1 1 Université de Technologie de Compiègne HEUDIASYC (UMR CNRS 6599) COST Action IC 0702 Working group C, Mallorca,

More information

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD Bayesian decision theory Nuno Vasconcelos ECE Department, UCSD Notation the notation in DHS is quite sloppy e.g. show that ( error = ( error z ( z dz really not clear what this means we will use the following

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Introduction. Basic Probability and Bayes Volkan Cevher, Matthias Seeger Ecole Polytechnique Fédérale de Lausanne 26/9/2011 (EPFL) Graphical Models 26/9/2011 1 / 28 Outline

More information

Data Fusion with Imperfect Implication Rules

Data Fusion with Imperfect Implication Rules Data Fusion with Imperfect Implication Rules J. N. Heendeni 1, K. Premaratne 1, M. N. Murthi 1 and M. Scheutz 2 1 Elect. & Comp. Eng., Univ. of Miami, Coral Gables, FL, USA, j.anuja@umiami.edu, kamal@miami.edu,

More information

A hierarchical fusion of expert opinion in the Transferable Belief Model (TBM) Minh Ha-Duong, CNRS, France

A hierarchical fusion of expert opinion in the Transferable Belief Model (TBM) Minh Ha-Duong, CNRS, France Ambiguity, uncertainty and climate change, UC Berkeley, September 17-18, 2009 A hierarchical fusion of expert opinion in the Transferable Belief Model (TBM) Minh Ha-Duong, CNRS, France Outline 1. Intro:

More information

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix

More information

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability

More information

BAYESIAN MULTI-TARGET TRACKING WITH SUPERPOSITIONAL MEASUREMENTS USING LABELED RANDOM FINITE SETS. Francesco Papi and Du Yong Kim

BAYESIAN MULTI-TARGET TRACKING WITH SUPERPOSITIONAL MEASUREMENTS USING LABELED RANDOM FINITE SETS. Francesco Papi and Du Yong Kim 3rd European Signal Processing Conference EUSIPCO BAYESIAN MULTI-TARGET TRACKING WITH SUPERPOSITIONAL MEASUREMENTS USING LABELED RANDOM FINITE SETS Francesco Papi and Du Yong Kim Department of Electrical

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Introduction: MLE, MAP, Bayesian reasoning (28/8/13) STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this

More information

Intelligent Systems I

Intelligent Systems I Intelligent Systems I 00 INTRODUCTION Stefan Harmeling & Philipp Hennig 24. October 2013 Max Planck Institute for Intelligent Systems Dptmt. of Empirical Inference Which Card? Opening Experiment Which

More information

SCUOLA DI SPECIALIZZAZIONE IN FISICA MEDICA. Sistemi di Elaborazione dell Informazione. Regressione. Ruggero Donida Labati

SCUOLA DI SPECIALIZZAZIONE IN FISICA MEDICA. Sistemi di Elaborazione dell Informazione. Regressione. Ruggero Donida Labati SCUOLA DI SPECIALIZZAZIONE IN FISICA MEDICA Sistemi di Elaborazione dell Informazione Regressione Ruggero Donida Labati Dipartimento di Informatica via Bramante 65, 26013 Crema (CR), Italy http://homes.di.unimi.it/donida

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Belief and Desire: On Information and its Value

Belief and Desire: On Information and its Value Belief and Desire: On Information and its Value Ariel Caticha Department of Physics University at Albany SUNY ariel@albany.edu Info-Metrics Institute 04/26/2013 1 Part 1: Belief 2 What is information?

More information

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem

More information

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction 15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive

More information

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

PILCO: A Model-Based and Data-Efficient Approach to Policy Search PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol

More information

Unsupervised Learning with Permuted Data

Unsupervised Learning with Permuted Data Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

Index. admissible matrix, 82, 83 admissible perturbation, 72 argument, 23, 24, 49, 57, 98

Index. admissible matrix, 82, 83 admissible perturbation, 72 argument, 23, 24, 49, 57, 98 References 1. RH. Bartlett, D.W. Roloff, RC. Cornell, A.F. Andrews, P.W. Dillon, and J.B. Zwischenberger, Extracorporeal Circulation in Neonatal Respiratory Failure: A Prospective Randomized Study, Pediatrics

More information

Joint Tracking and Classification of Airbourne Objects using Particle Filters and the Continuous Transferable Belief Model

Joint Tracking and Classification of Airbourne Objects using Particle Filters and the Continuous Transferable Belief Model Joint Tracking and Classification of Airbourne Objects using Particle Filters and the Continuous Transferable Belief Model Gavin Powell & David Marshall The Geometric Computing & Computer Vision Group,

More information

Lecture : Probabilistic Machine Learning

Lecture : Probabilistic Machine Learning Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Combining Belief Functions Issued from Dependent Sources

Combining Belief Functions Issued from Dependent Sources Combining Belief Functions Issued from Dependent Sources MARCO E.G.V. CATTANEO ETH Zürich, Switzerland Abstract Dempster s rule for combining two belief functions assumes the independence of the sources

More information

Multivariate statistical methods and data mining in particle physics

Multivariate statistical methods and data mining in particle physics Multivariate statistical methods and data mining in particle physics RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement of the problem Some general

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Intelligent Systems Statistical Machine Learning

Intelligent Systems Statistical Machine Learning Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2015/2016, Our model and tasks The model: two variables are usually present: - the first one is typically discrete

More information

Neutron inverse kinetics via Gaussian Processes

Neutron inverse kinetics via Gaussian Processes Neutron inverse kinetics via Gaussian Processes P. Picca Politecnico di Torino, Torino, Italy R. Furfaro University of Arizona, Tucson, Arizona Outline Introduction Review of inverse kinetics techniques

More information

Predictive Variance Reduction Search

Predictive Variance Reduction Search Predictive Variance Reduction Search Vu Nguyen, Sunil Gupta, Santu Rana, Cheng Li, Svetha Venkatesh Centre of Pattern Recognition and Data Analytics (PRaDA), Deakin University Email: v.nguyen@deakin.edu.au

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality

More information

The Limitation of Bayesianism

The Limitation of Bayesianism The Limitation of Bayesianism Pei Wang Department of Computer and Information Sciences Temple University, Philadelphia, PA 19122 pei.wang@temple.edu Abstract In the current discussion about the capacity

More information

13: Variational inference II

13: Variational inference II 10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Discrete Mathematics and Probability Theory Fall 2015 Lecture 21

Discrete Mathematics and Probability Theory Fall 2015 Lecture 21 CS 70 Discrete Mathematics and Probability Theory Fall 205 Lecture 2 Inference In this note we revisit the problem of inference: Given some data or observations from the world, what can we infer about

More information

Information Theory in Intelligent Decision Making

Information Theory in Intelligent Decision Making Information Theory in Intelligent Decision Making Adaptive Systems and Algorithms Research Groups School of Computer Science University of Hertfordshire, United Kingdom June 7, 2015 Information Theory

More information

Learning Gaussian Process Models from Uncertain Data

Learning Gaussian Process Models from Uncertain Data Learning Gaussian Process Models from Uncertain Data Patrick Dallaire, Camille Besse, and Brahim Chaib-draa DAMAS Laboratory, Computer Science & Software Engineering Department, Laval University, Canada

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes

More information

an introduction to bayesian inference

an introduction to bayesian inference with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena

More information

Decision theory. 1 We may also consider randomized decision rules, where δ maps observed data D to a probability distribution over

Decision theory. 1 We may also consider randomized decision rules, where δ maps observed data D to a probability distribution over Point estimation Suppose we are interested in the value of a parameter θ, for example the unknown bias of a coin. We have already seen how one may use the Bayesian method to reason about θ; namely, we

More information

Solving Classification Problems By Knowledge Sets

Solving Classification Problems By Knowledge Sets Solving Classification Problems By Knowledge Sets Marcin Orchel a, a Department of Computer Science, AGH University of Science and Technology, Al. A. Mickiewicza 30, 30-059 Kraków, Poland Abstract We propose

More information

Review: Statistical Model

Review: Statistical Model Review: Statistical Model { f θ :θ Ω} A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced the data. The statistical model

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Density Approximation Based on Dirac Mixtures with Regard to Nonlinear Estimation and Filtering

Density Approximation Based on Dirac Mixtures with Regard to Nonlinear Estimation and Filtering Density Approximation Based on Dirac Mixtures with Regard to Nonlinear Estimation and Filtering Oliver C. Schrempf, Dietrich Brunn, Uwe D. Hanebeck Intelligent Sensor-Actuator-Systems Laboratory Institute

More information

An introduction to Bayesian reasoning in particle physics

An introduction to Bayesian reasoning in particle physics An introduction to Bayesian reasoning in particle physics Graduiertenkolleg seminar, May 15th 2013 Overview Overview A concrete example Scientific reasoning Probability and the Bayesian interpretation

More information

Lecture Notes 1: Decisions and Data. In these notes, I describe some basic ideas in decision theory. theory is constructed from

Lecture Notes 1: Decisions and Data. In these notes, I describe some basic ideas in decision theory. theory is constructed from Topics in Data Analysis Steven N. Durlauf University of Wisconsin Lecture Notes : Decisions and Data In these notes, I describe some basic ideas in decision theory. theory is constructed from The Data:

More information

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm 1/29 EM & Latent Variable Models Gaussian Mixture Models EM Theory The Expectation-Maximization Algorithm Mihaela van der Schaar Department of Engineering Science University of Oxford MLE for Latent Variable

More information

Probability and Information Theory. Sargur N. Srihari

Probability and Information Theory. Sargur N. Srihari Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal

More information

Belief Update in CLG Bayesian Networks With Lazy Propagation

Belief Update in CLG Bayesian Networks With Lazy Propagation Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian

More information

Dimension Reduction (PCA, ICA, CCA, FLD,

Dimension Reduction (PCA, ICA, CCA, FLD, Dimension Reduction (PCA, ICA, CCA, FLD, Topic Models) Yi Zhang 10-701, Machine Learning, Spring 2011 April 6 th, 2011 Parts of the PCA slides are from previous 10-701 lectures 1 Outline Dimension reduction

More information

Probabilistic Machine Learning. Industrial AI Lab.

Probabilistic Machine Learning. Industrial AI Lab. Probabilistic Machine Learning Industrial AI Lab. Probabilistic Linear Regression Outline Probabilistic Classification Probabilistic Clustering Probabilistic Dimension Reduction 2 Probabilistic Linear

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Lecture 4: Probabilistic Learning

Lecture 4: Probabilistic Learning DD2431 Autumn, 2015 1 Maximum Likelihood Methods Maximum A Posteriori Methods Bayesian methods 2 Classification vs Clustering Heuristic Example: K-means Expectation Maximization 3 Maximum Likelihood Methods

More information

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Intelligent Systems Statistical Machine Learning

Intelligent Systems Statistical Machine Learning Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2014/2015, Our tasks (recap) The model: two variables are usually present: - the first one is typically discrete k

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

DS-Based Uncertain Implication Rules for Inference and Fusion Applications

DS-Based Uncertain Implication Rules for Inference and Fusion Applications DS-Based Uncertain Implication Rules for Inference and Fusion Applications Rafael C. Núñez, Ranga Dabarera, Matthias Scheutz, Otávio Bueno, Kamal Premaratne, Manohar N. Murthi Department of Electrical

More information

Is Entropy Enough to Evaluate the Probability Transformation Approach of Belief Function?

Is Entropy Enough to Evaluate the Probability Transformation Approach of Belief Function? Is Entropy Enough to Evaluate the Probability Transformation Approach of Belief Function? Deqiang Han Jean Dezert Chongzhao Han Yi Yang Originally published as Han D., Dezert J., Han C., Is Entropy Enough

More information

Nonparametric Density Estimation. October 1, 2018

Nonparametric Density Estimation. October 1, 2018 Nonparametric Density Estimation October 1, 2018 Introduction If we can t fit a distribution to our data, then we use nonparametric density estimation. Start with a histogram. But there are problems with

More information

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability

More information

Reasoning with Uncertainty

Reasoning with Uncertainty Reasoning with Uncertainty Representing Uncertainty Manfred Huber 2005 1 Reasoning with Uncertainty The goal of reasoning is usually to: Determine the state of the world Determine what actions to take

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

Bayesian Support Vector Machines for Feature Ranking and Selection

Bayesian Support Vector Machines for Feature Ranking and Selection Bayesian Support Vector Machines for Feature Ranking and Selection written by Chu, Keerthi, Ong, Ghahramani Patrick Pletscher pat@student.ethz.ch ETH Zurich, Switzerland 12th January 2006 Overview 1 Introduction

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information

2. A Basic Statistical Toolbox

2. A Basic Statistical Toolbox . A Basic Statistical Toolbo Statistics is a mathematical science pertaining to the collection, analysis, interpretation, and presentation of data. Wikipedia definition Mathematical statistics: concerned

More information

The Maximum Entropy Principle and Applications to MIMO Channel Modeling

The Maximum Entropy Principle and Applications to MIMO Channel Modeling The Maximum Entropy Principle and Applications to MIMO Channel Modeling Eurecom CM Talk, 16/02/2006 Maxime Guillaud maxime.guillaud@eurecom.fr (Joint work with Mérouane Debbah) The Maximum Entropy Principle

More information

Hierarchical Proportional Redistribution Principle for Uncertainty Reduction and BBA Approximation

Hierarchical Proportional Redistribution Principle for Uncertainty Reduction and BBA Approximation Hierarchical Proportional Redistribution Principle for Uncertainty Reduction and BBA Approximation Jean Dezert Deqiang Han Zhun-ga Liu Jean-Marc Tacnet Abstract Dempster-Shafer evidence theory is very

More information

Support Vector Machines and Bayes Regression

Support Vector Machines and Bayes Regression Statistical Techniques in Robotics (16-831, F11) Lecture #14 (Monday ctober 31th) Support Vector Machines and Bayes Regression Lecturer: Drew Bagnell Scribe: Carl Doersch 1 1 Linear SVMs We begin by considering

More information

Introduction to Bayesian Inference

Introduction to Bayesian Inference Introduction to Bayesian Inference p. 1/2 Introduction to Bayesian Inference September 15th, 2010 Reading: Hoff Chapter 1-2 Introduction to Bayesian Inference p. 2/2 Probability: Measurement of Uncertainty

More information

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018 CLASS NOTES Models, Algorithms and Data: Introduction to computing 208 Petros Koumoutsakos, Jens Honore Walther (Last update: June, 208) IMPORTANT DISCLAIMERS. REFERENCES: Much of the material (ideas,

More information

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions DD2431 Autumn, 2014 1 2 3 Classification with Probability Distributions Estimation Theory Classification in the last lecture we assumed we new: P(y) Prior P(x y) Lielihood x2 x features y {ω 1,..., ω K

More information

Lecture 2: Basic Concepts of Statistical Decision Theory

Lecture 2: Basic Concepts of Statistical Decision Theory EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture

More information

Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles

Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles Myungsoo Jun and Raffaello D Andrea Sibley School of Mechanical and Aerospace Engineering Cornell University

More information

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER Zhen Zhen 1, Jun Young Lee 2, and Abdus Saboor 3 1 Mingde College, Guizhou University, China zhenz2000@21cn.com 2 Department

More information