Bayesian Social Learning with Random Decision Making in Sequential Systems

Size: px
Start display at page:

Download "Bayesian Social Learning with Random Decision Making in Sequential Systems"

Transcription

1 Bayesian Social Learning with Random Decision Making in Sequential Systems Yunlong Wang supervised by Petar M. Djurić Department of Electrical and Computer Engineering Stony Brook University Stony Brook, NY 11794, USA September 15, 2014 Yunlong Wang Bayesian Social Learning with Random Decision Making 1/37

2 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 2/37

3 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 3/37

4 Interest We consider a large number of agents that get observations from an hypothesis that can be either H 0 and H 1. In a predefined order, they make decisions sequentially between the two hypotheses. The decisions of agents are made according to their beliefs on the true state of nature. For each agent, its decision can be observed by all the agents. The agents learn from the previous agents decisions and formulate their beliefs by using the Bayesian theory. Yunlong Wang Bayesian Social Learning with Random Decision Making 4/37

5 Interest Decision making in the proposed sequential system Yunlong Wang Bayesian Social Learning with Random Decision Making 5/37

6 Interest We want to find weather herd behavior and information cascade can take place in this system. We want to show that asymptotic learning can be achieved, i.e., that the probability an agent makes right decision converges to one. Yunlong Wang Bayesian Social Learning with Random Decision Making 6/37

7 Motivation An example of herd behavior Yunlong Wang Bayesian Social Learning with Random Decision Making 7/37

8 Our argument A large body of literature investigates this problem with the purpose of maximizing some utility function, where the agents are assumed to be selfish. Then once the belief of one selfish agent has been formed, the decision will be made deterministically. However, in some scenarios, the behaviors of agents may be by design random. For example, there is a chance that an agent refuses to behave as predicted because of some random disturbance such as mood, misleading information, etc. Our contribution lies in that we analyze the properties of such systems when the agents behave in a random manner. Yunlong Wang Bayesian Social Learning with Random Decision Making 8/37

9 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 9/37

10 The setup A network of N Bayesian agents make decisions between two hypotheses H 0 and H 1 in a predefined order. Each agent has its private observation generated either by H 1 or H 0. Every agent s decision can be observed by all the following agents. The agents form their beliefs on the state of nature by using Bayes rule. Yunlong Wang Bayesian Social Learning with Random Decision Making 10/37

11 The network model The sequential system can be depicted by the following diagram: Yunlong Wang Bayesian Social Learning with Random Decision Making 11/37

12 The observation model Each agent A n receives its private observation y n that is generated according to one of the following two hypotheses: H 1 : y n p(y n H 1 ) H 0 : y n p(y n H 0 ). Let w n (k) = p(y n H k ), k {0, 1} be the data likelihood for each hypothesis, then the log(w n (1) /w n (0) ) represents the log likelihood ratio (LLR) of data y n. The boundedness of the LLR will play an important role in social learning: a bounded LLR can be overwhelmed by a strong prior. Yunlong Wang Bayesian Social Learning with Random Decision Making 12/37

13 Let α n {0, 1} be the decision of agent A n and α 1:n be the decision sequence from A 1 to A n. Then agent A n can form its private belief β n = p(h 1 α 1:n 1, y n ) by using the Bayes rule: β n = π n 1 w n (1) π n 1 w n (1) + (1 π n 1 )w n (0) where π n 1 = p(h k α 1:n 1 ) denotes the social belief. The social belief π n 1 serves as the prior knowledge of agent A n before it has its private observation y n. For agent A 1, the initial social belief is defined to be π 0 = 1/2. Yunlong Wang Bayesian Social Learning with Random Decision Making 13/37

14 The decision policy Agent A n makes its decision from a certain distribution conditioned on its private belief β n, i.e. α n p(α n β n ). For example, agents can make decision 1 whenever β n > 0.5 and 0 otherwise. In this work, we address a specific random decision making policy. We propose that after agent A n obtaining its private belief, it makes its decision according to, { β n, if α n = 1, p(α n β n ) = 1 β n, O.W., where the decision α n of A n is a Bernoulli random variable parameterized by the A n s private belief β n. Yunlong Wang Bayesian Social Learning with Random Decision Making 14/37

15 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 15/37

16 Bayesian learning Learning from actions: Social belief update In this system, once agent A n makes its decision α n, all the following agents should update the social belief by Bayes rule from π n 1 = p(h 1 α 1:n 1 ) to p(h 1 α 1:n ). The social belief is updated according to π n = ( where k {0, 1}, π n 1 (1 l n (1) ) π n 1 (1 l n (1) ) + (1 π n 1 )(1 l (0) ( denotes the action likelihood. π n 1 l n (1) π n 1 l n (1) + (1 π n 1 )l n (0) l (k) n = p(α n = 1 α 1:n 1, H k ) n ) ) αn, ) 1 αn Yunlong Wang Bayesian Social Learning with Random Decision Making 16/37

17 Bayesian learning The action likelihood By marginalizing the private observation y n, it can be shown that, l (k) n = p(α n = 1 α 1:n 1, y n )p(y n H k )dy n. We point out that the above sequential learning algorithm belongs to the social learning filter. Yunlong Wang Bayesian Social Learning with Random Decision Making 17/37

18 Information cascade and herds Definitions We start the analysis on the proposed system with the two definitions: An agent A n herds on the public belief π n 1 if it makes its decision α n independently of its observation y n, i.e., p(α n = 1 α 1:n 1, y n ) = p(α n = 1 α 1:n 1 ). An information cascade occurs at time ñ if the public belief stops evolving after agent Añ, i.e., π n = πñ, n > ñ. Yunlong Wang Bayesian Social Learning with Random Decision Making 18/37

19 Information cascade and herds The action likelihood and herd behavior If agent A n herds on the public belief π n 1, then the action likelihood will be independent of the true state of nature, i.e., l n (0) = l n (1), where l n (k) = p(α n = 1 α 1:n 1, H k ). If l (0) n = l (1) n, it holds that π n = π n 1. If every agent after agent Añ herds on the public belief, there is an information cascade starting with the agent Añ. If l n (0) l n (1), the agent A n must not herd on the public belief, and its private information is partially conveyed by its decision α n. Yunlong Wang Bayesian Social Learning with Random Decision Making 19/37

20 Information cascade and herds The action likelihood and herd behavior(cont.) When the agents make their decisions according to the proposed model, the action likelihood can be written as l n (k) π n 1 w n (1) = π n 1 w n (1) + (1 π n 1 )w n (0) w n (k) dy n, by which we have the following lemma: Lemma With the proposed random policy, if the two distributions w n (1) and w n (0) are not identical everywhere and if 1 > π n > 0, n N +, we have that l n (1) l n (0) 0, n N +. The above lemma shows that no agent herds on the public belief. Yunlong Wang Bayesian Social Learning with Random Decision Making 20/37

21 Asymptotical properties Asymptotical learning In sum, we have shown that the social belief never stops evolving. We next show that the social belief will converge to the true state of nature by the following theorem: Theorem In the proposed sequential random decision making system, the expected value of social belief asymptotically converge to 1 when the true state of nature is H s = H 1, and 0 otherwise, i.e., lim Eπ n = n { 1, if H s = H 1, 0, if H s = H 0. Yunlong Wang Bayesian Social Learning with Random Decision Making 21/37

22 Asymptotical properties Asymptotical learning(cont.) Considering the symmetric structure of the proposed system, we assume that the true state of nature is H 1. Then it can be shown that Eπ n is increasing in terms of n, which is given by Eπ n Eπ n 1 = p(α 1:n 1 H 1 ) (α 1:n 1 ), α 1:n 1 A n 1 where (α 1:n 1 ) = π n 1 Z (1 π n 1) 2 (l (1) n l (0) n ) 2 > 0. By the above equations, the limit of Eπ n exists. Noting that l n (1) l n (0), the Eπ n Eπ n 1 is strictly greater than zero unless Eπ n = 1. Yunlong Wang Bayesian Social Learning with Random Decision Making 22/37

23 Asymptotical properties Asymptotical learning(cont.) Theorem Suppose that the agents implement the random decision policy. Let H s be the true state of nature, then s {0, 1}, we have lim p(α n = s) = 1. n The probability that the agents make the right decision converges to one even if the LLR is bounded. Yunlong Wang Bayesian Social Learning with Random Decision Making 23/37

24 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 24/37

25 Introduction In this part we want to further analyze the proposed random system by comparison with other two systems. First we describe the three systems, and then we explain some of their features. Due to lack of time, we do not show the detailed derivations on the evolution of beliefs in those systems. Yunlong Wang Bayesian Social Learning with Random Decision Making 25/37

26 Deterministic policy In cases where the agents implement the deterministic policy, for agent A n, its decision is made according to { 1, if β n > 0.5, α n = 0, O.W., By using this deterministic policy, the agents myopically optimize their personal utility. Yunlong Wang Bayesian Social Learning with Random Decision Making 26/37

27 Deterministic policy(cont.) When the LLR is bounded, the following statements hold: The space of social belief S = [0, 1] can be divided into three non empty subsets S 0,S r, and S 1. Once the social belief evolves into S 0 or S 1, all the following agents will make decision as zero or one with probability one, and information cascade will start. Social belief evolves to S 0 or S 1 in finite time with probability one. On the other hand, if the LLR is unbounded, we have that the error probability will converge to zero. Yunlong Wang Bayesian Social Learning with Random Decision Making 27/37

28 Random policy Compared with the system that uses deterministic policy, in a random system even if the LLR is bounded, the region S 0 and S 1 do not exist. This property prevents the information cascade. As was shown by Lemma 1, no agent ignores its private observation no matter what the social belief it uses as its prior in forming its private belief. Yunlong Wang Bayesian Social Learning with Random Decision Making 28/37

29 Hybrid policy We also propose that the agent A n can make its decision by a hybrid policy, where 1, if β n > c, α n = R n, if c > β n > 1 c, 0, O.W., with R n Ber(β n ) being a Bernoulli random variable parameterized by β n, and with 1 > c > 0.5 denoting a threshold for the agent to make a decision either randomly or deterministically. The deterministic and random policies are two special cases of the hybrid policy when c = 0.5 and c = 1, respectively. The asymptotical property of this policy is identical with that of the deterministic policy. Yunlong Wang Bayesian Social Learning with Random Decision Making 29/37

30 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 30/37

31 Experiment 1: Binomial model Consider the data model given by H 1 : y n B(m, p 1 ), H 0 : y n B(m, p 0 ), where B(m, p) denotes a binomial distribution parameterized by m and p. It can be shown that the LLR of this model is bounded. In the first experiment, we verified the analytical result of the expected social belief by implementing the three policies with the binomial data for 2,000 trials. In each trial, the parameters were set to be N = 1, 500, m = 5, p 0 = 0.6, and p 1 = 0.7. When using the hybrid policy, we set c = Yunlong Wang Bayesian Social Learning with Random Decision Making 31/37

32 Result 1: Evolution of social belief with the binomial model Average social belief from all the 2,000 trials plotted as a function of agent number for three policies Eπ n deterministric hybrid random agent number Yunlong Wang Bayesian Social Learning with Random Decision Making 32/37

33 Experiment 2: Gaussian model Consider a system where the agents get observations from the Guassian data model: H 1 : y n N (θ 1, σ 2 w ), H 0 : y n N (θ 0, σ 2 w ), where y n is a Gaussian random variable parameterized by θ 0 = 0 or θ 1 = θ. The parameters θ and σ w are known by the agents. The LLR of this Gaussian data model is unbounded. Same as in the first experiment, in experiment 2, we implemented the three policies with the Gaussian data for 2000 trials. The parameters here were set to be N = 1, 500, σ w = 2, θ = 1 and c = Yunlong Wang Bayesian Social Learning with Random Decision Making 33/37

34 Result 2: Evolution of social belief with Gaussian model The convergence of expected social belief can be achieved by all three policies Eπ n deterministric hybrid random agent number Yunlong Wang Bayesian Social Learning with Random Decision Making 34/37

35 Experiment 3: Outliers Evolution of social belief with the sequence of decisions where α 1:40 all equal to one except that α 10 = π n deterministic method, α 10 =0 hybrid method, α 10 =0 random method, α 10 = agent number Yunlong Wang Bayesian Social Learning with Random Decision Making 35/37

36 Outline Introduction Problem formulation Analysis Comparisons with other systems Simulation results Conclusions Yunlong Wang Bayesian Social Learning with Random Decision Making 36/37

37 Concluding remarks We presented a random model for the social learning system. The agents implement the Bayesian method to learn from previous agents actions and the private signal. After analyzing the herd behavior and information cascade, we showed the convergence of social belief. We demonstrated the performance of the proposed method by applying it to the binomial and Gaussian data models. We used Monte Carlo simulations to verify the analytical results. Yunlong Wang Bayesian Social Learning with Random Decision Making 37/37

Bayesian Learning in Social Networks

Bayesian Learning in Social Networks Bayesian Learning in Social Networks Asu Ozdaglar Joint work with Daron Acemoglu, Munther Dahleh, Ilan Lobel Department of Electrical Engineering and Computer Science, Department of Economics, Operations

More information

Preliminary Results on Social Learning with Partial Observations

Preliminary Results on Social Learning with Partial Observations Preliminary Results on Social Learning with Partial Observations Ilan Lobel, Daron Acemoglu, Munther Dahleh and Asuman Ozdaglar ABSTRACT We study a model of social learning with partial observations from

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics Bayesian Parameter Estimation Introduction to Bayesian Statistics Harvey Thornburg Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University Stanford, California

More information

Statistical Learning. Philipp Koehn. 10 November 2015

Statistical Learning. Philipp Koehn. 10 November 2015 Statistical Learning Philipp Koehn 10 November 2015 Outline 1 Learning agents Inductive learning Decision tree learning Measuring learning performance Bayesian learning Maximum a posteriori and maximum

More information

Lecture : Probabilistic Machine Learning

Lecture : Probabilistic Machine Learning Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning

More information

Rate of Convergence of Learning in Social Networks

Rate of Convergence of Learning in Social Networks Rate of Convergence of Learning in Social Networks Ilan Lobel, Daron Acemoglu, Munther Dahleh and Asuman Ozdaglar Massachusetts Institute of Technology Cambridge, MA Abstract We study the rate of convergence

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability

More information

BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS

BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS Petar M. Djurić Dept. of Electrical & Computer Engineering Stony Brook University Stony Brook, NY 11794, USA e-mail: petar.djuric@stonybrook.edu

More information

Statistical learning. Chapter 20, Sections 1 3 1

Statistical learning. Chapter 20, Sections 1 3 1 Statistical learning Chapter 20, Sections 1 3 Chapter 20, Sections 1 3 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Statistical learning. Chapter 20, Sections 1 4 1

Statistical learning. Chapter 20, Sections 1 4 1 Statistical learning Chapter 20, Sections 1 4 Chapter 20, Sections 1 4 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model Logis@cs CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Assignment

More information

CS 6140: Machine Learning Spring 2016

CS 6140: Machine Learning Spring 2016 CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa?on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Logis?cs Assignment

More information

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida Bayesian Statistical Methods Jeff Gill Department of Political Science, University of Florida 234 Anderson Hall, PO Box 117325, Gainesville, FL 32611-7325 Voice: 352-392-0262x272, Fax: 352-392-8127, Email:

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Performance Evaluation and Comparison

Performance Evaluation and Comparison Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation

More information

Social Learning with Endogenous Network Formation

Social Learning with Endogenous Network Formation Social Learning with Endogenous Network Formation Yangbo Song November 10, 2014 Abstract I study the problem of social learning in a model where agents move sequentially. Each agent receives a private

More information

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish

More information

Learning with Probabilities

Learning with Probabilities Learning with Probabilities CS194-10 Fall 2011 Lecture 15 CS194-10 Fall 2011 Lecture 15 1 Outline Bayesian learning eliminates arbitrary loss functions and regularizers facilitates incorporation of prior

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Decision Theory: Q-Learning

Decision Theory: Q-Learning Decision Theory: Q-Learning CPSC 322 Decision Theory 5 Textbook 12.5 Decision Theory: Q-Learning CPSC 322 Decision Theory 5, Slide 1 Lecture Overview 1 Recap 2 Asynchronous Value Iteration 3 Q-Learning

More information

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Generative Models Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1

More information

Bayesian RL Seminar. Chris Mansley September 9, 2008

Bayesian RL Seminar. Chris Mansley September 9, 2008 Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Valen E. Johnson Texas A&M University February 27, 2014 Valen E. Johnson Texas A&M University Uniformly most powerful Bayes

More information

Costly Social Learning and Rational Inattention

Costly Social Learning and Rational Inattention Costly Social Learning and Rational Inattention Srijita Ghosh Dept. of Economics, NYU September 19, 2016 Abstract We consider a rationally inattentive agent with Shannon s relative entropy cost function.

More information

Statistical learning. Chapter 20, Sections 1 3 1

Statistical learning. Chapter 20, Sections 1 3 1 Statistical learning Chapter 20, Sections 1 3 Chapter 20, Sections 1 3 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete

More information

From inductive inference to machine learning

From inductive inference to machine learning From inductive inference to machine learning ADAPTED FROM AIMA SLIDES Russel&Norvig:Artificial Intelligence: a modern approach AIMA: Inductive inference AIMA: Inductive inference 1 Outline Bayesian inferences

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and Estimation I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 22, 2015

More information

A REVERSE TO THE JEFFREYS LINDLEY PARADOX

A REVERSE TO THE JEFFREYS LINDLEY PARADOX PROBABILITY AND MATHEMATICAL STATISTICS Vol. 38, Fasc. 1 (2018), pp. 243 247 doi:10.19195/0208-4147.38.1.13 A REVERSE TO THE JEFFREYS LINDLEY PARADOX BY WIEBE R. P E S T M A N (LEUVEN), FRANCIS T U E R

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

1 Bayesian Linear Regression (BLR)

1 Bayesian Linear Regression (BLR) Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,

More information

Mixture of Gaussians Models

Mixture of Gaussians Models Mixture of Gaussians Models Outline Inference, Learning, and Maximum Likelihood Why Mixtures? Why Gaussians? Building up to the Mixture of Gaussians Single Gaussians Fully-Observed Mixtures Hidden Mixtures

More information

Hierarchical Models & Bayesian Model Selection

Hierarchical Models & Bayesian Model Selection Hierarchical Models & Bayesian Model Selection Geoffrey Roeder Departments of Computer Science and Statistics University of British Columbia Jan. 20, 2016 Contact information Please report any typos or

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning

More information

NBER WORKING PAPER SERIES BAYESIAN LEARNING IN SOCIAL NETWORKS. Daron Acemoglu Munther A. Dahleh Ilan Lobel Asuman Ozdaglar

NBER WORKING PAPER SERIES BAYESIAN LEARNING IN SOCIAL NETWORKS. Daron Acemoglu Munther A. Dahleh Ilan Lobel Asuman Ozdaglar NBER WORKING PAPER SERIES BAYESIAN LEARNING IN SOCIAL NETWORKS Daron Acemoglu Munther A. Dahleh Ilan Lobel Asuman Ozdaglar Working Paper 14040 http://www.nber.org/papers/w14040 NATIONAL BUREAU OF ECONOMIC

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Empirical Bayes, Hierarchical Bayes Mark Schmidt University of British Columbia Winter 2017 Admin Assignment 5: Due April 10. Project description on Piazza. Final details coming

More information

Computational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept

Computational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept Computational Biology Lecture #3: Probability and Statistics Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept 26 2005 L2-1 Basic Probabilities L2-2 1 Random Variables L2-3 Examples

More information

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 CS students: don t forget to re-register in CS-535D. Even if you just audit this course, please do register.

More information

10. Composite Hypothesis Testing. ECE 830, Spring 2014

10. Composite Hypothesis Testing. ECE 830, Spring 2014 10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters

More information

4.1 The basic model of herding

4.1 The basic model of herding 72 Monopoly pricing of a new good A monopoly who captures some consumer surplus will take into account the benefit of experimentation for the future. This problem is considered in Section 4.5.2. A monopoly

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Estimation of Dynamic Regression Models

Estimation of Dynamic Regression Models University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.

More information

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

Statistical Methods in Particle Physics Lecture 1: Bayesian methods

Statistical Methods in Particle Physics Lecture 1: Bayesian methods Statistical Methods in Particle Physics Lecture 1: Bayesian methods SUSSP65 St Andrews 16 29 August 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan

More information

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction 15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive

More information

Bayesian inference. Justin Chumbley ETH and UZH. (Thanks to Jean Denizeau for slides)

Bayesian inference. Justin Chumbley ETH and UZH. (Thanks to Jean Denizeau for slides) Bayesian inference Justin Chumbley ETH and UZH (Thanks to Jean Denizeau for slides) Overview of the talk Introduction: Bayesian inference Bayesian model comparison Group-level Bayesian model selection

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01 STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01 Nasser Sadeghkhani a.sadeghkhani@queensu.ca There are two main schools to statistical inference: 1-frequentist

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Bias-Variance Tradeoff

Bias-Variance Tradeoff What s learning, revisited Overfitting Generative versus Discriminative Logistic Regression Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University September 19 th, 2007 Bias-Variance Tradeoff

More information

Probability Models of Information Exchange on Networks Lecture 5

Probability Models of Information Exchange on Networks Lecture 5 Probability Models of Information Exchange on Networks Lecture 5 Elchanan Mossel (UC Berkeley) July 25, 2013 1 / 22 Informational Framework Each agent receives a private signal X i which depends on S.

More information

Machine Learning. Probability Basics. Marc Toussaint University of Stuttgart Summer 2014

Machine Learning. Probability Basics. Marc Toussaint University of Stuttgart Summer 2014 Machine Learning Probability Basics Basic definitions: Random variables, joint, conditional, marginal distribution, Bayes theorem & examples; Probability distributions: Binomial, Beta, Multinomial, Dirichlet,

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Logistic Regression Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Analog Circuit Verification by Statistical Model Checking

Analog Circuit Verification by Statistical Model Checking 16 th Asian Pacific Design Automation ti Conference Analog Circuit Verification by Statistical Model Checking Author Ying-Chih Wang (Presenter) Anvesh Komuravelli Paolo Zuliani Edmund M. Clarke Date 2011/1/26

More information

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012 Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood

More information

Inconsistency of Bayesian inference when the model is wrong, and how to repair it

Inconsistency of Bayesian inference when the model is wrong, and how to repair it Inconsistency of Bayesian inference when the model is wrong, and how to repair it Peter Grünwald Thijs van Ommen Centrum Wiskunde & Informatica, Amsterdam Universiteit Leiden June 3, 2015 Outline 1 Introduction

More information

Introductory Econometrics

Introductory Econometrics Session 4 - Testing hypotheses Roland Sciences Po July 2011 Motivation After estimation, delivering information involves testing hypotheses Did this drug had any effect on the survival rate? Is this drug

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

an introduction to bayesian inference

an introduction to bayesian inference with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric?

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/

More information

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Bayesian Inference. STA 121: Regression Analysis Artin Armagan Bayesian Inference STA 121: Regression Analysis Artin Armagan Bayes Rule...s! Reverend Thomas Bayes Posterior Prior p(θ y) = p(y θ)p(θ)/p(y) Likelihood - Sampling Distribution Normalizing Constant: p(y

More information

Inference for a Population Proportion

Inference for a Population Proportion Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

Bayesian Models in Machine Learning

Bayesian Models in Machine Learning Bayesian Models in Machine Learning Lukáš Burget Escuela de Ciencias Informáticas 2017 Buenos Aires, July 24-29 2017 Frequentist vs. Bayesian Frequentist point of view: Probability is the frequency of

More information

Chapter 5. Bayesian Statistics

Chapter 5. Bayesian Statistics Chapter 5. Bayesian Statistics Principles of Bayesian Statistics Anything unknown is given a probability distribution, representing degrees of belief [subjective probability]. Degrees of belief [subjective

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Jonathan Marchini Department of Statistics University of Oxford MT 2013 Jonathan Marchini (University of Oxford) BS2a MT 2013 1 / 27 Course arrangements Lectures M.2

More information

Bayes Correlated Equilibrium and Comparing Information Structures

Bayes Correlated Equilibrium and Comparing Information Structures Bayes Correlated Equilibrium and Comparing Information Structures Dirk Bergemann and Stephen Morris Spring 2013: 521 B Introduction game theoretic predictions are very sensitive to "information structure"

More information

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability

More information

Constructing Ensembles of Pseudo-Experiments

Constructing Ensembles of Pseudo-Experiments Constructing Ensembles of Pseudo-Experiments Luc Demortier The Rockefeller University, New York, NY 10021, USA The frequentist interpretation of measurement results requires the specification of an ensemble

More information

Bayesian statistics. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Bayesian statistics. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Bayesian statistics DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall15 Carlos Fernandez-Granda Frequentist vs Bayesian statistics In frequentist statistics

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Particle Filters. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

Particle Filters. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Particle Filters Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Motivation For continuous spaces: often no analytical formulas for Bayes filter updates

More information

Correlated Equilibrium in Games with Incomplete Information

Correlated Equilibrium in Games with Incomplete Information Correlated Equilibrium in Games with Incomplete Information Dirk Bergemann and Stephen Morris Econometric Society Summer Meeting June 2012 Robust Predictions Agenda game theoretic predictions are very

More information

2 Belief, probability and exchangeability

2 Belief, probability and exchangeability 2 Belief, probability and exchangeability We first discuss what properties a reasonable belief function should have, and show that probabilities have these properties. Then, we review the basic machinery

More information

EnM Probability and Random Processes

EnM Probability and Random Processes Historical Note: EnM 503 - Probability and Random Processes Probability has its roots in games of chance, which have been played since prehistoric time. Games and equipment have been found in Egyptian

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Revisiting the Exploration-Exploitation Tradeoff in Bandit Models

Revisiting the Exploration-Exploitation Tradeoff in Bandit Models Revisiting the Exploration-Exploitation Tradeoff in Bandit Models joint work with Aurélien Garivier (IMT, Toulouse) and Tor Lattimore (University of Alberta) Workshop on Optimization and Decision-Making

More information

The Value of Congestion

The Value of Congestion Laurens G. Debo Tepper School of Business Carnegie Mellon University Pittsburgh, PA 15213 The Value of Congestion Uday Rajan Ross School of Business University of Michigan Ann Arbor, MI 48109. February

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Bayesian Classification Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574

More information

Probability and Information Theory. Sargur N. Srihari

Probability and Information Theory. Sargur N. Srihari Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal

More information

Bayesian Learning. CSL603 - Fall 2017 Narayanan C Krishnan

Bayesian Learning. CSL603 - Fall 2017 Narayanan C Krishnan Bayesian Learning CSL603 - Fall 2017 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Bayes Theorem MAP Learners Bayes optimal classifier Naïve Bayes classifier Example text classification Bayesian networks

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Probabilities Marc Toussaint University of Stuttgart Winter 2018/19 Motivation: AI systems need to reason about what they know, or not know. Uncertainty may have so many sources:

More information

Bayesian Learning (II)

Bayesian Learning (II) Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning Dipendra Misra Cornell University dkm@cs.cornell.edu https://dipendramisra.wordpress.com/ Task Grasp the green cup. Output: Sequence of controller actions Setup from Lenz et. al.

More information

Bayesian Congestion Control over a Markovian Network Bandwidth Process

Bayesian Congestion Control over a Markovian Network Bandwidth Process Bayesian Congestion Control over a Markovian Network Bandwidth Process Parisa Mansourifard 1/30 Bayesian Congestion Control over a Markovian Network Bandwidth Process Parisa Mansourifard (USC) Joint work

More information

Approximate Bayesian Computation

Approximate Bayesian Computation Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki and Aalto University 1st December 2015 Content Two parts: 1. The basics of approximate

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 143 Part IV

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics Lecture 11 January 7, 2013 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline How to communicate the statistical uncertainty

More information

The knowledge gradient method for multi-armed bandit problems

The knowledge gradient method for multi-armed bandit problems The knowledge gradient method for multi-armed bandit problems Moving beyond inde policies Ilya O. Ryzhov Warren Powell Peter Frazier Department of Operations Research and Financial Engineering Princeton

More information

(1) Introduction to Bayesian statistics

(1) Introduction to Bayesian statistics Spring, 2018 A motivating example Student 1 will write down a number and then flip a coin If the flip is heads, they will honestly tell student 2 if the number is even or odd If the flip is tails, they

More information