What is to be done? Two attempts using Gaussian process priors

Size: px
Start display at page:

Download "What is to be done? Two attempts using Gaussian process priors"

Transcription

1 What is to be done? Two attempts using Gaussian process priors Maximilian Kasy Department of Economics, Harvard University Oct / 33

2 What questions should econometricians work on? Incentives of the publication process: Appeal to referees from the same subfield. Danger of self-referentiality, untethering from external relevance. Versus broader usefulness: Tools useful for empirical researchers, policy makers. Anchored in substantive applications, broader methodological considerations. One way to get there: Well defined decision problems. 2 / 33

3 Decision problems Objects to carefully choose: Objective function. Space of possible decisions / policy alternatives. Identifying assumptions. Prior information. Features the priors should be uninformative about. Once these are specified, coherent and well-behaved solutions can be derived. Useful tool for tractable solutions without functional form restrictions: Gaussian process priors. 3 / 33

4 Outline of this talk Brief introduction to Gaussian process regression Application 1: Optimal treatment assignment in experiments. Setting: Treatment assignment given baseline covariates General decision theory result: Non-random rules dominate random rules Prior for expectation of potential outcomes given covariates Expression for MSE of estimator for ATE to minimize by treatment assignment Application 2: Optimal insurance and taxation. Economic setting: Co-insurance rate for health insurance Statistical setting: prior for behavioral average response function Expression for posterior expected social welfare to maximize by choice of co-insurance rate 4 / 33

5 References Williams, C. and Rasmussen, C. (2006). Gaussian processes for machine learning. MIT Press, chapter 2. Kasy, M. (2016). Why experimenters might not always want to randomize, and what they could do instead. Political Analysis, 24(3): Kasy, M. (2017). Optimal taxation and insurance using machine learning. Working Paper, Harvard University. 5 / 33

6 Gaussian process regression Brief introduction to Gaussian process regression Suppose we observe n i.i.d. draws of (Y i,x i ), where Y i is real valued and X i is a k vector. Y i = f(x i ) + ε i ε i X,f( ) N(0,σ 2 ) Prior: f is distributed according to a Gaussian process, f X GP(0,C), where C is a covariance kernel, Cov(f(x),f(x ) X) = C(x,x ). We will leave conditioning on X implicit. 6 / 33

7 Gaussian process regression Posterior mean The joint distribution of (f(x),y ) is given by where ( ) ( ( )) f(x) C(x,x) c(x) N 0, Y c(x) C + σ 2, I n c(x) is the n vector with entries C(x,X i ), and C is the n n matrix with entries C i,j = C(X i,x j ). Therefore Read: f( ) = E[f( ) Y ] E[f(x) Y ] = c(x) (C + σ 2 I n ) 1 Y. is a linear combination of the functions C(,X i ) with weights ( C + σ 2 I n ) 1 Y. 7 / 33

8 Gaussian process regression Both applications use Gaussian process priors 1. Optimal experimental design How to assign treatment to minimize mean squared error for treatment effect estimators? Gaussian process prior for the conditional expectation of potential outcomes given covariates. 2. Optimal insurance and taxation How to choose a co-insurance rate or tax rate to maximize social welfare, given (quasi-)experimental data? Gaussian process prior for the behavioral response function mapping the co-insurance rate into the tax base. 8 / 33

9 Experimental design Application 1 Why experimenters might not always want to randomize Setup 1. Sampling: random sample of n units baseline survey vector of covariates X i 2. Treatment assignment: binary treatment assigned by D i = d i (X,U) X matrix of covariates; U randomization device 3. Realization of outcomes: Y i = D i Yi 1 + (1 D i )Yi 0 4. Estimation: estimator β of the (conditional) average treatment effect, β = 1 n i E[Yi 1 Yi 0 X i,θ] 9 / 33

10 Experimental design Questions How should we assign treatment? In particular, if X i has continuous or many discrete components? How should we estimate β? What is the role of prior information? 10 / 33

11 Experimental design Some intuition Compare apples with apples balance covariate distribution. Not just balance of means! We don t add random noise to estimators why add random noise to experimental designs? Identification requires controlled trials (CTs), but not randomized controlled trials (RCTs). 11 / 33

12 Experimental design General decision problem allowing for randomization General decision problem: State of the world θ, observed data X, randomization device U X, decision procedure δ(x,u), loss L(δ(X,U),θ). Conditional expected loss of decision procedure δ(x, U): R(δ,θ U = u) = E[L(δ(X,u),θ) θ] Bayes risk: R B (δ,π) = Minimax risk: R mm (δ) = R(δ,θ U = u)dπ(θ)dp(u) maxr(δ,θ U = u)dp(u) θ 12 / 33

13 Experimental design Theorem (Optimality of deterministic decisions) Consider a general decision problem. Let R equal R B or R mm. Then: 1. The optimal risk R (δ ), when considering only deterministic procedures δ(x), is no larger than the optimal risk when allowing for randomized procedures δ(x,u). 2. If the optimal deterministic procedure δ is unique, then it has strictly lower risk than any non-trivial randomized procedure. 13 / 33

14 Experimental design Proof Any probability distribution P(u) satisfies u P(u) = 1, P(u) 0 for all u. Thus u R u P(u) min u R u for any set of values R u. Let δ u (x) = δ(x,u). Then Similarly R B (δ,π) = u min u R mm (δ) = u min u R(δ u,θ)dπ(θ)p(u) R(δ u,θ)dπ(θ) = minr B (δ u,π). u maxr(δ u,θ)p(u) θ maxr(δ u,θ) = minr mm (δ u ). θ u 14 / 33

15 Experimental design Bayesian setup Back to experimental design setting. Conditional distribution of potential outcomes: for d = 0, 1 Gaussian process prior: Y d i X i = x N(f(x,d),σ 2 ). f GP(µ,C), E[f(x,d)] = µ(x,d) Cov(f(x 1,d 1 ),f(x 2,d 2 )) = C((x 1,d 1 ),(x 2,d 2 )) Conditional average treatment effect (CATE): β = 1 n i E[Y 1 i Y 0 i X i,θ] = 1 n i f(x i,1) f(x i,0). 15 / 33

16 Experimental design Notation Covariance matrix C, where C i,j = C((X i,d i ),(X j,d j )) Mean vector µ, components µ i = µ(x i,d i ) Covariance of observations with CATE, C i = Cov(Y i,β X,D) = 1 n j (C((X i,d i ),(X j,1)) C((X i,d i ),(X j,0))). 16 / 33

17 Experimental design Posterior expectation and risk The posterior expectation β of β equals β = µ β + C (C + σ 2 I) 1 (Y µ). The corresponding risk equals R B (d, β X) = Var(β X,Y ) = Var(β X) Var(E[β X,Y ] X) = Var(β X) C (C + σ 2 I) 1 C. 17 / 33

18 Experimental design Discrete optimization The optimal design solves Possible optimization algorithms: 1. Search over random d 2. greedy algorithm 3. simulated annealing max d C (C + σ 2 I) 1 C. 18 / 33

19 Experimental design Special case linear separable model Suppose f(x,d) = x γ + d β, γ N(0,Σ), and we estimate β using comparison of means. Bias of β equals (X 1 X 0 ) γ, prior expected squared bias (X 1 X 0 ) Σ (X 1 X 0 ). Mean squared error [ 1 MSE(d 1,...,d n ) = σ ] + (X 1 X 0 ) Σ (X 1 X 0 ). n 1 n 0 Risk is minimized by 1. choosing treatment and control arms of equal size, 2. and optimizing balance as measured by the difference in covariate means (X 1 X 0 ). 19 / 33

20 Optimal insurance Application 2 Optimal insurance and taxation using machine learning Economic setting Population of insured individuals i. Y i : health care expenditures of individual i. T i : share of health care expenditures covered by the insurance 1 T i : coinsurance rate; Y i (1 T i ): out-of-pocket expenditures Behavioral response to share covered: structural function Y i = g(t i,ε i ). Per capita expenditures under policy t: average structural function m(t) = E[g(t,ε i )]. 20 / 33

21 Optimal insurance Policy objective Insurance provider s expenditures per person: t m(t). Mechanical effect of increase in t (accounting): m(t)dt. Behavioral effect of increase in t (key empirical challenge): t m (t)dt. Utility of the insured: Mechanical effect of increase in t (accounting): m(t)dt. Behavioral effect: None, by envelope theorem. effect on utility = equivalent variation = mechanical effect Assign relative value λ > 1 to a marginal dollar for the sick vs. the insurer. 21 / 33

22 Optimal insurance Marginal effect of a change in t on social welfare: u (t) = (λ 1) m(t) t m (t) = λ m(t) (t m(t)). (1) t Integrating and imposing the normalization u(0) = 0: u(t) = λ t 0 m(x)dx t m(t). (2) Special case λ = 1: Harberger triangle (not the relevant case) 22 / 33

23 Optimal insurance Observed data and prior n i.i.d. draws of (Y i,t i ) T i was randomly assigned in an experiment, so that T i ε i, and E[Y i T i = t] = E[g(t,ε i ) T i = t] = E[g(t,ε i )] = m(t). Y i is normally distributed given T i, Y i T i = t N(m(t),σ 2 ). Gaussian process prior for m( ), m( ) GP(µ( ),C(, )). 23 / 33

24 Optimal insurance Prior moments Linear functions of normal vectors are normal. Linear operators of Gaussian processes are Gaussian processes. Prior moments: ν(t) = E[u(t)] = λ D(t,t ) = Cov(u(t),m(t ))) = λ t Var(u(t)) = λ 2 2λ t 0 t µ(x)dx t µ(t), C(x,t )dx t C(t,t ), 0 t t 0 0 t 0 C(x,x )dx dx C(x,t)dx + t 2 C(t,t). 24 / 33

25 Optimal insurance Posterior expectation of u( ) Covariance with data: D(t) = Cov(u(t),Y T ) = Cov(u(t),(m(T 1 ),...,m(t n )) T ) = (D(t,T 1 ),...,D(t,T n )). Posterior expectation of u(t): û(t) = E[u(t) Y,T ] = E[u(t) T ] + Cov(u(t),Y T ) Var(Y T ) 1 (Y E[Y T ]) = ν(t) + D(t) [C + σ ] 2 1 I (Y µ). 25 / 33

26 Optimal insurance Optimal policy choice Bayesian policy maker aims to maximize expected social welfare (note: different from expectation of maximizer of social welfare!) Thus First order condition t = t (Y,T ) argmax t t û( t ) = E[u ( t ) Y,T ] û(t). = ν ( t ) + B( t ) [C + σ 2 I ] 1 (Y µ) = 0, where B(t) = (B(t,T 1 ),...,B(t,T n )) and ) B(t,t ) = Cov( t u(t),m(t ) = t D(t,t ) = (λ 1) C(t,t ) t t C(t,t ). 26 / 33

27 Optimal insurance The RAND health insurance experiment (cf. Aron-Dine et al., 2013) Between 1974 and 1981 representative sample of 2000 households in six locations across the US families randomly assigned to plans with one of six consumer coinsurance rates 95, 50, 25, or 0 percent 2 more complicated plans (we drop those) Additionally: randomized Maximum Dollar Expenditure limits 5, 10, or 15 percent of family income, up to a maximum of $750 or $1,000 (we pool across those) 27 / 33

28 Optimal insurance Table: Expected spending for different coinsurance rates (1) (2) (3) (4) Share with Spending Share with Spending any in $ any in $ Free Care (0.006) (78.76) (0.006) (72.06) 25% Coinsurance (0.013) (130.5) (0.012) (115.2) 50% Coinsurance (0.018) (273.7) (0.016) (279.6) 95% Coinsurance (0.011) (95.40) (0.009) (88.48) family x month x site X X X X fixed effects covariates X X N / 33

29 Optimal insurance Assumptions 1. Model: The optimal insurance model as presented before 2. Prior: Gaussian process prior for m, squared exponential in distance, uninformative about level and slope 3. Relative value of funds for sick people vs contributors: λ = Pooling data: across levels of maximum dollar expenditure Under these assumptions we find: Optimal copay equals 18% (But free care is almost as good) 29 / 33

30 Optimal insurance / 33

31 Optimal insurance / 33

32 Optimal insurance Conclusion Explicit decision problems are useful to focus econometric research. Carefully choose: Objective function. Space of possible decisions / policy alternatives. Identifying assumptions. Prior information. Features the priors should be uninformative about. Gaussian process priors allow for tractable solutions. Two examples: 1. Optimal experimental design. 2. Optimal insurance and taxation. 32 / 33

33 Optimal insurance Thank you! 33 / 33

Using data to inform policy

Using data to inform policy Using data to inform policy Maximilian Kasy Department of Economics, Harvard University Maximilian Kasy (Harvard) data and policy 1 / 41 Introduction The roles of econometrics Forecasting: What will be?

More information

Why experimenters should not randomize, and what they should do instead

Why experimenters should not randomize, and what they should do instead Why experimenters should not randomize, and what they should do instead Maximilian Kasy Department of Economics, Harvard University Maximilian Kasy (Harvard) Experimental design 1 / 42 project STAR Introduction

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

Optimal taxation and insurance using machine learning sufficient statistics and beyond

Optimal taxation and insurance using machine learning sufficient statistics and beyond Optimal taxation and insurance using machine learning sufficient statistics and beyond Maximilian Kasy April 2, 218 Abstract How should one use (quasi-)experimental evidence when choosing policies such

More information

Using data to inform policy

Using data to inform policy Using data to inform policy Maximilian Kasy June 16, 2014 Abstract In this paper, a general framework is proposed for the use of (quasi-) experimental data when choosing policies such as tax rates or the

More information

Econ 2148, spring 2019 Statistical decision theory

Econ 2148, spring 2019 Statistical decision theory Econ 2148, spring 2019 Statistical decision theory Maximilian Kasy Department of Economics, Harvard University 1 / 53 Takeaways for this part of class 1. A general framework to think about what makes a

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Econ 2140, spring 2018, Part IIa Statistical Decision Theory

Econ 2140, spring 2018, Part IIa Statistical Decision Theory Econ 2140, spring 2018, Part IIa Maximilian Kasy Department of Economics, Harvard University 1 / 35 Examples of decision problems Decide whether or not the hypothesis of no racial discrimination in job

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Practical Bayesian Optimization of Machine Learning. Learning Algorithms Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

Minimax-Regret Sample Design in Anticipation of Missing Data, With Application to Panel Data. Jeff Dominitz RAND. and

Minimax-Regret Sample Design in Anticipation of Missing Data, With Application to Panel Data. Jeff Dominitz RAND. and Minimax-Regret Sample Design in Anticipation of Missing Data, With Application to Panel Data Jeff Dominitz RAND and Charles F. Manski Department of Economics and Institute for Policy Research, Northwestern

More information

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan Background: Global Optimization and Gaussian Processes The Geometry of Gaussian Processes and the Chaining Trick Algorithm

More information

Econometric Causality

Econometric Causality Econometric (2008) International Statistical Review, 76(1):1-27 James J. Heckman Spencer/INET Conference University of Chicago Econometric The econometric approach to causality develops explicit models

More information

Model Selection for Gaussian Processes

Model Selection for Gaussian Processes Institute for Adaptive and Neural Computation School of Informatics,, UK December 26 Outline GP basics Model selection: covariance functions and parameterizations Criteria for model selection Marginal

More information

Lecture : Probabilistic Machine Learning

Lecture : Probabilistic Machine Learning Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Predicting the Treatment Status

Predicting the Treatment Status Predicting the Treatment Status Nikolay Doudchenko 1 Introduction Many studies in social sciences deal with treatment effect models. 1 Usually there is a treatment variable which determines whether a particular

More information

The Generalized Roy Model and Treatment Effects

The Generalized Roy Model and Treatment Effects The Generalized Roy Model and Treatment Effects Christopher Taber University of Wisconsin November 10, 2016 Introduction From Imbens and Angrist we showed that if one runs IV, we get estimates of the Local

More information

Today. Probability and Statistics. Linear Algebra. Calculus. Naïve Bayes Classification. Matrix Multiplication Matrix Inversion

Today. Probability and Statistics. Linear Algebra. Calculus. Naïve Bayes Classification. Matrix Multiplication Matrix Inversion Today Probability and Statistics Naïve Bayes Classification Linear Algebra Matrix Multiplication Matrix Inversion Calculus Vector Calculus Optimization Lagrange Multipliers 1 Classical Artificial Intelligence

More information

Lecture 2: Basic Concepts of Statistical Decision Theory

Lecture 2: Basic Concepts of Statistical Decision Theory EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes CSci 8980: Advanced Topics in Graphical Models Gaussian Processes Instructor: Arindam Banerjee November 15, 2007 Gaussian Processes Outline Gaussian Processes Outline Parametric Bayesian Regression Gaussian

More information

Short Questions (Do two out of three) 15 points each

Short Questions (Do two out of three) 15 points each Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal

More information

1 Bayesian Linear Regression (BLR)

1 Bayesian Linear Regression (BLR) Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,

More information

Introduction to Econometrics

Introduction to Econometrics Introduction to Econometrics Lecture 3 : Regression: CEF and Simple OLS Zhaopeng Qu Business School,Nanjing University Oct 9th, 2017 Zhaopeng Qu (Nanjing University) Introduction to Econometrics Oct 9th,

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Recap from previous lecture

Recap from previous lecture Recap from previous lecture Learning is using past experience to improve future performance. Different types of learning: supervised unsupervised reinforcement active online... For a machine, experience

More information

Treatment Effects. Christopher Taber. September 6, Department of Economics University of Wisconsin-Madison

Treatment Effects. Christopher Taber. September 6, Department of Economics University of Wisconsin-Madison Treatment Effects Christopher Taber Department of Economics University of Wisconsin-Madison September 6, 2017 Notation First a word on notation I like to use i subscripts on random variables to be clear

More information

Gaussian Processes (10/16/13)

Gaussian Processes (10/16/13) STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs

More information

CS 195-5: Machine Learning Problem Set 1

CS 195-5: Machine Learning Problem Set 1 CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of

More information

Econ 2148, fall 2017 Instrumental variables I, origins and binary treatment case

Econ 2148, fall 2017 Instrumental variables I, origins and binary treatment case Econ 2148, fall 2017 Instrumental variables I, origins and binary treatment case Maximilian Kasy Department of Economics, Harvard University 1 / 40 Agenda instrumental variables part I Origins of instrumental

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Markus Haas LMU München Summer term 2011 15. Mai 2011 The Simple Linear Regression Model Considering variables x and y in a specific population (e.g., years of education and wage

More information

GAUSSIAN PROCESS REGRESSION

GAUSSIAN PROCESS REGRESSION GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Generated Covariates in Nonparametric Estimation: A Short Review.

Generated Covariates in Nonparametric Estimation: A Short Review. Generated Covariates in Nonparametric Estimation: A Short Review. Enno Mammen, Christoph Rothe, and Melanie Schienle Abstract In many applications, covariates are not observed but have to be estimated

More information

Prediction of double gene knockout measurements

Prediction of double gene knockout measurements Prediction of double gene knockout measurements Sofia Kyriazopoulou-Panagiotopoulou sofiakp@stanford.edu December 12, 2008 Abstract One way to get an insight into the potential interaction between a pair

More information

Overfitting, Bias / Variance Analysis

Overfitting, Bias / Variance Analysis Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian

More information

Principles Underlying Evaluation Estimators

Principles Underlying Evaluation Estimators The Principles Underlying Evaluation Estimators James J. University of Chicago Econ 350, Winter 2019 The Basic Principles Underlying the Identification of the Main Econometric Evaluation Estimators Two

More information

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

PILCO: A Model-Based and Data-Efficient Approach to Policy Search PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol

More information

Item Response Theory for Conjoint Survey Experiments

Item Response Theory for Conjoint Survey Experiments Item Response Theory for Conjoint Survey Experiments Devin Caughey Hiroto Katsumata Teppei Yamamoto Massachusetts Institute of Technology PolMeth XXXV @ Brigham Young University July 21, 2018 Conjoint

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Gaussian Processes Barnabás Póczos http://www.gaussianprocess.org/ 2 Some of these slides in the intro are taken from D. Lizotte, R. Parr, C. Guesterin

More information

Robust Predictions in Games with Incomplete Information

Robust Predictions in Games with Incomplete Information Robust Predictions in Games with Incomplete Information joint with Stephen Morris (Princeton University) November 2010 Payoff Environment in games with incomplete information, the agents are uncertain

More information

Statistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes

Statistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes Statistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes Lecturer: Drew Bagnell Scribe: Venkatraman Narayanan 1, M. Koval and P. Parashar 1 Applications of Gaussian

More information

Machine Learning. Bayesian Regression & Classification. Marc Toussaint U Stuttgart

Machine Learning. Bayesian Regression & Classification. Marc Toussaint U Stuttgart Machine Learning Bayesian Regression & Classification learning as inference, Bayesian Kernel Ridge regression & Gaussian Processes, Bayesian Kernel Logistic Regression & GP classification, Bayesian Neural

More information

Correlated Equilibrium in Games with Incomplete Information

Correlated Equilibrium in Games with Incomplete Information Correlated Equilibrium in Games with Incomplete Information Dirk Bergemann and Stephen Morris Econometric Society Summer Meeting June 2012 Robust Predictions Agenda game theoretic predictions are very

More information

Gibbs Sampling in Endogenous Variables Models

Gibbs Sampling in Endogenous Variables Models Gibbs Sampling in Endogenous Variables Models Econ 690 Purdue University Outline 1 Motivation 2 Identification Issues 3 Posterior Simulation #1 4 Posterior Simulation #2 Motivation In this lecture we take

More information

Probabilistic & Unsupervised Learning

Probabilistic & Unsupervised Learning Probabilistic & Unsupervised Learning Gaussian Processes Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College London

More information

GWAS V: Gaussian processes

GWAS V: Gaussian processes GWAS V: Gaussian processes Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS V: Gaussian processes Summer 2011

More information

Incentives and Nutrition for Rotten Kids: Intrahousehold Food Allocation in the Philippines

Incentives and Nutrition for Rotten Kids: Intrahousehold Food Allocation in the Philippines Incentives and Nutrition for Rotten Kids: Intrahousehold Food Allocation in the Philippines Pierre Dubois and Ethan Ligon presented by Rachel Heath November 3, 2006 Introduction Outline Introduction Modification

More information

Nonparametric Regression With Gaussian Processes

Nonparametric Regression With Gaussian Processes Nonparametric Regression With Gaussian Processes From Chap. 45, Information Theory, Inference and Learning Algorithms, D. J. C. McKay Presented by Micha Elsner Nonparametric Regression With Gaussian Processes

More information

Introduction to causal identification. Nidhiya Menon IGC Summer School, New Delhi, July 2015

Introduction to causal identification. Nidhiya Menon IGC Summer School, New Delhi, July 2015 Introduction to causal identification Nidhiya Menon IGC Summer School, New Delhi, July 2015 Outline 1. Micro-empirical methods 2. Rubin causal model 3. More on Instrumental Variables (IV) Estimating causal

More information

Econ 2148, fall 2017 Instrumental variables II, continuous treatment

Econ 2148, fall 2017 Instrumental variables II, continuous treatment Econ 2148, fall 2017 Instrumental variables II, continuous treatment Maximilian Kasy Department of Economics, Harvard University 1 / 35 Recall instrumental variables part I Origins of instrumental variables:

More information

ACM 116: Lectures 3 4

ACM 116: Lectures 3 4 1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance

More information

An Introduction to Gaussian Processes for Spatial Data (Predictions!)

An Introduction to Gaussian Processes for Spatial Data (Predictions!) An Introduction to Gaussian Processes for Spatial Data (Predictions!) Matthew Kupilik College of Engineering Seminar Series Nov 216 Matthew Kupilik UAA GP for Spatial Data Nov 216 1 / 35 Why? When evidence

More information

Prediction of Data with help of the Gaussian Process Method

Prediction of Data with help of the Gaussian Process Method of Data with help of the Gaussian Process Method R. Preuss, U. von Toussaint Max-Planck-Institute for Plasma Physics EURATOM Association 878 Garching, Germany March, Abstract The simulation of plasma-wall

More information

IDENTIFICATION OF TREATMENT EFFECTS WITH SELECTIVE PARTICIPATION IN A RANDOMIZED TRIAL

IDENTIFICATION OF TREATMENT EFFECTS WITH SELECTIVE PARTICIPATION IN A RANDOMIZED TRIAL IDENTIFICATION OF TREATMENT EFFECTS WITH SELECTIVE PARTICIPATION IN A RANDOMIZED TRIAL BRENDAN KLINE AND ELIE TAMER Abstract. Randomized trials (RTs) are used to learn about treatment effects. This paper

More information

Introduction to Machine Learning Midterm Exam Solutions

Introduction to Machine Learning Midterm Exam Solutions 10-701 Introduction to Machine Learning Midterm Exam Solutions Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes,

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

Empirical approaches in public economics

Empirical approaches in public economics Empirical approaches in public economics ECON4624 Empirical Public Economics Fall 2016 Gaute Torsvik Outline for today The canonical problem Basic concepts of causal inference Randomized experiments Non-experimental

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Lecture Notes 1: Decisions and Data. In these notes, I describe some basic ideas in decision theory. theory is constructed from

Lecture Notes 1: Decisions and Data. In these notes, I describe some basic ideas in decision theory. theory is constructed from Topics in Data Analysis Steven N. Durlauf University of Wisconsin Lecture Notes : Decisions and Data In these notes, I describe some basic ideas in decision theory. theory is constructed from The Data:

More information

Bayesian Interpretations of Regularization

Bayesian Interpretations of Regularization Bayesian Interpretations of Regularization Charlie Frogner 9.50 Class 15 April 1, 009 The Plan Regularized least squares maps {(x i, y i )} n i=1 to a function that minimizes the regularized loss: f S

More information

Introduction to Simple Linear Regression

Introduction to Simple Linear Regression Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department

More information

Lecture 1: Bayesian Framework Basics

Lecture 1: Bayesian Framework Basics Lecture 1: Bayesian Framework Basics Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de April 21, 2014 What is this course about? Building Bayesian machine learning models Performing the inference of

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information

Chapter 2 The Simple Linear Regression Model: Specification and Estimation

Chapter 2 The Simple Linear Regression Model: Specification and Estimation Chapter The Simple Linear Regression Model: Specification and Estimation Page 1 Chapter Contents.1 An Economic Model. An Econometric Model.3 Estimating the Regression Parameters.4 Assessing the Least Squares

More information

Intro to Probability. Andrei Barbu

Intro to Probability. Andrei Barbu Intro to Probability Andrei Barbu Some problems Some problems A means to capture uncertainty Some problems A means to capture uncertainty You have data from two sources, are they different? Some problems

More information

A Note on the Welfare Evaluation of Tax Reform with Non-Convex Preferences and Discrete Labor Supply

A Note on the Welfare Evaluation of Tax Reform with Non-Convex Preferences and Discrete Labor Supply A Note on the Welfare Evaluation of Tax Reform with Non-Convex Preferences and Discrete Labor Supply Henrik Jacobsen Kleven University of Copenhagen, EPRU, and CEPR Claus Thustrup Kreiner University of

More information

Gaussian Processes for Machine Learning

Gaussian Processes for Machine Learning Gaussian Processes for Machine Learning Carl Edward Rasmussen Max Planck Institute for Biological Cybernetics Tübingen, Germany carl@tuebingen.mpg.de Carlos III, Madrid, May 2006 The actual science of

More information

Lecture 2: Statistical Decision Theory (Part I)

Lecture 2: Statistical Decision Theory (Part I) Lecture 2: Statistical Decision Theory (Part I) Hao Helen Zhang Hao Helen Zhang Lecture 2: Statistical Decision Theory (Part I) 1 / 35 Outline of This Note Part I: Statistics Decision Theory (from Statistical

More information

Microeconomic theory focuses on a small number of concepts. The most fundamental concept is the notion of opportunity cost.

Microeconomic theory focuses on a small number of concepts. The most fundamental concept is the notion of opportunity cost. Microeconomic theory focuses on a small number of concepts. The most fundamental concept is the notion of opportunity cost. Opportunity Cost (or "Wow, I coulda had a V8!") The underlying idea is derived

More information

Stat 366 A1 (Fall 2006) Midterm Solutions (October 23) page 1

Stat 366 A1 (Fall 2006) Midterm Solutions (October 23) page 1 Stat 366 A1 Fall 6) Midterm Solutions October 3) page 1 1. The opening prices per share Y 1 and Y measured in dollars) of two similar stocks are independent random variables, each with a density function

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Reliability Monitoring Using Log Gaussian Process Regression

Reliability Monitoring Using Log Gaussian Process Regression COPYRIGHT 013, M. Modarres Reliability Monitoring Using Log Gaussian Process Regression Martin Wayne Mohammad Modarres PSA 013 Center for Risk and Reliability University of Maryland Department of Mechanical

More information

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 1: Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section

More information

Use of Matching Methods for Causal Inference in Experimental and Observational Studies. This Talk Draws on the Following Papers:

Use of Matching Methods for Causal Inference in Experimental and Observational Studies. This Talk Draws on the Following Papers: Use of Matching Methods for Causal Inference in Experimental and Observational Studies Kosuke Imai Department of Politics Princeton University April 27, 2007 Kosuke Imai (Princeton University) Matching

More information

Robust Bayesian Variable Selection for Modeling Mean Medical Costs

Robust Bayesian Variable Selection for Modeling Mean Medical Costs Robust Bayesian Variable Selection for Modeling Mean Medical Costs Grace Yoon 1,, Wenxin Jiang 2, Lei Liu 3 and Ya-Chen T. Shih 4 1 Department of Statistics, Texas A&M University 2 Department of Statistics,

More information

Notes for Math 324, Part 19

Notes for Math 324, Part 19 48 Notes for Math 324, Part 9 Chapter 9 Multivariate distributions, covariance Often, we need to consider several random variables at the same time. We have a sample space S and r.v. s X, Y,..., which

More information

Financial Econometrics

Financial Econometrics Financial Econometrics Estimation and Inference Gerald P. Dwyer Trinity College, Dublin January 2013 Who am I? Visiting Professor and BB&T Scholar at Clemson University Federal Reserve Bank of Atlanta

More information

Incentive Compatible Regression Learning

Incentive Compatible Regression Learning Incentive Compatible Regression Learning Ofer Dekel 1 Felix Fischer 2 Ariel D. Procaccia 1 1 School of Computer Science and Engineering The Hebrew University of Jerusalem 2 Institut für Informatik Ludwig-Maximilians-Universität

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

The risk of machine learning

The risk of machine learning / 33 The risk of machine learning Alberto Abadie Maximilian Kasy July 27, 27 2 / 33 Two key features of machine learning procedures Regularization / shrinkage: Improve prediction or estimation performance

More information

The cover page of the Encyclopedia of Health Economics (2014) Introduction to Econometric Application in Health Economics

The cover page of the Encyclopedia of Health Economics (2014) Introduction to Econometric Application in Health Economics PHPM110062 Teaching Demo The cover page of the Encyclopedia of Health Economics (2014) Introduction to Econometric Application in Health Economics Instructor: Mengcen Qian School of Public Health What

More information

Introduction to Machine Learning Midterm Exam

Introduction to Machine Learning Midterm Exam 10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but

More information

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,

More information

A Memory Based Model of Bounded Rationality

A Memory Based Model of Bounded Rationality A Memory Based Model of Bounded Rationality Sendhil Mullainathan published in Quarterly Journal of Economics, 2002 January 16, 2012 Introduction Can memory imperfections help explain observed economic

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems

More information

Machine Learning. Probability Basics. Marc Toussaint University of Stuttgart Summer 2014

Machine Learning. Probability Basics. Marc Toussaint University of Stuttgart Summer 2014 Machine Learning Probability Basics Basic definitions: Random variables, joint, conditional, marginal distribution, Bayes theorem & examples; Probability distributions: Binomial, Beta, Multinomial, Dirichlet,

More information

Multi-level Models: Idea

Multi-level Models: Idea Review of 140.656 Review Introduction to multi-level models The two-stage normal-normal model Two-stage linear models with random effects Three-stage linear models Two-stage logistic regression with random

More information

Name of the Student: Problems on Discrete & Continuous R.Vs

Name of the Student: Problems on Discrete & Continuous R.Vs Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan

More information

Bayesian Linear Regression [DRAFT - In Progress]

Bayesian Linear Regression [DRAFT - In Progress] Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory

More information