Hierarchical Bayes. Peter Lenk. Stephen M Ross School of Business at the University of Michigan September 2004
|
|
- Caren Bradford
- 6 years ago
- Views:
Transcription
1 Herarchcal Bayes Peter Lenk Stephen M Ross School of Busness at the Unversty of Mchgan September 2004
2 Outlne Bayesan Decson Theory Smple Bayes and Shrnkage Estmates Herarchcal Bayes Numercal Methods Battng Averages HB Interacton Model 2
3 Bayesan Decson Model Decson Model Actons States Consequences Inference Model Pror Lkelhood Separable Parts: Don t mx them Updatng Bayes Theorem Posteror Loss Functon Integraton Result Optmal Decson: Marketng Acton & Inference 3
4 Bayes Theorem Model for the data gven parameters f(y θ) were θ = unknown parameters E.g. Y = µ + ε and θ = (µ,σ) Lkelhood l(θ) = f(y θ) f(y 2 θ) f(y n θ) Pror dstrbuton of parameters p(θ) Update pror p(θ Data ) = l(θ)p(θ)/f(y) f(y) = margnal dstrbuton of data 4
5 Easy Example: Estmate mean from a normal dstrbuton. Y = µ + ε Error terms {ε } are d normal Mean s zero Standard devaton of error terms s σ. Assume that σ s known 5
6 Conjugate Pror for Mean Pror dstrbuton for µ s normal Pror mean s m 0 Pror varance s v 0 2 Precson s / v 0 2 6
7 7 Posteror Dstrbuton and 0 ) ( v n v w v n n w m w wy m n n + = < < + = + = σ σ σ Observe n data ponts Posteror dstrbuton s normal Mean s m n Varance s v n 2
8 Shrnkage Estmators Bayes estmators combnes pror guesses wth sample estmates If the pror precson s larger than sample precson (pror has more nformaton), then put more weght on pror mean. If the sample precson s larger the pror precson (sample has more nformaton), then put more weght on sample average 8
9 Example Y s normal wth mean 0 and Varance 6 Normal pror for the populaton mean Mean = 5 & Varance = 2 Pror s nformatve and way off Data n = 5, Average = 0.9, Varance = 4.7 Posteror s normal Mean = 7.4 and varance s.2 9
10 Pror & Posteror n=5 Pror Posteror Lkelhood Pror Mean Posteror Mean Mean Sample Mean 0
11 Pror & Posteror n=50 Pror Posteror Lkelhood Mean
12 Use Less Informatve Pror Y s normal wth mean 0 and Varance 6 Normal pror for the populaton mean Mean = 5 & Varance = 0 nstead of 2 Pror s flatter Data n = 5, Average = 0.9, Varance = 4.7 Posteror s normal Mean = 9.6 and varance s 2.3 2
13 Pror & Posteror n=5 Pror Posteror Lkelhood Mean 3
14 Pror & Posteror n=50 Pror Posteror Lkelhood Mean 4
15 Summary Pror has less effect as sample sze ncreases Very nformatve prors gve good results wth smaller samples f pror nformaton s correct If you really don t know, then use flatter or less nformatve prors 5
16 What about Marketng? HB revoluton n how we thnk about customers 6
17 Henry Ford All Customers are the Same 7
18 Alfred Sloan Several Common Preferences 8
19 Contnuous Heterogenety 9
20 Proft Maxmzaton 20
21 It Can Get Wld! 2
22 HB Model for Weekly Spendng Wthn-subject model: Y,j = µ + ε,j and var(ε,j ) = σ 2 Heterogenety n mean weekly spendng or between-subjects µ = θ + δ and var(δ ) = τ 2 Pror Dstrbuton θ s N(u 0,v 02 ) Varances are known 22
23 Varances & Covarances Var(Y,j µ ) = σ 2 (known µ ) Var(Y,j θ) = τ 2 + σ 2 (unknown µ ) Cov(Y,j, Y,k ) = τ 2 for j not equal to k Observatons from dfferent subjects are ndependent 23
24 24 Precsons = /Varance ( ) ( ) ( ) ( ) ( ) ( ) n Y n Y Y Y v j j , 2, and Pr Pr and Pr Pr Pr pror precson s Pr σ τ θ σ µ σ τ θ σ µ τ θ µ θ + = = + = = = =
25 25 Jont Dstrbuton ( ) ( ) ( ) ( ) = = = n j j N y f g v u h P Y 2, ,,,,, σ µ τ θ µ θ θ µ Pror Between Subjects Wthn Subjects
26 Bayes Theorem P ( µ, θ Y ) = (, µ, θ ) (, µ, θ ) P Y P Y d µ d θ P P ( ) ( ) P Y, µ, θ µ, θ Y = P( Y ) ( µ, θ Y ) P( Y, µ, θ ) Constant because Y s fxed & known 26
27 Bayes Estmator Posteror means are optmal under squared error loss E(µ Y) and E(θ Y) Measure of accuracy s posteror varance var(µ Y) and var(θ Y) 27
28 Posteror Dstrbuton of θ Normal dstrbuton Posteror mean s u N Posteror varance s v N 2 Posteror precson s Pr(θ Y) = /v N 2 28
29 Pr Posteror Precson of θ Pr = Precson = /Varance ( ) ( ) θ Y Pr( θ ) + Pr Y θ N = = Pr ( ) ( ) θ = and Pr Y θ v 2 0 = τ 2 + σ 2 n 29
30 Posteror Mean of θ N u N = w u = wy w 0 = Pr Pr( θ ) ( θ Y ) and w = Pr Pr ( ) θ Y ( θ Y ) 30
31 Updatng of θ Pror Mean Posteror Mean u 0 u N Pror Var Posteror Var v 0 2 v N 2 3
32 32 Posteror Mean of µ [ ] ( ) ( ) ( ) ( ) Pr Pr Pr N n n Y Y u Y Y E σ τ σ µ θ µ µ α α α µ + = + = + =
33 Between-Subject Heterogenety n Mean Household Spendng 0.4 Heterogenety Dstrbuton Spendng 33
34 Between & Wthn Subjects Dstrbutons Heterogenety Subject Subject 2 Subject Dstrbuton Spendng 34
35 2 Observatons per Subject Heterogenety Subject Subject 2 Subject Dstrbuton Spendng 35
36 Subject Averages Heterogenety Subject Subject 2 Subject Dstrbuton Spendng 36
37 Pooled Estmate of Mean Heterogenety Subject Subject 2 Subject 3 Dstrbuton Spendng Pooled estmate of populaton average spendng 37
38 Sample Estmates Dsaggregate estmate of µ only uses the observatons for subject. Super f 30 or more observatons per subject Pooled or aggregate estmator Y of θ Y Smaller samplng error Ignores ndvdual dfference 38
39 HB Shrnkage Estmator Take combnaton of ndvdual average and pooled average wy What are the correct weghts? HB automatcally gves optmal weghts based on Pror varance of µ Number of observatons for subject Varance of past spendng for subject Number of subjects ( w )Y + Amount of heterogenety n household means 39
40 Shrnkage Estmates Heterogenety Subject Subject 2 Subject Dstrbuton Spendng 40
41 20 Observatons per Subject Heterogenety Subject Subject 2 Subject Dstrbuton Spendng 4
42 Bayes & Shrnkage Estmates Bayes estmators automatcally determne the optmal amount of shrnkage to mnmze MSE for true parameters and predctons Borrows strength from all subjects Tradeoff some bas for varance reducton 42
43 Good & Bad News Only smple models result n equatons Models we use n marketng requre numercal methods to compute posteror mean, posteror standard devatons, predctons and so on. 43
44 Numercal Integraton Compute posteror mean of functon T(θ). E[ T ( θ ) y] = T ( θ ) p( θ y) dθ 44
45 T(x) and f(x) X T(x)f(x) X 45
46 Trapezod Rule T(x)f(x) X 46
47 Grd Methods Very accurate wth few functonal evaluatons Need to know where the acton s Does not scale well to hgher dmensons You need to be very smart to make t work 47
48 Monte Carlo Generate random draws θ, θ 2,, θ m from posteror dstrbuton usng a random number generator. [ ( ) ] ( ) θ y T θ E T m j= What happened to the densty of θ? m j 48
49 Good & Bad News If your computer has a random number generator for the posteror dstrbuton, Monte Carlo s a snap to do. Your computer almost never has the correct random number generator. 49
50 Importance Samplng Would lke to sample from densty f You have a good random number generator for the densty g Importance samplng lets you generate random devates from g to evaluate expectatons wth respect to f. Generate φ,, φ m from g 50
51 5 Importance Samplng Approxmaton ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) = = = = = = = m j j m m m r r w g f r w T r r T d g g f T d f T and ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ ϕ θ θ θ
52 Markov Chan Monte Carlo Extenson of Monte Carlo Random draws are not ndependent Jont dstrbuton f(β,φ) does not have a convenent random number generator. Condtonal dstrbutons g(φ β) and h(β φ) are easy to generate from. 52
53 Iteratve Generaton from Full Condtonals Start at φ 0 Generate β from h(β φ 0 ). Generate φ from g(φ β ). Generate β m+ from h(β φ m ) Generate φ m+ from g(φ β m+ ) 53
54 Baseball Example 90 MLB Players n 2000 season. Observe at bats (AB) and hts (BA) n May Infer dstrbuton of battng averages across players. Predct battng averages n October usng data from May. 54
55 Baseball Battng Averages The Cleveland Indans May October BA AB BA AB Murray Belle Vzquel Pena
56 Estmatng a Probablty n at bats n May X = number of hts p = battng average for season X has a bnomal dstrbuton mean np varance np(-p) 56
57 0.4 n = 50 p = 0.3 Bnomal Dstrbuton mean = 5.00 STD = Probablty Sample Successes 57
58 Need Pror for Battng Average p 0 < p < Beta dstrbuton s popular choce It has two parameters: α and β Densty s proportonal to p α- (-p) β- Pror Mean = α/(α+β) 58
59 59 Beta Pror for p [ ] ( ) ( ) ( ) ( ) 0 ), (, Γ Γ + Γ = = p for p p p Beta p f β α β α β α β α β α ( ) = Γ 0 dx e x x α α
60 Mean and Varance α E( p) = α + β E( p) [ E( p) ] V ( p) = α + β + 60
61 Beta Dstrbuton alpha = beta = mean = 0.50 Densty P 6
62 Beta Dstrbuton 6 alpha = 5 beta = mean = Densty P 62
63 Beta Dstrbuton Densty alpha = 5 beta = 0 mean = P 63
64 Beta Dstrbuton 2 alpha = 0.5 beta = 2 mean = Densty P 64
65 Beta Dstrbuton 5 alpha = 0.5 beta = 0.5 mean = Densty P 65
66 66 Bayes Theorem: Update pror for p after observng n and x ( ) ), ( ], [ ] Pr[ ], [ ] Pr[ ], [ ] Pr[ ],, [ 0 x n x p Beta p p p f p x dq q f q x p f p x x p f x n x + + = = + + β α β α β α β α β α β α
67 Inference About P: Posteror s also Beta Pror Parameters Posteror Parameters α α+x β β+n-x 67
68 Posteror Mean of p: Its another shrnkage estmator w pˆ n n + ( w n ) ( Pror Mean) x n pˆ n = n and w n = α + β + n 68
69 Densty pror mean = 0.25 alpha = 5 beta = Pror p=beta & x=bnomal N=50 & X = 8 Posteror P 69
70 Herarchcal Bayes Model Varaton wthn batter : X gven p has a bnomal dstrbuton Varaton among batters: p s a beta dstrbuton wth parameters α and β. Pror dstrbuton for α and β Gamma (ch-square) dstrbuton 70
71 7 Gamma Dstrbuton s Y E Y V and s r Y E y for e y r s s r y G y g sy r r ) ( ) ( ) ( 0 ) ( ), ( ) ( = = > Γ = =
72 0.2 Gamma Dstrbuton alpha = beta = 0. mean = 0.00 STD = De nsty X 72
73 0.07 Gamma Dstrbuton alpha = 0 beta = 0.5 mean = STD = De ns ty X 73
74 Specfy Pror Parameters: r, s, u & v Prors: α s G(r,s) & β s G(u,v). E(α) = r/s and V(α) = E(α)/s. s determnes varance relatve to mean. I used s = 0.25 or the varance s four tmes larger than the mean. Same for v. 74
75 75 [ ] 0 ), ( ) ( p u r r p E E p E = + = = β α [ ] ( ) ), ( = = u r p p p V E c β α ( ) ( ) ( ) = = c p p p u and c p p p r
76 Specfy Pror Parameters Guess a mean of all battng averages: p 0 = 0.25 Measure of my uncertanty of that guess: c = 0.0 Parameter r = 4.4 Parameter u =
77 MCMC for Battng Averages Need full condtonals for p gve α and β Beta dstrbuton Need full condtonals for α and β gven p. Unknown dstrbuton Use Metropols algorthm 77
78 78 MCMC: Full Condtonals for Player Battng Average p [ ] ( ) ( ) ( ) ( ) x n x a x n x p Beta p p p f p x x p f + + = + + β α β α β α β,, Pr,,
79 79 MCMC: Full Condtonal for α and β ( ) ( ) ( ) ( ) = n n n v u g s r g p p p p x x g,,,,,, β α β α β α K K
80 Metropols Algorthm Want to generate θ from f Instead, generate canddate value φ from g(. θ) Densty g can depend on θ eg Random walk: φ = θ + δ Wth probablty α(θ,φ) accept φ as the new value of θ Wth probablty α(θ,φ) keep θ 80
81 Transton Probablty α ( θ, ϕ) ( ϕ) g( θ ϕ) ( θ ) g( ϕ θ ) f = max, f f s the full condtonal densty of θ Ratos: do not need to know constants Usually compute α on log scale. Works f denstes are not zero Works better f g s close to f 8
82 Alpha and Beta vs Iteraton Parameter BETA ALPHA MCMC Iteraton 82
83 Posteror of Alpha Alpha Frequency
84 Posteror of Beta Beta Frequency
85 Parameters Estmates α (std) β (std) Pror 7.8 (8.4) 53.2 (4.6) Posteror 26.2 (4.6) 68.2 (.7) 85
86 Dstrbuton of Battng Averages Densty Posteror after May Pror Battng Average 86
87 Predcton of Season Averages RMSE MAPE MLE % Bayes % 87
88 Battng Averages Bayes Shrnks MLE Season MLE Bayes May
89 HB Conjont Lenk, DeSarbo, Green, Young (996) Evaluated computer profles on a 0 to 0 scale for lkelhood to purchase 0 = Would not buy 0 = Would defntely buy Desgn 78 subjects 3 attrbutes wth two levels each 20 profles per subject 89
90 Attrbutes: Effect Codng. Hotlne support 2. RAM 3. Screen Sze 4. CPU 5. Hard Dsk 6. Multmeda 7. Cache 8. Color 9. Retal Store 0.Warrantee.Software 2.Guarantee 3.Prce 90
91 Subject-Covarates Female: f female and 0 f male Years: # years of work experence Own: f has computer & 0 else Nerd: f techncal background & 0 else Apply: # of software applcatons Expert: self-report expertse ratng 9
92 Summary Statstcs for Covarates Varable Mean Std Dev MIN MAX FEMALE YEARS OWN NERD APPLY EXPERT
93 Interacton Model Wthn-Subjects Y [ ] ( ) ε = N ε 0, σ 2 I = X β + ε for =, K, m m n Between-Subjects Heterogenety β = Θ z + [ δ ] = N ( δ 0, ) p δ 93
94 Average over Posteror Means and Std Dev of Partworths Across Subjects PostMean PostSTD PostMean PostSTD CNST Cache HotLne Color RAM Dstrbtn ScrnSz Wrrnty CPU Sftwr HrdDsk Grntee MultMd Prce
95 Impact of Covarates on Partworths CNST RAM CPU Dstrbtn Wrrnty Grntee Prce CNST FEMALE YEARS -0. OWN NERD APPLY 0.0 EXPERT
96 Summary HB allows ndvdual-level coeffcents Two level model Wth-n subjects Between subjects (heterogenety) HB shrnks unstable, subject-level estmators to populaton mean 96
97 Summary BDT provdes ntegrated framework for makng decsons and nference Good models consder all sources of uncertanty Good methods keep track of all sources of uncertanty Bayes does both 97
Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement
Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs
More informationCS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements
CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before
More informationParametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010
Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More informationMLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012
MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:
More informationEM and Structure Learning
EM and Structure Learnng Le Song Machne Learnng II: Advanced Topcs CSE 8803ML, Sprng 2012 Partally observed graphcal models Mxture Models N(μ 1, Σ 1 ) Z X N N(μ 2, Σ 2 ) 2 Gaussan mxture model Consder
More informationMotion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong
Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal
More informationA New Method for Estimating Overdispersion. David Fletcher and Peter Green Department of Mathematics and Statistics
A New Method for Estmatng Overdsperson Davd Fletcher and Peter Green Department of Mathematcs and Statstcs Byron Morgan Insttute of Mathematcs, Statstcs and Actuaral Scence Unversty of Kent, England Overvew
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed
More informationMATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)
1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons
More informationConjugacy and the Exponential Family
CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationBIO Lab 2: TWO-LEVEL NORMAL MODELS with school children popularity data
Lab : TWO-LEVEL NORMAL MODELS wth school chldren popularty data Purpose: Introduce basc two-level models for normally dstrbuted responses usng STATA. In partcular, we dscuss Random ntercept models wthout
More informationWeb Appendix B Estimation. We base our sampling procedure on the method of data augmentation (e.g., Tanner and Wong,
Web Appendx B Estmaton Lkelhood and Data Augmentaton We base our samplng procedure on the method of data augmentaton (eg anner and Wong 987) here e treat the unobserved ndvdual choces as parameters Specfcally
More informationThe Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD
he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s
More informationFirst Year Examination Department of Statistics, University of Florida
Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve
More informationMACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression
11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING
More informationAdvances in Longitudinal Methods in the Social and Behavioral Sciences. Finite Mixtures of Nonlinear Mixed-Effects Models.
Advances n Longtudnal Methods n the Socal and Behavoral Scences Fnte Mxtures of Nonlnear Mxed-Effects Models Jeff Harrng Department of Measurement, Statstcs and Evaluaton The Center for Integrated Latent
More informationOutline for today. Markov chain Monte Carlo. Example: spatial statistics (Christensen and Waagepetersen 2001)
Markov chan Monte Carlo Rasmus Waagepetersen Department of Mathematcs Aalborg Unversty Denmark November, / Outlne for today MCMC / Condtonal smulaton for hgh-dmensonal U: Markov chan Monte Carlo Consder
More informationChapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the
More informationβ0 + β1xi and want to estimate the unknown
SLR Models Estmaton Those OLS Estmates Estmators (e ante) v. estmates (e post) The Smple Lnear Regresson (SLR) Condtons -4 An Asde: The Populaton Regresson Functon B and B are Lnear Estmators (condtonal
More informationStat 543 Exam 2 Spring 2016
Stat 543 Exam 2 Sprng 206 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of questons. Do at least 0 of the parts of the man exam. I wll score
More informationDepartment of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6
Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.
More informationOutline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.
Outlne 3. Multple Regresson Analyss: Estmaton I. Motvaton II. Mechancs and Interpretaton of OLS Read Wooldrdge (013), Chapter 3. III. Expected Values of the OLS IV. Varances of the OLS V. The Gauss Markov
More informationSmall Area Interval Estimation
.. Small Area Interval Estmaton Partha Lahr Jont Program n Survey Methodology Unversty of Maryland, College Park (Based on jont work wth Masayo Yoshmor, Former JPSM Vstng PhD Student and Research Fellow
More informationStatistics for Economics & Business
Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable
More informationPopulation Design in Nonlinear Mixed Effects Multiple Response Models: extension of PFIM and evaluation by simulation with NONMEM and MONOLIX
Populaton Desgn n Nonlnear Mxed Effects Multple Response Models: extenson of PFIM and evaluaton by smulaton wth NONMEM and MONOLIX May 4th 007 Carolne Bazzol, Sylve Retout, France Mentré Inserm U738 Unversty
More informationBOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu
BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com
More informationLecture 3: Probability Distributions
Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the
More informationLINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables
LINEAR REGRESSION ANALYSIS MODULE VIII Lecture - 7 Indcator Varables Dr. Shalabh Department of Maematcs and Statstcs Indan Insttute of Technology Kanpur Indcator varables versus quanttatve explanatory
More informationMachine learning: Density estimation
CS 70 Foundatons of AI Lecture 3 Machne learnng: ensty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square ata: ensty estmaton {.. n} x a vector of attrbute values Objectve: estmate the model of
More informationIntroduction to Regression
Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes
More informationP R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /
Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons
More informationStatistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation
Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear
More informationSee Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)
Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes
More informationMIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU
Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern
More informationChapter 8 Indicator Variables
Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n
More information1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands
Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of
More informationMaximum Likelihood Estimation
Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?
More informationAdvanced Statistical Methods: Beyond Linear Regression
Advanced Statstcal Methods: Beyond Lnear Regresson John R. Stevens Utah State Unversty Notes 2. Statstcal Methods I Mathematcs Educators Workshop 28 March 2009 1 http://www.stat.usu.edu/~rstevens/pcm 2
More informationLab 4: Two-level Random Intercept Model
BIO 656 Lab4 009 Lab 4: Two-level Random Intercept Model Data: Peak expratory flow rate (pefr) measured twce, usng two dfferent nstruments, for 17 subjects. (from Chapter 1 of Multlevel and Longtudnal
More informationLecture 6: Introduction to Linear Regression
Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6
More informatione i is a random error
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown
More information4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA
4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected
More informationStat 543 Exam 2 Spring 2016
Stat 543 Exam 2 Sprng 2016 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of 11 questons. Do at least 10 of the 11 parts of the man exam.
More informationECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics
ECOOMICS 35*-A Md-Term Exam -- Fall Term 000 Page of 3 pages QUEE'S UIVERSITY AT KIGSTO Department of Economcs ECOOMICS 35* - Secton A Introductory Econometrcs Fall Term 000 MID-TERM EAM ASWERS MG Abbott
More informationArtificial Intelligence Bayesian Networks
Artfcal Intellgence Bayesan Networks Adapted from sldes by Tm Fnn and Mare desjardns. Some materal borrowed from Lse Getoor. 1 Outlne Bayesan networks Network structure Condtonal probablty tables Condtonal
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More informationChapter 12 Analysis of Covariance
Chapter Analyss of Covarance Any scentfc experment s performed to know somethng that s unknown about a group of treatments and to test certan hypothess about the correspondng treatment effect When varablty
More informationLinear regression. Regression Models. Chapter 11 Student Lecture Notes Regression Analysis is the
Chapter 11 Student Lecture Notes 11-1 Lnear regresson Wenl lu Dept. Health statstcs School of publc health Tanjn medcal unversty 1 Regresson Models 1. Answer What Is the Relatonshp Between the Varables?.
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationStatistics for Business and Economics
Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear
More informationSTAT 511 FINAL EXAM NAME Spring 2001
STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte
More informationTesting for seasonal unit roots in heterogeneous panels
Testng for seasonal unt roots n heterogeneous panels Jesus Otero * Facultad de Economía Unversdad del Rosaro, Colomba Jeremy Smth Department of Economcs Unversty of arwck Monca Gulett Aston Busness School
More informationExtending the Generalized Multinomial Logit Model: Error Scale and Decision-Maker Characteristics
Unversty of Pennsylvana ScholarlyCommons Marketng Papers Wharton Faculty Research 11-9 Extendng the Generaled Multnomal Logt Model: Error Scale and Decson-Maker Characterstcs Eleanor M. Fet Unversty of
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationBayesian decision theory. Nuno Vasconcelos ECE Department, UCSD
Bayesan decson theory Nuno Vasconcelos ECE Department UCSD Notaton the notaton n DHS s qute sloppy e.. show that error error z z dz really not clear what ths means we wll use the follown notaton subscrpts
More informationPrimer on High-Order Moment Estimators
Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc
More informationOn Outlier Robust Small Area Mean Estimate Based on Prediction of Empirical Distribution Function
On Outler Robust Small Area Mean Estmate Based on Predcton of Emprcal Dstrbuton Functon Payam Mokhtaran Natonal Insttute of Appled Statstcs Research Australa Unversty of Wollongong Small Area Estmaton
More informationThe EM Algorithm (Dempster, Laird, Rubin 1977) The missing data or incomplete data setting: ODL(φ;Y ) = [Y;φ] = [Y X,φ][X φ] = X
The EM Algorthm (Dempster, Lard, Rubn 1977 The mssng data or ncomplete data settng: An Observed Data Lkelhood (ODL that s a mxture or ntegral of Complete Data Lkelhoods (CDL. (1a ODL(;Y = [Y;] = [Y,][
More informationChecking Pairwise Relationships. Lecture 19 Biostatistics 666
Checkng Parwse Relatonshps Lecture 19 Bostatstcs 666 Last Lecture: Markov Model for Multpont Analyss X X X 1 3 X M P X 1 I P X I P X 3 I P X M I 1 3 M I 1 I I 3 I M P I I P I 3 I P... 1 IBD states along
More informationChapter 3 Describing Data Using Numerical Measures
Chapter 3 Student Lecture Notes 3-1 Chapter 3 Descrbng Data Usng Numercal Measures Fall 2006 Fundamentals of Busness Statstcs 1 Chapter Goals To establsh the usefulness of summary measures of data. The
More informationComparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method
Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method
More informationENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition
EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30
More informationExplaining the Stein Paradox
Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationClassification as a Regression Problem
Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationJAB Chain. Long-tail claims development. ASTIN - September 2005 B.Verdier A. Klinger
JAB Chan Long-tal clams development ASTIN - September 2005 B.Verder A. Klnger Outlne Chan Ladder : comments A frst soluton: Munch Chan Ladder JAB Chan Chan Ladder: Comments Black lne: average pad to ncurred
More informationBAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup
BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS Darusz Bskup 1. Introducton The paper presents a nonparaetrc procedure for estaton of an unknown functon f n the regresson odel y = f x + ε = N. (1) (
More informationTarget tracking example Filtering: Xt. (main interest) Smoothing: X1: t. (also given with SIS)
Target trackng example Flterng: Xt Y1: t (man nterest) Smoothng: X1: t Y1: t (also gven wth SIS) However as we have seen, the estmate of ths dstrbuton breaks down when t gets large due to the weghts becomng
More informationEvaluation for sets of classes
Evaluaton for Tet Categorzaton Classfcaton accuracy: usual n ML, the proporton of correct decsons, Not approprate f the populaton rate of the class s low Precson, Recall and F 1 Better measures 21 Evaluaton
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More informationMaximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models
ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationLimited Dependent Variables and Panel Data. Tibor Hanappi
Lmted Dependent Varables and Panel Data Tbor Hanapp 30.06.2010 Lmted Dependent Varables Dscrete: Varables that can take onl a countable number of values Censored/Truncated: Data ponts n some specfc range
More informationLecture 3 Stat102, Spring 2007
Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More informationImproved conditional approximations of the population Fisher Information Matrix
Improved condtonal approxmatons of the populaton Fsher Informaton Matrx Joakm Nyberg, Sebastan Ueckert and Andrew C. Hooker Pharmacometrcs research group Department of Pharmaceutcal Boscences Uppsala Unversty
More information8 : Learning in Fully Observed Markov Networks. 1 Why We Need to Learn Undirected Graphical Models. 2 Structural Learning for Completely Observed MRF
10-708: Probablstc Graphcal Models 10-708, Sprng 2014 8 : Learnng n Fully Observed Markov Networks Lecturer: Erc P. Xng Scrbes: Meng Song, L Zhou 1 Why We Need to Learn Undrected Graphcal Models In the
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Experment-I MODULE VIII LECTURE - 34 ANALYSIS OF VARIANCE IN RANDOM-EFFECTS MODEL AND MIXED-EFFECTS EFFECTS MODEL Dr Shalabh Department of Mathematcs and Statstcs Indan
More informationMarkov Chain Monte Carlo Lecture 6
where (x 1,..., x N ) X N, N s called the populaton sze, f(x) f (x) for at least one {1, 2,..., N}, and those dfferent from f(x) are called the tral dstrbutons n terms of mportance samplng. Dfferent ways
More informationThe conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above
The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ
More informationGaussian Mixture Models
Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous
More informationBias-correction under a semi-parametric model for small area estimation
Bas-correcton under a sem-parametrc model for small area estmaton Laura Dumtrescu, Vctora Unversty of Wellngton jont work wth J. N. K. Rao, Carleton Unversty ICORS 2017 Workshop on Robust Inference for
More informationMultiple Choice. Choose the one that best completes the statement or answers the question.
ECON 56 Homework Multple Choce Choose the one that best completes the statement or answers the queston ) The probablty of an event A or B (Pr(A or B)) to occur equals a Pr(A) Pr(B) b Pr(A) + Pr(B) f A
More information8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore
8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø
More informationProbability Theory (revisited)
Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted
More information( ) ( ) ( ) ( ) STOCHASTIC SIMULATION FOR BLOCKED DATA. Monte Carlo simulation Rejection sampling Importance sampling Markov chain Monte Carlo
SOCHASIC SIMULAIO FOR BLOCKED DAA Stochastc System Analyss and Bayesan Model Updatng Monte Carlo smulaton Rejecton samplng Importance samplng Markov chan Monte Carlo Monte Carlo smulaton Introducton: If
More informationLossy Compression. Compromise accuracy of reconstruction for increased compression.
Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost
More informationBasic Business Statistics, 10/e
Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson
More informationComputing MLE Bias Empirically
Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.
More informationBayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County
Smart Home Health Analytcs Sprng 2018 Bayesan Learnng Nrmalya Roy Department of Informaton Systems Unversty of Maryland Baltmore ounty www.umbc.edu Bayesan Learnng ombnes pror knowledge wth evdence to
More informationOn an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1
On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool
More informationStatistical inference for generalized Pareto distribution based on progressive Type-II censored data with random removals
Internatonal Journal of Scentfc World, 2 1) 2014) 1-9 c Scence Publshng Corporaton www.scencepubco.com/ndex.php/ijsw do: 10.14419/jsw.v21.1780 Research Paper Statstcal nference for generalzed Pareto dstrbuton
More informationHere is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)
Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,
More informationLet the Shape Speak - Discriminative Face Alignment using Conjugate Priors
Let the Shape Speak - Dscrmnatve Face Algnment usng Conjugate Prors Pedro Martns, Ru Casero, João F. Henrques, Jorge Batsta http://www.sr.uc.pt/~pedromartns Insttute of Systems and Robotcs Unversty of
More information