Prediction and Change Detection

Size: px
Start display at page:

Download "Prediction and Change Detection"

Transcription

1 Predcton and Change Detecton Mark Steyvers Scott Brown Unversty of Calforna, Irvne Unversty of Calforna, Irvne Irvne, CA Irvne, CA Abstract We measure the ablty of human observers to predct the next datum n a sequence that s generated by a smple statstcal process undergong change at random ponts n tme. Accurate performance n ths task requres the dentfcaton of changeponts. We assess ndvdual dfferences between observers both emprcally, and usng two knds of models: a Bayesan approach for change detecton and a famly of cogntvely plausble fast and frugal models. Some ndvduals detect too many changes and hence perform sub-optmally due to excess varablty. Other ndvduals do not detect enough changes, and perform sub-optmally because they fal to notce short-term temporal trends. Introducton Decson-makng often requres a rapd response to change. For example, stock analysts need to quckly detect changes n the market n order to adjust nvestment strateges. Coaches need to track changes n a player s performance n order to adjust strategy. When trackng changes, there are costs nvolved when ether more or less changes are observed than actually occurred. For example, when usng an overly conservatve change detecton crteron, a stock analyst mght mss mportant short-term trends and nterpret them as random fluctuatons nstead. On the other hand, a change may also be detected too readly. For example, n basketball, a player who makes a seres of consecutve baskets s often dentfed as a hot hand player whose underlyng ablty s perceved to have suddenly ncreased [,2]. Ths mght lead to sub-optmal passng strateges, based on random fluctuatons. We are nterested n explanng ndvdual dfferences n a sequental predcton task. Observers are shown stmul generated from a smple statstcal process wth the task of predctng the next datum n the sequence. The latent parameters of the statstcal process change dscretely at random ponts n tme. Performance n ths task depends on the accurate detecton of those changeponts, as well as nference about future outcomes based on the outcomes that followed the most recent nferred changepont. There s much pror research n statstcs on the problem of dentfyng changeponts [3,4,5]. In ths paper, we adopt a Bayesan approach to the changepont dentfcaton problem and develop a smple nference procedure to predct the next datum n a sequence. The Bayesan model serves as an deal observer model and s useful to characterze the ways n whch ndvduals devate from optmalty.

2 The plan of the paper s as follows. We frst ntroduce the sequental predcton task and dscuss a Bayesan analyss of ths predcton problem. We then dscuss the results from a few ndvduals n ths predcton task and show how the Bayesan approach can capture ndvdual dfferences wth a sngle twtchness parameter that descrbes how readly changes are perceved n random sequences. We wll show that some ndvduals are too twtchy: ther performance s too varable because they base ther predctons on too lttle of the recent data. Other ndvduals are not twtchy enough, and they fal to capture fast changes n the data. We also show how behavor can be explaned wth a set of fast and frugal models [6]. These are cogntvely realstc models that operate under plausble computatonal constrants. 2 A predcton task wth multple changeponts In the predcton task, stmul are presented sequentally and the task s to predct the next stmulus n the sequence. After t trals, the observer has been presented wth stmul y, y 2,, y t and the task s to make a predcton about y t+. After the predcton s made, the actual outcome y t+ s revealed and the next tral proceeds to the predcton of y t+2. Ths procedure starts wth y and s repeated for T trals. The observatons y t are D-dmensonal vectors wth elements sampled from bnomal dstrbutons. The parameters of those dstrbutons change dscretely at random ponts n tme such that the mean ncreases or decreases after a change pont. Ths generates a sequence of observaton vectors, y, y 2,, y T, where each y t = {y t, y t,d }. Each of the y t,d s sampled from a bnomal dstrbuton Bn(θ t,d,k), so y t,d K. The parameter vector θ t ={θ t, θ t,d } changes dependng on the locatons of the changeponts. At each tme step, x t s a bnary ndcator for the occurrence of a changepont occurrng at tme t+. The parameter α determnes the probablty of a change occurrng n the sequence. The generatve model s specfed by the followng algorthm:. For d=..d sample θ,d from a Unform(,) dstrbuton 2. For t=2..t, (a) Sample x t- from a Bernoull(α) dstrbuton (b) If x t- =, then θ t =θ t-, else for d=..d sample θ t,d from a Unform(,) dstrbuton (c) for d=..d, sample y t from a Bn(θ t,d,k) dstrbuton Table shows some data generated from the changepont model wth T=2, α=.,and D=. In the predcton task, y wll be observed, but x and θ are not. Table : Example data t x θ y

3 3 A Bayesan predcton model In both our Bayesan and fast-and-frugal analyses, the predcton task s decomposed nto two nference procedures. Frst, the changepont locatons are dentfed. Ths s followed by predctve nference for the next outcome based on the most recent changepont locatons. Several Bayesan approaches have been developed for changepont problems nvolvng sngle or multple changeponts [3,5]. We apply a Markov Chan Monte Carlo (MCMC) analyss to approxmate the jont posteror dstrbuton over changepont assgnments x whle ntegratng out θ. Gbbs samplng wll be used to sample from ths posteror margnal dstrbuton. The samples can then be used to predct the next outcome n the sequence. 3. Inference for changepont assgnments. To apply Gbbs samplng, we evaluate the condtonal probablty of assgnng a changepont at tme, gven all other changepont assgnments and the current α value. By ntegratng out θ, the condtonal probablty s (,, α) (, θ, α, ) P x x y = P x x y () θ where x represents all swtch pont assgnments except x. Ths can be smplfed by consderng the locaton of the most recent changepont precedng and followng tme L and the outcomes occurrng between these locatons. Let n be the number of tme steps from the last changepont up to and ncludng the current tme step such that x = and x = for <j< L R L n. Smlarly, let n n L n + j be the number of tme steps that follow tme step up to the next changepont such that x = and x = for + + R <j< R L n. Let y = L y and R n k y < k = y. The update equaton for the k< k + n R k changepont assgnment can then be smplfed to ( m x ) P x = R n D L R L R L R ( y, j y, j) ( Kn Kn y, j y, j) L R j= Γ ( 2 + Kn + Kn ) L L L R R R ( yj, ) ( Kn yj, ) ( yj, ) ( Kn yj, ) L R Γ ( 2+ Kn ) Γ ( 2+ Kn ) Γ + + Γ + + ( α ) m = D Γ + Γ + Γ + Γ + α j= m = We ntalze the Gbbs sampler by samplng each x t from a Bernoull(α) dstrbuton. All changepont assgnments are then updated sequentally by the Gbbs samplng equaton above. The sampler s run for M teratons after whch one set of changepont assgnments s saved. The Gbbs sampler s then restarted multple tmes untl S samples have been collected. Although we could have ncluded an update equaton for α, n ths analyss we treat α as a known constant. Ths wll be useful when characterzng the dfferences between human observers n terms of dfferences n α. n j (2)

4 3.2 Predctve nference The next latent parameter value θ t+ and outcome y t+ can be predcted on the bass of observed outcomes that occurred after the last nferred changepont: θ t ( θ ) = y / K, y = round K (3) t+, j, j t+, j t+, j * = t + where t * s the locaton of the most recent change pont. By consderng multple Gbbs samples, we get a dstrbuton over outcomes y t+. We base the model predctons on the mean of ths dstrbuton. 3.3 Illustraton of model performance Fgure llustrates the performance of the model on a one dmensonal sequence (D=) generated from the changepont model wth T=6, α=.5, and K=. The Gbbs sampler was run for M=3 teratons and S=2 samples were collected. The top panel shows the actual changeponts (trangles) and the dstrbuton of changepont assgnments averaged over samples. The bottom panel shows the observed data y (thn lnes) as well as the θ values n the generatve model (rescaled between and ). At locatons wth large changes between observatons, the margnal changepont probablty s qute hgh. At other locatons, the true change n the mean s very small, and the model s less lkely to put n a changepont. The lower rght panel shows the dstrbuton over predcted θ t+ values. x t.5 θ t+ y t Fgure. Results of model smulaton. 4 Predcton experment We tested performance of 9 human observers n the predcton task. The observers ncluded the authors, a vstor, and one student who were aware of the statstcal nature of the task as well as naïve students. The observers were seated n front of an LCD touch screen dsplayng a two-dmensonal grd of x buttons. The changepont model was used to generate a sequence of T=5 stmul for two bnomal varables y and y 2 (D=2, K=). The change probablty α was set to.. The two varables y and y 2 specfed the two-dmensonal button locaton. The same sequence was used for all observers. On each tral, the observer touched a button on the grd dsplayed on the touch screen. Followng each button press, the button correspondng to the next {y,y 2 } outcome n the sequence was hghlghted. Observers were nstructed to press the button that best predcted the next locaton of the hghlghted button. The 5 trals were dvded nto

5 three blocks of 5 trals. Breaks were allowed between blocks. The whole experment lasted between 5 and 3 mnutes. Fgure 2 shows the frst 5 trals from the thrd block of the experment. The top and bottom panels show the actual outcomes for the y and y 2 button grd coordnates as well as the predctons for two observers (SB and MY). The fgure shows that at tral 5, the y and y 2 coordnates show a large shft followed by an mmedate shft n observer s MY predctons (on tral 6). Observer SB wats untl tral 7 to make a shft. 5 5 outcomes SB predctons MY predctons Tral Fgure 2. Tral by tral predctons from two observers. 4. Task error We assessed predcton performance by comparng the predcton wth the actual outcome n the sequence. Task error was measured by normalzed cty-block dstance task error= T O O yt, yt, + yt,2 y (4) t,2 ( ) T t= 2 where y O represents the observer s predcton. Note that the very frst tral s excluded from ths calculaton. Even though more sutable probablstc measures for predcton error could have been adopted, we wanted to allow comparson of observer s performance wth both probablstc and non-probablstc models. Task error ranged from 2.8 (for partcpant MY) to 3.3 (for ML). We also assessed the performance of fve models ther task errors ranged from 2.78 to 3.2. The Bayesan models (Secton 3) had the lowest task errors, just below 2.8. Ths fts wth our defnton of the Bayesan models as deal observer models ther task error s lower than any other model s and any human observer s task error. The fast and frugal models (Secton 5) had task errors rangng from 2.85 to Modelng Results We wll refer to the models wth the followng letter codes: B=Bayesan Model, LB=lmted Bayesan model, FF..3=fast and frugal models..3. We assessed model ft by comparng the model s predcton aganst the human observers predctons, agan usng a normalzed cty-block dstance

6 model error= ( T ) M O M O yt, yt, + yt,2 y (5) t,2 T t= 2 where y M represents the model s predcton. The model error for each ndvdual observer s shown n Fgure 3. It s mportant to note that because each model s assocated wth a set of free parameters, the parameters optmzed for task error and model error are dfferent. For Fgure 3, the parameters were optmzed to mnmze Equaton (5) for each ndvdual observer, showng the extent to whch these models can capture the performance of ndvdual observers, not necessarly provdng the best task performance. 2 B LB FF FF2 FF3 Model Error.5.5 MY MS MM EJ PH NP DN SB ML Fgure 3. Model error for each ndvdual observer. 5. Bayesan predcton models At each tral t, the model was provded wth the sequence of all prevous outcomes. The Gbbs samplng and nference procedures from Eq. (2) and (3) were appled wth M=3 teratons and S=2 samples. The change probablty α was a free parameter. In the full Bayesan model, the whole sequence of observatons up to the current tral s avalable for predcton, leadng to a memory requrement of up to T=5 trals a psychologcally unreasonable assumpton. We therefore also smulated a lmted Bayesan model (LB) where the observed sequence was truncated to the last outcomes. The LB model showed almost no decrement n task performance compared to the full Bayesan model. Fgure 3 also shows that t ft human data qute well. 5.2 Indvdual Dfferences The rght-hand panel of Fgure 4 plots each observer s task error as a functon of the mean cty-block dstance between ther subsequent button presses. Ths shows a clear U-shaped functon. Observers wth very varable predctons (e.g., ML and DN) had large average changes between successve button pushes, and also had large task error: These observers were too twtchy. Observers wth very small average button changes (e.g., SB and NP) were not twtchy enough, and also had large task error. Observers n the mddle had the lowest task error (e.g., MS and MY). The left-hand panel of Fgure 4 shows the same data, but wth the x-axs based on the Bayesan model fts. Instead of usng mean button change dstance to ndex twtchness (as n Error bars ndcate bootstrapped 95% confdence ntervals.

7 the rght-hand panel), the left-hand panel uses the estmated α parameters from the Bayesan model. A smlar U-shaped pattern s observed: ndvduals wth too large or too small α estmates have large task errors. 3.3 SB ML 3.3 SB ML 3.2 NP DN 3.2 NP DN Task Error 3. 3 EJ MM PH Task Error 3. 3 EJ MM PH B MS MY MS MY α Mean Button Change Fgure 4. Task error vs. twtchness. Left-hand panel ndexes twtchness usng estmated α parameters from Bayesan model fts. Rght-hand panel uses mean dstance between successve predctons. 5.3 Fast-and-Frugal (FF) predcton models These models perform the predcton task usng smple heurstcs that are cogntvely plausble. The FF models keep a short memory of prevous stmulus values and make predctons usng the same two-step process as the Bayesan model. Frst, a decson s made as to whether the latent parameter θ has changed. Second, remembered stmulus values that occurred after the most recently detected changepont are used to generate the next predcton. A smple heurstc s used to detect changeponts: If the dstance between the most recent observaton and predcton s greater than some threshold amount, a change s nferred. We defned the dstance between a predcton (p) and an observaton (y) as the dfference between the log-lkelhoods of y assumng θ=p and θ=y. Thus, f f B (. θ, K) s the bnomal densty wth parameters θ and K, the dstance between observaton y and predcton p s defned as d(y,p)=log(f B (y y,k))-log(f B (y p,k)). A changepont on tme step t+ s nferred whenever d(y t,p t )>C. The parameter C governs the twtchness of the model predctons. If C s large, only very dramatc changeponts wll be detected, and the model wll be too conservatve. If C s small, the model wll be too twtchy, and wll detect changeponts on the bass of small random fluctuatons. Predctons are based on the most recent M observatons, whch are kept n memory, unless a changepont has been detected n whch case only those observatons occurrng after the changepont are used for predcton. The predcton for tme step t+ s smply the mean of these observatons, say p. Human observers were retcent to make predctons very close to the boundares. Ths was modeled by allowng the FF model to change ts predcton for the next tme step, y t+, towards the mean predcton (.5). Ths change reflects a two-way bet. If the probablty of a change occurrng s α, the best guess wll be.5 f that change occurs, or the mean p f the change does not occur. Thus, the predcton made s actually y t+ =/2 α+(-α)p. Note that we do not allow perfect knowledge of the probablty of a changepont, α. Instead, an estmated value of α s used based on the number of changeponts detected n the data seres up to tme t.

8 The FF model nests two smpler FF models that are psychologcally nterestng. If the twtchness threshold parameter C becomes arbtrarly large, the model never detects a change and nstead becomes a contnuous runnng average model. Predctons from ths model are smply a boxcar smooth of the data. Alternatvely, f we assume no memory the model must based each predcton on only the prevous stmulus (.e., M=). Above, n Fgure 3, we labeled the complete FF model as FF, the boxcar model as FF2 and the memoryless model FF3. Fgure 3 showed that the complete FF model (FF) ft the data from all observers sgnfcantly better than ether the boxcar model (FF2) or the memoryless model (FF3). Exceptons were observers PH, DN and ML, for whom all three FF model ft equally well. Ths result suggests that our observers were (mostly) dong more than just keepng a runnng average of the data, or usng only the most recent observaton. The FF model ft the data about as well as the Bayesan models for all observers except MY and MS. Note that, n general, the FF and Bayesan model fts are very good: the average cty block dstance between the human data and the model predcton s around.75 (out of ) buttons on both the x- and y-axes. 6 Concluson We used an onlne predcton task to study changepont detecton. Human observers had to predct the next observaton n stochastc sequences contanng random changeponts. We showed that some observers are too twtchy : They perform poorly on the predcton task because they see changes where only random fluctuaton exsts. Other observers are not twtchy enough, and they perform poorly because they fal to see small changes. We developed a Bayesan changepont detecton model that performed the task optmally, and also provded a good ft to human data when sub-optmal parameter settngs were used. Fnally, we developed a fast-and-frugal model that showed how partcpants may be able to perform well at the task usng mnmal nformaton and smple decson heurstcs. Acknowledgments We thank Erc-Jan Wagenmakers and Mke Y for useful dscussons related to ths work. Ths work was supported n part by a grant from the US Ar Force Offce of Scentfc Research (AFOSR grant number FA ). References [] Glovch, T., Vallone, R. and Tversky, A. (985). The hot hand n basketball: on the mspercepton of random sequences. Cogntve Psychology7, [2] Albrght, S.C. (993a). A statstcal analyss of httng streaks n baseball. Journal of the Amercan Statstcal Assocaton, 88, [3] Stephens, D.A. (994). Bayesan retrospectve multple changepont dentfcaton. Appled Statstcs 43(), [4] Carln, B.P., Gelfand, A.E., & Smth, A.F.M. (992). Herarchcal Bayesan analyss of changepont problems. Appled Statstcs 4(2), [5] Green, P.J. (995). Reversble jump Markov chan Monte Carlo computaton and Bayesan model determnaton. Bometrka 82(4), [6] Ggerenzer, G., & Goldsten, D.G. (996). Reasonng the fast and frugal way: Models of bounded ratonalty. Psychologcal Revew, 3,

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Hidden Markov Models

Hidden Markov Models CM229S: Machne Learnng for Bonformatcs Lecture 12-05/05/2016 Hdden Markov Models Lecturer: Srram Sankararaman Scrbe: Akshay Dattatray Shnde Edted by: TBD 1 Introducton For a drected graph G we can wrte

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study

More information

Definition. Measures of Dispersion. Measures of Dispersion. Definition. The Range. Measures of Dispersion 3/24/2014

Definition. Measures of Dispersion. Measures of Dispersion. Definition. The Range. Measures of Dispersion 3/24/2014 Measures of Dsperson Defenton Range Interquartle Range Varance and Standard Devaton Defnton Measures of dsperson are descrptve statstcs that descrbe how smlar a set of scores are to each other The more

More information

Artificial Intelligence Bayesian Networks

Artificial Intelligence Bayesian Networks Artfcal Intellgence Bayesan Networks Adapted from sldes by Tm Fnn and Mare desjardns. Some materal borrowed from Lse Getoor. 1 Outlne Bayesan networks Network structure Condtonal probablty tables Condtonal

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

Chapter 8 Indicator Variables

Chapter 8 Indicator Variables Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n

More information

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30 STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Convergence of random processes

Convergence of random processes DS-GA 12 Lecture notes 6 Fall 216 Convergence of random processes 1 Introducton In these notes we study convergence of dscrete random processes. Ths allows to characterze phenomena such as the law of large

More information

Bayesian Planning of Hit-Miss Inspection Tests

Bayesian Planning of Hit-Miss Inspection Tests Bayesan Plannng of Ht-Mss Inspecton Tests Yew-Meng Koh a and Wllam Q Meeker a a Center for Nondestructve Evaluaton, Department of Statstcs, Iowa State Unversty, Ames, Iowa 5000 Abstract Although some useful

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Online Appendix to: Axiomatization and measurement of Quasi-hyperbolic Discounting

Online Appendix to: Axiomatization and measurement of Quasi-hyperbolic Discounting Onlne Appendx to: Axomatzaton and measurement of Quas-hyperbolc Dscountng José Lus Montel Olea Tomasz Strzaleck 1 Sample Selecton As dscussed before our ntal sample conssts of two groups of subjects. Group

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Copyright 2017 by Taylor Enterprises, Inc., All Rights Reserved. Adjusted Control Limits for P Charts. Dr. Wayne A. Taylor

Copyright 2017 by Taylor Enterprises, Inc., All Rights Reserved. Adjusted Control Limits for P Charts. Dr. Wayne A. Taylor Taylor Enterprses, Inc. Control Lmts for P Charts Copyrght 2017 by Taylor Enterprses, Inc., All Rghts Reserved. Control Lmts for P Charts Dr. Wayne A. Taylor Abstract: P charts are used for count data

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Chapter 5 Multilevel Models

Chapter 5 Multilevel Models Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level

More information

Polynomial Regression Models

Polynomial Regression Models LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance

More information

Chapter 15 - Multiple Regression

Chapter 15 - Multiple Regression Chapter - Multple Regresson Chapter - Multple Regresson Multple Regresson Model The equaton that descrbes how the dependent varable y s related to the ndependent varables x, x,... x p and an error term

More information

This column is a continuation of our previous column

This column is a continuation of our previous column Comparson of Goodness of Ft Statstcs for Lnear Regresson, Part II The authors contnue ther dscusson of the correlaton coeffcent n developng a calbraton for quanttatve analyss. Jerome Workman Jr. and Howard

More information

Statistics Chapter 4

Statistics Chapter 4 Statstcs Chapter 4 "There are three knds of les: les, damned les, and statstcs." Benjamn Dsrael, 1895 (Brtsh statesman) Gaussan Dstrbuton, 4-1 If a measurement s repeated many tmes a statstcal treatment

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Joint Statistical Meetings - Biopharmaceutical Section

Joint Statistical Meetings - Biopharmaceutical Section Iteratve Ch-Square Test for Equvalence of Multple Treatment Groups Te-Hua Ng*, U.S. Food and Drug Admnstraton 1401 Rockvlle Pke, #200S, HFM-217, Rockvlle, MD 20852-1448 Key Words: Equvalence Testng; Actve

More information

Uncertainty as the Overlap of Alternate Conditional Distributions

Uncertainty as the Overlap of Alternate Conditional Distributions Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant

More information

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics ECOOMICS 35*-A Md-Term Exam -- Fall Term 000 Page of 3 pages QUEE'S UIVERSITY AT KIGSTO Department of Economcs ECOOMICS 35* - Secton A Introductory Econometrcs Fall Term 000 MID-TERM EAM ASWERS MG Abbott

More information

CS-433: Simulation and Modeling Modeling and Probability Review

CS-433: Simulation and Modeling Modeling and Probability Review CS-433: Smulaton and Modelng Modelng and Probablty Revew Exercse 1. (Probablty of Smple Events) Exercse 1.1 The owner of a camera shop receves a shpment of fve cameras from a camera manufacturer. Unknown

More information

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS Darusz Bskup 1. Introducton The paper presents a nonparaetrc procedure for estaton of an unknown functon f n the regresson odel y = f x + ε = N. (1) (

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

Introduction to Regression

Introduction to Regression Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014 COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture #16 Scrbe: Yannan Wang Aprl 3, 014 1 Introducton The goal of our onlne learnng scenaro from last class s C comparng wth best expert and

More information

Chapter 1. Probability

Chapter 1. Probability Chapter. Probablty Mcroscopc propertes of matter: quantum mechancs, atomc and molecular propertes Macroscopc propertes of matter: thermodynamcs, E, H, C V, C p, S, A, G How do we relate these two propertes?

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Statistical Evaluation of WATFLOOD

Statistical Evaluation of WATFLOOD tatstcal Evaluaton of WATFLD By: Angela MacLean, Dept. of Cvl & Envronmental Engneerng, Unversty of Waterloo, n. ctober, 005 The statstcs program assocated wth WATFLD uses spl.csv fle that s produced wth

More information

A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS

A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS HCMC Unversty of Pedagogy Thong Nguyen Huu et al. A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS Thong Nguyen Huu and Hao Tran Van Department of mathematcs-nformaton,

More information

Generative classification models

Generative classification models CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

Lecture 3 Stat102, Spring 2007

Lecture 3 Stat102, Spring 2007 Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture

More information

Checking Pairwise Relationships. Lecture 19 Biostatistics 666

Checking Pairwise Relationships. Lecture 19 Biostatistics 666 Checkng Parwse Relatonshps Lecture 19 Bostatstcs 666 Last Lecture: Markov Model for Multpont Analyss X X X 1 3 X M P X 1 I P X I P X 3 I P X M I 1 3 M I 1 I I 3 I M P I I P I 3 I P... 1 IBD states along

More information

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected.

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected. ANSWERS CHAPTER 9 THINK IT OVER thnk t over TIO 9.: χ 2 k = ( f e ) = 0 e Breakng the equaton down: the test statstc for the ch-squared dstrbuton s equal to the sum over all categores of the expected frequency

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

= z 20 z n. (k 20) + 4 z k = 4

= z 20 z n. (k 20) + 4 z k = 4 Problem Set #7 solutons 7.2.. (a Fnd the coeffcent of z k n (z + z 5 + z 6 + z 7 + 5, k 20. We use the known seres expanson ( n+l ( z l l z n below: (z + z 5 + z 6 + z 7 + 5 (z 5 ( + z + z 2 + z + 5 5

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

More metrics on cartesian products

More metrics on cartesian products More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear

More information

Non-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT

Non-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT Malaysan Journal of Mathematcal Scences 8(S): 37-44 (2014) Specal Issue: Internatonal Conference on Mathematcal Scences and Statstcs 2013 (ICMSS2013) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES Journal

More information

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1 C/CS/Phy9 Problem Set 3 Solutons Out: Oct, 8 Suppose you have two qubts n some arbtrary entangled state ψ You apply the teleportaton protocol to each of the qubts separately What s the resultng state obtaned

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors Stat60: Bayesan Modelng and Inference Lecture Date: February, 00 Reference Prors Lecturer: Mchael I. Jordan Scrbe: Steven Troxler and Wayne Lee In ths lecture, we assume that θ R; n hgher-dmensons, reference

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

Laboratory 1c: Method of Least Squares

Laboratory 1c: Method of Least Squares Lab 1c, Least Squares Laboratory 1c: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y) Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,

More information

Probability Theory (revisited)

Probability Theory (revisited) Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted

More information

Explaining the Stein Paradox

Explaining the Stein Paradox Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten

More information

International Journal of Industrial Engineering Computations

International Journal of Industrial Engineering Computations Internatonal Journal of Industral Engneerng Computatons 4 (03) 47 46 Contents lsts avalable at GrowngScence Internatonal Journal of Industral Engneerng Computatons homepage: www.growngscence.com/ec Modelng

More information

Gravitational Acceleration: A case of constant acceleration (approx. 2 hr.) (6/7/11)

Gravitational Acceleration: A case of constant acceleration (approx. 2 hr.) (6/7/11) Gravtatonal Acceleraton: A case of constant acceleraton (approx. hr.) (6/7/11) Introducton The gravtatonal force s one of the fundamental forces of nature. Under the nfluence of ths force all objects havng

More information

Introduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:

Introduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law: CE304, Sprng 2004 Lecture 4 Introducton to Vapor/Lqud Equlbrum, part 2 Raoult s Law: The smplest model that allows us do VLE calculatons s obtaned when we assume that the vapor phase s an deal gas, and

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Statistics for Business and Economics

Statistics for Business and Economics Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear

More information

THE SUMMATION NOTATION Ʃ

THE SUMMATION NOTATION Ʃ Sngle Subscrpt otaton THE SUMMATIO OTATIO Ʃ Most of the calculatons we perform n statstcs are repettve operatons on lsts of numbers. For example, we compute the sum of a set of numbers, or the sum of the

More information

Appendix B: Resampling Algorithms

Appendix B: Resampling Algorithms 407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

SPANC -- SPlitpole ANalysis Code User Manual

SPANC -- SPlitpole ANalysis Code User Manual Functonal Descrpton of Code SPANC -- SPltpole ANalyss Code User Manual Author: Dale Vsser Date: 14 January 00 Spanc s a code created by Dale Vsser for easer calbratons of poston spectra from magnetc spectrometer

More information

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

Lossy Compression. Compromise accuracy of reconstruction for increased compression. Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost

More information

One-sided finite-difference approximations suitable for use with Richardson extrapolation

One-sided finite-difference approximations suitable for use with Richardson extrapolation Journal of Computatonal Physcs 219 (2006) 13 20 Short note One-sded fnte-dfference approxmatons sutable for use wth Rchardson extrapolaton Kumar Rahul, S.N. Bhattacharyya * Department of Mechancal Engneerng,

More information

x yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting.

x yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting. The Practce of Statstcs, nd ed. Chapter 14 Inference for Regresson Introducton In chapter 3 we used a least-squares regresson lne (LSRL) to represent a lnear relatonshp etween two quanttatve explanator

More information

28. SIMPLE LINEAR REGRESSION III

28. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III Ftted Values and Resduals US Domestc Beers: Calores vs. % Alcohol To each observed x, there corresponds a y-value on the ftted lne, y ˆ = βˆ + βˆ x. The are called ftted

More information

Introduction to Generalized Linear Models

Introduction to Generalized Linear Models INTRODUCTION TO STATISTICAL MODELLING TRINITY 00 Introducton to Generalzed Lnear Models I. Motvaton In ths lecture we extend the deas of lnear regresson to the more general dea of a generalzed lnear model

More information

FREQUENCY DISTRIBUTIONS Page 1 of The idea of a frequency distribution for sets of observations will be introduced,

FREQUENCY DISTRIBUTIONS Page 1 of The idea of a frequency distribution for sets of observations will be introduced, FREQUENCY DISTRIBUTIONS Page 1 of 6 I. Introducton 1. The dea of a frequency dstrbuton for sets of observatons wll be ntroduced, together wth some of the mechancs for constructng dstrbutons of data. Then

More information

EM and Structure Learning

EM and Structure Learning EM and Structure Learnng Le Song Machne Learnng II: Advanced Topcs CSE 8803ML, Sprng 2012 Partally observed graphcal models Mxture Models N(μ 1, Σ 1 ) Z X N N(μ 2, Σ 2 ) 2 Gaussan mxture model Consder

More information

Information Geometry of Gibbs Sampler

Information Geometry of Gibbs Sampler Informaton Geometry of Gbbs Sampler Kazuya Takabatake Neuroscence Research Insttute AIST Central 2, Umezono 1-1-1, Tsukuba JAPAN 305-8568 k.takabatake@ast.go.jp Abstract: - Ths paper shows some nformaton

More information