Prediction and Change Detection
|
|
- Brittney Bates
- 5 years ago
- Views:
Transcription
1 Predcton and Change Detecton Mark Steyvers Scott Brown Unversty of Calforna, Irvne Unversty of Calforna, Irvne Irvne, CA Irvne, CA Abstract We measure the ablty of human observers to predct the next datum n a sequence that s generated by a smple statstcal process undergong change at random ponts n tme. Accurate performance n ths task requres the dentfcaton of changeponts. We assess ndvdual dfferences between observers both emprcally, and usng two knds of models: a Bayesan approach for change detecton and a famly of cogntvely plausble fast and frugal models. Some ndvduals detect too many changes and hence perform sub-optmally due to excess varablty. Other ndvduals do not detect enough changes, and perform sub-optmally because they fal to notce short-term temporal trends. Introducton Decson-makng often requres a rapd response to change. For example, stock analysts need to quckly detect changes n the market n order to adjust nvestment strateges. Coaches need to track changes n a player s performance n order to adjust strategy. When trackng changes, there are costs nvolved when ether more or less changes are observed than actually occurred. For example, when usng an overly conservatve change detecton crteron, a stock analyst mght mss mportant short-term trends and nterpret them as random fluctuatons nstead. On the other hand, a change may also be detected too readly. For example, n basketball, a player who makes a seres of consecutve baskets s often dentfed as a hot hand player whose underlyng ablty s perceved to have suddenly ncreased [,2]. Ths mght lead to sub-optmal passng strateges, based on random fluctuatons. We are nterested n explanng ndvdual dfferences n a sequental predcton task. Observers are shown stmul generated from a smple statstcal process wth the task of predctng the next datum n the sequence. The latent parameters of the statstcal process change dscretely at random ponts n tme. Performance n ths task depends on the accurate detecton of those changeponts, as well as nference about future outcomes based on the outcomes that followed the most recent nferred changepont. There s much pror research n statstcs on the problem of dentfyng changeponts [3,4,5]. In ths paper, we adopt a Bayesan approach to the changepont dentfcaton problem and develop a smple nference procedure to predct the next datum n a sequence. The Bayesan model serves as an deal observer model and s useful to characterze the ways n whch ndvduals devate from optmalty.
2 The plan of the paper s as follows. We frst ntroduce the sequental predcton task and dscuss a Bayesan analyss of ths predcton problem. We then dscuss the results from a few ndvduals n ths predcton task and show how the Bayesan approach can capture ndvdual dfferences wth a sngle twtchness parameter that descrbes how readly changes are perceved n random sequences. We wll show that some ndvduals are too twtchy: ther performance s too varable because they base ther predctons on too lttle of the recent data. Other ndvduals are not twtchy enough, and they fal to capture fast changes n the data. We also show how behavor can be explaned wth a set of fast and frugal models [6]. These are cogntvely realstc models that operate under plausble computatonal constrants. 2 A predcton task wth multple changeponts In the predcton task, stmul are presented sequentally and the task s to predct the next stmulus n the sequence. After t trals, the observer has been presented wth stmul y, y 2,, y t and the task s to make a predcton about y t+. After the predcton s made, the actual outcome y t+ s revealed and the next tral proceeds to the predcton of y t+2. Ths procedure starts wth y and s repeated for T trals. The observatons y t are D-dmensonal vectors wth elements sampled from bnomal dstrbutons. The parameters of those dstrbutons change dscretely at random ponts n tme such that the mean ncreases or decreases after a change pont. Ths generates a sequence of observaton vectors, y, y 2,, y T, where each y t = {y t, y t,d }. Each of the y t,d s sampled from a bnomal dstrbuton Bn(θ t,d,k), so y t,d K. The parameter vector θ t ={θ t, θ t,d } changes dependng on the locatons of the changeponts. At each tme step, x t s a bnary ndcator for the occurrence of a changepont occurrng at tme t+. The parameter α determnes the probablty of a change occurrng n the sequence. The generatve model s specfed by the followng algorthm:. For d=..d sample θ,d from a Unform(,) dstrbuton 2. For t=2..t, (a) Sample x t- from a Bernoull(α) dstrbuton (b) If x t- =, then θ t =θ t-, else for d=..d sample θ t,d from a Unform(,) dstrbuton (c) for d=..d, sample y t from a Bn(θ t,d,k) dstrbuton Table shows some data generated from the changepont model wth T=2, α=.,and D=. In the predcton task, y wll be observed, but x and θ are not. Table : Example data t x θ y
3 3 A Bayesan predcton model In both our Bayesan and fast-and-frugal analyses, the predcton task s decomposed nto two nference procedures. Frst, the changepont locatons are dentfed. Ths s followed by predctve nference for the next outcome based on the most recent changepont locatons. Several Bayesan approaches have been developed for changepont problems nvolvng sngle or multple changeponts [3,5]. We apply a Markov Chan Monte Carlo (MCMC) analyss to approxmate the jont posteror dstrbuton over changepont assgnments x whle ntegratng out θ. Gbbs samplng wll be used to sample from ths posteror margnal dstrbuton. The samples can then be used to predct the next outcome n the sequence. 3. Inference for changepont assgnments. To apply Gbbs samplng, we evaluate the condtonal probablty of assgnng a changepont at tme, gven all other changepont assgnments and the current α value. By ntegratng out θ, the condtonal probablty s (,, α) (, θ, α, ) P x x y = P x x y () θ where x represents all swtch pont assgnments except x. Ths can be smplfed by consderng the locaton of the most recent changepont precedng and followng tme L and the outcomes occurrng between these locatons. Let n be the number of tme steps from the last changepont up to and ncludng the current tme step such that x = and x = for <j< L R L n. Smlarly, let n n L n + j be the number of tme steps that follow tme step up to the next changepont such that x = and x = for + + R <j< R L n. Let y = L y and R n k y < k = y. The update equaton for the k< k + n R k changepont assgnment can then be smplfed to ( m x ) P x = R n D L R L R L R ( y, j y, j) ( Kn Kn y, j y, j) L R j= Γ ( 2 + Kn + Kn ) L L L R R R ( yj, ) ( Kn yj, ) ( yj, ) ( Kn yj, ) L R Γ ( 2+ Kn ) Γ ( 2+ Kn ) Γ + + Γ + + ( α ) m = D Γ + Γ + Γ + Γ + α j= m = We ntalze the Gbbs sampler by samplng each x t from a Bernoull(α) dstrbuton. All changepont assgnments are then updated sequentally by the Gbbs samplng equaton above. The sampler s run for M teratons after whch one set of changepont assgnments s saved. The Gbbs sampler s then restarted multple tmes untl S samples have been collected. Although we could have ncluded an update equaton for α, n ths analyss we treat α as a known constant. Ths wll be useful when characterzng the dfferences between human observers n terms of dfferences n α. n j (2)
4 3.2 Predctve nference The next latent parameter value θ t+ and outcome y t+ can be predcted on the bass of observed outcomes that occurred after the last nferred changepont: θ t ( θ ) = y / K, y = round K (3) t+, j, j t+, j t+, j * = t + where t * s the locaton of the most recent change pont. By consderng multple Gbbs samples, we get a dstrbuton over outcomes y t+. We base the model predctons on the mean of ths dstrbuton. 3.3 Illustraton of model performance Fgure llustrates the performance of the model on a one dmensonal sequence (D=) generated from the changepont model wth T=6, α=.5, and K=. The Gbbs sampler was run for M=3 teratons and S=2 samples were collected. The top panel shows the actual changeponts (trangles) and the dstrbuton of changepont assgnments averaged over samples. The bottom panel shows the observed data y (thn lnes) as well as the θ values n the generatve model (rescaled between and ). At locatons wth large changes between observatons, the margnal changepont probablty s qute hgh. At other locatons, the true change n the mean s very small, and the model s less lkely to put n a changepont. The lower rght panel shows the dstrbuton over predcted θ t+ values. x t.5 θ t+ y t Fgure. Results of model smulaton. 4 Predcton experment We tested performance of 9 human observers n the predcton task. The observers ncluded the authors, a vstor, and one student who were aware of the statstcal nature of the task as well as naïve students. The observers were seated n front of an LCD touch screen dsplayng a two-dmensonal grd of x buttons. The changepont model was used to generate a sequence of T=5 stmul for two bnomal varables y and y 2 (D=2, K=). The change probablty α was set to.. The two varables y and y 2 specfed the two-dmensonal button locaton. The same sequence was used for all observers. On each tral, the observer touched a button on the grd dsplayed on the touch screen. Followng each button press, the button correspondng to the next {y,y 2 } outcome n the sequence was hghlghted. Observers were nstructed to press the button that best predcted the next locaton of the hghlghted button. The 5 trals were dvded nto
5 three blocks of 5 trals. Breaks were allowed between blocks. The whole experment lasted between 5 and 3 mnutes. Fgure 2 shows the frst 5 trals from the thrd block of the experment. The top and bottom panels show the actual outcomes for the y and y 2 button grd coordnates as well as the predctons for two observers (SB and MY). The fgure shows that at tral 5, the y and y 2 coordnates show a large shft followed by an mmedate shft n observer s MY predctons (on tral 6). Observer SB wats untl tral 7 to make a shft. 5 5 outcomes SB predctons MY predctons Tral Fgure 2. Tral by tral predctons from two observers. 4. Task error We assessed predcton performance by comparng the predcton wth the actual outcome n the sequence. Task error was measured by normalzed cty-block dstance task error= T O O yt, yt, + yt,2 y (4) t,2 ( ) T t= 2 where y O represents the observer s predcton. Note that the very frst tral s excluded from ths calculaton. Even though more sutable probablstc measures for predcton error could have been adopted, we wanted to allow comparson of observer s performance wth both probablstc and non-probablstc models. Task error ranged from 2.8 (for partcpant MY) to 3.3 (for ML). We also assessed the performance of fve models ther task errors ranged from 2.78 to 3.2. The Bayesan models (Secton 3) had the lowest task errors, just below 2.8. Ths fts wth our defnton of the Bayesan models as deal observer models ther task error s lower than any other model s and any human observer s task error. The fast and frugal models (Secton 5) had task errors rangng from 2.85 to Modelng Results We wll refer to the models wth the followng letter codes: B=Bayesan Model, LB=lmted Bayesan model, FF..3=fast and frugal models..3. We assessed model ft by comparng the model s predcton aganst the human observers predctons, agan usng a normalzed cty-block dstance
6 model error= ( T ) M O M O yt, yt, + yt,2 y (5) t,2 T t= 2 where y M represents the model s predcton. The model error for each ndvdual observer s shown n Fgure 3. It s mportant to note that because each model s assocated wth a set of free parameters, the parameters optmzed for task error and model error are dfferent. For Fgure 3, the parameters were optmzed to mnmze Equaton (5) for each ndvdual observer, showng the extent to whch these models can capture the performance of ndvdual observers, not necessarly provdng the best task performance. 2 B LB FF FF2 FF3 Model Error.5.5 MY MS MM EJ PH NP DN SB ML Fgure 3. Model error for each ndvdual observer. 5. Bayesan predcton models At each tral t, the model was provded wth the sequence of all prevous outcomes. The Gbbs samplng and nference procedures from Eq. (2) and (3) were appled wth M=3 teratons and S=2 samples. The change probablty α was a free parameter. In the full Bayesan model, the whole sequence of observatons up to the current tral s avalable for predcton, leadng to a memory requrement of up to T=5 trals a psychologcally unreasonable assumpton. We therefore also smulated a lmted Bayesan model (LB) where the observed sequence was truncated to the last outcomes. The LB model showed almost no decrement n task performance compared to the full Bayesan model. Fgure 3 also shows that t ft human data qute well. 5.2 Indvdual Dfferences The rght-hand panel of Fgure 4 plots each observer s task error as a functon of the mean cty-block dstance between ther subsequent button presses. Ths shows a clear U-shaped functon. Observers wth very varable predctons (e.g., ML and DN) had large average changes between successve button pushes, and also had large task error: These observers were too twtchy. Observers wth very small average button changes (e.g., SB and NP) were not twtchy enough, and also had large task error. Observers n the mddle had the lowest task error (e.g., MS and MY). The left-hand panel of Fgure 4 shows the same data, but wth the x-axs based on the Bayesan model fts. Instead of usng mean button change dstance to ndex twtchness (as n Error bars ndcate bootstrapped 95% confdence ntervals.
7 the rght-hand panel), the left-hand panel uses the estmated α parameters from the Bayesan model. A smlar U-shaped pattern s observed: ndvduals wth too large or too small α estmates have large task errors. 3.3 SB ML 3.3 SB ML 3.2 NP DN 3.2 NP DN Task Error 3. 3 EJ MM PH Task Error 3. 3 EJ MM PH B MS MY MS MY α Mean Button Change Fgure 4. Task error vs. twtchness. Left-hand panel ndexes twtchness usng estmated α parameters from Bayesan model fts. Rght-hand panel uses mean dstance between successve predctons. 5.3 Fast-and-Frugal (FF) predcton models These models perform the predcton task usng smple heurstcs that are cogntvely plausble. The FF models keep a short memory of prevous stmulus values and make predctons usng the same two-step process as the Bayesan model. Frst, a decson s made as to whether the latent parameter θ has changed. Second, remembered stmulus values that occurred after the most recently detected changepont are used to generate the next predcton. A smple heurstc s used to detect changeponts: If the dstance between the most recent observaton and predcton s greater than some threshold amount, a change s nferred. We defned the dstance between a predcton (p) and an observaton (y) as the dfference between the log-lkelhoods of y assumng θ=p and θ=y. Thus, f f B (. θ, K) s the bnomal densty wth parameters θ and K, the dstance between observaton y and predcton p s defned as d(y,p)=log(f B (y y,k))-log(f B (y p,k)). A changepont on tme step t+ s nferred whenever d(y t,p t )>C. The parameter C governs the twtchness of the model predctons. If C s large, only very dramatc changeponts wll be detected, and the model wll be too conservatve. If C s small, the model wll be too twtchy, and wll detect changeponts on the bass of small random fluctuatons. Predctons are based on the most recent M observatons, whch are kept n memory, unless a changepont has been detected n whch case only those observatons occurrng after the changepont are used for predcton. The predcton for tme step t+ s smply the mean of these observatons, say p. Human observers were retcent to make predctons very close to the boundares. Ths was modeled by allowng the FF model to change ts predcton for the next tme step, y t+, towards the mean predcton (.5). Ths change reflects a two-way bet. If the probablty of a change occurrng s α, the best guess wll be.5 f that change occurs, or the mean p f the change does not occur. Thus, the predcton made s actually y t+ =/2 α+(-α)p. Note that we do not allow perfect knowledge of the probablty of a changepont, α. Instead, an estmated value of α s used based on the number of changeponts detected n the data seres up to tme t.
8 The FF model nests two smpler FF models that are psychologcally nterestng. If the twtchness threshold parameter C becomes arbtrarly large, the model never detects a change and nstead becomes a contnuous runnng average model. Predctons from ths model are smply a boxcar smooth of the data. Alternatvely, f we assume no memory the model must based each predcton on only the prevous stmulus (.e., M=). Above, n Fgure 3, we labeled the complete FF model as FF, the boxcar model as FF2 and the memoryless model FF3. Fgure 3 showed that the complete FF model (FF) ft the data from all observers sgnfcantly better than ether the boxcar model (FF2) or the memoryless model (FF3). Exceptons were observers PH, DN and ML, for whom all three FF model ft equally well. Ths result suggests that our observers were (mostly) dong more than just keepng a runnng average of the data, or usng only the most recent observaton. The FF model ft the data about as well as the Bayesan models for all observers except MY and MS. Note that, n general, the FF and Bayesan model fts are very good: the average cty block dstance between the human data and the model predcton s around.75 (out of ) buttons on both the x- and y-axes. 6 Concluson We used an onlne predcton task to study changepont detecton. Human observers had to predct the next observaton n stochastc sequences contanng random changeponts. We showed that some observers are too twtchy : They perform poorly on the predcton task because they see changes where only random fluctuaton exsts. Other observers are not twtchy enough, and they perform poorly because they fal to see small changes. We developed a Bayesan changepont detecton model that performed the task optmally, and also provded a good ft to human data when sub-optmal parameter settngs were used. Fnally, we developed a fast-and-frugal model that showed how partcpants may be able to perform well at the task usng mnmal nformaton and smple decson heurstcs. Acknowledgments We thank Erc-Jan Wagenmakers and Mke Y for useful dscussons related to ths work. Ths work was supported n part by a grant from the US Ar Force Offce of Scentfc Research (AFOSR grant number FA ). References [] Glovch, T., Vallone, R. and Tversky, A. (985). The hot hand n basketball: on the mspercepton of random sequences. Cogntve Psychology7, [2] Albrght, S.C. (993a). A statstcal analyss of httng streaks n baseball. Journal of the Amercan Statstcal Assocaton, 88, [3] Stephens, D.A. (994). Bayesan retrospectve multple changepont dentfcaton. Appled Statstcs 43(), [4] Carln, B.P., Gelfand, A.E., & Smth, A.F.M. (992). Herarchcal Bayesan analyss of changepont problems. Appled Statstcs 4(2), [5] Green, P.J. (995). Reversble jump Markov chan Monte Carlo computaton and Bayesan model determnaton. Bometrka 82(4), [6] Ggerenzer, G., & Goldsten, D.G. (996). Reasonng the fast and frugal way: Models of bounded ratonalty. Psychologcal Revew, 3,
Bayesian predictive Configural Frequency Analysis
Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse
More informationComparison of Regression Lines
STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence
More informationMotion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong
Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal
More informationNegative Binomial Regression
STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...
More informationSimulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests
Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More informationDepartment of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6
Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.
More informationMarkov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement
Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs
More information1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands
Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of
More informationStatistics for Economics & Business
Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationLecture 6: Introduction to Linear Regression
Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6
More informationComparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method
Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationHidden Markov Models
CM229S: Machne Learnng for Bonformatcs Lecture 12-05/05/2016 Hdden Markov Models Lecturer: Srram Sankararaman Scrbe: Akshay Dattatray Shnde Edted by: TBD 1 Introducton For a drected graph G we can wrte
More informationLecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding
Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an
More information4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA
4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected
More informationLecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding
Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study
More informationDefinition. Measures of Dispersion. Measures of Dispersion. Definition. The Range. Measures of Dispersion 3/24/2014
Measures of Dsperson Defenton Range Interquartle Range Varance and Standard Devaton Defnton Measures of dsperson are descrptve statstcs that descrbe how smlar a set of scores are to each other The more
More informationArtificial Intelligence Bayesian Networks
Artfcal Intellgence Bayesan Networks Adapted from sldes by Tm Fnn and Mare desjardns. Some materal borrowed from Lse Getoor. 1 Outlne Bayesan networks Network structure Condtonal probablty tables Condtonal
More informationParametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010
Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More informationChapter 8 Indicator Variables
Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n
More informationSTATS 306B: Unsupervised Learning Spring Lecture 10 April 30
STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationConvergence of random processes
DS-GA 12 Lecture notes 6 Fall 216 Convergence of random processes 1 Introducton In these notes we study convergence of dscrete random processes. Ths allows to characterze phenomena such as the law of large
More informationBayesian Planning of Hit-Miss Inspection Tests
Bayesan Plannng of Ht-Mss Inspecton Tests Yew-Meng Koh a and Wllam Q Meeker a a Center for Nondestructve Evaluaton, Department of Statstcs, Iowa State Unversty, Ames, Iowa 5000 Abstract Although some useful
More informationNUMERICAL DIFFERENTIATION
NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the
More informationOnline Appendix to: Axiomatization and measurement of Quasi-hyperbolic Discounting
Onlne Appendx to: Axomatzaton and measurement of Quas-hyperbolc Dscountng José Lus Montel Olea Tomasz Strzaleck 1 Sample Selecton As dscussed before our ntal sample conssts of two groups of subjects. Group
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More informationSee Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)
Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes
More informationCopyright 2017 by Taylor Enterprises, Inc., All Rights Reserved. Adjusted Control Limits for P Charts. Dr. Wayne A. Taylor
Taylor Enterprses, Inc. Control Lmts for P Charts Copyrght 2017 by Taylor Enterprses, Inc., All Rghts Reserved. Control Lmts for P Charts Dr. Wayne A. Taylor Abstract: P charts are used for count data
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationChapter 5 Multilevel Models
Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level
More informationPolynomial Regression Models
LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance
More informationChapter 15 - Multiple Regression
Chapter - Multple Regresson Chapter - Multple Regresson Multple Regresson Model The equaton that descrbes how the dependent varable y s related to the ndependent varables x, x,... x p and an error term
More informationThis column is a continuation of our previous column
Comparson of Goodness of Ft Statstcs for Lnear Regresson, Part II The authors contnue ther dscusson of the correlaton coeffcent n developng a calbraton for quanttatve analyss. Jerome Workman Jr. and Howard
More informationStatistics Chapter 4
Statstcs Chapter 4 "There are three knds of les: les, damned les, and statstcs." Benjamn Dsrael, 1895 (Brtsh statesman) Gaussan Dstrbuton, 4-1 If a measurement s repeated many tmes a statstcal treatment
More informationThe Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction
ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also
More informationJoint Statistical Meetings - Biopharmaceutical Section
Iteratve Ch-Square Test for Equvalence of Multple Treatment Groups Te-Hua Ng*, U.S. Food and Drug Admnstraton 1401 Rockvlle Pke, #200S, HFM-217, Rockvlle, MD 20852-1448 Key Words: Equvalence Testng; Actve
More informationUncertainty as the Overlap of Alternate Conditional Distributions
Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant
More informationECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics
ECOOMICS 35*-A Md-Term Exam -- Fall Term 000 Page of 3 pages QUEE'S UIVERSITY AT KIGSTO Department of Economcs ECOOMICS 35* - Secton A Introductory Econometrcs Fall Term 000 MID-TERM EAM ASWERS MG Abbott
More informationCS-433: Simulation and Modeling Modeling and Probability Review
CS-433: Smulaton and Modelng Modelng and Probablty Revew Exercse 1. (Probablty of Smple Events) Exercse 1.1 The owner of a camera shop receves a shpment of fve cameras from a camera manufacturer. Unknown
More informationBAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup
BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS Darusz Bskup 1. Introducton The paper presents a nonparaetrc procedure for estaton of an unknown functon f n the regresson odel y = f x + ε = N. (1) (
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationConjugacy and the Exponential Family
CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the
More informationIntroduction to Regression
Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationStatistical analysis using matlab. HY 439 Presented by: George Fortetsanakis
Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationCOS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014
COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture #16 Scrbe: Yannan Wang Aprl 3, 014 1 Introducton The goal of our onlne learnng scenaro from last class s C comparng wth best expert and
More informationChapter 1. Probability
Chapter. Probablty Mcroscopc propertes of matter: quantum mechancs, atomc and molecular propertes Macroscopc propertes of matter: thermodynamcs, E, H, C V, C p, S, A, G How do we relate these two propertes?
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationStatistical Evaluation of WATFLOOD
tatstcal Evaluaton of WATFLD By: Angela MacLean, Dept. of Cvl & Envronmental Engneerng, Unversty of Waterloo, n. ctober, 005 The statstcs program assocated wth WATFLD uses spl.csv fle that s produced wth
More informationA PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS
HCMC Unversty of Pedagogy Thong Nguyen Huu et al. A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS Thong Nguyen Huu and Hao Tran Van Department of mathematcs-nformaton,
More informationGenerative classification models
CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationLecture 3 Stat102, Spring 2007
Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture
More informationChecking Pairwise Relationships. Lecture 19 Biostatistics 666
Checkng Parwse Relatonshps Lecture 19 Bostatstcs 666 Last Lecture: Markov Model for Multpont Analyss X X X 1 3 X M P X 1 I P X I P X 3 I P X M I 1 3 M I 1 I I 3 I M P I I P I 3 I P... 1 IBD states along
More informationANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected.
ANSWERS CHAPTER 9 THINK IT OVER thnk t over TIO 9.: χ 2 k = ( f e ) = 0 e Breakng the equaton down: the test statstc for the ch-squared dstrbuton s equal to the sum over all categores of the expected frequency
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationComputation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models
Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More information= z 20 z n. (k 20) + 4 z k = 4
Problem Set #7 solutons 7.2.. (a Fnd the coeffcent of z k n (z + z 5 + z 6 + z 7 + 5, k 20. We use the known seres expanson ( n+l ( z l l z n below: (z + z 5 + z 6 + z 7 + 5 (z 5 ( + z + z 2 + z + 5 5
More informationj) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1
Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons
More informationMore metrics on cartesian products
More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of
More information/ n ) are compared. The logic is: if the two
STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence
More informationStatistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation
Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear
More informationNon-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT
Malaysan Journal of Mathematcal Scences 8(S): 37-44 (2014) Specal Issue: Internatonal Conference on Mathematcal Scences and Statstcs 2013 (ICMSS2013) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES Journal
More informationC/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1
C/CS/Phy9 Problem Set 3 Solutons Out: Oct, 8 Suppose you have two qubts n some arbtrary entangled state ψ You apply the teleportaton protocol to each of the qubts separately What s the resultng state obtaned
More informationStat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors
Stat60: Bayesan Modelng and Inference Lecture Date: February, 00 Reference Prors Lecturer: Mchael I. Jordan Scrbe: Steven Troxler and Wayne Lee In ths lecture, we assume that θ R; n hgher-dmensons, reference
More informationNumerical Heat and Mass Transfer
Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and
More informationLaboratory 1c: Method of Least Squares
Lab 1c, Least Squares Laboratory 1c: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly
More informationDepartment of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution
Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable
More informationHere is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)
Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,
More informationProbability Theory (revisited)
Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted
More informationExplaining the Stein Paradox
Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten
More informationInternational Journal of Industrial Engineering Computations
Internatonal Journal of Industral Engneerng Computatons 4 (03) 47 46 Contents lsts avalable at GrowngScence Internatonal Journal of Industral Engneerng Computatons homepage: www.growngscence.com/ec Modelng
More informationGravitational Acceleration: A case of constant acceleration (approx. 2 hr.) (6/7/11)
Gravtatonal Acceleraton: A case of constant acceleraton (approx. hr.) (6/7/11) Introducton The gravtatonal force s one of the fundamental forces of nature. Under the nfluence of ths force all objects havng
More informationIntroduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:
CE304, Sprng 2004 Lecture 4 Introducton to Vapor/Lqud Equlbrum, part 2 Raoult s Law: The smplest model that allows us do VLE calculatons s obtaned when we assume that the vapor phase s an deal gas, and
More informationHidden Markov Models & The Multivariate Gaussian (10/26/04)
CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models
More informationStatistics for Business and Economics
Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear
More informationTHE SUMMATION NOTATION Ʃ
Sngle Subscrpt otaton THE SUMMATIO OTATIO Ʃ Most of the calculatons we perform n statstcs are repettve operatons on lsts of numbers. For example, we compute the sum of a set of numbers, or the sum of the
More informationAppendix B: Resampling Algorithms
407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationDurban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications
Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More informationSPANC -- SPlitpole ANalysis Code User Manual
Functonal Descrpton of Code SPANC -- SPltpole ANalyss Code User Manual Author: Dale Vsser Date: 14 January 00 Spanc s a code created by Dale Vsser for easer calbratons of poston spectra from magnetc spectrometer
More informationLossy Compression. Compromise accuracy of reconstruction for increased compression.
Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost
More informationOne-sided finite-difference approximations suitable for use with Richardson extrapolation
Journal of Computatonal Physcs 219 (2006) 13 20 Short note One-sded fnte-dfference approxmatons sutable for use wth Rchardson extrapolaton Kumar Rahul, S.N. Bhattacharyya * Department of Mechancal Engneerng,
More informationx yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting.
The Practce of Statstcs, nd ed. Chapter 14 Inference for Regresson Introducton In chapter 3 we used a least-squares regresson lne (LSRL) to represent a lnear relatonshp etween two quanttatve explanator
More information28. SIMPLE LINEAR REGRESSION III
8. SIMPLE LINEAR REGRESSION III Ftted Values and Resduals US Domestc Beers: Calores vs. % Alcohol To each observed x, there corresponds a y-value on the ftted lne, y ˆ = βˆ + βˆ x. The are called ftted
More informationIntroduction to Generalized Linear Models
INTRODUCTION TO STATISTICAL MODELLING TRINITY 00 Introducton to Generalzed Lnear Models I. Motvaton In ths lecture we extend the deas of lnear regresson to the more general dea of a generalzed lnear model
More informationFREQUENCY DISTRIBUTIONS Page 1 of The idea of a frequency distribution for sets of observations will be introduced,
FREQUENCY DISTRIBUTIONS Page 1 of 6 I. Introducton 1. The dea of a frequency dstrbuton for sets of observatons wll be ntroduced, together wth some of the mechancs for constructng dstrbutons of data. Then
More informationEM and Structure Learning
EM and Structure Learnng Le Song Machne Learnng II: Advanced Topcs CSE 8803ML, Sprng 2012 Partally observed graphcal models Mxture Models N(μ 1, Σ 1 ) Z X N N(μ 2, Σ 2 ) 2 Gaussan mxture model Consder
More informationInformation Geometry of Gibbs Sampler
Informaton Geometry of Gbbs Sampler Kazuya Takabatake Neuroscence Research Insttute AIST Central 2, Umezono 1-1-1, Tsukuba JAPAN 305-8568 k.takabatake@ast.go.jp Abstract: - Ths paper shows some nformaton
More information