A Comparison on Neural Network Forecasting
|
|
- Melvyn Brooks
- 5 years ago
- Views:
Transcription
1 011 nternatonal Conference on Crcuts, System and Smulaton PCST vol.7 (011) (011) ACST Press, Sngapore A Comparson on Neural Networ Forecastng Hong-Choon Ong 1, a and Shn-Yue Chan 1, b 1 School of Mathematcal Scences, Unverst Sans Malaysa, USM, Penang, Malaysa. a hcong@cs.usm.my, b shnyue_alu@yahoo.com Abstract. Ths study compares the effectveness of the Box-Jenns model and neural networs model n mang a forecast. An eghteen years bmonthly water consumpton n Penang data set s analyzed. Multlayer perceptron (MLP) n neural networ wth sngle hdden layer and double hdden layers usng errorbacpropagaton algorthm models are used. Four MLP programs wth double hdden layers whch are the orgnal data (O), the deseasonalzed wth the smoothng movng average method (DS), the lnear detrended model (DTL) and both deseasonalzed and detrended model (DSTL) and a sngle hdden layer MLP wth both deseasonalzed and detrended (DSTL) were smulated. For tme seres analyss, a Box-Jenns SARMA model s generated. The performance of the models are measured usng three types of error measurement, whch are mean absolute error (MAE), absolute percentage error (MAPE) and root mean squared error (RMSE). The model wth smallest value of MAE, MAPE and RMSE stands out to be the best model n predctng the water consumpton n Penang for the year 008. The results showed that both SARMA and double hdden layers MLP models perform relatvely well. Furthermore, double hdden layers MLP model shows an mprovement n predcton as compared wth the sngle hdden layer MLP model. Keywords: neural networ; mult-layer perceptron; Box-Jenns SARMA model. 1. ntroducton Neural networ has the ablty to represent both lnear and non-lnear relatonshps and the networ can learn these relatonshps drectly from the raw data. Mult-layered Perceptron (MLP) model s the most common model n neural networ and s a feed-forward model (Looney, 1997). Each node n one layer connects wth a certan weght to every other node n the followng layer. n general, more hdden layers can prevent the necessty of usng unnecessary neurons to acheve hghly non-lnear classfcaton. The error bac-propagaton algorthm used adjusts the weghts to obtan more accurate forecast values. A tme seres s a sequence of data ponts (observatons), collected typcally at successve tmes, spaced at often unform tme ntervals. Data collected nconsstently or only once may not be consdered as a tme seres. Tme seres analyss conssts of methods that attempt to understand the specfc tme seres, dentfyng the underlyng context of the observatons and ft a sutable model to mae predctons. Tme seres forecastng means the use of a model to predct future events based on the past data; to forecast future data ponts before they are measured. Tme seres can be appled nto varous felds, such as economc forecastng, sales forecastng, stoc maret analyss, process and qualty control, nventory forecastng, producton plannng and many more (Chatfeld, 004).. A Bref Lterature Revew n recent years, artfcal neural networs have been wdely used as a useful tool to model many areas of engneerng applcatons. Sarıdemr (008) presented the models n artfcal neural networs (ANN) for predctng compressve strength of concretes contanng metaaoln and slca fume that had been developed at the age of 1, 3, 7, 8, 56, 90 and 180 days. The mult-layered feed forward neural networs models were fed wth these parameters to predct the compressve strength values of concretes contanng metaaoln and slca fume. The results of ths study showed that neural networs have strong potental for predcton, whch proved that artfcal neural networs are capable of learnng and generalzng from examples and 56
2 experences. Ong, et al. (008) gave on a functonal approxmaton comparson between neural networs and polynomal regresson. The results showed that approxmaton usng polynomal regresson generates lower fracton of varance unexplaned (FVU) value as compared to approxmaton usng neural networs wth sngle hdden layer mult-layered perceptron (MLP) and double hdden layer MLP except for the complcated functons. Moreover, double hdden layer MLP showed better estmaton results than sngle hdden layer MLP wthout consderng the number of parameters beng used. On the other hand, statstcal models have commonly been used n tme seres data analyss and forecastng. Zou and Yang (003) proposed combnng tme seres model for forecastng for a better performance. Lam, et al. (008) used autoregressve ntegrated movng average (ARMA) model to measure the nterventon effects and the asymptotc change n the smulaton results of the busness process reengneerng that s based on the actvty model analyss. A case study of a purchasng process of a household applance manufacturng enterprse that nvolves 0 purchasng actvtes was used as the tranng data. The results ndcated that the changes can be explctly quantfed and the effects of busness process reengneerng can be measured. n general, many tme seres are asymptotcally unstable and ntrnscally non-statonary. Box-Jenns models solve these problems by mposng on data transformatons, such as dfferencng and logarthm. Grllenzon (1998) dscussed on a method for modelng tme seres wth unstable roots and changng parameters. A method of adaptve forecastng whch s based on the optmzaton of recursve estmators was appled to well-nown data sets. He had demonstrated ts valdty n several mplementatons and for dfferent model structures. 3. Methodology 3.1. Neural networ mult-layered perceptron Mult-layered perceptron s the most common neural networ model. t s abltes n achevng hghly nonlnear classfcaton and nose tolerance mae t sutable to be used n ths study. Bpolar sgmod functon s appled n ths study. n a sngle hdden layer model, the perceptron computes assocated outputs from multple real-valued nputs by formng a lnear combnaton accordng to ts nput weghts. The data set s treated as nputs and denoted as { x }, where = number of data and has an assocated set of exemplar outputs target vector. A random ntal synaptc weght set for the hdden and output layers are selected. Fnally, the current weghts are re-terated and updated usng the error bac-propagaton algorthm to force each of the nput exemplar feature vectors to be mapped closer to the target vector that dentfes a class n the nput space. n the double hdden layered MLP, the perceptrons also computes assocated outputs from multple real-valued nputs by formng a lnear combnaton accordng to ts nput weghts and the data set s treated as nputs. The man dfference s that the synaptc weght set has an addtonal second hdden layer. Error bac-propagaton s used to update the error computed n the networ wth respect to the networ's modfable weghts. The error value whch s the dfference between output value and desred value s then computed. The errors wll be propagated bac nto the networ usng a gradent descent technque so that weghts adjustment can be done. 3.. Desgn of Box-Jenns model There are three prmary stages n buldng a Box-Jenns tme seres model, whch are model dentfcaton, model estmaton and model dagnostcs. 1) Model dentfcaton: There s a need to determne f the seres s statonary, has a trend component and f there s any sgn of seasonalty that needs to be modeled. Statonarty, trend and seasonalty can be detected by plottng the tme seres graph. f there s trend component present n the data, the graph wll not be n a random form. Statonarty can also be checed va sample auto-correlaton (ACF). n a statonary seres, the auto-covarance functon, γ and the auto-correlaton functon, ρ show the followng propertes: 1. γ γ 0.. ρ 0 = 1 and -1< ρ <1 for γ = γ and ρ = ρ for all. Seasonalty can be accessed from sample auto-correlaton (ACF) at varous lag and also sample partal auto-correlaton (PACF) at varous lag of the process. The behavor of the ACF and PACF s shown n Table 1. After statonarty and seasonalty have been addressed, the next step s to dentfy the order (the p and q) of the autoregressve and movng average terms. The behavor of ACF and PACF wll ndcate the 57
3 order form of the autoregressve and movng average terms. TABLE. BEHAVOR OF ACF AND PACF AR(p) MA(q) ARMA(p,q) ACF Tals off Cuts off after lag q Tals off PACF Cuts off after lag p Tals off Tals off ) Model estmaton: Parameters estmaton for the Box-Jenns models s a qute complcated nonlnear estmaton problem. Therefore, hgh qualty software program that fts Box-Jenns model should be used to estmate the parameter (Anderson, 1994). Consder Y t s statonary tme seres and { ε t } are ndependently and dentcally dstrbuted wth N ( 0, σ ε ), an ARMA(p,q) model wth unnown parameters ϕ = { ϕ1, ϕ,... ϕp}, θ = { θ1, θ,... θ q } and σ ( ε = E εt ) s gven as: Y = ϕy + ϕ Y ϕ Y + ε θε θ ε... θ ε t 1 t 1 t p t p t 1 t 1 t q t q The man approaches to fttng Box-Jenns models are maxmum lelhood estmaton and least squares estmaton. 3) Model dagnostcs: The ACF plot and PACF plot of the estmated resdual { ε t } have the ablty to ndcate whether the resduals are ndependent and normally dstrbuted. f the ACF plot and PACF plot show the ponts are dstrbuted randomly or do not show any pattern, then ths ndcates that the resduals are ndependent and normally dstrbuted. However, f these assumptons are not satsfed, a more approprate model needs to be ftted. Ths means that, the model dentfcaton step has to be redone to re-develop a better model that satsfes the dagnostc test. Although a good model may not be obtaned on the frst tral, however a new and better model can be easly derved (We, 1990). Correlogram of resduals and correlogram of resduals squared are suffcent for determnng the status of whte nose. Q-statstcs s useful to test normalty of the resduals wthn a range of an ndependent varable. n other words, t served as a tool to test whether the seres s whte nose. The last two columns n the correlogram represent Ljung-Box Q- statstcs and ther p-value respectvely. The Q-statstcs at lag s a test statstcs for the null hypothess where there s no auto-correlaton up to order. t s computed as follows: Q = T( T + ) ( T j) r where rj represents the j-th auto-correlaton and T represents number of observatons. f the p-value s less than 0.05, then t s sgnfcant. Therefore, the null hypothess s rejected, and the error terms are correlated. f the p-values are greater or equal to 0.05, then the p-values are not sgnfcant at all lags. Thus, the null hypothess s not rejected, and the resduals are not correlated (Harvey, 001) Crtera comparson by error measurement Three types of error measurement are used n ths study, whch are the mean absolute error (MAE), the percentage error (MAPE) and the root mean squared error (RMSE). Mean Absolute Error (MAE) MAE = Average of the absolute values of forecastng error = 1 = ˆ where = 1,,,, = orgnal data, j = 1 1 j Mean Absolute Percentage Error (MAPE) MAPE = Percentage of the mean rato of the error to the orgnal data 58
4 = ˆ = 1 where = 1,,,, 100 = orgnal data, Root Mean Squared Error (RMSE) RMSE = Square root of the average of the summng square forecastng errors = n = 1 ( ) where = 1,,,, 4. Result and Dscusson = orgnal data, The total of 108 data collected s dvded nto two groups, namely tranng data and testng data. The tranng data whch conssts of the frst 10 data s used to forecast the next 6 forecast value. The last 6 data s used as testng set, where the man purpose s to mae comparson wth the forecastng value generated by tranng set n order to calculate the error. The model wth smallest error value of the forecast s the best ftted forecast model. Fnally, the next sx forecast ponts are forecasted usng the selected best model. The neural networ models are wrtten n C++ programmng language whle Box-Jenns model was evaluated n MNTAB 14 and E-VEW. Table shows the target value, forecasted value from SARMA model and forecasted value from neural networ models. TABLE. FORECASTED VALUE FROM SARMA MODEL AND NEURAL NETWORK MODELS. Neural Networ Models (double hdden layers) Neural Networ Model (sngle hdden layer) t orgnal data O DTL DS DSTL DSTL SARMA Model Crtera comparson by error measurement The error between forecasted value and the orgnal value s compared between estmated models. MAE, MAPE and RMSE are used as error measurement. Table 3 shows the error measurements for SARMA model, four Neural Networ models wth double hdden layers and one Neural Networ model wth sngle hdden layer. TABLE. ERROR MEASUREMENT FOR SARMA MODEL AND NEURAL NETWORK MODELS Model Type MAE MAPE RMSE SARMA Neural Networ models wth double hdden layers O DTL DS DSTL Neural Networ model wth sngle hdden layer DSTL When comparsons were made, both detrendng and deseasonalzed model (DSTL) wth double hdden layers gves the smallest error n MAE and MAPE compared to other models. However, SARMA model 59
5 stands out to be a better model f RMSE s beng compared. Snce the errors are squared before they are averaged, the RMSE gves a relatvely hgh weght to large errors. Although, the MAE s very smlar to the RMSE but t s less senstve to large forecast errors. Therefore, both DSTL wth double hdden layers model and SARMA model tends to have ts own good characterstcs n generatng smallest error durng forecastng. The orgnal model (O) wth double hdden layers appears to generate the hghest MAE, MAPE and RMSE, follow by deseasonalzed model (DS) wth double hdden layers, detrendng model (DTL) wth double hdden layers and both deseasonalzed and detrendng model (DSTL) wth sngle hdden layer. When DSTL wth double hdden layers model s compared wth DSTL wth sngle hdden layer model, the double hdden layers model outperforms sngle hdden layer model. 4.. Forecastng n water consumpton SARMA model and DSTL wth double hdden layers model are both the best model n forecastng the water consumpton n Penang and the next sx data are forecasted. Ths sx data s the bmonthly water consumpton n Penang data startng from January 008 to December 008. Table 4 and 5 show the forecasted sx data of bmonthly water consumpton n Penang from January 008 to December 008 by usng SARMA model and DSTL wth double hdden layers mult-layered perceptron model respectvely. TABLE V. FORECAST OF BMONTHLY WATER CONSUMPTON N PENANG USNG SARMA MODEL Bmonthly Jan-Feb Mar-Apr May-Jun Jul-Aug Sept-Oct Nov-Dec Forecast TABLE V. FORECAST OF BMONTHLY WATER CONSUMPTON N PENANG USNG DSTL 5. Concluson Bmonthly Jan-Feb Mar-Apr May-Jun Jul-Aug Sept-Oct Nov-Dec Forecast From the results generated, t s clearly shown that both SARMA model and DSTL are the best models used. For the neural networ models, t s obvous that preprocessng helps to mprove the forecast. Furthermore, t s also shown that the double hdden layered MLP that the DSTL s able to get a better result as compared to the sngle hdden layered MLP. Acnowledgment We would le to than the Drector of the Water Wors Department of Penang, Malaysa, for provdng us the data. Ths wor was supported n part by the U.S.M. nsentve grant no. 1001/PMATH/8184. References [1] T. W. Anderson, The statstcal analyss of tme seres, Unted States, John Wley and Sons, nc., [] C. Chatfeld, The analyss of tme seres: an ntroducton, 6th ed., Florda, CRC Press, 004. [3] C. Grllenzon, Forecastng unstable and nonstatonary tme seres, Journal of Forecastng, vol. 14(4), 1st December 1998, pp [4] A. C. Harvey, Forecastng, structural tme seres models and the Kalman flter, New Yor, Cambrdge Unversty Press, [5] C. Y. Lam, W. H. p and C. W. Lau, A busness process actvty model and performance measurement usng a tme seres ARMA nterventon analyss, Expert Systems wth Applcatons, vol. 36(3), Part, Aprl 009, pp [6] M. Sandemr, Predcton of compressve strength of concretes contanng metaaoln and slca fume by artfcal neural networs, Advances n Engneerng Software, vol. 40 (5), pp [7] We and W. S. Wllam, Tme seres analyss: unvarate and multvarate methods, Calforna: Addson-Wesley Publshng Company, nc., [8] H. Zou and Y. H. Yang, Combnng tme seres models for forecastng, Journal of Forecastng, vol. 0(1), January-March 005, pp
Comparison of Regression Lines
STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationNegative Binomial Regression
STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...
More informationA LINEAR PROGRAM TO COMPARE MULTIPLE GROSS CREDIT LOSS FORECASTS. Dr. Derald E. Wentzien, Wesley College, (302) ,
A LINEAR PROGRAM TO COMPARE MULTIPLE GROSS CREDIT LOSS FORECASTS Dr. Derald E. Wentzen, Wesley College, (302) 736-2574, wentzde@wesley.edu ABSTRACT A lnear programmng model s developed and used to compare
More informationStatistics for Economics & Business
Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable
More informationDepartment of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6
Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More informationChapter 6. Supplemental Text Material
Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationChapter 5 Multilevel Models
Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More informationShort Term Load Forecasting using an Artificial Neural Network
Short Term Load Forecastng usng an Artfcal Neural Network D. Kown 1, M. Km 1, C. Hong 1,, S. Cho 2 1 Department of Computer Scence, Sangmyung Unversty, Seoul, Korea 2 Department of Energy Grd, Sangmyung
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed
More informationDurban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications
Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department
More informationChapter 15 Student Lecture Notes 15-1
Chapter 15 Student Lecture Notes 15-1 Basc Busness Statstcs (9 th Edton) Chapter 15 Multple Regresson Model Buldng 004 Prentce-Hall, Inc. Chap 15-1 Chapter Topcs The Quadratc Regresson Model Usng Transformatons
More informationPolynomial Regression Models
LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance
More informationDepartment of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution
Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable
More informationBoostrapaggregating (Bagging)
Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More informationMultilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata
Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,
More informationSTAT 3008 Applied Regression Analysis
STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,
More information2016 Wiley. Study Session 2: Ethical and Professional Standards Application
6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationLecture 6: Introduction to Linear Regression
Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6
More informationPERFORMANCE COMPARISON BETWEEN BACK PROPAGATION, RPE AND MRPE ALGORITHMS FOR TRAINING MLP NETWORKS
PERFORMANCE COMPARISON BETWEEN BACK PROPAGATION, RPE AND MRPE ALGORITHMS FOR TRAINING MLP NETWORKS Mohd Yusoff Mashor School of Electrcal and Electronc Engneerng, Unversty Scence Malaysa, Pera Branch Campus,
More informationStatistical Evaluation of WATFLOOD
tatstcal Evaluaton of WATFLD By: Angela MacLean, Dept. of Cvl & Envronmental Engneerng, Unversty of Waterloo, n. ctober, 005 The statstcs program assocated wth WATFLD uses spl.csv fle that s produced wth
More information/ n ) are compared. The logic is: if the two
STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence
More informationx i1 =1 for all i (the constant ).
Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by
More information18. SIMPLE LINEAR REGRESSION III
8. SIMPLE LINEAR REGRESSION III US Domestc Beers: Calores vs. % Alcohol Ftted Values and Resduals To each observed x, there corresponds a y-value on the ftted lne, y ˆ ˆ = α + x. The are called ftted values.
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More informationModeling and Simulation NETW 707
Modelng and Smulaton NETW 707 Lecture 5 Tests for Random Numbers Course Instructor: Dr.-Ing. Magge Mashaly magge.ezzat@guc.edu.eg C3.220 1 Propertes of Random Numbers Random Number Generators (RNGs) must
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More informationThe Ordinary Least Squares (OLS) Estimator
The Ordnary Least Squares (OLS) Estmator 1 Regresson Analyss Regresson Analyss: a statstcal technque for nvestgatng and modelng the relatonshp between varables. Applcatons: Engneerng, the physcal and chemcal
More information28. SIMPLE LINEAR REGRESSION III
8. SIMPLE LINEAR REGRESSION III Ftted Values and Resduals US Domestc Beers: Calores vs. % Alcohol To each observed x, there corresponds a y-value on the ftted lne, y ˆ = βˆ + βˆ x. The are called ftted
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationSee Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)
Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes
More informationPhase I Monitoring of Nonlinear Profiles
Phase I Montorng of Nonlnear Profles James D. Wllams Wllam H. Woodall Jeffrey B. Brch May, 003 J.D. Wllams, Bll Woodall, Jeff Brch, Vrgna Tech 003 Qualty & Productvty Research Conference, Yorktown Heghts,
More informationIntroduction to Regression
Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes
More informationBOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu
BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationTopic 23 - Randomized Complete Block Designs (RCBD)
Topc 3 ANOVA (III) 3-1 Topc 3 - Randomzed Complete Block Desgns (RCBD) Defn: A Randomzed Complete Block Desgn s a varant of the completely randomzed desgn (CRD) that we recently learned. In ths desgn,
More information4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA
4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationA Particle Filter Algorithm based on Mixing of Prior probability density and UKF as Generate Importance Function
Advanced Scence and Technology Letters, pp.83-87 http://dx.do.org/10.14257/astl.2014.53.20 A Partcle Flter Algorthm based on Mxng of Pror probablty densty and UKF as Generate Importance Functon Lu Lu 1,1,
More informationDERIVATION OF THE PROBABILITY PLOT CORRELATION COEFFICIENT TEST STATISTICS FOR THE GENERALIZED LOGISTIC DISTRIBUTION
Internatonal Worshop ADVANCES IN STATISTICAL HYDROLOGY May 3-5, Taormna, Italy DERIVATION OF THE PROBABILITY PLOT CORRELATION COEFFICIENT TEST STATISTICS FOR THE GENERALIZED LOGISTIC DISTRIBUTION by Sooyoung
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationOnline Classification: Perceptron and Winnow
E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng
More informationBasically, if you have a dummy dependent variable you will be estimating a probability.
ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy
More informationis the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors
Multple Lnear and Polynomal Regresson wth Statstcal Analyss Gven a set of data of measured (or observed) values of a dependent varable: y versus n ndependent varables x 1, x, x n, multple lnear regresson
More informationLab 2e Thermal System Response and Effective Heat Transfer Coefficient
58:080 Expermental Engneerng 1 OBJECTIVE Lab 2e Thermal System Response and Effectve Heat Transfer Coeffcent Warnng: though the experment has educatonal objectves (to learn about bolng heat transfer, etc.),
More information1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands
Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of
More informationStatistics for Business and Economics
Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear
More informationThailand s Para Rubber Production Forecasting Comparison
Proceedngs of the Internatonal MultConference of Engneers and Computer Scentsts 2014 Vol II,, March 12-14, 2014, Hong Kong Thaland s Para Rubber Producton Forecastng Comparson Onuma Kosanan and Nantacha
More informationIssues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question
Issues To Consder when Estmatng Health Care Costs wth Generalzed Lnear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Queston ISPOR 20th Annual Internatonal Meetng May 19, 2015 Jalpa
More informationA Comparative Study for Estimation Parameters in Panel Data Model
A Comparatve Study for Estmaton Parameters n Panel Data Model Ahmed H. Youssef and Mohamed R. Abonazel hs paper examnes the panel data models when the regresson coeffcents are fxed random and mxed and
More informationThe Chaotic Robot Prediction by Neuro Fuzzy Algorithm (2) = θ (3) = ω. Asin. A v. Mana Tarjoman, Shaghayegh Zarei
The Chaotc Robot Predcton by Neuro Fuzzy Algorthm Mana Tarjoman, Shaghayegh Zare Abstract In ths paper an applcaton of the adaptve neurofuzzy nference system has been ntroduced to predct the behavor of
More informationIdentification of Linear Partial Difference Equations with Constant Coefficients
J. Basc. Appl. Sc. Res., 3(1)6-66, 213 213, TextRoad Publcaton ISSN 29-434 Journal of Basc and Appled Scentfc Research www.textroad.com Identfcaton of Lnear Partal Dfference Equatons wth Constant Coeffcents
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationSpatial Modelling of Peak Frequencies of Brain Signals
Malaysan Journal of Mathematcal Scences 3(1): 13-6 (9) Spatal Modellng of Peak Frequences of Bran Sgnals 1 Mahendran Shtan, Hernando Ombao, 1 Kok We Lng 1 Department of Mathematcs, Faculty of Scence, and
More informationChapter 12 Analysis of Covariance
Chapter Analyss of Covarance Any scentfc experment s performed to know somethng that s unknown about a group of treatments and to test certan hypothess about the correspondng treatment effect When varablty
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationMulti-Step-Ahead Prediction of Stock Price Using a New Architecture of Neural Networks
Journal of Computer & Robotcs 8(), 05 47-56 47 Mult-Step-Ahead Predcton of Stoc Prce Usng a New Archtecture of Neural Networs Mohammad Taleb Motlagh *, Hamd Khaloozadeh Department of Systems and Control,
More informationA new Approach for Solving Linear Ordinary Differential Equations
, ISSN 974-57X (Onlne), ISSN 974-5718 (Prnt), Vol. ; Issue No. 1; Year 14, Copyrght 13-14 by CESER PUBLICATIONS A new Approach for Solvng Lnear Ordnary Dfferental Equatons Fawz Abdelwahd Department of
More informationNUMERICAL DIFFERENTIATION
NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the
More informationInternet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks
Internet Engneerng Jacek Mazurkewcz, PhD Softcomputng Part 3: Recurrent Artfcal Neural Networks Self-Organsng Artfcal Neural Networks Recurrent Artfcal Neural Networks Feedback sgnals between neurons Dynamc
More informationSTAT 511 FINAL EXAM NAME Spring 2001
STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte
More informationChapter 8 Indicator Variables
Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n
More informationChapter 6. Supplemental Text Material. Run, i X i1 X i2 X i1 X i2 Response total (1) a b ab
Chapter 6. Supplemental Text Materal 6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.
More informationLearning Objectives for Chapter 11
Chapter : Lnear Regresson and Correlaton Methods Hldebrand, Ott and Gray Basc Statstcal Ideas for Managers Second Edton Learnng Objectves for Chapter Usng the scatterplot n regresson analyss Usng the method
More informationIII. Econometric Methodology Regression Analysis
Page Econ07 Appled Econometrcs Topc : An Overvew of Regresson Analyss (Studenmund, Chapter ) I. The Nature and Scope of Econometrcs. Lot s of defntons of econometrcs. Nobel Prze Commttee Paul Samuelson,
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationThe Expectation-Maximization Algorithm
The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationApplication research on rough set -neural network in the fault diagnosis system of ball mill
Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(4):834-838 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 Applcaton research on rough set -neural network n the
More informationHidden Markov Models & The Multivariate Gaussian (10/26/04)
CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationThe Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction
ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also
More informationLecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding
Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an
More informationLecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding
Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study
More informationBasic Business Statistics, 10/e
Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson
More informationT E C O L O T E R E S E A R C H, I N C.
T E C O L O T E R E S E A R C H, I N C. B rdg n g En g neern g a nd Econo mcs S nce 1973 THE MINIMUM-UNBIASED-PERCENTAGE ERROR (MUPE) METHOD IN CER DEVELOPMENT Thrd Jont Annual ISPA/SCEA Internatonal Conference
More informationStatistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation
Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear
More informationSTAT 405 BIOSTATISTICS (Fall 2016) Handout 15 Introduction to Logistic Regression
STAT 45 BIOSTATISTICS (Fall 26) Handout 5 Introducton to Logstc Regresson Ths handout covers materal found n Secton 3.7 of your text. You may also want to revew regresson technques n Chapter. In ths handout,
More informationMidterm Examination. Regression and Forecasting Models
IOMS Department Regresson and Forecastng Models Professor Wllam Greene Phone: 22.998.0876 Offce: KMC 7-90 Home page: people.stern.nyu.edu/wgreene Emal: wgreene@stern.nyu.edu Course web page: people.stern.nyu.edu/wgreene/regresson/outlne.htm
More informationChapter 14 Simple Linear Regression
Chapter 4 Smple Lnear Regresson Chapter 4 - Smple Lnear Regresson Manageral decsons often are based on the relatonshp between two or more varables. Regresson analss can be used to develop an equaton showng
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationLecture 23: Artificial neural networks
Lecture 23: Artfcal neural networks Broad feld that has developed over the past 20 to 30 years Confluence of statstcal mechancs, appled math, bology and computers Orgnal motvaton: mathematcal modelng of
More informationCorrelation and Regression. Correlation 9.1. Correlation. Chapter 9
Chapter 9 Correlaton and Regresson 9. Correlaton Correlaton A correlaton s a relatonshp between two varables. The data can be represented b the ordered pars (, ) where s the ndependent (or eplanator) varable,
More informationPop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing
Advanced Scence and Technology Letters, pp.164-168 http://dx.do.org/10.14257/astl.2013 Pop-Clc Nose Detecton Usng Inter-Frame Correlaton for Improved Portable Audtory Sensng Dong Yun Lee, Kwang Myung Jeon,
More informationMultigradient for Neural Networks for Equalizers 1
Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT
More informationMulti-layer neural networks
Lecture 0 Mult-layer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Lnear regresson w Lnear unts f () Logstc regresson T T = w = p( y =, w) = g( w ) w z f () = p ( y = ) w d w d Gradent
More informationon the improved Partial Least Squares regression
Internatonal Conference on Manufacturng Scence and Engneerng (ICMSE 05) Identfcaton of the multvarable outlers usng T eclpse chart based on the mproved Partal Least Squares regresson Lu Yunlan,a X Yanhu,b
More informationStatistical Hypothesis Testing for Returns to Scale Using Data Envelopment Analysis
Statstcal Hypothess Testng for Returns to Scale Usng Data nvelopment nalyss M. ukushge a and I. Myara b a Graduate School of conomcs, Osaka Unversty, Osaka 560-0043, apan (mfuku@econ.osaka-u.ac.p) b Graduate
More informationwhere I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).
11.4.1 Estmaton of Multple Regresson Coeffcents In multple lnear regresson, we essentally solve n equatons for the p unnown parameters. hus n must e equal to or greater than p and n practce n should e
More informationFirst Year Examination Department of Statistics, University of Florida
Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve
More informationResource Allocation with a Budget Constraint for Computing Independent Tasks in the Cloud
Resource Allocaton wth a Budget Constrant for Computng Independent Tasks n the Cloud Wemng Sh and Bo Hong School of Electrcal and Computer Engneerng Georga Insttute of Technology, USA 2nd IEEE Internatonal
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationRBF Neural Network Model Training by Unscented Kalman Filter and Its Application in Mechanical Fault Diagnosis
Appled Mechancs and Materals Submtted: 24-6-2 ISSN: 662-7482, Vols. 62-65, pp 2383-2386 Accepted: 24-6- do:.428/www.scentfc.net/amm.62-65.2383 Onlne: 24-8- 24 rans ech Publcatons, Swtzerland RBF Neural
More information