A Comparison on Neural Network Forecasting

Size: px
Start display at page:

Download "A Comparison on Neural Network Forecasting"

Transcription

1 011 nternatonal Conference on Crcuts, System and Smulaton PCST vol.7 (011) (011) ACST Press, Sngapore A Comparson on Neural Networ Forecastng Hong-Choon Ong 1, a and Shn-Yue Chan 1, b 1 School of Mathematcal Scences, Unverst Sans Malaysa, USM, Penang, Malaysa. a hcong@cs.usm.my, b shnyue_alu@yahoo.com Abstract. Ths study compares the effectveness of the Box-Jenns model and neural networs model n mang a forecast. An eghteen years bmonthly water consumpton n Penang data set s analyzed. Multlayer perceptron (MLP) n neural networ wth sngle hdden layer and double hdden layers usng errorbacpropagaton algorthm models are used. Four MLP programs wth double hdden layers whch are the orgnal data (O), the deseasonalzed wth the smoothng movng average method (DS), the lnear detrended model (DTL) and both deseasonalzed and detrended model (DSTL) and a sngle hdden layer MLP wth both deseasonalzed and detrended (DSTL) were smulated. For tme seres analyss, a Box-Jenns SARMA model s generated. The performance of the models are measured usng three types of error measurement, whch are mean absolute error (MAE), absolute percentage error (MAPE) and root mean squared error (RMSE). The model wth smallest value of MAE, MAPE and RMSE stands out to be the best model n predctng the water consumpton n Penang for the year 008. The results showed that both SARMA and double hdden layers MLP models perform relatvely well. Furthermore, double hdden layers MLP model shows an mprovement n predcton as compared wth the sngle hdden layer MLP model. Keywords: neural networ; mult-layer perceptron; Box-Jenns SARMA model. 1. ntroducton Neural networ has the ablty to represent both lnear and non-lnear relatonshps and the networ can learn these relatonshps drectly from the raw data. Mult-layered Perceptron (MLP) model s the most common model n neural networ and s a feed-forward model (Looney, 1997). Each node n one layer connects wth a certan weght to every other node n the followng layer. n general, more hdden layers can prevent the necessty of usng unnecessary neurons to acheve hghly non-lnear classfcaton. The error bac-propagaton algorthm used adjusts the weghts to obtan more accurate forecast values. A tme seres s a sequence of data ponts (observatons), collected typcally at successve tmes, spaced at often unform tme ntervals. Data collected nconsstently or only once may not be consdered as a tme seres. Tme seres analyss conssts of methods that attempt to understand the specfc tme seres, dentfyng the underlyng context of the observatons and ft a sutable model to mae predctons. Tme seres forecastng means the use of a model to predct future events based on the past data; to forecast future data ponts before they are measured. Tme seres can be appled nto varous felds, such as economc forecastng, sales forecastng, stoc maret analyss, process and qualty control, nventory forecastng, producton plannng and many more (Chatfeld, 004).. A Bref Lterature Revew n recent years, artfcal neural networs have been wdely used as a useful tool to model many areas of engneerng applcatons. Sarıdemr (008) presented the models n artfcal neural networs (ANN) for predctng compressve strength of concretes contanng metaaoln and slca fume that had been developed at the age of 1, 3, 7, 8, 56, 90 and 180 days. The mult-layered feed forward neural networs models were fed wth these parameters to predct the compressve strength values of concretes contanng metaaoln and slca fume. The results of ths study showed that neural networs have strong potental for predcton, whch proved that artfcal neural networs are capable of learnng and generalzng from examples and 56

2 experences. Ong, et al. (008) gave on a functonal approxmaton comparson between neural networs and polynomal regresson. The results showed that approxmaton usng polynomal regresson generates lower fracton of varance unexplaned (FVU) value as compared to approxmaton usng neural networs wth sngle hdden layer mult-layered perceptron (MLP) and double hdden layer MLP except for the complcated functons. Moreover, double hdden layer MLP showed better estmaton results than sngle hdden layer MLP wthout consderng the number of parameters beng used. On the other hand, statstcal models have commonly been used n tme seres data analyss and forecastng. Zou and Yang (003) proposed combnng tme seres model for forecastng for a better performance. Lam, et al. (008) used autoregressve ntegrated movng average (ARMA) model to measure the nterventon effects and the asymptotc change n the smulaton results of the busness process reengneerng that s based on the actvty model analyss. A case study of a purchasng process of a household applance manufacturng enterprse that nvolves 0 purchasng actvtes was used as the tranng data. The results ndcated that the changes can be explctly quantfed and the effects of busness process reengneerng can be measured. n general, many tme seres are asymptotcally unstable and ntrnscally non-statonary. Box-Jenns models solve these problems by mposng on data transformatons, such as dfferencng and logarthm. Grllenzon (1998) dscussed on a method for modelng tme seres wth unstable roots and changng parameters. A method of adaptve forecastng whch s based on the optmzaton of recursve estmators was appled to well-nown data sets. He had demonstrated ts valdty n several mplementatons and for dfferent model structures. 3. Methodology 3.1. Neural networ mult-layered perceptron Mult-layered perceptron s the most common neural networ model. t s abltes n achevng hghly nonlnear classfcaton and nose tolerance mae t sutable to be used n ths study. Bpolar sgmod functon s appled n ths study. n a sngle hdden layer model, the perceptron computes assocated outputs from multple real-valued nputs by formng a lnear combnaton accordng to ts nput weghts. The data set s treated as nputs and denoted as { x }, where = number of data and has an assocated set of exemplar outputs target vector. A random ntal synaptc weght set for the hdden and output layers are selected. Fnally, the current weghts are re-terated and updated usng the error bac-propagaton algorthm to force each of the nput exemplar feature vectors to be mapped closer to the target vector that dentfes a class n the nput space. n the double hdden layered MLP, the perceptrons also computes assocated outputs from multple real-valued nputs by formng a lnear combnaton accordng to ts nput weghts and the data set s treated as nputs. The man dfference s that the synaptc weght set has an addtonal second hdden layer. Error bac-propagaton s used to update the error computed n the networ wth respect to the networ's modfable weghts. The error value whch s the dfference between output value and desred value s then computed. The errors wll be propagated bac nto the networ usng a gradent descent technque so that weghts adjustment can be done. 3.. Desgn of Box-Jenns model There are three prmary stages n buldng a Box-Jenns tme seres model, whch are model dentfcaton, model estmaton and model dagnostcs. 1) Model dentfcaton: There s a need to determne f the seres s statonary, has a trend component and f there s any sgn of seasonalty that needs to be modeled. Statonarty, trend and seasonalty can be detected by plottng the tme seres graph. f there s trend component present n the data, the graph wll not be n a random form. Statonarty can also be checed va sample auto-correlaton (ACF). n a statonary seres, the auto-covarance functon, γ and the auto-correlaton functon, ρ show the followng propertes: 1. γ γ 0.. ρ 0 = 1 and -1< ρ <1 for γ = γ and ρ = ρ for all. Seasonalty can be accessed from sample auto-correlaton (ACF) at varous lag and also sample partal auto-correlaton (PACF) at varous lag of the process. The behavor of the ACF and PACF s shown n Table 1. After statonarty and seasonalty have been addressed, the next step s to dentfy the order (the p and q) of the autoregressve and movng average terms. The behavor of ACF and PACF wll ndcate the 57

3 order form of the autoregressve and movng average terms. TABLE. BEHAVOR OF ACF AND PACF AR(p) MA(q) ARMA(p,q) ACF Tals off Cuts off after lag q Tals off PACF Cuts off after lag p Tals off Tals off ) Model estmaton: Parameters estmaton for the Box-Jenns models s a qute complcated nonlnear estmaton problem. Therefore, hgh qualty software program that fts Box-Jenns model should be used to estmate the parameter (Anderson, 1994). Consder Y t s statonary tme seres and { ε t } are ndependently and dentcally dstrbuted wth N ( 0, σ ε ), an ARMA(p,q) model wth unnown parameters ϕ = { ϕ1, ϕ,... ϕp}, θ = { θ1, θ,... θ q } and σ ( ε = E εt ) s gven as: Y = ϕy + ϕ Y ϕ Y + ε θε θ ε... θ ε t 1 t 1 t p t p t 1 t 1 t q t q The man approaches to fttng Box-Jenns models are maxmum lelhood estmaton and least squares estmaton. 3) Model dagnostcs: The ACF plot and PACF plot of the estmated resdual { ε t } have the ablty to ndcate whether the resduals are ndependent and normally dstrbuted. f the ACF plot and PACF plot show the ponts are dstrbuted randomly or do not show any pattern, then ths ndcates that the resduals are ndependent and normally dstrbuted. However, f these assumptons are not satsfed, a more approprate model needs to be ftted. Ths means that, the model dentfcaton step has to be redone to re-develop a better model that satsfes the dagnostc test. Although a good model may not be obtaned on the frst tral, however a new and better model can be easly derved (We, 1990). Correlogram of resduals and correlogram of resduals squared are suffcent for determnng the status of whte nose. Q-statstcs s useful to test normalty of the resduals wthn a range of an ndependent varable. n other words, t served as a tool to test whether the seres s whte nose. The last two columns n the correlogram represent Ljung-Box Q- statstcs and ther p-value respectvely. The Q-statstcs at lag s a test statstcs for the null hypothess where there s no auto-correlaton up to order. t s computed as follows: Q = T( T + ) ( T j) r where rj represents the j-th auto-correlaton and T represents number of observatons. f the p-value s less than 0.05, then t s sgnfcant. Therefore, the null hypothess s rejected, and the error terms are correlated. f the p-values are greater or equal to 0.05, then the p-values are not sgnfcant at all lags. Thus, the null hypothess s not rejected, and the resduals are not correlated (Harvey, 001) Crtera comparson by error measurement Three types of error measurement are used n ths study, whch are the mean absolute error (MAE), the percentage error (MAPE) and the root mean squared error (RMSE). Mean Absolute Error (MAE) MAE = Average of the absolute values of forecastng error = 1 = ˆ where = 1,,,, = orgnal data, j = 1 1 j Mean Absolute Percentage Error (MAPE) MAPE = Percentage of the mean rato of the error to the orgnal data 58

4 = ˆ = 1 where = 1,,,, 100 = orgnal data, Root Mean Squared Error (RMSE) RMSE = Square root of the average of the summng square forecastng errors = n = 1 ( ) where = 1,,,, 4. Result and Dscusson = orgnal data, The total of 108 data collected s dvded nto two groups, namely tranng data and testng data. The tranng data whch conssts of the frst 10 data s used to forecast the next 6 forecast value. The last 6 data s used as testng set, where the man purpose s to mae comparson wth the forecastng value generated by tranng set n order to calculate the error. The model wth smallest error value of the forecast s the best ftted forecast model. Fnally, the next sx forecast ponts are forecasted usng the selected best model. The neural networ models are wrtten n C++ programmng language whle Box-Jenns model was evaluated n MNTAB 14 and E-VEW. Table shows the target value, forecasted value from SARMA model and forecasted value from neural networ models. TABLE. FORECASTED VALUE FROM SARMA MODEL AND NEURAL NETWORK MODELS. Neural Networ Models (double hdden layers) Neural Networ Model (sngle hdden layer) t orgnal data O DTL DS DSTL DSTL SARMA Model Crtera comparson by error measurement The error between forecasted value and the orgnal value s compared between estmated models. MAE, MAPE and RMSE are used as error measurement. Table 3 shows the error measurements for SARMA model, four Neural Networ models wth double hdden layers and one Neural Networ model wth sngle hdden layer. TABLE. ERROR MEASUREMENT FOR SARMA MODEL AND NEURAL NETWORK MODELS Model Type MAE MAPE RMSE SARMA Neural Networ models wth double hdden layers O DTL DS DSTL Neural Networ model wth sngle hdden layer DSTL When comparsons were made, both detrendng and deseasonalzed model (DSTL) wth double hdden layers gves the smallest error n MAE and MAPE compared to other models. However, SARMA model 59

5 stands out to be a better model f RMSE s beng compared. Snce the errors are squared before they are averaged, the RMSE gves a relatvely hgh weght to large errors. Although, the MAE s very smlar to the RMSE but t s less senstve to large forecast errors. Therefore, both DSTL wth double hdden layers model and SARMA model tends to have ts own good characterstcs n generatng smallest error durng forecastng. The orgnal model (O) wth double hdden layers appears to generate the hghest MAE, MAPE and RMSE, follow by deseasonalzed model (DS) wth double hdden layers, detrendng model (DTL) wth double hdden layers and both deseasonalzed and detrendng model (DSTL) wth sngle hdden layer. When DSTL wth double hdden layers model s compared wth DSTL wth sngle hdden layer model, the double hdden layers model outperforms sngle hdden layer model. 4.. Forecastng n water consumpton SARMA model and DSTL wth double hdden layers model are both the best model n forecastng the water consumpton n Penang and the next sx data are forecasted. Ths sx data s the bmonthly water consumpton n Penang data startng from January 008 to December 008. Table 4 and 5 show the forecasted sx data of bmonthly water consumpton n Penang from January 008 to December 008 by usng SARMA model and DSTL wth double hdden layers mult-layered perceptron model respectvely. TABLE V. FORECAST OF BMONTHLY WATER CONSUMPTON N PENANG USNG SARMA MODEL Bmonthly Jan-Feb Mar-Apr May-Jun Jul-Aug Sept-Oct Nov-Dec Forecast TABLE V. FORECAST OF BMONTHLY WATER CONSUMPTON N PENANG USNG DSTL 5. Concluson Bmonthly Jan-Feb Mar-Apr May-Jun Jul-Aug Sept-Oct Nov-Dec Forecast From the results generated, t s clearly shown that both SARMA model and DSTL are the best models used. For the neural networ models, t s obvous that preprocessng helps to mprove the forecast. Furthermore, t s also shown that the double hdden layered MLP that the DSTL s able to get a better result as compared to the sngle hdden layered MLP. Acnowledgment We would le to than the Drector of the Water Wors Department of Penang, Malaysa, for provdng us the data. Ths wor was supported n part by the U.S.M. nsentve grant no. 1001/PMATH/8184. References [1] T. W. Anderson, The statstcal analyss of tme seres, Unted States, John Wley and Sons, nc., [] C. Chatfeld, The analyss of tme seres: an ntroducton, 6th ed., Florda, CRC Press, 004. [3] C. Grllenzon, Forecastng unstable and nonstatonary tme seres, Journal of Forecastng, vol. 14(4), 1st December 1998, pp [4] A. C. Harvey, Forecastng, structural tme seres models and the Kalman flter, New Yor, Cambrdge Unversty Press, [5] C. Y. Lam, W. H. p and C. W. Lau, A busness process actvty model and performance measurement usng a tme seres ARMA nterventon analyss, Expert Systems wth Applcatons, vol. 36(3), Part, Aprl 009, pp [6] M. Sandemr, Predcton of compressve strength of concretes contanng metaaoln and slca fume by artfcal neural networs, Advances n Engneerng Software, vol. 40 (5), pp [7] We and W. S. Wllam, Tme seres analyss: unvarate and multvarate methods, Calforna: Addson-Wesley Publshng Company, nc., [8] H. Zou and Y. H. Yang, Combnng tme seres models for forecastng, Journal of Forecastng, vol. 0(1), January-March 005, pp

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

A LINEAR PROGRAM TO COMPARE MULTIPLE GROSS CREDIT LOSS FORECASTS. Dr. Derald E. Wentzien, Wesley College, (302) ,

A LINEAR PROGRAM TO COMPARE MULTIPLE GROSS CREDIT LOSS FORECASTS. Dr. Derald E. Wentzien, Wesley College, (302) , A LINEAR PROGRAM TO COMPARE MULTIPLE GROSS CREDIT LOSS FORECASTS Dr. Derald E. Wentzen, Wesley College, (302) 736-2574, wentzde@wesley.edu ABSTRACT A lnear programmng model s developed and used to compare

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

Chapter 6. Supplemental Text Material

Chapter 6. Supplemental Text Material Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Chapter 5 Multilevel Models

Chapter 5 Multilevel Models Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

Short Term Load Forecasting using an Artificial Neural Network

Short Term Load Forecasting using an Artificial Neural Network Short Term Load Forecastng usng an Artfcal Neural Network D. Kown 1, M. Km 1, C. Hong 1,, S. Cho 2 1 Department of Computer Scence, Sangmyung Unversty, Seoul, Korea 2 Department of Energy Grd, Sangmyung

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

Chapter 15 Student Lecture Notes 15-1

Chapter 15 Student Lecture Notes 15-1 Chapter 15 Student Lecture Notes 15-1 Basc Busness Statstcs (9 th Edton) Chapter 15 Multple Regresson Model Buldng 004 Prentce-Hall, Inc. Chap 15-1 Chapter Topcs The Quadratc Regresson Model Usng Transformatons

More information

Polynomial Regression Models

Polynomial Regression Models LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

Boostrapaggregating (Bagging)

Boostrapaggregating (Bagging) Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

2016 Wiley. Study Session 2: Ethical and Professional Standards Application 6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

PERFORMANCE COMPARISON BETWEEN BACK PROPAGATION, RPE AND MRPE ALGORITHMS FOR TRAINING MLP NETWORKS

PERFORMANCE COMPARISON BETWEEN BACK PROPAGATION, RPE AND MRPE ALGORITHMS FOR TRAINING MLP NETWORKS PERFORMANCE COMPARISON BETWEEN BACK PROPAGATION, RPE AND MRPE ALGORITHMS FOR TRAINING MLP NETWORKS Mohd Yusoff Mashor School of Electrcal and Electronc Engneerng, Unversty Scence Malaysa, Pera Branch Campus,

More information

Statistical Evaluation of WATFLOOD

Statistical Evaluation of WATFLOOD tatstcal Evaluaton of WATFLD By: Angela MacLean, Dept. of Cvl & Envronmental Engneerng, Unversty of Waterloo, n. ctober, 005 The statstcs program assocated wth WATFLD uses spl.csv fle that s produced wth

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

18. SIMPLE LINEAR REGRESSION III

18. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III US Domestc Beers: Calores vs. % Alcohol Ftted Values and Resduals To each observed x, there corresponds a y-value on the ftted lne, y ˆ ˆ = α + x. The are called ftted values.

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

Modeling and Simulation NETW 707

Modeling and Simulation NETW 707 Modelng and Smulaton NETW 707 Lecture 5 Tests for Random Numbers Course Instructor: Dr.-Ing. Magge Mashaly magge.ezzat@guc.edu.eg C3.220 1 Propertes of Random Numbers Random Number Generators (RNGs) must

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

The Ordinary Least Squares (OLS) Estimator

The Ordinary Least Squares (OLS) Estimator The Ordnary Least Squares (OLS) Estmator 1 Regresson Analyss Regresson Analyss: a statstcal technque for nvestgatng and modelng the relatonshp between varables. Applcatons: Engneerng, the physcal and chemcal

More information

28. SIMPLE LINEAR REGRESSION III

28. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III Ftted Values and Resduals US Domestc Beers: Calores vs. % Alcohol To each observed x, there corresponds a y-value on the ftted lne, y ˆ = βˆ + βˆ x. The are called ftted

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Phase I Monitoring of Nonlinear Profiles

Phase I Monitoring of Nonlinear Profiles Phase I Montorng of Nonlnear Profles James D. Wllams Wllam H. Woodall Jeffrey B. Brch May, 003 J.D. Wllams, Bll Woodall, Jeff Brch, Vrgna Tech 003 Qualty & Productvty Research Conference, Yorktown Heghts,

More information

Introduction to Regression

Introduction to Regression Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes

More information

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Topic 23 - Randomized Complete Block Designs (RCBD)

Topic 23 - Randomized Complete Block Designs (RCBD) Topc 3 ANOVA (III) 3-1 Topc 3 - Randomzed Complete Block Desgns (RCBD) Defn: A Randomzed Complete Block Desgn s a varant of the completely randomzed desgn (CRD) that we recently learned. In ths desgn,

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

A Particle Filter Algorithm based on Mixing of Prior probability density and UKF as Generate Importance Function

A Particle Filter Algorithm based on Mixing of Prior probability density and UKF as Generate Importance Function Advanced Scence and Technology Letters, pp.83-87 http://dx.do.org/10.14257/astl.2014.53.20 A Partcle Flter Algorthm based on Mxng of Pror probablty densty and UKF as Generate Importance Functon Lu Lu 1,1,

More information

DERIVATION OF THE PROBABILITY PLOT CORRELATION COEFFICIENT TEST STATISTICS FOR THE GENERALIZED LOGISTIC DISTRIBUTION

DERIVATION OF THE PROBABILITY PLOT CORRELATION COEFFICIENT TEST STATISTICS FOR THE GENERALIZED LOGISTIC DISTRIBUTION Internatonal Worshop ADVANCES IN STATISTICAL HYDROLOGY May 3-5, Taormna, Italy DERIVATION OF THE PROBABILITY PLOT CORRELATION COEFFICIENT TEST STATISTICS FOR THE GENERALIZED LOGISTIC DISTRIBUTION by Sooyoung

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

Online Classification: Perceptron and Winnow

Online Classification: Perceptron and Winnow E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng

More information

Basically, if you have a dummy dependent variable you will be estimating a probability.

Basically, if you have a dummy dependent variable you will be estimating a probability. ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy

More information

is the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors

is the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors Multple Lnear and Polynomal Regresson wth Statstcal Analyss Gven a set of data of measured (or observed) values of a dependent varable: y versus n ndependent varables x 1, x, x n, multple lnear regresson

More information

Lab 2e Thermal System Response and Effective Heat Transfer Coefficient

Lab 2e Thermal System Response and Effective Heat Transfer Coefficient 58:080 Expermental Engneerng 1 OBJECTIVE Lab 2e Thermal System Response and Effectve Heat Transfer Coeffcent Warnng: though the experment has educatonal objectves (to learn about bolng heat transfer, etc.),

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Statistics for Business and Economics

Statistics for Business and Economics Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear

More information

Thailand s Para Rubber Production Forecasting Comparison

Thailand s Para Rubber Production Forecasting Comparison Proceedngs of the Internatonal MultConference of Engneers and Computer Scentsts 2014 Vol II,, March 12-14, 2014, Hong Kong Thaland s Para Rubber Producton Forecastng Comparson Onuma Kosanan and Nantacha

More information

Issues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question

Issues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question Issues To Consder when Estmatng Health Care Costs wth Generalzed Lnear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Queston ISPOR 20th Annual Internatonal Meetng May 19, 2015 Jalpa

More information

A Comparative Study for Estimation Parameters in Panel Data Model

A Comparative Study for Estimation Parameters in Panel Data Model A Comparatve Study for Estmaton Parameters n Panel Data Model Ahmed H. Youssef and Mohamed R. Abonazel hs paper examnes the panel data models when the regresson coeffcents are fxed random and mxed and

More information

The Chaotic Robot Prediction by Neuro Fuzzy Algorithm (2) = θ (3) = ω. Asin. A v. Mana Tarjoman, Shaghayegh Zarei

The Chaotic Robot Prediction by Neuro Fuzzy Algorithm (2) = θ (3) = ω. Asin. A v. Mana Tarjoman, Shaghayegh Zarei The Chaotc Robot Predcton by Neuro Fuzzy Algorthm Mana Tarjoman, Shaghayegh Zare Abstract In ths paper an applcaton of the adaptve neurofuzzy nference system has been ntroduced to predct the behavor of

More information

Identification of Linear Partial Difference Equations with Constant Coefficients

Identification of Linear Partial Difference Equations with Constant Coefficients J. Basc. Appl. Sc. Res., 3(1)6-66, 213 213, TextRoad Publcaton ISSN 29-434 Journal of Basc and Appled Scentfc Research www.textroad.com Identfcaton of Lnear Partal Dfference Equatons wth Constant Coeffcents

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

Spatial Modelling of Peak Frequencies of Brain Signals

Spatial Modelling of Peak Frequencies of Brain Signals Malaysan Journal of Mathematcal Scences 3(1): 13-6 (9) Spatal Modellng of Peak Frequences of Bran Sgnals 1 Mahendran Shtan, Hernando Ombao, 1 Kok We Lng 1 Department of Mathematcs, Faculty of Scence, and

More information

Chapter 12 Analysis of Covariance

Chapter 12 Analysis of Covariance Chapter Analyss of Covarance Any scentfc experment s performed to know somethng that s unknown about a group of treatments and to test certan hypothess about the correspondng treatment effect When varablty

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Multi-Step-Ahead Prediction of Stock Price Using a New Architecture of Neural Networks

Multi-Step-Ahead Prediction of Stock Price Using a New Architecture of Neural Networks Journal of Computer & Robotcs 8(), 05 47-56 47 Mult-Step-Ahead Predcton of Stoc Prce Usng a New Archtecture of Neural Networs Mohammad Taleb Motlagh *, Hamd Khaloozadeh Department of Systems and Control,

More information

A new Approach for Solving Linear Ordinary Differential Equations

A new Approach for Solving Linear Ordinary Differential Equations , ISSN 974-57X (Onlne), ISSN 974-5718 (Prnt), Vol. ; Issue No. 1; Year 14, Copyrght 13-14 by CESER PUBLICATIONS A new Approach for Solvng Lnear Ordnary Dfferental Equatons Fawz Abdelwahd Department of

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks Internet Engneerng Jacek Mazurkewcz, PhD Softcomputng Part 3: Recurrent Artfcal Neural Networks Self-Organsng Artfcal Neural Networks Recurrent Artfcal Neural Networks Feedback sgnals between neurons Dynamc

More information

STAT 511 FINAL EXAM NAME Spring 2001

STAT 511 FINAL EXAM NAME Spring 2001 STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte

More information

Chapter 8 Indicator Variables

Chapter 8 Indicator Variables Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n

More information

Chapter 6. Supplemental Text Material. Run, i X i1 X i2 X i1 X i2 Response total (1) a b ab

Chapter 6. Supplemental Text Material. Run, i X i1 X i2 X i1 X i2 Response total (1) a b ab Chapter 6. Supplemental Text Materal 6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.

More information

Learning Objectives for Chapter 11

Learning Objectives for Chapter 11 Chapter : Lnear Regresson and Correlaton Methods Hldebrand, Ott and Gray Basc Statstcal Ideas for Managers Second Edton Learnng Objectves for Chapter Usng the scatterplot n regresson analyss Usng the method

More information

III. Econometric Methodology Regression Analysis

III. Econometric Methodology Regression Analysis Page Econ07 Appled Econometrcs Topc : An Overvew of Regresson Analyss (Studenmund, Chapter ) I. The Nature and Scope of Econometrcs. Lot s of defntons of econometrcs. Nobel Prze Commttee Paul Samuelson,

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Application research on rough set -neural network in the fault diagnosis system of ball mill

Application research on rough set -neural network in the fault diagnosis system of ball mill Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(4):834-838 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 Applcaton research on rough set -neural network n the

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Multilayer Perceptron (MLP)

Multilayer Perceptron (MLP) Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study

More information

Basic Business Statistics, 10/e

Basic Business Statistics, 10/e Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson

More information

T E C O L O T E R E S E A R C H, I N C.

T E C O L O T E R E S E A R C H, I N C. T E C O L O T E R E S E A R C H, I N C. B rdg n g En g neern g a nd Econo mcs S nce 1973 THE MINIMUM-UNBIASED-PERCENTAGE ERROR (MUPE) METHOD IN CER DEVELOPMENT Thrd Jont Annual ISPA/SCEA Internatonal Conference

More information

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear

More information

STAT 405 BIOSTATISTICS (Fall 2016) Handout 15 Introduction to Logistic Regression

STAT 405 BIOSTATISTICS (Fall 2016) Handout 15 Introduction to Logistic Regression STAT 45 BIOSTATISTICS (Fall 26) Handout 5 Introducton to Logstc Regresson Ths handout covers materal found n Secton 3.7 of your text. You may also want to revew regresson technques n Chapter. In ths handout,

More information

Midterm Examination. Regression and Forecasting Models

Midterm Examination. Regression and Forecasting Models IOMS Department Regresson and Forecastng Models Professor Wllam Greene Phone: 22.998.0876 Offce: KMC 7-90 Home page: people.stern.nyu.edu/wgreene Emal: wgreene@stern.nyu.edu Course web page: people.stern.nyu.edu/wgreene/regresson/outlne.htm

More information

Chapter 14 Simple Linear Regression

Chapter 14 Simple Linear Regression Chapter 4 Smple Lnear Regresson Chapter 4 - Smple Lnear Regresson Manageral decsons often are based on the relatonshp between two or more varables. Regresson analss can be used to develop an equaton showng

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Lecture 23: Artificial neural networks

Lecture 23: Artificial neural networks Lecture 23: Artfcal neural networks Broad feld that has developed over the past 20 to 30 years Confluence of statstcal mechancs, appled math, bology and computers Orgnal motvaton: mathematcal modelng of

More information

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9 Chapter 9 Correlaton and Regresson 9. Correlaton Correlaton A correlaton s a relatonshp between two varables. The data can be represented b the ordered pars (, ) where s the ndependent (or eplanator) varable,

More information

Pop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing

Pop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing Advanced Scence and Technology Letters, pp.164-168 http://dx.do.org/10.14257/astl.2013 Pop-Clc Nose Detecton Usng Inter-Frame Correlaton for Improved Portable Audtory Sensng Dong Yun Lee, Kwang Myung Jeon,

More information

Multigradient for Neural Networks for Equalizers 1

Multigradient for Neural Networks for Equalizers 1 Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT

More information

Multi-layer neural networks

Multi-layer neural networks Lecture 0 Mult-layer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Lnear regresson w Lnear unts f () Logstc regresson T T = w = p( y =, w) = g( w ) w z f () = p ( y = ) w d w d Gradent

More information

on the improved Partial Least Squares regression

on the improved Partial Least Squares regression Internatonal Conference on Manufacturng Scence and Engneerng (ICMSE 05) Identfcaton of the multvarable outlers usng T eclpse chart based on the mproved Partal Least Squares regresson Lu Yunlan,a X Yanhu,b

More information

Statistical Hypothesis Testing for Returns to Scale Using Data Envelopment Analysis

Statistical Hypothesis Testing for Returns to Scale Using Data Envelopment Analysis Statstcal Hypothess Testng for Returns to Scale Usng Data nvelopment nalyss M. ukushge a and I. Myara b a Graduate School of conomcs, Osaka Unversty, Osaka 560-0043, apan (mfuku@econ.osaka-u.ac.p) b Graduate

More information

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X). 11.4.1 Estmaton of Multple Regresson Coeffcents In multple lnear regresson, we essentally solve n equatons for the p unnown parameters. hus n must e equal to or greater than p and n practce n should e

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

Resource Allocation with a Budget Constraint for Computing Independent Tasks in the Cloud

Resource Allocation with a Budget Constraint for Computing Independent Tasks in the Cloud Resource Allocaton wth a Budget Constrant for Computng Independent Tasks n the Cloud Wemng Sh and Bo Hong School of Electrcal and Computer Engneerng Georga Insttute of Technology, USA 2nd IEEE Internatonal

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

RBF Neural Network Model Training by Unscented Kalman Filter and Its Application in Mechanical Fault Diagnosis

RBF Neural Network Model Training by Unscented Kalman Filter and Its Application in Mechanical Fault Diagnosis Appled Mechancs and Materals Submtted: 24-6-2 ISSN: 662-7482, Vols. 62-65, pp 2383-2386 Accepted: 24-6- do:.428/www.scentfc.net/amm.62-65.2383 Onlne: 24-8- 24 rans ech Publcatons, Swtzerland RBF Neural

More information