A Comparative Study of Parametric and Nonparametric Regressions

Size: px
Start display at page:

Download "A Comparative Study of Parametric and Nonparametric Regressions"

Transcription

1 Iranian Econoic Review, Vol.16, No.30, Fall 011 A Coparative Study of Paraetric and Nonparaetric Regressions Shahra Fattahi Received: 010/08/8 Accepted: 011/04/4 Abstract his paper evaluates inflation forecasts ade by paraetric and T nonparaetric odels. The results revealed that the neural network odel yields better estiates of inflation rate than do paraetric autoregressive integrated oving average (ARIMA) and linear odels. Furtherore, the neural network odel outperfored nonparaetric odels (except MARS). Keywords: ARIMA, AM, MARS, PPR, NN, Inflation Forecast. 1- Introduction The paraetric regression approach is based on the prior knowledge of the functional for relationship. If the knowledge is correct, the paraetric ethod can odel ost data sets well. However, if the wrong functional for is chosen a priori, this will result in larger bias as copared to copetitive odels (Fan and Yao, 003). Paraetric linear odels, as a type of paraetric regression, are frequently used to describe the association between the dependent variable and explanatory variables. They require the estiation of a finite nuber of paraeters. Furtherore, paraetric linear dynaic odels which are based on a theoretical or data-driven approach will be eployed. Why is nonparaetric regression iportant? Over the last decade, increasing attention has been devoted to nonparaetric regression as a new. Assistant Professor, Departent of Econoics, Razi University.

2 0/ A Coparative Study of Paraetric and Nonparaetric Regressions technique for estiation and forecasting in different sciences including econoics. Nonparaetric regression analysis relaxes the assuption of the linearity in the regression analysis and enables one to explore the data ore flexibly. However, in high diensions variance of the estiates rapidly increases, called as curse of diensionality, due to the sparseness of data. To overcoe this proble, soe nonparaetric ethods have been proposed such as additive odel (AM), ultiple adaptive regression splines (MARS), proection-pursuit regression (PPR) and neural networks (NN). This study copares paraetric and nonparaetric ethods. The agents are assued to use an optial paraetric autoregressive oving average (ARMA) odel or nonparaetric odels including Additive odel (AD), Multiple Adaptive Regression Splines (MARS), Proection-Pursuit Regression (PPR), and Neural Networks (NN) for forecasting. In fact, outof-saple estiates of inflation generated by the paraetric and nonparaetric odels will be copared. The paper is divided into four sections. Following introduction in section 1 we discuss paraetric and nonparaetric ethods in section. Section 3 reports the epirical results of this study. Finally, section 4 concludes. - Statistical predictors -1- Paraetric prediction odels A brief review of ARIMA odeling is presented (Chatfield, 000). Autoregressive integrated oving average (ARIMA) or (Box-Jenkins) odels are the basis of any fundaental ideas in tie-series analysis. In order to analyze a tie series, it ust be assued that the structure of the stochastic process which generates the observations is essentially invariant through tie. The iportant assuption is that of stationarity, which requires the process to be in a particular state of statistical equilibriu (Box and Jenkins, 1976). An autoregressive oving average process: ARMA (p,q) is obtained by cobining p autoregressive ters and q oving average ters and can be written as ( L) X ( L) t t p ( L) 1 1L L... pl with AR polynoial and

3 Fattahi, SH. /1 MA polynoial q ( L) 1 1L L... ql. An ARMA odel is stationary provided that the roots of ( L) 0 lie outside the unite circle. This process is invertible if the roots of ( L) 0 lie outside the unite circle. Low order ARMA odels are of uch interest since any real data sets are well approxiated by the rather than by a pure AR or pure MA odel. In general, ARMA odels need fewer paraeters to describe the process. In ost cases econoic tie series are non-stationary and therefore we cannot apply ARMA odels directly. One possible way to reove the proble is to take difference so as to ake the stationary. Non-stationary series often becoe stationary after taking first difference ( X t X t1 (1 L) X ). If the original tie series is differenced d t ties, then the odel is said to be an ARIMA (p, d, q) where I stands for integrated and d denotes the nuber of differences taken. Such a odel is described by ( L)(1 L) d X ( L) t The cobined AR operator is now ( L)(1 L) d. The polynoials () z and () z have all their roots outside the unit circle. The odel is called integrated of order d and the process is said to have d unit roots. t -- Nonparaetric prediction odels --1- Nonparaetric Soothers The general nonparaetric regression odel (Fox, 000, 005) is as follows: ' y f ( X ) i i i f ( x, x,..., x ) i1 i ik i i ~ NID(0, ) The regression function is f (.) unspecified in advance and is estiated directly. In fact, there is no paraeter to estiate. It is iplicitly assued that f (.) is a sooth, continuous function. If there is only one predictor yi f ( xi ) i then it is called scatter plot soothing because it traces a sooth curve through a scatter plot of y against x.

4 / A Coparative Study of Paraetric and Nonparaetric Regressions There are several soothers such as local averaging, kernel soother, Weighted Scatterplot Soothing (Lowess) and spline Soother that fit a linear or polynoial regression to the data points in the vicinity of x and then use the soothed value as the predicted value at x Local Averaging In local averaging procedures, we ove a window continuously over the data, averaging the observations that fall in the window. The estiated values f ˆ( x ) at a nuber of focal values of x are calculated and connected. It is possible to use a window of fixed width or to adust the width of window to include a constant nuber of observations. Local averages are usually subect to boundary bias, roughness and distortion (when outliers fall in the window) Kernel Soother A Kernel soother is an extension of local averaging and usually produces a soother result. At the focal value x 0, it is of the for n fˆ( x ) 0 i1 xi x yk i b 0 n i1 xi x K b 0 where b is a bandwidth paraeter, and K a kernel function. The Gaussian kernel ( KN ( z)) and the tricube kernel ( KT ( z)) are popular choices of kernel functions. 1 z KN () z e 3 3 K () T z (1- z ) for z <1 0 for z 1 For the Gaussian kernel the bandwidth b is the standard deviation of a noral distribution and for the tricube kernel b is the half-width of a window enclosing the observations for the local regression. Although the kernel

5 Fattahi, SH. /3 soother has a better perforance as copared to the local average regression, it is still subect to boundary bias. It is iplicitly assued that the bandwidth b is fixed, but it is possible for kernel soothers to be adapted to nearest-neighbor bandwidths. We can adust b (x) so that a constant nuber of observations are included in the window. The fraction /n is called the span of the kernel soother and is chosen based on a cross-validation approach. The kernel estiator can produce soother results using larger bandwidths. In fact, there is a direct relationship between the span and soothing degree: the larger the span, the soother the result Lowess Soother As entioned above, the kernel estiation has soe probles. Local polynoial regression tries to overcoe these difficulties and provides a generally adequate ethod of nonparaetric regression which extends to additive regression (Fox, 005). An ipleentation of local polynoial regression is lowess (Cleveland, 1979). The algorith used by lowess soothing applies robust locally linear fits. It is siilar to local averaging but the data points that lie in the window are weighted so that nearby points get the ost weight and a robust weighted regression is used. We can exaine local polynoial regression in two cases: siple regression and ultiple regression. Siple Regression: suppose we want to estiate the siple regression yi f ( xi ) i at a particular x-value, for exaple x 0. Local polynoial regression extends kernel estiation to a polynoial fit at x 0, using local kernel weights, wi K[( xi x0) / b]. We ipleent a pth - order weighted-least-squares polynoial regression of y on x, y ( x x ) ( x x )... ( x x ) p e i 1 i 0 i 0 p i 0 i to iniize the weighted residual su of squares, we. This procedure is repeated for representative values of x. As in kernel regression, the bandwidth b can either be fixed or variable, b(x), and the span of the local-regression soother is selected based on a cross-validation approach. ' Multiple Regression: in this case, y f ( X ), we need to define a i i i n i1 i i

6 4/ A Coparative Study of Paraetric and Nonparaetric Regressions ' a ultivariate neighborhood around a focal point x0 ( x01, x0,..., x0k ). Furtherore, Euclidean distance is eployed in the lowess function as: k D( x, x ) ( z z ) where the z i are the standardized predictors, i 0 i 0 1 zi xi x s, x is the ean of the th predictor and s is its standard deviation. Calculating weights are based on the scaled distances: D( xi, x0) wi W b Where w (.) is a weight function. In soe cases, b needs to be adusted to define a neighborhood including the [ns] nearest neighbors of x 0 (where the square brackets denote rounding to the nearest integer). As a siple exaple, a local linear fit takes the for: y ( x x ) ( x x )... ( x x ) e i 1 i1 01 i 0 k ik 0k i The cobinations of predictor values are used repeatedly to create the regression surface Spline Soother Suppose we have n pairs( x, y ) considered as n xax '' ( ) i ( i) ( ) xin i1 i i. A soothing spline equation is ss h y f x h f x dx The equation consists of two ters. The first ter is the residual su of squares and the second ter is a roughness penalty. The obect is to find the function f ˆ( x) with two continuous derivatives that iniized the penalized su of squares. Here h is a soothing paraeter. For h=0, f ˆ( x) will interpolate the data if the xi are distinct; this is siilar to a localregression estiate with span=1/n. If h is very large, then ˆf will be selected " so that f ˆ ( x) is everywhere 0, which iplies globally linear least-squares

7 Fattahi, SH. /5 fit to the data. This is again siilar to local regression with infinite neighborhoods. The Spline Soother is ore attractive than local regression because there is an explicit obective-function to optiize. But it is not easy to generalize splines to ultiple regression. Generally, the soothing paraeter h is selected indirectly by setting the equivalent nuber of paraeters for the soother.both soothing-spline and local-regression fits with the sae degree of freedo are usually very siilar. --- Nonparaetric Models Additive Model (AM) Nonparaetric regression based on kernel and soothing spline estiates in high diensions faces two probles, that is, the curse of diensionality and interpretability. Stone (1985) proposed the additive odel to overcoe these probles. In this odel, since each of the individual additive ters is estiated using a univariate soother, the curse of diensionality is avoided. Furtherore, while the nonparaetric for akes the odel ore flexible, the additivity allows us to interpret the estiates of the individual ters. Hastie and Tibshirani (1990) proposed generalized additive odels for a wide range of distribution failies. These odels allow the response variable distribution to be any eber of the exponential faily of distributions. We can apply additive odels to Gaussian response data, logistic regression odels for binary data, and loglinear or log-additive odels for Poisson count data. A generalized additive odel has the for Y f ( X ) f ( X )... f ( X ) 1 1 where f (.) are unspecified sooth (partial-regression)functions. We fit each function using a scatterplot soother and provide an algorith for siultaneously estiating all functions. Here an additive odel is applied to a logistic regression odel as a generalized additive odel. Consider a logistic regression odel for binary data. The ean of the binary response ( X ) Pr( Y 1 X ) is related to the explanatory variables via a linear regression odel and the logit link functions: p p

8 6/ A Coparative Study of Paraetric and Nonparaetric Regressions ( X ) log 1X1... X 1 ( X ) The additive logistic odel replaces each linear ter by a ore general functional for ( ) log X f1( X1)... f ( X ) 1 ( X ) In general, the conditional ean ( X ) of a response Y is related to an additive function of the explanatory variables via a link function g: g[ ( X)] f1( X1)... f ( X ) The functions f are estiated in a flexible way using the backfitting algorith. This algorith fits an additive odel using regression-type fitting echaniss. Consider the th set of partial residuals Y ( fk ( X k )) k Then E( X ) f ( X ). This observation provides a way for estiating each (.) f given estiates f ˆ (.), i for all the others. The iterative process is called the backfitting algorith (Friedan and Stuetzle, 1981) Multiple Adaptive Regression Splines (MARS) This approach (Friedan (1991), Hastie et al (001)) fits a weighted su of ultivariate spline basis functions and is well suited for high-diensional probles, where the curse of diensionality would likely create probles for other ethods. The MARS uses the basis functions ( x ) and( t ) in the following way t x ( xt) ( tx) x-t if x>t 0 otherwise t-x if x<t 0 otherwise

9 Fattahi, SH. /7 The + denotes positive part. Each function is piecewise linear or linear spline, with a knot at value t. These functions are called a reflected pair for each input X with knots at each observed value x i of that input, and then the set of basis functions is defined as C ( X t),( t X ) The strategy for odel-building is a forward stepwise linear regression using functions fro the set C and their products. Thus the MARS odel has the for M 0 1 f ( X ) h ( X ) where the coefficients are estiated by iniizing the residual suof-squares and each h ( X ) is a function in C. By setting h ( X) 1 (constant function), the other ultivariate splines are products 0 of univariate spline basis functions: k h ( X ) h( x t ) 1 k i( s, ) s, s1 where the subscript i( s, ) eans a particular explanatory variable, and the basis spline in that variable has a knot at t s,. k is the level of interactions between i( s, ) variables and the values of, k1, k,..., k, are the knot sets. Explanatory variables in the odel can be linearly or nonlinearly and are chosen for inclusion adaptively fro the data. The odel will be additive if the order of interactions equals one ( k 1). A backward deletion procedure is used in the MARS odel to prevent overfitting. The basis functions which have little contributions to the accuracy of fit are deleted fro the odel at each stage, producing an estiated best odel f ˆ( ) of each size λ. We can apply a generalized cross-validation criterion to estiate the optial value of λ in the following way N ˆ ( y ( )) 1 i f x i i GCV ( ) (1 M( )/ N) The value of M (λ) includes the nuber of basis functions and the nuber of paraeters used in selecting the optial positions of the knots.

10 8/ A Coparative Study of Paraetric and Nonparaetric Regressions Proection-Pursuit Regression (PPR) If the explanatory vector X is of high diension, the additive odel does not cover the effect of interactions between the independent variables. Proection-Pursuit Regression (Friedan and Stuetzle, 1981) applies an additive odel to proected variables, proecting predictor variables X in M, as follows M T ( ) E( ) 0,var( ) 1 Y g w X where w are unit p-vectors of unknown paraeters. The functions g are unspecified and estiated along with the direction w using soe flexible soothing ethod. The PPR odel eploys the backfitting algorith and Gauss-Newton search to fit Y. T The functions g( wx ) are called the ridge functions because they are constant in all but one direction. They vary only in the direction defined T by the vector w. The scalar variable V ( wx ) is the proection X onto the unit vector w. The ai is to find w to yield the best fit to the data. If M is chosen large enough then the PPR odel can approxiate arbitrary continuous function of X (Diaconis and Shahshahani, 1984). However, in this case there is a proble of interpretation of the fitted odel since each input enters into the odel in a coplex and ultifaceted way (Hastie et al, 001). As a result, the PPR odel is a good option only for forecasting. To fit a PPR odel, we need to iniize the error function N M T i i i1 1 E [ y g ( w x )] over functions g and direction vectors by iteration. Iposing coplexity constraints on the w. The g and w are estiated g is needed to avoid overfitting. There are two stages to estiate g and w. First, to obtain an estiate of g, suppose there is one ter (M=1). We can for the derived T variables vi w x for any value of w. This iplies a one-diensional i soothing proble and any scatterplot soother such as soothing spline can be used to estiate g. Second, we iniize E over w for any value of g. These two steps are iterated until convergence. If there is ore than one ter

11 Fattahi, SH. /9 in the PPR odel then the odel is built in a forward stage-wise anner that at each stage a pair ( w, g ) is added Neural Networks Many recent ethods to developing data-driven odels have been inspired by the learning abilities of biological systes. For instance, ost adults drive a car without knowledge of the underlying laws of physics and huans as well as anials can recognize patterns for the tasks such as face, voice or sell recognition. They learn the only through data-driven interaction with the environent. The field of pattern recognition considers such abilities and tries to build artificial pattern recognition systes that can iitate huan brain. The interest to such systes led to extensive studies about neural networks in the id-1980s (Cherkassky and Mulier, 007). Why use Neural Networks? Neural network odeling has seen an explosion of interest as a new technique for estiation and forecasting in econoics over the last decades. They are able to learn fro experience in order to iprove their perforance and to adapt theselves to changes in the environent. In fact, they can derive trends and detect patterns fro coplicated or iprecise data, and then odel coplex relationships between explanatory variables (inputs) and dependent variables (outputs). They are resistance to noisy data due to a assively parallel distributed processing. Learning in neural network odel Stochastic approxiation (or gradient descent) is one of the basic nonlinear optiization strategies coonly used in statistical and neural network ethods (Cherkassky and Mulier, 007). The gradient-descent ethods are based on the first order Taylor expansion of a risk functional R( w) L( y, f ( x, w)) p( x, y) dxdy (1) where Rw ( ) is the risk functional, L( y, f ( x, w)) the loss function and p( x, y) the oint probability density function. For regression, a coon loss function is the squared error

12 30/ A Coparative Study of Paraetric and Nonparaetric Regressions L( y, f ( x, w)) ( y f ( x, w)) () Learning is then defined as the process of estiating the function f ( x, w0 ) that iniizes the risk functional R( w) ( y f ( x, w)) p( x, y) dxdy using only the training data. Although the gradient-descent ethods are coputationally rather slow, their siplicity has ade the popular in neural networks. We will exaine two cases to describe such ethods: linear paraeter estiation and nonlinear paraeter estiation. Linear Paraeter Estiation Consider a linear (in paraeters) approxiating function and the loss function specified above. For the task of regression, it can be shown that the epirical risk is as follows 1 1 R w L x y w y f x w n n ep ( ) ( i, i, ) ( i ( i, )) (3) n i1 n i1 This function is to be iniized with respect to the vector of paraeters w. Here the approxiating function is a linear cobination of fixed basis functions yˆ f ( x, w) w g ( x) (4) 1 For soe (fixed). The updating equation for iniizing R ( ) ep w with respect to w is w( k 1) w( k) k L( x( k), y( k), w) w where xk ( ) and yk ( ) are the sequences of input and output data saples presented at iteration step k. The gradient above can be written as (5)

13 Fattahi, SH. /31 L yˆ L( x, y, w) ( yˆ y) g ( x) w yˆ w (6) Now the local iniu of the epirical risk can be coputed using the gradient (6). Let us start with soe initial values w (0). The stochastic approxiation ethod for paraeter updating during each presentation of kth training saple is as: Step 1: Forward pass coputations. z ( k) g ( x( k)), 1,.., (7) yˆ( k) w ( k) z ( k). (8) 1 Step : Backward pass coputations. ( k) yˆ ( k) y( k) (9) w ( k 1) w ( k) ( k) z ( k) (10) k k is a sall positive nuber decreasing with k. where the learning rate In the forward pass, the output of the approxiating function is coputed whereas in the backward pass, the error ter (9), which is called delta in neural network literature, for the presented saple is calculated and utilized to odify the paraeters. The paraeter updating equation (10), known as delta rule, updates paraeters with every training saple. Nonlinear Paraeter Estiation The standard ethod used in the neural network literature is the back propagation algorith which is an exaple of stochastic approxiation strategy for nonlinear approxiating functions. As it was considered already, the apping fro inputs to output given by a single layer of hidden units is as follows

14 3/ A Coparative Study of Paraetric and Nonparaetric Regressions n f ( x, w, V ) w w g( v x v ) 0 0 i i 1 i1 d (11) In contrast to (4), the set of functions is nonlinear in the paraeters V. We seek values for the unknown paraeters (weights) V and w that ake the odel fit the training data well. To do so, the su of squared errors as a easure of fit ust be iniized: n ep ( ( i,, ) i) (1) i1 R f x w V y The stochastic approxiation procedure for iniizing respect to the paraeters V and w is Rep with V ( k 1) V ( k) L( x( k), y( k), V ( k), w( k)), (13) k V w( k 1) w( k) L( x( k), y( k), V ( k), w( k)), k 1,..., n, k w (14) where xk ( ) and yk ( ) are the kth training saples, presented at iteration step k. The loss function L is 1 L( x( k), y( k), V ( k), w( k)) ( f ( x, w, V ) y) (15) where the factor ½ is included only for siplifying gradient calculations in the learning algorith. We need to decopose the approxiation function (11) for coputations of the gradient of loss function (15) as follows a z d x v (16) i i i0 g( a ), (17) z, yˆ w z (18)

15 Fattahi, SH. /33 For siplicity, we drop the iteration step k, consider calculation/ paraeter update for one saple at a tie and incorporate the ters w 0 and v into the suations ( x0 1). The relevant gradients, based on the 0 chain rule of derivatives, are R R yˆ a v yˆ a v i i R R yˆ w yˆ w, (19). (0) In order to calculate each of the partial derivatives, we need equations (15) to (18). Therefore, R y ˆ y yˆ yˆ ' g ( a ) w a a v i yˆ w x i z (1) () (3) (4) If we plug the partial derivatives (1)-(4) into (19) and (0), the gradient equations are R ' ( yˆ y) g ( a ) w x i (5) v i R w ( yˆ y) z (6)

16 34/ A Coparative Study of Paraetric and Nonparaetric Regressions Using these gradients and the updating equations, we can construct a coputational ethod to iniize the epirical risk. Starting with soe initial values w (0) and V (0), the stochastic approxiation ethod updates weights upon presentation of a saple (x (k), y (k)) at iteration step k with learning rate as k Step 1: Forward pass coputations. Hidden layer d a ( k) x ( k) v ( k), (7) i i i0 z ( k) g( a ( k)), (8) Output layer z ( k) 1 0 yˆ( k) w ( k) z ( k). (9) 0 Step : Backward pass coputations. Output layer ˆ 0( k) y( k) y( k) (30) w ( k 1) w ( k) ( k) z ( k) (31) k 0 Hidden layer ' 1( k) 0( k) g ( a ( k)) w( k 1) (3) v ( k 1) v ( k) ( k) x ( k) (33) i i k 1 i In the forward pass, the output of the approxiating function is coputed whereas in the backward pass, the error ter for the presented saple is calculated and utilized to odify the paraeters in the output layer. Since it is possible to propagate the error at the output back to an error at each of the internal nodes a through the chain rule of derivatives, the procedure is called error backpropagation. In fact it is a propagation of the error signals fro the output layer to the input layer.

17 Fattahi, SH. /35 The updating steps for output layer are siilar to those for the linear case. Besides, the updating rule for the hidden layer is the sae as the linear one but for the delta ter (3). For this reason, backpropagation update rules (3) and (33) are usually called the generalized delta rule. The paraeter updating algorith holds if the saple size is large (infinite). However, if the nuber of training saples is finite, the asyptotic conditions of stochastic approxiation are (approxiately) satisfied by the repeated presentation of the finite training saple to the training algorith. This is called recycling and the nuber of such repeated training saples is called the nuber of cycles (or epochs). It is possible to use the backpropagation algorith for networks with several output layers and networks with several hidden layers. For instance, if additional layers are added to the approxiation function, then errors are propagated fro layer to layer by repeated application of generalized delta rule. It should be noted that a neural network odel can be identified as a pursuit proection regression (PPR) odel (Hastie et al, 001). In fact, the neural network with one hidden layer has the exactly the sae for as the PPR odel. The only difference is that the PPR odel uses nonparaetric functions ( g() v ) while the neural network eploys a sipler function which is based on a sigoid transfer function. 3- Epirical Results 3-1- Background The Iranian econoy is oil-reliant so that any change in oil price can directly affect all econoic sectors. It should be noted that Iran ranks second in the world in natural gas reserves and third in oil reserves. It is also OPEC s second largest oil exporter. The econoic sectors include the services sector, industry (oil, ining and anufacturing) and the agricultural sector. During the recent decades, the services sector has contributed the largest percentage of the GNP, followed by industry and agricultural sectors. The share of the services sector was 51 percent of GNP in 003 while those of the industry and agricultural sectors were 35.1 and 13.9 percent of GNP respectively. The Iranian econoy has experienced a relatively high inflation averaging about 15 percent per year. The inflation rate has even been ore

18 36/ A Coparative Study of Paraetric and Nonparaetric Regressions than 1 percent on average after the 1973 oil crisis. Furtherore, there is a general agreeent over the underestiating of the easured inflation due to price controls and governent subsidies. The epirical evidence iplies that inflation is persistent in Iran. In other words, the effects of a shock to inflation result in a changed level of inflation for an extended period. To see this, the inflation rate is regressed on its own lags. rgnpi 0.44rgnpi 0.49rgnpi t t1 t (t-value) (3.15) (3.4) As the su of coefficients on lagged inflation (0.93) is close to one, shocks to inflation have long-lasting effects on inflation. 3-- Results This study copares the perforance of paraetric and nonparaetric regressions in the Iranian econoy over the period Different odels will be applied to the inflation series and then the predicted values will be evaluated. In fact, out-of-saple estiates of inflation generated by the paraetric and nonparaetric odels will be copared. The epirical results of the Augented Dicky-Fuller test indicated that all the variables eployed are stationary, and thus this issue helps us to avoid the proble of the spurious relationships (see Appendix for the data source and definitions). It is assued agents use the lagged values of inflation and real GNP growth to forecast inflation. Figure 1.a and Figure 1.b deonstrate a local linear regression fit of inflation rate (rgnpi), defined as the rate of change of GNP deflator, on the lagged inflation rate (rgnpilag1) and lagged real GNP growth rate (rgnplag1) using the Lowess function for a variety of spans. If the fitted regression looks too rough, then we try to increase the span but if it looks sooth, then we will exaine whether the span can be decreased without aking the fit too rough (Fox, 005). The obective is to find the sallest value of span (s) that provides a sooth fit. A trial and error procedure suggests that the span s=0.5 is suitable and it sees to provide a reasonable coproise between soothness and fidelity to the data.

19 rgnpi rgnpi rgnpi rgnpi rgnpi rgnpi rgnpi rgnpi rgnpi rgnpi Fattahi, SH. /37 s = 0.1 s = 0.3 s = rgnpilag rgnpilag rgnpilag1 s = 0.7 s = rgnpilag rgnpilag1 Figure 1-a: Local Linear Regression fit of Inflation Rate (Rgnpi) on the Lagged Inflation Rate (Rgnpilag1) Using Lowess Function for a Variety of Spans s = 0.1 s = 0.3 s = rgnplag rgnplag rgnplag1 s = 0.7 s = rgnplag rgnplag1 Figure 1-b: Local linear regression fit of inflation rate (rgnpi) on the lagged real GNP growth rate (rgnplag1) using Lowess function for a variety of spans

20 38/ A Coparative Study of Paraetric and Nonparaetric Regressions A test of nonlinearity is perfored by contrasting the nonparaetric regression odel with the linear siple-regression odel. We regress inflation on rgnpilag1 (Case 1) and rgnplag1 (Case) separately. As a linear odel is a special case of a nonlinear odel, two odels are nested. An F- test is forulated by coparing alternative nested odels. The results is as follows Linear odel vs Nonparaetric regression (Case1): F=8.78(pvalue=0.008) Linear odel vs Nonparaetric regression (Case): F=6.48(pvalue=0.04) It is obvious that the relationship between the dependent variable and explanatory variables are significantly nonlinear. It should be noted that the variable rgnplags1 will not be significant if a linear regression is considered. Since nonparaetric regression based on soothing functions faces the curse of diensionality, the additive odel has been proposed (Stone, 1985). The result of fitting an additive odel using Lowess soother can be written as rgnpi S(rgnpilag1) S(rgnplag1) F (4.13) (4.43) p-value (0.01) (0.03) where S denotes the Lowess soother function. It is obvious that both soothers are significantly eaningful. Furtherore, the linear odel is nested by the additive odel with p-value being equal to Figure illustrates plots of the estiated partial-regression functions for the additive regression odel. The points in each graph are partial residuals for the corresponding explanatory variable, reoving the effect of the other explanatory variable. The broken lines deonstrate pointwise 95-percent confidence envelopes for the partial fits.

21 rgnpi rgnpi Fattahi, SH. / rgnplag rgnpilag1 Figure : Plots of the Estiated Partial-Regression Functions for the Additive Regression of the Inflation Rate (Rgnpi) on the Lagged Real GNP growth Rate (rgnplag1) and the Lagged Inflation Rate (Rgnpilag1) We use MARS odel to fit a piecewise linear odel with additive ters to the data. The results indicate that pairwise interaction ters (by degree= and degree=3) ake little difference to the effectiveness of explanatory variables. The additive odel sees to be too flexible and it is not able to cover the effect of interactions between explanatory variables. To reove this proble, the Proection- Pursuit Regression odel has been proposed. The PPR odel applies an additive odel to proected variables (Friedan and Stuetzle, 1981). Figure 3 shows plots of the ridge functions for the three twoter proection pursuit regressions fitted to the data.

22 / A Coparative Study of Paraetric and Nonparaetric Regressions ter ter ter ter ter ter Figure 3: Plots of the Ridge regression for three two-ter Proection Pursuit Regressions Fitted to the Data. Although MARS odel is an accurate ethod, it is sensitive to concurvity. Neural networks do not share this proble and are better able to predict in this situation. In fact, as neural networks are nonlinear proection ethods and tend to overparaeterize, they are not subect to concurvity (Hastie et al, 001). We exained several neural network odels and the results indicate that a -3-1 network has a better perforance. The Wilcoxon test has been used to copare the squared error of a neural network odel and a rival odel. In fact, out of saple forecasting has been eployed to evaluate different odels. The perforance of PPR and additive odels appears to differ fro the neural network odel, iplying that the NN odel can significantly outperfor the PPR odel and it has a better perforance than the additive odel, but not by uch. Furtherore, the NN odel is significantly better than the linear odel (LM). However, there is no possibility that the NN odel can outperfor the MARS odel (see Table 1).

23 Fattahi, SH. /41 Table 1: Out of Saple Forecasting based on Wilcoxon test p-value PPR vs. NN 0.01 LM vs. NN 0.00 MARS vs. NN 1 AM vs. NN 0.38 Now we copare the NN odel to the paraetric autoregressive oving average (ARMA) odel for inflation. A collection of ARMA (p, q) odels, for different orders of p and q, have been estiated and then the best odel was selected according to the Akaike inforation criterion (AIC) and the Schwarz inforation criterion (SIC). Exaining the ARMA odels for the inflation series indicated that ARMA (1, 1) is the best-fitting odel. Diagnostic checking, the correlogra (autocorrelations) of inflation fro the regression tests was exained and confired the results. The last 5 observations are used for coparing the ex post forecasts generated by the two odels. Furtherore, the Root Mean Square Error (RMSE) is used to evaluate ex post forecasts. We apply the feed-forward backpropagation as learning algorith where only lagged inflation is used as input. The results iply that the forecasting perforance of the NN odel (RMSE=0.05) is significantly better than that of the ARMA odel (RMSE=11.73). It should be noted that the results fro the inflation lags exceeding one and ore nuber of hidden layers are alost the sae. Therefore, the NN odel outperfors the paraetric ARMA odel. 3- Conclusions This study exained the forecast perforance of paraetric and nonparaetric odels. The agents were assued to use a paraetric autoregressive oving average (ARMA) odel or nonparaetric odels to forecast inflation. The results revealed that the neural network odel yields better estiates of inflation rate than do paraetric autoregressive oving average (ARMA) and linear odels. Coparing to the nonparaetric alternatives, the results of Wilcoxon tests deonstrated that the forecasting perforance of Proection-Pursuit Regression and Additive odels appeared to differ fro the Neural Network odel, iplying that the Neural Network

24 4/ A Coparative Study of Paraetric and Nonparaetric Regressions odel can significantly outperfor Proection-Pursuit Regression and Additive odels. However, there was no possibility that the Neural Network odel can outperfor the Multiple Adaptive Regression Splines odel. References 1- Barucci E and Landi L(1998), Nonlinear Versus Linear Devices: A Procedural Perspective, Coputational Econoics, Vol. 1, pp Box, G.E.P and Jenkins, G (1976) Tie Series Analysis: Forecasting and Control. San Francisco: Holden Day. 3- Chatfield C (000), Tie-Series Forecasting, Chapan & Hall/CRC 4- Cherkassky, V and Mulier F (007), Learning fro Data: Concepts, Theory, and Methods, John Wiley & Sons, Second Edition 5- Cleveland, W.S. (1979), Robust Locally Weighted Regression and Soothing Scatterplots, Journal of the Aerican Statistical Association 74 (368): Diaconis P and Shahshahani M (1984), On nonlinear functions of linear cobinations, SIAM Journal of Scientific and Statistical Coputing, Vol. 5, No. 1, pp Fan, J, Yao, Q (003), Nonlinear Tie Series: Nonparaetric and Paraetric Methods, Springer-Verlag: Berlin Heidelberg and New York. 8- Fox, J (000), Multivariate Generalized Nonparaetric Regression, Sage Publications. 9- Fox, J (005), Introduction to Nonparaetric Regression, McMaster University, Canada. 10- Friedan, JH, Stuetzle, W (1981), Proection Pursuit Regression, Journal of the Aerican Statistical Association, Vol. 76, No. 376, pp Friedan, HF (1991), Multivariate Adaptive Regression Splines, The Annals of Statistics, Vol. 19, No. 1, pp Hastie, TJ, Tibshirani, RJ (1990), Generalized Additive Models, Chapan & Hall. 13- Hastie, TJ, Tibshirani, R, Friedan, J (001), The Eleents of Statistical Learning: Data Mining, Inference, and Prediction, Springer-Verlag: Berlin Heidelberg and New York. 14- Stone, C.J. (1985), Additive Regression and Other Nonparaetric Models, Annals of Statistics, 13,

25 Fattahi, SH. /43 Appendix: Data source and definitions The data are annually for the period and are collected fro the Central Bank of Iran. gnp = real GNP (at the constant 1997 prices) gnpi = GNP deflator (1997=100)

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Machine Learning Basics: Estimators, Bias and Variance

Machine Learning Basics: Estimators, Bias and Variance Machine Learning Basics: Estiators, Bias and Variance Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Basics

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS A Thesis Presented to The Faculty of the Departent of Matheatics San Jose State University In Partial Fulfillent of the Requireents

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Keywords: Estimator, Bias, Mean-squared error, normality, generalized Pareto distribution

Keywords: Estimator, Bias, Mean-squared error, normality, generalized Pareto distribution Testing approxiate norality of an estiator using the estiated MSE and bias with an application to the shape paraeter of the generalized Pareto distribution J. Martin van Zyl Abstract In this work the norality

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

Bootstrapping Dependent Data

Bootstrapping Dependent Data Bootstrapping Dependent Data One of the key issues confronting bootstrap resapling approxiations is how to deal with dependent data. Consider a sequence fx t g n t= of dependent rando variables. Clearly

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS An Iproved Method for Bandwidth Selection When Estiating ROC Curves Peter G Hall and Rob J Hyndan Working Paper 11/00 An iproved

More information

Modeling the Structural Shifts in Real Exchange Rate with Cubic Spline Regression (CSR). Turkey

Modeling the Structural Shifts in Real Exchange Rate with Cubic Spline Regression (CSR). Turkey International Journal of Business and Social Science Vol. 2 No. 17 www.ijbssnet.co Modeling the Structural Shifts in Real Exchange Rate with Cubic Spline Regression (CSR). Turkey 1987-2008 Dr. Bahar BERBEROĞLU

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words)

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words) 1 A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine (1900 words) Contact: Jerry Farlow Dept of Matheatics Univeristy of Maine Orono, ME 04469 Tel (07) 866-3540 Eail: farlow@ath.uaine.edu

More information

Ensemble Based on Data Envelopment Analysis

Ensemble Based on Data Envelopment Analysis Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807

More information

Support Vector Machines MIT Course Notes Cynthia Rudin

Support Vector Machines MIT Course Notes Cynthia Rudin Support Vector Machines MIT 5.097 Course Notes Cynthia Rudin Credit: Ng, Hastie, Tibshirani, Friedan Thanks: Şeyda Ertekin Let s start with soe intuition about argins. The argin of an exaple x i = distance

More information

A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION

A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION A eshsize boosting algorith in kernel density estiation A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION C.C. Ishiekwene, S.M. Ogbonwan and J.E. Osewenkhae Departent of Matheatics, University

More information

arxiv: v1 [cs.ds] 3 Feb 2014

arxiv: v1 [cs.ds] 3 Feb 2014 arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Probability Distributions

Probability Distributions Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples

More information

Extension of CSRSM for the Parametric Study of the Face Stability of Pressurized Tunnels

Extension of CSRSM for the Parametric Study of the Face Stability of Pressurized Tunnels Extension of CSRSM for the Paraetric Study of the Face Stability of Pressurized Tunnels Guilhe Mollon 1, Daniel Dias 2, and Abdul-Haid Soubra 3, M.ASCE 1 LGCIE, INSA Lyon, Université de Lyon, Doaine scientifique

More information

Soft Computing Techniques Help Assign Weights to Different Factors in Vulnerability Analysis

Soft Computing Techniques Help Assign Weights to Different Factors in Vulnerability Analysis Soft Coputing Techniques Help Assign Weights to Different Factors in Vulnerability Analysis Beverly Rivera 1,2, Irbis Gallegos 1, and Vladik Kreinovich 2 1 Regional Cyber and Energy Security Center RCES

More information

Online Publication Date: 19 April 2012 Publisher: Asian Economic and Social Society

Online Publication Date: 19 April 2012 Publisher: Asian Economic and Social Society Online Publication Date: April Publisher: Asian Econoic and Social Society Using Entropy Working Correlation Matrix in Generalized Estiating Equation for Stock Price Change Model Serpilılıç (Faculty of

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical IEEE TRANSACTIONS ON INFORMATION THEORY Large Alphabet Source Coding using Independent Coponent Analysis Aichai Painsky, Meber, IEEE, Saharon Rosset and Meir Feder, Fellow, IEEE arxiv:67.7v [cs.it] Jul

More information

are equal to zero, where, q = p 1. For each gene j, the pairwise null and alternative hypotheses are,

are equal to zero, where, q = p 1. For each gene j, the pairwise null and alternative hypotheses are, Page of 8 Suppleentary Materials: A ultiple testing procedure for ulti-diensional pairwise coparisons with application to gene expression studies Anjana Grandhi, Wenge Guo, Shyaal D. Peddada S Notations

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna

More information

Pattern Classification using Simplified Neural Networks with Pruning Algorithm

Pattern Classification using Simplified Neural Networks with Pruning Algorithm Pattern Classification using Siplified Neural Networks with Pruning Algorith S. M. Karuzzaan 1 Ahed Ryadh Hasan 2 Abstract: In recent years, any neural network odels have been proposed for pattern classification,

More information

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians Using EM To Estiate A Probablity Density With A Mixture Of Gaussians Aaron A. D Souza adsouza@usc.edu Introduction The proble we are trying to address in this note is siple. Given a set of data points

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Support Vector Machines. Goals for the lecture

Support Vector Machines. Goals for the lecture Support Vector Machines Mark Craven and David Page Coputer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Soe of the slides in these lectures have been adapted/borrowed fro aterials developed

More information

Analyzing Simulation Results

Analyzing Simulation Results Analyzing Siulation Results Dr. John Mellor-Cruey Departent of Coputer Science Rice University johnc@cs.rice.edu COMP 528 Lecture 20 31 March 2005 Topics for Today Model verification Model validation Transient

More information

OBJECTIVES INTRODUCTION

OBJECTIVES INTRODUCTION M7 Chapter 3 Section 1 OBJECTIVES Suarize data using easures of central tendency, such as the ean, edian, ode, and idrange. Describe data using the easures of variation, such as the range, variance, and

More information

1 Bounding the Margin

1 Bounding the Margin COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Comparison of Stability of Selected Numerical Methods for Solving Stiff Semi- Linear Differential Equations

Comparison of Stability of Selected Numerical Methods for Solving Stiff Semi- Linear Differential Equations International Journal of Applied Science and Technology Vol. 7, No. 3, Septeber 217 Coparison of Stability of Selected Nuerical Methods for Solving Stiff Sei- Linear Differential Equations Kwaku Darkwah

More information

A Note on the Applied Use of MDL Approximations

A Note on the Applied Use of MDL Approximations A Note on the Applied Use of MDL Approxiations Daniel J. Navarro Departent of Psychology Ohio State University Abstract An applied proble is discussed in which two nested psychological odels of retention

More information

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS Jochen Till, Sebastian Engell, Sebastian Panek, and Olaf Stursberg Process Control Lab (CT-AST), University of Dortund,

More information

Ph 20.3 Numerical Solution of Ordinary Differential Equations

Ph 20.3 Numerical Solution of Ordinary Differential Equations Ph 20.3 Nuerical Solution of Ordinary Differential Equations Due: Week 5 -v20170314- This Assignent So far, your assignents have tried to failiarize you with the hardware and software in the Physics Coputing

More information

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science Proceedings of the 6th WSEAS International Conference on Applied Coputer Science, Tenerife, Canary Islands, Spain, Deceber 16-18, 2006 183 Qualitative Modelling of Tie Series Using Self-Organizing Maps:

More information

E. Alpaydın AERFAISS

E. Alpaydın AERFAISS E. Alpaydın AERFAISS 00 Introduction Questions: Is the error rate of y classifier less than %? Is k-nn ore accurate than MLP? Does having PCA before iprove accuracy? Which kernel leads to highest accuracy

More information

Testing equality of variances for multiple univariate normal populations

Testing equality of variances for multiple univariate normal populations University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Inforation Sciences 0 esting equality of variances for ultiple univariate

More information

Stochastic Subgradient Methods

Stochastic Subgradient Methods Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

Polygonal Designs: Existence and Construction

Polygonal Designs: Existence and Construction Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G

More information

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,

More information

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs BULLETIN OF THE POLISH ACADEMY OF SCIENCES TECHNICAL SCIENCES Vol. 57, No. 3, 2009 Algoriths for parallel processor scheduling with distinct due windows and unit-tie obs A. JANIAK 1, W.A. JANIAK 2, and

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

The Weierstrass Approximation Theorem

The Weierstrass Approximation Theorem 36 The Weierstrass Approxiation Theore Recall that the fundaental idea underlying the construction of the real nubers is approxiation by the sipler rational nubers. Firstly, nubers are often deterined

More information

Graphical Models in Local, Asymmetric Multi-Agent Markov Decision Processes

Graphical Models in Local, Asymmetric Multi-Agent Markov Decision Processes Graphical Models in Local, Asyetric Multi-Agent Markov Decision Processes Ditri Dolgov and Edund Durfee Departent of Electrical Engineering and Coputer Science University of Michigan Ann Arbor, MI 48109

More information

Best Arm Identification: A Unified Approach to Fixed Budget and Fixed Confidence

Best Arm Identification: A Unified Approach to Fixed Budget and Fixed Confidence Best Ar Identification: A Unified Approach to Fixed Budget and Fixed Confidence Victor Gabillon Mohaad Ghavazadeh Alessandro Lazaric INRIA Lille - Nord Europe, Tea SequeL {victor.gabillon,ohaad.ghavazadeh,alessandro.lazaric}@inria.fr

More information

Estimation of Korean Monthly GDP with Mixed-Frequency Data using an Unobserved Component Error Correction Model

Estimation of Korean Monthly GDP with Mixed-Frequency Data using an Unobserved Component Error Correction Model 100Econoic Papers Vol.11 No.1 Estiation of Korean Monthly GDP with Mixed-Frequency Data using an Unobserved Coponent Error Correction Model Ki-Ho Ki* Abstract Since GDP is announced on a quarterly basis,

More information

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space Journal of Machine Learning Research 3 (2003) 1333-1356 Subitted 5/02; Published 3/03 Grafting: Fast, Increental Feature Selection by Gradient Descent in Function Space Sion Perkins Space and Reote Sensing

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory Proble sets 5 and 6 Due: Noveber th Please send your solutions to learning-subissions@ttic.edu Notations/Definitions Recall the definition of saple based Radeacher

More information

Forecasting Financial Indices: The Baltic Dry Indices

Forecasting Financial Indices: The Baltic Dry Indices International Journal of Maritie, Trade & Econoic Issues pp. 109-130 Volue I, Issue (1), 2013 Forecasting Financial Indices: The Baltic Dry Indices Eleftherios I. Thalassinos 1, Mike P. Hanias 2, Panayiotis

More information

Introduction to Machine Learning. Recitation 11

Introduction to Machine Learning. Recitation 11 Introduction to Machine Learning Lecturer: Regev Schweiger Recitation Fall Seester Scribe: Regev Schweiger. Kernel Ridge Regression We now take on the task of kernel-izing ridge regression. Let x,...,

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

Ufuk Demirci* and Feza Kerestecioglu**

Ufuk Demirci* and Feza Kerestecioglu** 1 INDIRECT ADAPTIVE CONTROL OF MISSILES Ufuk Deirci* and Feza Kerestecioglu** *Turkish Navy Guided Missile Test Station, Beykoz, Istanbul, TURKEY **Departent of Electrical and Electronics Engineering,

More information

Support Vector Machines. Maximizing the Margin

Support Vector Machines. Maximizing the Margin Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the

More information

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

Topic 5a Introduction to Curve Fitting & Linear Regression

Topic 5a Introduction to Curve Fitting & Linear Regression /7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline

More information

Distributed Subgradient Methods for Multi-agent Optimization

Distributed Subgradient Methods for Multi-agent Optimization 1 Distributed Subgradient Methods for Multi-agent Optiization Angelia Nedić and Asuan Ozdaglar October 29, 2007 Abstract We study a distributed coputation odel for optiizing a su of convex objective functions

More information

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter

More information

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x)

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x) 7Applying Nelder Mead s Optiization Algorith APPLYING NELDER MEAD S OPTIMIZATION ALGORITHM FOR MULTIPLE GLOBAL MINIMA Abstract Ştefan ŞTEFĂNESCU * The iterative deterinistic optiization ethod could not

More information

TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES

TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES S. E. Ahed, R. J. Tokins and A. I. Volodin Departent of Matheatics and Statistics University of Regina Regina,

More information

ZISC Neural Network Base Indicator for Classification Complexity Estimation

ZISC Neural Network Base Indicator for Classification Complexity Estimation ZISC Neural Network Base Indicator for Classification Coplexity Estiation Ivan Budnyk, Abdennasser Сhebira and Kurosh Madani Iages, Signals and Intelligent Systes Laboratory (LISSI / EA 3956) PARIS XII

More information

A Jackknife Correction to a Test for Cointegration Rank

A Jackknife Correction to a Test for Cointegration Rank Econoetrics 205, 3, 355-375; doi:0.3390/econoetrics3020355 OPEN ACCESS econoetrics ISSN 2225-46 www.dpi.co/journal/econoetrics Article A Jackknife Correction to a Test for Cointegration Rank Marcus J.

More information

Feedforward Networks

Feedforward Networks Feedforward Neural Networks - Backpropagation Feedforward Networks Gradient Descent Learning and Backpropagation CPSC 533 Fall 2003 Christian Jacob Dept.of Coputer Science,University of Calgary Feedforward

More information

Inflation Forecasts: An Empirical Re-examination. Swarna B. Dutt University of West Georgia. Dipak Ghosh Emporia State University

Inflation Forecasts: An Empirical Re-examination. Swarna B. Dutt University of West Georgia. Dipak Ghosh Emporia State University Southwest Business and Econoics Journal/2006-2007 Inflation Forecasts: An Epirical Re-exaination Swarna B. Dutt University of West Georgia Dipak Ghosh Eporia State University Abstract Inflation forecasts

More information

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters journal of ultivariate analysis 58, 96106 (1996) article no. 0041 The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Paraeters H. S. Steyn

More information

Multi-view Discriminative Manifold Embedding for Pattern Classification

Multi-view Discriminative Manifold Embedding for Pattern Classification Multi-view Discriinative Manifold Ebedding for Pattern Classification X. Wang Departen of Inforation Zhenghzou 450053, China Y. Guo Departent of Digestive Zhengzhou 450053, China Z. Wang Henan University

More information

Feedforward Networks

Feedforward Networks Feedforward Networks Gradient Descent Learning and Backpropagation Christian Jacob CPSC 433 Christian Jacob Dept.of Coputer Science,University of Calgary CPSC 433 - Feedforward Networks 2 Adaptive "Prograing"

More information

Nonlinear Log-Periodogram Regression for Perturbed Fractional Processes

Nonlinear Log-Periodogram Regression for Perturbed Fractional Processes Nonlinear Log-Periodogra Regression for Perturbed Fractional Processes Yixiao Sun Departent of Econoics Yale University Peter C. B. Phillips Cowles Foundation for Research in Econoics Yale University First

More information

NBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University

NBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University NBN Algorith Bogdan M. Wilaoswki Auburn University Hao Yu Auburn University Nicholas Cotton Auburn University. Introduction. -. Coputational Fundaentals - Definition of Basic Concepts in Neural Network

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Soft-margin SVM can address linearly separable problems with outliers

Soft-margin SVM can address linearly separable problems with outliers Non-linear Support Vector Machines Non-linearly separable probles Hard-argin SVM can address linearly separable probles Soft-argin SVM can address linearly separable probles with outliers Non-linearly

More information

Lost-Sales Problems with Stochastic Lead Times: Convexity Results for Base-Stock Policies

Lost-Sales Problems with Stochastic Lead Times: Convexity Results for Base-Stock Policies OPERATIONS RESEARCH Vol. 52, No. 5, Septeber October 2004, pp. 795 803 issn 0030-364X eissn 1526-5463 04 5205 0795 infors doi 10.1287/opre.1040.0130 2004 INFORMS TECHNICAL NOTE Lost-Sales Probles with

More information

C na (1) a=l. c = CO + Clm + CZ TWO-STAGE SAMPLE DESIGN WITH SMALL CLUSTERS. 1. Introduction

C na (1) a=l. c = CO + Clm + CZ TWO-STAGE SAMPLE DESIGN WITH SMALL CLUSTERS. 1. Introduction TWO-STGE SMPLE DESIGN WITH SMLL CLUSTERS Robert G. Clark and David G. Steel School of Matheatics and pplied Statistics, University of Wollongong, NSW 5 ustralia. (robert.clark@abs.gov.au) Key Words: saple

More information

Deflation of the I-O Series Some Technical Aspects. Giorgio Rampa University of Genoa April 2007

Deflation of the I-O Series Some Technical Aspects. Giorgio Rampa University of Genoa April 2007 Deflation of the I-O Series 1959-2. Soe Technical Aspects Giorgio Rapa University of Genoa g.rapa@unige.it April 27 1. Introduction The nuber of sectors is 42 for the period 1965-2 and 38 for the initial

More information

Predicting FTSE 100 Close Price Using Hybrid Model

Predicting FTSE 100 Close Price Using Hybrid Model SAI Intelligent Systes Conference 2015 Noveber 10-11, 2015 London, UK Predicting FTSE 100 Close Price Using Hybrid Model Bashar Al-hnaity, Departent of Electronic and Coputer Engineering, Brunel University

More information

A Theoretical Analysis of a Warm Start Technique

A Theoretical Analysis of a Warm Start Technique A Theoretical Analysis of a War Start Technique Martin A. Zinkevich Yahoo! Labs 701 First Avenue Sunnyvale, CA Abstract Batch gradient descent looks at every data point for every step, which is wasteful

More information

EE5900 Spring Lecture 4 IC interconnect modeling methods Zhuo Feng

EE5900 Spring Lecture 4 IC interconnect modeling methods Zhuo Feng EE59 Spring Parallel LSI AD Algoriths Lecture I interconnect odeling ethods Zhuo Feng. Z. Feng MTU EE59 So far we ve considered only tie doain analyses We ll soon see that it is soeties preferable to odel

More information

Figure 1: Equivalent electric (RC) circuit of a neurons membrane

Figure 1: Equivalent electric (RC) circuit of a neurons membrane Exercise: Leaky integrate and fire odel of neural spike generation This exercise investigates a siplified odel of how neurons spike in response to current inputs, one of the ost fundaental properties of

More information

Feedforward Networks. Gradient Descent Learning and Backpropagation. Christian Jacob. CPSC 533 Winter 2004

Feedforward Networks. Gradient Descent Learning and Backpropagation. Christian Jacob. CPSC 533 Winter 2004 Feedforward Networks Gradient Descent Learning and Backpropagation Christian Jacob CPSC 533 Winter 2004 Christian Jacob Dept.of Coputer Science,University of Calgary 2 05-2-Backprop-print.nb Adaptive "Prograing"

More information

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Proc. of the IEEE/OES Seventh Working Conference on Current Measureent Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Belinda Lipa Codar Ocean Sensors 15 La Sandra Way, Portola Valley, CA 98 blipa@pogo.co

More information

ANALYSIS OF HALL-EFFECT THRUSTERS AND ION ENGINES FOR EARTH-TO-MOON TRANSFER

ANALYSIS OF HALL-EFFECT THRUSTERS AND ION ENGINES FOR EARTH-TO-MOON TRANSFER IEPC 003-0034 ANALYSIS OF HALL-EFFECT THRUSTERS AND ION ENGINES FOR EARTH-TO-MOON TRANSFER A. Bober, M. Guelan Asher Space Research Institute, Technion-Israel Institute of Technology, 3000 Haifa, Israel

More information

MSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE

MSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE Proceeding of the ASME 9 International Manufacturing Science and Engineering Conference MSEC9 October 4-7, 9, West Lafayette, Indiana, USA MSEC9-8466 MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL

More information