Robust Estimator with the SCAD Function in Penalized Linear Regression

Size: px
Start display at page:

Download "Robust Estimator with the SCAD Function in Penalized Linear Regression"

Transcription

1 The SIJ Trasactios o Computer Sciece Egieerig & its Applicatios (CSEA), Vol. 2, No. 4, Jue 24 Robust Estimator with the SCAD Fuctio i Pealize Liear Regressio Kag-Mo Jug* *Professor, Departmet of Statistics a Computer Sciece, Kusa Natioal Uiversity, Kusa, Chobuk, SOUTH KOREA. kmjug@kusa.ac.kr Abstract Pealize regressio proceures have recetly see a lot of attetio, because it ca yiel both estimatio a variable selectio simultaeously. There are icreasig applicatios i bioiformatics research which treats large umber of variables. However, their performace ca be severely ifluece by outliers i either the respose or the covariate space. This paper proposes a weighte regressio estimator with the Smoothly Clippe Absolute Deviatio (SCAD) fuctio, because it has two avatages, sparse system a ubiaseess for large coefficiets. It eals with robust variable selectio a robust estimatio. We evelop a uifie algorithm for the propose estimator icluig the SCAD estimate a the tuig parameter, base o the Local Quaratic Approximatio (LQA) a the Local Liear Approximatio (LLA) of the o-covex SCAD pealty fuctio. We compare the robustess of the propose algorithm with other pealize regressio estimators. Numerical simulatio results show that the propose estimator is effective to aalyze a cotamiate ata. Keywors Liear Regressio; Local Liear Approximatio; Local Quaratic Approximatio; Robust; Pealize Fuctio; Smoothly Clippe Absolute Deviatio; Tuig Parameter; Weight. Abbreviatios Akaike Iformatio Criterio (AIC); Bayesia Iformatio Criterio (BIC); Least Absolute Deviatio (LAD); LAD a the L -type pealty (LAD-L); LAD a the SCAD pealty (LAD-SCAD); Least Absolute Shrikage a Selectio Operator (LASSO); Local Liear Approximatio (LLA); Local Quaratic Approximatio (LQA); Least Squares Estimator (LSE); Smoothly Clippe Absolute Deviatio (SCAD); the Weighte LAD a the SCAD pealty (WLAD-SCAD). I. INTRODUCTION RECENT ays we ca obtai easily large sample high imesioal ata sets from cheap sesors. However the growth of variables ca prevet us from costructig a parsimoious moel which provies goo iterpretatio about the system. Oe importat stream of statistical research requires effective variable selectio proceures to improve both accuracy a iterpretability of the learig techique [Kitter, 986]. Variable selectio is a importat research topic i liear regressio especially for moel selectio i high-imesioal ata situatio [Jug, 28]. Tibshirai (996) propose the least absolute shrikage a selectio operator (LASSO), which ca simultaeously select valuable covariates a estimate regressio parameters. Traitioal moel selectio criteria such as Akaike Iformatio Criterio (AIC) [Akaike, 973] a Bayesia Iformatio Criterio (BIC) [Schwarz, 978] have major rawbacks that parameter estimatio a moel selectio are two separate processes. The LASSO is a regularisatio with l -type pealizatio a it becomes extremely popular, because it shriks the regressio coefficiets towar zero with the possibility of settig some coefficiets equal to zero, resultig i a simultaeous estimatio a variable selectio. May literatures show successful applicatios usig the LASSO. However, the LASSO ca be biase for the coefficiets whose absolute values are large. Fa & Li (2) propose a pealize regressio with the Smoothly Clippe Absolute Deviatio (SCAD) pealty fuctio a showe that it has better theoretical properties tha the LASSO with L -type pealty. The pealize regressio with SCAD ot oly selects importat covariates cosistetly but also prouces parameter estimators as efficiet as if the true moel were kow, i.e., the oracle property. The LASSO oes ot satisfy the oracle property. The SCAD fuctio is a o-covex pealty fuctio to make up for the eficiecies of the LASSO. The pealize regressio cosists of a loss fuctio a a pealty fuctio. I traitioal regressio settig it is well kow that the least squares metho is sesitive to eve sigle outlier. Alterative to the least squares metho is the Least Absolute Deviatio (LAD) metho. Wag et al., (27) evelope a robust algorithm with the LAD loss fuctio a L -type pealty (LAD-L). They showe that the LAD-L is resistat to o-ormal error istributios a outliers. Jug (27, 22) propose a robust metho with the LAD loss fuctio a the SCAD pealty (LAD-SCAD) estimate. It shows that the SCAD pealty fuctio is more efficiet tha the LAD-L. ISSN: Publishe by The Staar Iteratioal Jourals (The SIJ) 56

2 The SIJ Trasactios o Computer Sciece Egieerig & its Applicatios (CSEA), Vol. 2, No. 4, Jue 24 Recetly statisticias ofte treat ata sets with a oormal respose variable or covariates that may cotai multiple outliers or leverage poits. Eve though the LAD is more robust tha the least squares metho, the uboue loss fuctio of the LAD affects strogly the LAD estimator. I this paper we cosier a weight metho for the boue loss fuctio. The weight i our algorithm atteuates the ifluece of outliers o the estimator a works the same ifluece of o-outliers as the u-weighte LAD-SCAD metho. The weighte LAD loss fuctio with the SCAD pealty fuctio (WLAD-SCAD) improves the error performace of the LAD-SCAD. The propose metho combies the robustess of the weighte LAD a the oracle property of the SCAD pealty. The tuig parameter cotrols the moel complexity a plays a importat role i the variable selectio proceure. We propose a BIC-type tuig parameter selector usig a ata-rive metho, the WLAD- SCAD with the BIC tuig parameter ca ietify the most parsimoious correct moel. The paper is orgaize as follows. Sectio 2 escribes the previously relate works. Sectio 3 provies our propose algorithm of the weighte LAD with the SCAD pealty fuctio. Sice the SCAD fuctio is ot covex, we use a approximatio algorithm to solve the o-ifferetiable a o-covex objective fuctio i pealize regressio with the SCAD pealty. We use approximatio methos such as the Local Quaratic Approximatio (LQA) a the Local Liear Approximatio (LLA) to solve the optimizatio problem for the o-covex SCAD pealty fuctio. We provie two results, the solver of a Newto-Raphso solutio a the LAD criterio. Sectio 4 illustrates simulatio results. It shows that the propose algorithm has superior to other methos from the view of robustess a parsimoious moel. II. RELATED WORKS Cosier the liear regressio moel y i = α + x T i β + ε i, i =,,, where x i = (x i,, x i ) T is the -imesioal covariate, β = (β,, β ) T, is the umber of covariates a is the umber of observatios. Let y = (y,, y ) T a let X be a T ( + ) matrix whose i-th row, x i. The Least Squares Estimator (LSE) (X T X) X T y which miimizes the sum of squares of resiuals ca be istorte by the heay-taile probability istributio of errors or eve sigle outlier [Rousseeuw & Leroy, 987]. The criterio ca be writte by ( y i α + x T i β ) 2. Oe alterative to the least squares metho is the LAD metho which miimises the criterio fuctio, the sum of absolute eviatios of the errors y i ( α + x T i β). The major avatage of the LAD metho lies i its robustess relative to the LSE. The LAD estimates are less affecte by the presece of a few outliers or ifluetial observatios. However both the LSE a the LAD caot be useful for moel selectio whe especially the umber of covariates is very large. Tibshirai (996) propose a pealty base o the L orm for automatically eletig uecessary covariates. The LASSO criterio is a simply pealize least squares with the L pealty ( y i α + x T i β ) 2 + λ, () where λ > is the tuig parameter which cotrols the traeoff betwee moel fittig a moel sparsity. Whe the tuig parameter is large, the criterio focuses o moel sparsity. Traitioally i moel selectio, crossvaliatio a iformatio criteria-icluig the AIC [Akaike, 973] a BIC [Schwarz, 978]-are wiely applie. Shao (997) showe that the BIC ca ietify the true moel cosistetly i liear regressio with fixe imesioal covariates, but the AIC may fail ue to over-fittig. Yag (25) showe that cross-valiatio is asymptotically equivalet to the AIC [Yag, 25] a so they behave similarly. Leg et al., (26) showe that the LASSO is ot asymptotically cosistet a so the LASSO ca be biase for large coefficiets. Fa & Li (2) aresse this problem a propose the SCAD pealty fuctio. They escribe the coitios of a goo pealty fuctio (a) ubiaseess: the resultig estimator is early ubiase whe the true ukow parameter is large; (b) sparsity: the resultig estimator is a thresholig rule, which automatically sets small estimate coefficiets to be zero; (c) cotiuity: the resultig estimator is cotiuous i the ata. The LSE with the SCAD pealty fuctio miimizes the criterio fuctio ( y i α + x T i β ) 2 + p λ ( ), (2) where λ β, if β < λ a 2 λ 2 ( β aλ) 2, p λ ( β ) = 2(a ) if λ β < aλ 2 (a + )2 λ 2, a so its erivative becomes if β > aλ λ, if β < λ p λ β = aλ β, a if λ β < aλ, if β > aλ where a ca be chose usig cross-valiatio or geeralize cross-valiatio. However, the simulatio of Fa & Li (2) gives us a = 3.7 which is approximately optimal. I this article we set a = 3.7. Similar to that the LAD is more robust tha the LSE i o-pealize regressio moel, Wag et al., (27) propose the LAD with the SCAD pealty which miimises the criterio fuctio ISSN: Publishe by The Staar Iteratioal Jourals (The SIJ) 57

3 The SIJ Trasactios o Computer Sciece Egieerig & its Applicatios (CSEA), Vol. 2, No. 4, Jue 24 y i α + x T i β + p λ ( ). (3) Eve though the LAD is robust, its breakow poit is also / which is equivalet to the LSE. Jug (22) propose the weighte LAD with the SCAD pealty, because the simulatio results of Giloi et al., (26) show that i o-pealize liear regressio the performace of the weighte LAD estimator is competitive with that of high breakow regressio estimators, particularly i the presece of outliers locate at leverage poits. Jug (2) propose a weighte LAD pealize estimator which combies the weighte LAD estimator with the L pealty. Jug (22) use the criterio fuctio y i α + x T i β + p λ ( ), (4) where the weight epes o the space of covariates. We call the solutio of (4) the weighte LAD with the SCAD (WLAD-SCAD). The propose metho uses the weight for robustess which is resistat to leverage poits or ifluetial observatios, because the weight reuces the effects of the observatios havig large eviatios or the leverage poits. The objective fuctio will give the robustess of weight methos a the avatages of the SCAD pealize fuctio. III. METHODS The solutio of (4) ca be obtaie by a staar optimizatio program if the criterio fuctio is covex a ifferetiable. Ufortuately the absolute fuctio i (4) is ot ifferetiable at zero a the SCAD pealty fuctio is ot covex i β. Approximatio of the absolute fuctio a the SCAD fuctio trasforms the objective fuctio ito liear equatios a so we ca obtai efficietly a iterative solutio of the WLAD-SCAD estimator. The absolute fuctio u u2 ear u gives + 2 u 2 u y i α + x i T β (y i α + x i T β ) 2 2 y i α + x i T β + 2 y i α + x i T β, for ozero u for iitial values α, β ear the miimisatio of (3). Also the Taylor expasio at the o-zero yiels p λ p λ + p λ β 2 2 j, (6) 2 a we set = if is ear zero. Assume that the loglikelihoo fuctio is smooth with respect to β a its first two partial erivatives are cotiuous. The liear quaratic approximatio (LQA) is a moificatio of the Newto- Raphso algorithm [Fa & Li, 2]. The LQA is broaly useful for solvig optimizatio problems with the oifferetiable criterio fuctio. The the criterio fuctio (4) becomes (5) (y i α + x i T β ) 2 y i α + x i T β + [ p λ y i α + x i T β + p λ β 2 2 j ]. 2 a we obtai the criterio fuctio with up to costats (y X β)t W y X β + 2 βt Qβ, (8) where iag( p λ β = (α, β T ) T, W = iag y i α +x i T β (7), Q = ) a X is the ( + ) ata matrix whose first colum vector is the vector of oes with legth. Thus the Newto-Raphso solutio yiels the iterative solutio of (8) β (l+) = (X T W l X + Q (l) ) X T W l y, (9) for the lth solutio β (l), the matrix Q l a the lth weight matrix W l. Whe W = I a Q l = Q, the solutio (2) reuces to the LSE-SCAD estimate which is the solutio of the criterio (3). For the case the lth weight matrix W l = iag iag( p λ y i α +x i T β a the matrix Q l = ), the solutio of (9) becomes the LAD- SCAD. A for the lth weight matrix W l = iag iag( p λ y i α +x i T β a the matrix Q l = ), the solutio of (9) is calle the WLAD- SCAD. I Sectio 4 the iteratio stops whe the maximum ifferece amog betwee the previous a the curret solutio is less tha 4. We will provie aother approach to solve the miimizatio of the criterio fuctio (4). By the Taylor expasio of p λ ( ) we obtai the relatioship as follows p λ p λ + p λ, for β () j. The costat terms i () ca ot affect the fuctio of (3) or (4) a so they ca be elimiate. The criterio fuctio (4) ca be writte by y i α + x T i β + p λ ( β j ), () whose regularizatio part is the liearize SCAD pealty fuctio [Zou & Li, 28]. It is a local liear approximatio (LLA) metho for the SCAD pealty fuctio. Computatioally, it is very easy to fi the solutio of (). Specifically, we ca costruct the augmete ata set { y i, x i } as ISSN: Publishe by The Staar Iteratioal Jourals (The SIJ) 58

4 The SIJ Trasactios o Computer Sciece Egieerig & its Applicatios (CSEA), Vol. 2, No. 4, Jue 24 y i, x i = y i, x i for i =,,, p λ β i e i for i = +,, + p, where e j is the uit vector havig except the jth elemet oe [Wag et al., 27]. The the WLAD-SCAD estimator of () ca be obtaie by miimisig the criterio fuctio +p y T i (α + x i β). (2) Cosequetly we ca use a staar LAD program (the fuctio rq i R program) without computatioal effort. To fi a goo tuig parameter is a importat proceure i pealize estimatio methos. Fa & Li (2) fou the values of tuig parameters by optimizig the performace via cross-valiatio a geeralize cross valiatio. Zou (26) i LASSO use the tuig parameters set by the reciprocal of the absolute value of the LSE. Wag et al., (27) use the tuig parameter by miimisig a BICtype criterio fuctio. I this paper we propose the tuig parameter which ca be obtaie by miimizig (y i α + x T i β λ ) 2 / GCV λ = ( f(λ)/) 2, Where f λ = tr[ X(X T X + Q) X T ], X is the ata matrix excluig the costat term, a Q = iag( p λ β λj ). For liear moels, the geeralize cross-valiatio is asymptotically equivalet to the Mallows C p, AIC a leaveoe-out cross valiatio [Hastie et al., 2]. IV. SIMULATION RESULTS This sectio emostrates simulatios i various situatios to show the robustess of the metho propose i Sectio III. We umerically compare the propose metho WLAD- SCAD estimates with the LASSO estimates [Thibshirai, 996], the LSE with the SCAD pealize fuctio [Fa & Li, 2] a the LAD-SCAD estimate [Jug, 27]. All simulatios are performe by R program. We cosier the liear regressio moel y i = x T i β + σε i, i =,,, Where β T = (3,,,, 2,,) a ε i are staar ormally istribute. The covariate vector follows the multivariate ormal istributio with the zero mea vector a the correlatio matrix whose (j, k)th elemet is.5 j k for j, k =,,7. This is a similar to Tibshirai (996) a Fa & Li (2). We set several situatios. The sample sizes are give by = 2, 4,. Two ifferet values for the error staar eviatio σ are give by σ =, 2. The error istributios are the staar ormal istributio, the ouble expoetial istributio a the t istributio with egrees of 2. The last two istributios have thick-taile probability istributio fuctios. To show the robustess of our propose algorithm we cosier the cotamiate ata with leverage poits about 2%. β λj The simulatio ata cosist of the traiig ata a iepeet test ata. The regressio coefficiets ca be obtaie o the traiig ata, a the performace is evaluate o the test ata of sample size geeratig from the ormal istributio with the above efie matrix. We coucte 2 simulatio iteratios to evaluate the performace of the fitess of the algorithm, which ca be calculate by the Average of the Mea Absolute Deviatios () o the test ata. The performace of the sparseess of the algorithm ca be evaluate by the average umber of correctly estimate zero coefficiets which is the colum labelle i the tables. Aalogously the colum labelle Icorrect eotes the umber of zero estimates which are ot zero coefficiets, which meas the iaccuracy for each algorithm. The umber i the parethesis is the sample staar eviatio for each algorithm. Thus the moel is best i case the lesser of the a the Icorrect terms. For the true moel the umber of o-zero coefficiets is 5, a thus the term shoul be close to 5 if the algorithm is compatible. Table summarize the simulatio results for the ata without outliers. We cosiere three error moels, a summarize oly the results for the ouble expoetial error istributio. It shows that eve though LASSO is the best moel i the sparseess of the moel, it has large moel errors. As expecte the staar eviatio of the values become larger whe σ becomes larger. Whe, the values of the WLAD-SCAD is little large tha that of the LASSO. It meas that the propose estimator yiels a sparse moel whe the error istributio is ot ormal. Table : Simulatio Results with Double Expoetial Errors without Outliers DE LAD Methos LASSO SCAD WLAD Errors -SCAD -SCAD (.94) (.85) (.77) (.78) = σ = (.43) (.66) (.76) (.75).... Icorrect (.) (.) (.) (.) (.354) (.346) (.364) (.393) = (.846) (.845) (.95) (.838) Icorrect (.549) (.538) (.57) (.54) Table 2 summarize the simulatio results for the staar ormal errors a 2% leverage poits. Seeig term implies that i moel sparsity the WLAD- SCAD is the best amog the total estimators. The ifferece betwee the WLAD-SCAD a other estimator is meaigful eve the icorrectess of the propose metho is somewhat large. Table 2: Simulatio Results with the Normal Errors with 2% Outliers Normal Methos LASSO SCAD LAD- WLAD ISSN: Publishe by The Staar Iteratioal Jourals (The SIJ) 59

5 The SIJ Trasactios o Computer Sciece Egieerig & its Applicatios (CSEA), Vol. 2, No. 4, Jue 24 Errors SCAD -SCAD (.25) (.88) (.239) (.29) = σ = (.48) (.842) (.966) (.929) Icorrect (.749) (.6) (.77) (.742) (.226) (.98) (.257) (.25) = (.45) (.837) (.995) (.868) Icorrect (.83) (.639) (.724) (.728) Table 3 summarize the simulatio results for the ouble expoetial error istributio a 2% leverage poits. The WLAD-SCAD is the best algorithm amog 4 estimators i the view of variable selectio, because the value of the WLAD-SCAD is the closest to 5 for all estimators. Table 3 shows that i moel error the WLAD-SCAD is the most efficiet estimatio regarless of outliers a the sprea of errors a the sample size, sice we cosiere the estimatio metho which reuces the ifluece of leverage poits. Table 3: Simulatio Results with the Double Expoetial Errors with 2% Outliers Normal LAD- WLAD Methos LASSO SCAD Errors SCAD -SCAD (.225) (.26) (.278) (.22) = σ = (.68) (.86) (.97) (.878) Icorrect (.77) (.598) (.642) (.73) (.) (.98) (.29) (.29) = σ = (.58) (.824) (.884) (.582) Icorrect (.657) (.532) (.63) (.5) (.272) (.255) (.34) (.88) = (.7) (.85) (.) (.77) Icorrect (.82) (.636) (.742) (.628) V. CONCLUSION I this paper we propose a robust algorithm for the pealize regressio moel with the LAD loss fuctio a the SCAD pealty fuctio. We use a weight fuctio for the robust loss fuctio a improve the effectiveess of the propose algorithm through umerical simulatios. We erive two approximatios for objective fuctios to treat the o-covex optimizatio problem. Oe is LLA a the other is LQA. Sice the former is liear, it is so easy to implemet it. Both two methos are robust to outliers or ifluetial observatios. The umerical simulatios show that the propose metho is more robust tha other methos from the poit view of fiig exact o-zero coefficiets. Thus the propose metho gives a metho of variable selectio from thousas of iput variables appeare i the applicatios of biometrical experimets. For the further stuy we will cosier the Huber fuctio for the loss fuctio with the SCAD pealty fuctio. ACKNOWLEDGEMENTS This research was supporte by Basic Sciece Research Program through the Natioal Research Fouatio of Korea (NRF) fue by the Miistry of Eucatio, Sciece a Techology (NRF-22RAA4A4594). REFERENCES [] H. Akaike (973), Iformatio Theory a a Extesio of the Maximum Likelihoo Priciple, Proceeigs of the Seco Iteratioal Symposium o Iformatio Theory, Eitors: B.N. Petrov & F. Csàki, Akaemiai Kiao, Buapest. [2] G. Schwarz (978), Estimatig the Dimesio of a Moel, Aals of Statistics, Vol. 6, Pp [3] J. Kitter (986), Feature Selectio a Extractio, Habook of Patter Recogitio a Image Processig, Eitors: T.Y. Youg a K.-S. Fu, Acaemic Press, New York. [4] P.J. Rousseeuw & A.M. Leroy (987), Robust Regressio a Outlier Detectio, Joh Wiley, New York. [5] R. Tibshirai (996), Regressio Shrikage a Selectio via the LASSO, Joural of the Royal Statistical Society, Series B, Vol. 58, Pp [6] J. Shao (997), A Asymptotic Theory for Liear Moel Selectio, Statistical Siica, Vol. 7, Pp [7] T. Hastie, R. Tibshirai & J. Friema (2), The Elemets of Statistical Learig: Data Miig, Iferece, a Preictio, Spriger, New York. [8] J. Fa & R. Li (2), Variable Selectio via Nococave Pealize Likelihoo a its Oracle Properties, Joural of the America Statistical Associatio, Vol. 96, Pp [9] Y. Yag (25), Ca the Stregths of AIC a BIC be Share? A Coflict betwee Moel Ietificatio a Regressio Estimatio, Biometrika, Vol. 92, Pp [] H. Zou (26), The Aaptive Lasso a its Oracle Properties, Joural of the America Statistical Associatio, Vol., Pp [] C. Leg, Y. Li & G. Wahba (26), A Note o the LASSO a Relate Proceures i Moel Selectio, Statistical Siica, Vol. 6, Pp [2] A. Giloi, J.S. Simooff & B. Segupta (26), Robust Weighte LAD Regressio, Computatioal Statistics a Data Aalysis, Vol. 5, Pp [3] H. Wag, G. Li & G. Jiag (27), Robust Regressio Shrikage a Cosistet Variable Selectio through the LAD- Lasso, Joural of Busiess & Ecoomic Statistics, Vol. 25, Pp [4] K.-M. Jug (27), A Robust Estimator i Rige Regressio, Joural of the Korea Data Aalysis Society, Vol. 9, Pp [5] K.-M. Jug (28), Robust Statistical Methos i Variable Selectio, Joural of the Korea Data Aalysis Society, Vol., Pp [6] H. Zou & R. Li (28), Oe-step Sparse Estimates i Nococave Pealize Likelihoo Moels, Aals of Statistics, Vol. 36, Pp [7] K.-M. Jug (2), Weighte Least Absolute Deviatio Lasso Estimator, Commuicatios of the Korea Statistical Society, Vol. 8, Pp [8] K.-M. Jug (22), Weighte Least Absolute Deviatio Regressio Estimator with the SCAD Fuctio, Joural of the Korea Data Aalysis Society, Vol. 4, Pp ISSN: Publishe by The Staar Iteratioal Jourals (The SIJ) 6

Robust Algorithm for Multiclass Weighted Support Vector Machine

Robust Algorithm for Multiclass Weighted Support Vector Machine he SIJ rasactios o Iustrial Fiacial & Busiess Maagemet (IFBM) Vol. 2 No. 3 May 24 Robust Algorithm for Multiclass Weighte Support Vector Machie ag-mo Jug* *Professor Departmet of Statistics a Computer

More information

Lecture 24: Variable selection in linear models

Lecture 24: Variable selection in linear models Lecture 24: Variable selectio i liear models Cosider liear model X = Z β + ε, β R p ad Varε = σ 2 I. Like the LSE, the ridge regressio estimator does ot give 0 estimate to a compoet of β eve if that compoet

More information

A COMPUTATIONAL STUDY UPON THE BURR 2-DIMENSIONAL DISTRIBUTION

A COMPUTATIONAL STUDY UPON THE BURR 2-DIMENSIONAL DISTRIBUTION TOME VI (year 8), FASCICULE 1, (ISSN 1584 665) A COMPUTATIONAL STUDY UPON THE BURR -DIMENSIONAL DISTRIBUTION MAKSAY Ştefa, BISTRIAN Diaa Alia Uiversity Politehica Timisoara, Faculty of Egieerig Hueoara

More information

A Note on Adaptive Group Lasso

A Note on Adaptive Group Lasso A Note o Adaptive Group Lasso Hasheg Wag ad Chelei Leg Pekig Uiversity & Natioal Uiversity of Sigapore July 7, 2006. Abstract Group lasso is a atural extesio of lasso ad selects variables i a grouped maer.

More information

Lecture 6 Testing Nonlinear Restrictions 1. The previous lectures prepare us for the tests of nonlinear restrictions of the form:

Lecture 6 Testing Nonlinear Restrictions 1. The previous lectures prepare us for the tests of nonlinear restrictions of the form: Eco 75 Lecture 6 Testig Noliear Restrictios The previous lectures prepare us for the tests of oliear restrictios of the form: H 0 : h( 0 ) = 0 versus H : h( 0 ) 6= 0: () I this lecture, we cosier Wal,

More information

Chapter 2 Transformations and Expectations

Chapter 2 Transformations and Expectations Chapter Trasformatios a Epectatios Chapter Distributios of Fuctios of a Raom Variable Problem: Let be a raom variable with cf F ( ) If we efie ay fuctio of, say g( ) g( ) is also a raom variable whose

More information

Variable Selection in High-dimensional Varying-coefficient Models with Global Optimality

Variable Selection in High-dimensional Varying-coefficient Models with Global Optimality Joural of Machie Learig Research 13 2012 1973-1998 Submitte 3/11; Revise 1/12; Publishe 6/12 Variable Selectio i High-imesioal Varyig-coefficiet Moels with Global Optimality La Xue Departmet of Statistics

More information

Double Stage Shrinkage Estimator of Two Parameters. Generalized Exponential Distribution

Double Stage Shrinkage Estimator of Two Parameters. Generalized Exponential Distribution Iteratioal Mathematical Forum, Vol., 3, o. 3, 3-53 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.9/imf.3.335 Double Stage Shrikage Estimator of Two Parameters Geeralized Expoetial Distributio Alaa M.

More information

Tarek M. Omara. Department of Statistics, Mathematics and Insurance, Faculty of Commerce, Kafrelsheikh University, Kafrelsheikh, Egypt

Tarek M. Omara. Department of Statistics, Mathematics and Insurance, Faculty of Commerce, Kafrelsheikh University, Kafrelsheikh, Egypt America Joural of Mathematics ad Statistics 7, 7(): 7-77 DOI:.593/.ams.77.3 Weighted Robust Lasso ad Adaptive Elastic Net Method for Regularizatio ad Variable Selectio i Robust Regressio ith Optimal Scalig

More information

Machine Learning Regression I Hamid R. Rabiee [Slides are based on Bishop Book] Spring

Machine Learning Regression I Hamid R. Rabiee [Slides are based on Bishop Book] Spring Machie Learig Regressio I Hamid R. Rabiee [Slides are based o Bishop Book] Sprig 015 http://ce.sharif.edu/courses/93-94//ce717-1 Liear Regressio Liear regressio: ivolves a respose variable ad a sigle predictor

More information

(average number of points per unit length). Note that Equation (9B1) does not depend on the

(average number of points per unit length). Note that Equation (9B1) does not depend on the EE603 Class Notes 9/25/203 Joh Stesby Appeix 9-B: Raom Poisso Poits As iscusse i Chapter, let (t,t 2 ) eote the umber of Poisso raom poits i the iterval (t, t 2 ]. The quatity (t, t 2 ) is a o-egative-iteger-value

More information

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample. Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized

More information

6.867 Machine learning, lecture 11 (Jaakkola)

6.867 Machine learning, lecture 11 (Jaakkola) 6.867 Machie learig, lecture 11 (Jaakkola) 1 Lecture topics: moel selectio criteria Miimum escriptio legth (MDL) Feature (subset) selectio Moel selectio criteria: Miimum escriptio legth (MDL) The miimum

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

Quantile regression with multilayer perceptrons.

Quantile regression with multilayer perceptrons. Quatile regressio with multilayer perceptros. S.-F. Dimby ad J. Rykiewicz Uiversite Paris 1 - SAMM 90 Rue de Tolbiac, 75013 Paris - Frace Abstract. We cosider oliear quatile regressio ivolvig multilayer

More information

Overview. Structured learning for feature selection and prediction. Motivation for feature selection. Outline. Part III:

Overview. Structured learning for feature selection and prediction. Motivation for feature selection. Outline. Part III: Overview Structured learig for feature selectio ad predictio Yookyug Lee Departmet of Statistics The Ohio State Uiversity Part I: Itroductio to Kerel methods Part II: Learig with Reproducig Kerel Hilbert

More information

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates Iteratioal Joural of Scieces: Basic ad Applied Research (IJSBAR) ISSN 2307-4531 (Prit & Olie) http://gssrr.org/idex.php?joural=jouralofbasicadapplied ---------------------------------------------------------------------------------------------------------------------------

More information

The Method of Least Squares. To understand least squares fitting of data.

The Method of Least Squares. To understand least squares fitting of data. The Method of Least Squares KEY WORDS Curve fittig, least square GOAL To uderstad least squares fittig of data To uderstad the least squares solutio of icosistet systems of liear equatios 1 Motivatio Curve

More information

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise)

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise) Lecture 22: Review for Exam 2 Basic Model Assumptios (without Gaussia Noise) We model oe cotiuous respose variable Y, as a liear fuctio of p umerical predictors, plus oise: Y = β 0 + β X +... β p X p +

More information

A New Multivariate Markov Chain Model with Applications to Sales Demand Forecasting

A New Multivariate Markov Chain Model with Applications to Sales Demand Forecasting Iteratioal Coferece o Idustrial Egieerig ad Systems Maagemet IESM 2007 May 30 - Jue 2 BEIJING - CHINA A New Multivariate Markov Chai Model with Applicatios to Sales Demad Forecastig Wai-Ki CHING a, Li-Mi

More information

Complex Algorithms for Lattice Adaptive IIR Notch Filter

Complex Algorithms for Lattice Adaptive IIR Notch Filter 4th Iteratioal Coferece o Sigal Processig Systems (ICSPS ) IPCSIT vol. 58 () () IACSIT Press, Sigapore DOI:.7763/IPCSIT..V58. Complex Algorithms for Lattice Adaptive IIR Notch Filter Hog Liag +, Nig Jia

More information

Some Exponential Ratio-Product Type Estimators using information on Auxiliary Attributes under Second Order Approximation

Some Exponential Ratio-Product Type Estimators using information on Auxiliary Attributes under Second Order Approximation ; [Formerly kow as the Bulleti of Statistics & Ecoomics (ISSN 097-70)]; ISSN 0975-556X; Year: 0, Volume:, Issue Number: ; It. j. stat. eco.; opyright 0 by ESER Publicatios Some Expoetial Ratio-Product

More information

Machine Learning Brett Bernstein

Machine Learning Brett Bernstein Machie Learig Brett Berstei Week 2 Lecture: Cocept Check Exercises Starred problems are optioal. Excess Risk Decompositio 1. Let X = Y = {1, 2,..., 10}, A = {1,..., 10, 11} ad suppose the data distributio

More information

Statistica Sinica Preprint No: SS R3

Statistica Sinica Preprint No: SS R3 Statistica Siica Preprit No: SS-2015-0179R3 Title Sparse ad robust liear regressio: a optimizatio algorithm ad its statistical properties Mauscript ID SS-2015-0179 URL http://www.stat.siica.edu.tw/statistica/

More information

MOMENT-METHOD ESTIMATION BASED ON CENSORED SAMPLE

MOMENT-METHOD ESTIMATION BASED ON CENSORED SAMPLE Vol. 8 o. Joural of Systems Sciece ad Complexity Apr., 5 MOMET-METHOD ESTIMATIO BASED O CESORED SAMPLE I Zhogxi Departmet of Mathematics, East Chia Uiversity of Sciece ad Techology, Shaghai 37, Chia. Email:

More information

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So,

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So, 0 2. OLS Part II The OLS residuals are orthogoal to the regressors. If the model icludes a itercept, the orthogoality of the residuals ad regressors gives rise to three results, which have limited practical

More information

Response Variable denoted by y it is the variable that is to be predicted measure of the outcome of an experiment also called the dependent variable

Response Variable denoted by y it is the variable that is to be predicted measure of the outcome of an experiment also called the dependent variable Statistics Chapter 4 Correlatio ad Regressio If we have two (or more) variables we are usually iterested i the relatioship betwee the variables. Associatio betwee Variables Two variables are associated

More information

Regression with quadratic loss

Regression with quadratic loss Regressio with quadratic loss Maxim Ragisky October 13, 2015 Regressio with quadratic loss is aother basic problem studied i statistical learig theory. We have a radom couple Z = X,Y, where, as before,

More information

A Central Limit Theorem for Local Polynomial Backfitting Estimators

A Central Limit Theorem for Local Polynomial Backfitting Estimators Joural of Multivariate Aalysis 70, 5765 (1999) Article ID mva.1999.1812, available olie at httpwww.iealibrary.com o A Cetral Limit Theorem for Local Polyomial Backfittig Estimators M. P. Wa Harvar School

More information

Summary and Discussion on Simultaneous Analysis of Lasso and Dantzig Selector

Summary and Discussion on Simultaneous Analysis of Lasso and Dantzig Selector Summary ad Discussio o Simultaeous Aalysis of Lasso ad Datzig Selector STAT732, Sprig 28 Duzhe Wag May 4, 28 Abstract This is a discussio o the work i Bickel, Ritov ad Tsybakov (29). We begi with a short

More information

Outline. Linear regression. Regularization functions. Polynomial curve fitting. Stochastic gradient descent for regression. MLE for regression

Outline. Linear regression. Regularization functions. Polynomial curve fitting. Stochastic gradient descent for regression. MLE for regression REGRESSION 1 Outlie Liear regressio Regularizatio fuctios Polyomial curve fittig Stochastic gradiet descet for regressio MLE for regressio Step-wise forward regressio Regressio methods Statistical techiques

More information

Information-based Feature Selection

Information-based Feature Selection Iformatio-based Feature Selectio Farza Faria, Abbas Kazeroui, Afshi Babveyh Email: {faria,abbask,afshib}@staford.edu 1 Itroductio Feature selectio is a topic of great iterest i applicatios dealig with

More information

11 Correlation and Regression

11 Correlation and Regression 11 Correlatio Regressio 11.1 Multivariate Data Ofte we look at data where several variables are recorded for the same idividuals or samplig uits. For example, at a coastal weather statio, we might record

More information

Expectation maximization

Expectation maximization Motivatio Expectatio maximizatio Subhrasu Maji CMSCI 689: Machie Learig 14 April 015 Suppose you are builig a aive Bayes spam classifier. After your are oe your boss tells you that there is o moey to label

More information

FROM SPECIFICATION TO MEASUREMENT: THE BOTTLENECK IN ANALOG INDUSTRIAL TESTING

FROM SPECIFICATION TO MEASUREMENT: THE BOTTLENECK IN ANALOG INDUSTRIAL TESTING FROM SPECIFICATION TO MEASUREMENT: THE BOTTLENECK IN ANALOG INDUSTRIAL TESTING R.J. va Rijsige, A.A.R.M. Haggeburg, C. e Vries Philips Compoets Busiess Uit Cosumer IC Gerstweg 2, 6534 AE Nijmege The Netherlas

More information

Bull. Korean Math. Soc. 36 (1999), No. 3, pp. 451{457 THE STRONG CONSISTENCY OF NONLINEAR REGRESSION QUANTILES ESTIMATORS Seung Hoe Choi and Hae Kyung

Bull. Korean Math. Soc. 36 (1999), No. 3, pp. 451{457 THE STRONG CONSISTENCY OF NONLINEAR REGRESSION QUANTILES ESTIMATORS Seung Hoe Choi and Hae Kyung Bull. Korea Math. Soc. 36 (999), No. 3, pp. 45{457 THE STRONG CONSISTENCY OF NONLINEAR REGRESSION QUANTILES ESTIMATORS Abstract. This paper provides suciet coditios which esure the strog cosistecy of regressio

More information

REGRESSION WITH QUADRATIC LOSS

REGRESSION WITH QUADRATIC LOSS REGRESSION WITH QUADRATIC LOSS MAXIM RAGINSKY Regressio with quadratic loss is aother basic problem studied i statistical learig theory. We have a radom couple Z = X, Y ), where, as before, X is a R d

More information

Commonly Used Distributions and Parameter Estimation

Commonly Used Distributions and Parameter Estimation Commoly Use Distributios a Parameter stimatio Berli Che Departmet of Computer Sciece & Iformatio gieerig Natioal Taiwa Normal Uiversity Referece:. W. Navii. Statistics for gieerig a Scietists. Chapter

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

Moment closure for biochemical networks

Moment closure for biochemical networks Momet closure for biochemical etworks João Hespaha Departmet of Electrical a Computer Egieerig Uiversity of Califoria, Sata Barbara 9-9 email: hespaha@ece.ucsb.eu Abstract Momet closure is a techique use

More information

Ω ). Then the following inequality takes place:

Ω ). Then the following inequality takes place: Lecture 8 Lemma 5. Let f : R R be a cotiuously differetiable covex fuctio. Choose a costat δ > ad cosider the subset Ωδ = { R f δ } R. Let Ωδ ad assume that f < δ, i.e., is ot o the boudary of f = δ, i.e.,

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Patter Recogitio Classificatio: No-Parametric Modelig Hamid R. Rabiee Jafar Muhammadi Sprig 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Ageda Parametric Modelig No-Parametric Modelig

More information

Confidence Interval for Standard Deviation of Normal Distribution with Known Coefficients of Variation

Confidence Interval for Standard Deviation of Normal Distribution with Known Coefficients of Variation Cofidece Iterval for tadard Deviatio of Normal Distributio with Kow Coefficiets of Variatio uparat Niwitpog Departmet of Applied tatistics, Faculty of Applied ciece Kig Mogkut s Uiversity of Techology

More information

Machine Learning. Ilya Narsky, Caltech

Machine Learning. Ilya Narsky, Caltech Machie Learig Ilya Narsky, Caltech Lecture 4 Multi-class problems. Multi-class versios of Neural Networks, Decisio Trees, Support Vector Machies ad AdaBoost. Reductio of a multi-class problem to a set

More information

Linear Regression Demystified

Linear Regression Demystified Liear Regressio Demystified Liear regressio is a importat subject i statistics. I elemetary statistics courses, formulae related to liear regressio are ofte stated without derivatio. This ote iteds to

More information

6.3.3 Parameter Estimation

6.3.3 Parameter Estimation 130 CHAPTER 6. ARMA MODELS 6.3.3 Parameter Estimatio I this sectio we will iscuss methos of parameter estimatio for ARMAp,q assumig that the orers p a q are kow. Metho of Momets I this metho we equate

More information

Lecture #3. Math tools covered today

Lecture #3. Math tools covered today Toay s Program:. Review of previous lecture. QM free particle a particle i a bo. 3. Priciple of spectral ecompositio. 4. Fourth Postulate Math tools covere toay Lecture #3. Lear how to solve separable

More information

Linear Regression Models

Linear Regression Models Liear Regressio Models Dr. Joh Mellor-Crummey Departmet of Computer Sciece Rice Uiversity johmc@cs.rice.edu COMP 528 Lecture 9 15 February 2005 Goals for Today Uderstad how to Use scatter diagrams to ispect

More information

k=1 s k (x) (3) and that the corresponding infinite series may also converge; moreover, if it converges, then it defines a function S through its sum

k=1 s k (x) (3) and that the corresponding infinite series may also converge; moreover, if it converges, then it defines a function S through its sum 0. L Hôpital s rule You alreay kow from Lecture 0 that ay sequece {s k } iuces a sequece of fiite sums {S } through S = s k, a that if s k 0 as k the {S } may coverge to the it k= S = s s s 3 s 4 = s k.

More information

Using the IML Procedure to Examine the Efficacy of a New Control Charting Technique

Using the IML Procedure to Examine the Efficacy of a New Control Charting Technique Paper 2894-2018 Usig the IML Procedure to Examie the Efficacy of a New Cotrol Chartig Techique Austi Brow, M.S., Uiversity of Norther Colorado; Bryce Whitehead, M.S., Uiversity of Norther Colorado ABSTRACT

More information

The Chi Squared Distribution Page 1

The Chi Squared Distribution Page 1 The Chi Square Distributio Page Cosier the istributio of the square of a score take from N(, The probability that z woul have a value less tha is give by z / g ( ( e z if > F π, if < z where ( e g e z

More information

Strong consistency of log-likelihood-based information criterion in high-dimensional canonical correlation analysis

Strong consistency of log-likelihood-based information criterion in high-dimensional canonical correlation analysis Strog cosistecy of log-likelihood-based iformatio criterio i high-dimesioal caoical correlatio aalysis Ryoya Oda, Hirokazu Yaagihara ad Yasuori Fujikoshi Departmet of Mathematics, Graduate School of Sciece,

More information

Study the bias (due to the nite dimensional approximation) and variance of the estimators

Study the bias (due to the nite dimensional approximation) and variance of the estimators 2 Series Methods 2. Geeral Approach A model has parameters (; ) where is ite-dimesioal ad is oparametric. (Sometimes, there is o :) We will focus o regressio. The fuctio is approximated by a series a ite

More information

CMSE 820: Math. Foundations of Data Sci.

CMSE 820: Math. Foundations of Data Sci. Lecture 17 8.4 Weighted path graphs Take from [10, Lecture 3] As alluded to at the ed of the previous sectio, we ow aalyze weighted path graphs. To that ed, we prove the followig: Theorem 6 (Fiedler).

More information

Notes on iteration and Newton s method. Iteration

Notes on iteration and Newton s method. Iteration Notes o iteratio ad Newto s method Iteratio Iteratio meas doig somethig over ad over. I our cotet, a iteratio is a sequece of umbers, vectors, fuctios, etc. geerated by a iteratio rule of the type 1 f

More information

Automatic polynomial wavelet regression

Automatic polynomial wavelet regression Statistics ad Computig 14: 337 341, 2004 C 2004 Kluwer Academic Publishers. Maufactured i The Netherlads. Automatic polyomial wavelet regressio THOMAS C. M. LEE ad HEE-SEOK OH Departmet of Statistics,

More information

Estimation of Population Mean Using Co-Efficient of Variation and Median of an Auxiliary Variable

Estimation of Population Mean Using Co-Efficient of Variation and Median of an Auxiliary Variable Iteratioal Joural of Probability ad Statistics 01, 1(4: 111-118 DOI: 10.593/j.ijps.010104.04 Estimatio of Populatio Mea Usig Co-Efficiet of Variatio ad Media of a Auxiliary Variable J. Subramai *, G. Kumarapadiya

More information

Rational Functions. Rational Function. Example. The degree of q x. If n d then f x is an improper rational function. x x. We have three forms

Rational Functions. Rational Function. Example. The degree of q x. If n d then f x is an improper rational function. x x. We have three forms Ratioal Fuctios We have three forms R Ratioal Fuctio a a1 a p p 0 b b1 b0 q q p a a1 a0 q b b1 b0 The egree of p The egree of q is is If the f is a improper ratioal fuctio Compare forms Epae a a a b b1

More information

Taylor polynomial solution of difference equation with constant coefficients via time scales calculus

Taylor polynomial solution of difference equation with constant coefficients via time scales calculus TMSCI 3, o 3, 129-135 (2015) 129 ew Treds i Mathematical Scieces http://wwwtmscicom Taylor polyomial solutio of differece equatio with costat coefficiets via time scales calculus Veysel Fuat Hatipoglu

More information

Study on Coal Consumption Curve Fitting of the Thermal Power Based on Genetic Algorithm

Study on Coal Consumption Curve Fitting of the Thermal Power Based on Genetic Algorithm Joural of ad Eergy Egieerig, 05, 3, 43-437 Published Olie April 05 i SciRes. http://www.scirp.org/joural/jpee http://dx.doi.org/0.436/jpee.05.34058 Study o Coal Cosumptio Curve Fittig of the Thermal Based

More information

THE DATA-BASED CHOICE OF BANDWIDTH FOR KERNEL QUANTILE ESTIMATOR OF VAR

THE DATA-BASED CHOICE OF BANDWIDTH FOR KERNEL QUANTILE ESTIMATOR OF VAR Iteratioal Joural of Iovative Maagemet, Iformatio & Productio ISME Iteratioal c2013 ISSN 2185-5439 Volume 4, Number 1, Jue 2013 PP. 17-24 THE DATA-BASED CHOICE OF BANDWIDTH FOR KERNEL QUANTILE ESTIMATOR

More information

Probability in Medical Imaging

Probability in Medical Imaging Chapter P Probability i Meical Imagig Cotets Itrouctio P1 Probability a isotropic emissios P2 Raioactive ecay statistics P4 Biomial coutig process P4 Half-life P5 Poisso process P6 Determiig activity of

More information

Variable selection and estimation in generalized linear models with the seamless L 0 penalty

Variable selection and estimation in generalized linear models with the seamless L 0 penalty Variable selectio ad estimatio i geeralized liear models with the seamless L 0 pealty Zili Li, Sijia Wag 2 ad Xihog Li 3 Departmet of Mathematics, Tsighua Uiversity, Beijig Chia 2 Departmet of Biostatistics

More information

A new iterative algorithm for reconstructing a signal from its dyadic wavelet transform modulus maxima

A new iterative algorithm for reconstructing a signal from its dyadic wavelet transform modulus maxima ol 46 No 6 SCIENCE IN CHINA (Series F) December 3 A ew iterative algorithm for recostructig a sigal from its dyadic wavelet trasform modulus maxima ZHANG Zhuosheg ( u ), LIU Guizhog ( q) & LIU Feg ( )

More information

Stat 421-SP2012 Interval Estimation Section

Stat 421-SP2012 Interval Estimation Section Stat 41-SP01 Iterval Estimatio Sectio 11.1-11. We ow uderstad (Chapter 10) how to fid poit estimators of a ukow parameter. o However, a poit estimate does ot provide ay iformatio about the ucertaity (possible

More information

The Method of Particular Solutions (MPS) for Solving One- Dimensional Hyperbolic Telegraph Equation

The Method of Particular Solutions (MPS) for Solving One- Dimensional Hyperbolic Telegraph Equation ISS 746-7659, Egla, UK Joural of Iformatio a Computig Sciece Vol., o. 3, 05, pp. 99-08 The Metho of Particular Solutios (MPS) for Solvig Oe- Dimesioal Hyperbolic Telegraph Equatio LigDe Su,,, ZiWu Jiag

More information

Mathematical Modeling of Optimum 3 Step Stress Accelerated Life Testing for Generalized Pareto Distribution

Mathematical Modeling of Optimum 3 Step Stress Accelerated Life Testing for Generalized Pareto Distribution America Joural of Theoretical ad Applied Statistics 05; 4(: 6-69 Published olie May 8, 05 (http://www.sciecepublishiggroup.com/j/ajtas doi: 0.648/j.ajtas.05040. ISSN: 6-8999 (Prit; ISSN: 6-9006 (Olie Mathematical

More information

ACCURATE DICTIONARY LEARNING WITH DIRECT SPARSITY CONTROL. Hongyu Mou, Adrian Barbu

ACCURATE DICTIONARY LEARNING WITH DIRECT SPARSITY CONTROL. Hongyu Mou, Adrian Barbu ACCURATE DICTIONARY LEARNING WITH DIRECT SPARSITY CONTROL Hogyu Mou, Adria Barbu Statistics Departmet, Florida State Uiversity Tallahassee FL 32306 ABSTRACT Dictioary learig is a popular method for obtaiig

More information

M-Estimators in Regression Models

M-Estimators in Regression Models M-Estimators i Regressio Models MuthukrishaR Departmet of Statistics, Bharathiar Uiversity Coimbatore-641 046, Tamiladu, Idia E-mail: muthukrisha70@rediffmailcom RadhaM Departmet of Statistics, Bharathiar

More information

1.010 Uncertainty in Engineering Fall 2008

1.010 Uncertainty in Engineering Fall 2008 MIT OpeCourseWare http://ocw.mit.edu.00 Ucertaity i Egieerig Fall 2008 For iformatio about citig these materials or our Terms of Use, visit: http://ocw.mit.edu.terms. .00 - Brief Notes # 9 Poit ad Iterval

More information

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ.

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ. 2 5. Weighted umber of late jobs 5.1. Release dates ad due dates: maximimizig the weight of o-time jobs Oce we add release dates, miimizig the umber of late jobs becomes a sigificatly harder problem. For

More information

BIOINF 585: Machine Learning for Systems Biology & Clinical Informatics

BIOINF 585: Machine Learning for Systems Biology & Clinical Informatics BIOINF 585: Machie Learig for Systems Biology & Cliical Iformatics Lecture 14: Dimesio Reductio Jie Wag Departmet of Computatioal Medicie & Bioiformatics Uiversity of Michiga 1 Outlie What is feature reductio?

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

Power of Mean Chart under Second Order Auto- Correlation with Known Coefficient of Variation

Power of Mean Chart under Second Order Auto- Correlation with Known Coefficient of Variation Iteratioal Joural of Scietific a Research ublicatios Volume Issue December 0 ISSN 50-5 ower of Mea Chart uer Seco Orer Auto- Correlatio with Kow Coefficiet of Variatio Sigh D. a Sigh J.R Vikram uiversity

More information

REGRESSION (Physics 1210 Notes, Partial Modified Appendix A)

REGRESSION (Physics 1210 Notes, Partial Modified Appendix A) REGRESSION (Physics 0 Notes, Partial Modified Appedix A) HOW TO PERFORM A LINEAR REGRESSION Cosider the followig data poits ad their graph (Table I ad Figure ): X Y 0 3 5 3 7 4 9 5 Table : Example Data

More information

Bi-Responses Nonparametric Regression Model. Using MARS and Its Properties

Bi-Responses Nonparametric Regression Model. Using MARS and Its Properties Applied Mathematical Scieces, Vol. 9, 05, o. 9, 47-47 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/0.988/ams.05.57 Bi-Resposes Noparametric Regressio Model Usig MARS ad Its Properties Ayub Parli Ampulembag

More information

Support vector machine revisited

Support vector machine revisited 6.867 Machie learig, lecture 8 (Jaakkola) 1 Lecture topics: Support vector machie ad kerels Kerel optimizatio, selectio Support vector machie revisited Our task here is to first tur the support vector

More information

Global Convergence of a New Conjugate Gradient Method with Wolfe Type Line Search +

Global Convergence of a New Conjugate Gradient Method with Wolfe Type Line Search + ISSN 76-7659, Ela, UK Joural of Iformatio a Computi Sciece Vol 7, No,, pp 67-7 Global Coverece of a New Cojuate Graiet Metho with Wolfe ype Lie Search + Yuayua Che, School of Maaemet, Uiversity of Shahai

More information

4.5 Multiple Imputation

4.5 Multiple Imputation 45 ultiple Imputatio Itroductio Assume a parametric model: y fy x; θ We are iterested i makig iferece about θ I Bayesia approach, we wat to make iferece about θ from fθ x, y = πθfy x, θ πθfy x, θdθ where

More information

Intermittent demand forecasting by using Neural Network with simulated data

Intermittent demand forecasting by using Neural Network with simulated data Proceedigs of the 011 Iteratioal Coferece o Idustrial Egieerig ad Operatios Maagemet Kuala Lumpur, Malaysia, Jauary 4, 011 Itermittet demad forecastig by usig Neural Network with simulated data Nguye Khoa

More information

Supporting Information

Supporting Information Supportig Iformatio Kirkpatrick et al. 0.073/pas.68354 Raom Patters I this sectio we show that usig EWC it is possible to recover a power-law ecay for the SNR of raom patters. The task cosists of associatig

More information

WEIGHTED LEAST SQUARES - used to give more emphasis to selected points in the analysis. Recall, in OLS we minimize Q =! % =!

WEIGHTED LEAST SQUARES - used to give more emphasis to selected points in the analysis. Recall, in OLS we minimize Q =! % =! WEIGHTED LEAST SQUARES - used to give more emphasis to selected poits i the aalysis What are eighted least squares?! " i=1 i=1 Recall, i OLS e miimize Q =! % =!(Y - " - " X ) or Q = (Y_ - X "_) (Y_ - X

More information

Composite Hermite and Anti-Hermite Polynomials

Composite Hermite and Anti-Hermite Polynomials Avaces i Pure Mathematics 5 5 87-87 Publishe Olie December 5 i SciRes. http://www.scirp.org/joural/apm http://.oi.org/.436/apm.5.5476 Composite Hermite a Ati-Hermite Polyomials Joseph Akeyo Omolo Departmet

More information

10/2/ , 5.9, Jacob Hays Amit Pillay James DeFelice

10/2/ , 5.9, Jacob Hays Amit Pillay James DeFelice 0//008 Liear Discrimiat Fuctios Jacob Hays Amit Pillay James DeFelice 5.8, 5.9, 5. Miimum Squared Error Previous methods oly worked o liear separable cases, by lookig at misclassified samples to correct

More information

3. Calculus with distributions

3. Calculus with distributions 6 RODICA D. COSTIN 3.1. Limits of istributios. 3. Calculus with istributios Defiitio 4. A sequece of istributios {u } coverges to the istributio u (all efie o the same space of test fuctios) if (φ, u )

More information

STA6938-Logistic Regression Model

STA6938-Logistic Regression Model Dr. Yig Zhag STA6938-Logistic Regressio Model Topic -Simple (Uivariate) Logistic Regressio Model Outlies:. Itroductio. A Example-Does the liear regressio model always work? 3. Maximum Likelihood Curve

More information

LASSO TUNING PARAMETER SELECTION

LASSO TUNING PARAMETER SELECTION Proceedigs of the 57th Aual Coferece of SASA (2015), 49 56 49 LASSO TUNING PARAMETER SELECTION Lisa-A Kirklad, Fras Kafer ad Sollie Millard Uiversity of Pretoria Key words: LASSO, model selectio, pealized

More information

Comparison of Methods for Estimation of Sample Sizes under the Weibull Distribution

Comparison of Methods for Estimation of Sample Sizes under the Weibull Distribution Iteratioal Joural of Applied Egieerig Research ISSN 0973-4562 Volume 12, Number 24 (2017) pp. 14273-14278 Research Idia Publicatios. http://www.ripublicatio.com Compariso of Methods for Estimatio of Sample

More information

NEW IDENTIFICATION AND CONTROL METHODS OF SINE-FUNCTION JULIA SETS

NEW IDENTIFICATION AND CONTROL METHODS OF SINE-FUNCTION JULIA SETS Joural of Applied Aalysis ad Computatio Volume 5, Number 2, May 25, 22 23 Website:http://jaac-olie.com/ doi:.948/252 NEW IDENTIFICATION AND CONTROL METHODS OF SINE-FUNCTION JULIA SETS Jie Su,2, Wei Qiao

More information

Root Finding COS 323

Root Finding COS 323 Root Fidig COS 323 Remider Sig up for Piazza Assigmet 0 is posted, due Tue 9/25 Last time.. Floatig poit umbers ad precisio Machie epsilo Sources of error Sesitivity ad coditioig Stability ad accuracy

More information

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 12

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 12 Machie Learig Theory Tübige Uiversity, WS 06/07 Lecture Tolstikhi Ilya Abstract I this lecture we derive risk bouds for kerel methods. We will start by showig that Soft Margi kerel SVM correspods to miimizig

More information

Numerical Conformal Mapping via a Fredholm Integral Equation using Fourier Method ABSTRACT INTRODUCTION

Numerical Conformal Mapping via a Fredholm Integral Equation using Fourier Method ABSTRACT INTRODUCTION alaysia Joural of athematical Scieces 3(1): 83-93 (9) umerical Coformal appig via a Fredholm Itegral Equatio usig Fourier ethod 1 Ali Hassa ohamed urid ad Teh Yua Yig 1, Departmet of athematics, Faculty

More information

A NEW METHOD FOR CONSTRUCTING APPROXIMATE CONFIDENCE INTERVALS FOR M-ESTU1ATES. Dennis D. Boos

A NEW METHOD FOR CONSTRUCTING APPROXIMATE CONFIDENCE INTERVALS FOR M-ESTU1ATES. Dennis D. Boos .- A NEW METHOD FOR CONSTRUCTING APPROXIMATE CONFIDENCE INTERVALS FOR M-ESTU1ATES by Deis D. Boos Departmet of Statistics North Carolia State Uiversity Istitute of Statistics Mimeo Series #1198 September,

More information

10-701/ Machine Learning Mid-term Exam Solution

10-701/ Machine Learning Mid-term Exam Solution 0-70/5-78 Machie Learig Mid-term Exam Solutio Your Name: Your Adrew ID: True or False (Give oe setece explaatio) (20%). (F) For a cotiuous radom variable x ad its probability distributio fuctio p(x), it

More information

Goodness-Of-Fit For The Generalized Exponential Distribution. Abstract

Goodness-Of-Fit For The Generalized Exponential Distribution. Abstract Goodess-Of-Fit For The Geeralized Expoetial Distributio By Amal S. Hassa stitute of Statistical Studies & Research Cairo Uiversity Abstract Recetly a ew distributio called geeralized expoetial or expoetiated

More information

Expectation-Maximization Algorithm.

Expectation-Maximization Algorithm. Expectatio-Maximizatio Algorithm. Petr Pošík Czech Techical Uiversity i Prague Faculty of Electrical Egieerig Dept. of Cyberetics MLE 2 Likelihood.........................................................................................................

More information

A steepest descent method with a Wolf type line search

A steepest descent method with a Wolf type line search ISSN 1746-7659, Ela, UK Joural of Iformatio a Computi Sciece Vol 9, No 4, 14, pp 5-61 A steepest escet metho with a Wolf type lie search Ju Ji-jie,Pa De-ya,Du Shou-qia Collee of Mathematics, Qiao Uiversity,Qiao

More information

Final Examination Solutions 17/6/2010

Final Examination Solutions 17/6/2010 The Islamic Uiversity of Gaza Faculty of Commerce epartmet of Ecoomics ad Political Scieces A Itroductio to Statistics Course (ECOE 30) Sprig Semester 009-00 Fial Eamiatio Solutios 7/6/00 Name: I: Istructor:

More information

A collocation method for singular integral equations with cosecant kernel via Semi-trigonometric interpolation

A collocation method for singular integral equations with cosecant kernel via Semi-trigonometric interpolation Iteratioal Joural of Mathematics Research. ISSN 0976-5840 Volume 9 Number 1 (017) pp. 45-51 Iteratioal Research Publicatio House http://www.irphouse.com A collocatio method for sigular itegral equatios

More information

A Risk Comparison of Ordinary Least Squares vs Ridge Regression

A Risk Comparison of Ordinary Least Squares vs Ridge Regression Joural of Machie Learig Research 14 (2013) 1505-1511 Submitted 5/12; Revised 3/13; Published 6/13 A Risk Compariso of Ordiary Least Squares vs Ridge Regressio Paramveer S. Dhillo Departmet of Computer

More information