Array Variate Random Variables with Multiway Kronecker Delta Covariance Matrix Structure

Size: px
Start display at page:

Download "Array Variate Random Variables with Multiway Kronecker Delta Covariance Matrix Structure"

Transcription

1 Array Variate Randm Variables with Multiway Krnecker Delta Cvariance Matrix Structure Deniz Akdemir Department f Statistics University f Central Flrida Orland, FL Arjun K. Gupta Department f Mathematics and Statistics Bwling Green State University Bwling Green, Ohi February 27, 2011 Abstract Standard statistical methds applied t matrix randm variables ften fail t describe the underlying structure in multiway data sets. After a review f the essential backgrund material, this paper intrduces the ntin f array variate randm variable. A nrmal array variate randm variable is defined and a methd fr estimating the parameters f array variate nrmal distributin is given. We intrduce a technique called slicing fr estimating the cvariance matrix f high dimensinal data. Finally, principal cmpnent analysis and classificatin techniques are develped fr array variate bservatins and high dimensinal data. AMS 2000 Subject Classificatin: Primary 62H10, Secndary 62H05. Keywrds & Phrases: Nrmal Distributin, Multivariate Distributin, Matrix Variate Nrmal Distributin, Array Variate Randm Variable, Array Variate Nrmal Distributin, Multilevel Data Analysis, Repeated Measures, Classificatin, Dimensin Reductin 1 Intrductin A data structure is a particular way f string and rganizing data. Different kinds f data structures are suited fr different kinds f applicatins. In this paper, we will study cntinuus prbability distributins fr data that has array structure. A ne dimensinal array is a scalar, a tw dimensinal array is a matrix, a three dimensinal array is a stack f matrices, a fur dimensinal array is a stack 1

2 f three dimensinal arrays, etc. Type f data that statistical analysis entails is usually best rganized in array frm. T stre ne bservatin f a m 1 -variate randm variable, we use a m 1 1 matrix ( a vectr), a tw dimensinal array. If we have a randm sample f m 2 m 1 -variate bservatins, we can rganize this in a m 1 m 2 matrix, again a tw dimensinal array. It is als very cmmn that a data describes the bservatin f m 1 variables, n m 2 individuals at m 3 time pints. This kind f data set is suitable t be rganized in a three dimensinal array f dimensins m 1 m 2 m 3. If bth m 3 time pints and m 4 different places are bserved then this can be stred in a fur dimensinal array f dimensins m 1 m 2 m 3 m 4. The assciated randm variable in each f these cases has array structure. The array variate randm variable up t 2 dimensins has been studied intensively in [Gupta and Nagar, 2000] and by many thers. Hwever fr arrays bservatins f 3, 4 r in general i dimensins n apprpriate prbability mdels have been prpsed. Figure 1 illustrates N bservatins f a 4 way array variable. Repeated measurements (acrss time, acrss space r bth) can be assumed t have array structure. Mst applicatins in spatial statistics invlve mdeling f cmplex spatial tempral dependency structures, and many f the prblems f space and time mdeling and seperatin f effects can be slved by using the results in this paper. In the literature, this kind f data is usually refered as multiway data [Krnenberg and Crpratin, 2008] and we assume that multiway data is in array frm N ,2,3,4 1,2,3,4,..,10 1,2,3,4,5 Figure 1: The figure illustrates N bservatins f a 4 way array variate variable. In Sectin 2, we first study the algebra f arrays, and als intrduce the cncept f an array variable randm variable and a nrmal mdel fr array variables. In sectin 3, estimatin methd fr the nrmal mdel is described, 2

3 als describe a technique called slicing fr nnsingular estimatin f high dimensinal cvariance matrices when N < p. Then, in Sectins 4 and 5, we give applicatins f the mdel t principal cmpnent analysis and classificatin. 2 Array Algebra and Array Variate Randm Variables 2.1 Array Algebra In this paper we will nly study arrays with real elements. We will write X t say that X is an array. When it is necessary we can write the dimensins f the array as subindices, e.g., if X is a m1 m 2 m 3 m 4 dimensinal array in R m1 m2... mi, then we can write X m1 m 2 m 3 m 4. Arrays with the usual elementwise summatin and scalar multiplicatin peratins can be shwn t be a vectr space. T refer t an element f an array X m1 m 2 m 3 m 4, we write the psitin f the element as a subindex t the array name in parenthesis, ( X) r1r 2r 3r 4. If we wanttrefertaspecificclumnvectrbtainedbykeepingallbutanindicated dimensin cnstant, we indicate the cnstant dimensins as befre but we will put : fr the nn cnstant dimensin, e.g., fr X m1 m 2 m 3 m 4, ( X) r1r 2:r 4 refers t the the clumn vectr ((X) r1r 21r 4,(X) r1r 22r 4,...,(X) r1r 2m 3r 4 ). We will nw review sme basic principles and techniques f array algebra. These results and their prfs can be fund in Rauhala [Rauhala, 1974], [Rauhala, 1980] and Blaha [Blaha, 1977]. Definitin 2.1. Inverse Krnecker prduct f tw matrices A and B f dimensins p q and r s crrespndingly is written as A i B and is defined as A i B = [A(B) jk ] pr qs = B A, where represents the rdinary Krnecker prduct. The fllwing prperties f the inverse Krnecker prduct are useful: 0 i A = A i 0 = 0. (A 1 +A 2 ) i B = A 1 i B +A 2 i B. A i (B 1 +B 2 ) = A i B 1 +A i B 2. αa i βb = αβa i B. (A 1 i B 1 )(A 2 i B 2 ) = A 1 A 2 i B 1 B 2. (A i B) 1 = (A 1 i B 1 ). (A i B) + = (A + i B + ), where A + is the Mre-Penrse inverse f A. (A i B) = (A i B ), where A is the l-inverse f A defined as A = (A A) 1 A. 3

4 If {λ i } and {µ j } are the eigenvalues with the crrespnding eigenvectrs {x i } and {y j } fr matrices A and B respectively, then A i B has eigenvalues {λ i µ j } with crrespnding eigenvectrs {x i i y j }. Given tw matrices A n n and B m m A i B = A m B n, tr(a i B) = tr(a)tr(b). A i B = B A = U 1 A BU 2, fr sme permutatin matrices U 1 and U 2. It is well knwn that a matrix equatin AXB = C can be rewritten in its mnlinear frm as Furthermre, the matrix equality A i Bvec(X) = vec(c). (1) A i BXC = E btained by stacking equatins f the frm (1) can be written in its mnlinear frm as (A i B i C)vec(X) = vec(e). This prcess f stacking equatins culd be cntinued and R-matrix multiplicatin peratin intrduced by Rauhala [Rauhala, 1974] prvides a cmpact way f representing these equatins in array frm: Definitin 2.2. R-Matrix Multiplicatin is defined elementwise: ((A 1 ) 1 (A 2 ) 2...(A i ) i Xm1 m 2... m i ) q1q 2...q i m 1 m 2 m 3 m i = (A 1 ) q1r 1 (A 2 ) q2r 2 (A 3 ) q3r 3... (A i ) qir i ( X) r1r 2...r i. r 1=1 r 2=1 r 3=1 R-Matrix multiplicatin generalizes the matrix multiplicatin (array multiplicatin in tw dimensins)t the case f k-dimensinal arrays. The fllwing useful prperties f the R-Matrix multiplicatin are reviewed by Blaha [Blaha, 1977]: (A) 1 B = AB. (A 1 ) 1 (A 2 ) 2 C = A 1 CA 2. Ỹ = (I)1 (I) 2...(I) i Ỹ. r i=1 ((A 1) 1 (A 2) 2...(A i) i )((B 1) 1 (B 2) 2...(B i) i )Ỹ = (A1B1)1 (A 2B 2) 2...(A ib i) i Ỹ. The peratr rvec describes the relatinship between X m1 m 2...m i and its mnlinear frm x m1m 2...m i 1. 4

5 Definitin 2.3. rvec( X m1 m 2...m i ) = x m1m 2...m i 1 where x is the clumn vectr btained by stacking the elements f the array X in the rder f its dimensins; i.e., ( X) j1j 2...j i = (x) j where j = (j i 1)n i 1 n i 2...n 1 + (j i 2)n i 2 n i 3...n (j 2 1)n 1 +j 1. Let L m1 m 2...m i = (A 1 ) 1 (A 2 ) 2...(A i ) i X where (Aj ) j is an m j n j matrix fr j = 1,2,...,i and X is an n 1 n 2... n i array. Write l = rvec( L) and x = rvec( X). Then, l = A 1 i A 2 i... i A i x. Therefre, there is an equivalent expressin f the array equatin in mnlinear frm. Definitin 2.4. The square nrm f Xm1 m 2...m i is defined as X 2 = m 1 m 2 j 1=1j 2=1 m i... (( X) j1j 2...j i ) 2. Definitin 2.5. The distance f X 1m1 m2...m i frm X 2m1 m2...m i is defined as X 1 X 2 2. Example 2.1. Let Ỹ = (A 1) 1 (A 2 ) 2...(A i ) i X + Ẽ. Then Ẽ 2 is minimized fr X = (A 1 ) 1 (A 2 )2...(A i )i Ỹ. j i=1 2.2 Array Variate Randm Variables Arrays can be cnstant arrays, i.e. if ( X) r1r 2...r i R are cnstants fr all r j, j = 1,2,...,m j and j = 1,2,...,i then the array X is a cnstant array. Array variate randm variables are arrays with all elements ( X) r1r 2...r i R randm variables. If the sample space fr the randm utcme s is S, ( X) r1r 2...r i = ( X(s)) r1r 2...r i whereeach f( X(s)) r1r 2...r i is arealvalued functin frm S t R. If X is an array variate randm variable, its density (if it exists) is a scalar functin f X( X) such that: f X( X) 0; X f X( X)d X = 1; P( X A) = A f X( X)d X, where A is a subset f the space f realizatins fr X. A scalar functin f X, Ỹ ( X,Ỹ) defines a jint (bi-array variate) prbability density functin if f X, Ỹ ( X,Ỹ) 0; Ỹ X f X, Ỹ ( X,Ỹ)d XdỸ = 1; 5

6 P(( X,Ỹ) A) = A f X, Ỹ ( X,Ỹ)d XdỸ, where A is a subset f the space f realizatins fr ( X,Ỹ). The marginal prbability density functin f X is defined by f X( X) = f X, Ỹ ( X,Ỹ)dỸ, Ỹ and the cnditinal prbability density functin f X given Ỹ is defined by f X Ỹ ( X Ỹ) = f X, Ỹ ( X,Ỹ) fỹ(ỹ), where fỹ(ỹ) > 0. Tw randm arrays X and Ỹ are independent if and nly if f X, Ỹ ( X,Ỹ) = f X( X)fỸ(Ỹ). Therem 2.1. Let (A 1 ) 1, (A 2 ) 2,..., (A i ) i be m 1, m 2,..., m i dimensinal psitive definite matrices. The Jacbian J( X Z) f the transfrmatin X = (A 1 ) 1 (A 2 ) 2...(A i ) i Z + M is ( A 1 j 1 mj A 2 j 2 mj... A i j i mj ) 1. Prf. The result is prven using the equivalence f mnlinear frm btained thrugh the rvec( X) and array X. Let L m1 m 2...m i = (A 1 ) 1 (A 2 ) 2...(A i ) i Z where (A j ) j is an m j n j matrix fr j = 1,2,...,i and X is an n 1 n 2... n i array. Write l = rvec( L) and z = rvec( Z). Then, l = A 1 i A 2 i... i A i z. The resultfllws frmntingthatj(l z) = A 1 i A 2 i... i A i 1,andusinginductin with the rule A i B = A m B n fr n n matrix A and m m matrix B t shw that A 1 i A 2 i... i A i 1 = ( A 1 j 1 m j A 2 j 2 m j... A i j i m j ) 1. Crllary 2.1. Let Z f Z( Z). Define X = (A 1 ) 1 (A 2 ) 2...(A i ) i Z + M where (A 1 ) 1,(A 2 ) 2,...,(A i ) i be m 1,m 2,...,m i dimensinal psitive definite matrices. The pdf f X is given by f X( X;(A 1) 1,(A 2) 2,...,(A i) i, M) = f(a 1 1 )1 (A 1 2 )2...(A 1 i ) i ( X M)) A 1 j 1 m j A 2 j 2 m j... A. i j i m j The main advantage in chsing a Krnecker structure is the decrease in the number f parameters. In Sectin 3, we use Krnecker delta cvariance structures t prvide a regularized estimatr f cvariance structure f an array. 6

7 2.2.1 Array Variate Nrmal Distributin By using the results in the previus sectin n array algebra, mainly the relatinship f the arrays t their mnlinear frms described by Definitin 2.3, we can write the density f the standard nrmal array variable. Definitin 2.6. If Z N m1 m 2... m i ( M = 0,Λ = I m1m 2...m i ), then Z has array variate standard nrmal distributin. The pdf f Z is given by f Z( Z) = exp( 1 2 Z 2 ) (2π) m1m2...mi/2. (2) Fr the scalar case, the density fr the standard nrmal variable z R 1 is given as φ 1 (z) = 1 exp( 1 (2π) z2 ). Fr the m 1 dimensinal standard nrmal vectr z R m1, the density is given by 1 φ m1 (z) = exp( 1 (2π) m z z). Finally the m 1 m 2 standard matrix variate variable Z R m1 m2 has the density 1 φ m1 m 2 (Z) = exp( 1 2 trace(z Z)). (2π) m 1 m 2 2 With the abve definitin, we have generalized the ntin f nrmal randm variable t the array variate case. Definitin 2.7. We write X N m1 m 2... m i ( M,Λ m1m 2...m i ) if rvec( X) N m1m 2...m i (rvec( M),Λ m1m 2...m i ). Here, M is the expected value f X, and Λm1m 2...m i is the cvariance matrix f the m 1 m 2...m i -variate randm variable rvec( X). The family f nrmal densities with Krnecker Delta Cvariance Structure are btained by cnsidering the densities btained by the lcatin-scale transfrmatins f the standard nrmal variables. This kind f mdel is defined in the next therem. Therem 2.2. Let Z N m1 m 2... m i ( M = 0,Λ = I m1m 2...m i ). Define X = (A 1 ) 1 (A 2 ) 2...(A i ) i Z + M where A1,A 2,...,A i are nn singular matrices f rders m 1,m 2,...,m i. Then the pdf f X is given by φ( X; M,A exp( 1 2 1,A 2,...A i) = (A 1 1 )1 (A 1 2 )2...(A 1 i ) i ( X M) 2 ) (2π) m 1m 2...m i /2 A 1 j 1 m j A 2 j 2 m j... A. i j i m j (3) 7

8 Array variate densities in the elliptical family are easily cnstructed using Crrlary 2.1. Fr example, the fllwing definitin prvides a generalizatin f the Student s t distributin t the array variate case. Definitin 2.8. Let A 1,A 2,...,A i be nn singular matrices with rders m 1, m 2,..., m i and M be a m 1 m 2... m i cnstant array. Then the pdf f an m 1 m 2... m i array variate t randm variable, T, with degrees f freedm k is given by f( T; M,A 1,A 2,...A i) = c (1+ (A 1 1 )1 (A 1 2 )2...(A 1 i ) i ( T M) 2 ) (k+m 1m 2...m i )/2 A 1 j 1 m j A 2 j 2 m j... A i j i m j where c = (kπ)m 1 m 2...m i /2 Γ((k+m 1m 2...m i)/2) Γ(k/2). Distributinal prperties f a array nrmal variable with density in the frm f Therem 2.2 can btained by using the equivalent mnlinear representatin. The mments, the marginal and cnditinal distributins, independence f variates shuld be studied cnsidering the equivalent mnlinear frm f the array variable and the well knwn prperties f the multivariate nrmal randm variable. 3 Estimatin In this sectin we prvide an heuristic methd f estimating the mdel parameters. The ptimality f these estimatrs are nt prven but merely checked by simulatin studies. Inference abut the parameters f the mdel in Therem 2.2 fr the matrix variate case has been cnsidered in the statistical literature (Ry and Khattree [2003], [Ry and Leiva, 2008], Lu and Zimmerman [2005], [Srivastava et al., 2008], etc.). In these papers, the unique maximum likelihd estimatrs f the parameters f the mdel in Therem 2.2 fr the matrix variate case are btained under different assumptins fr the cvariance parameters. Sme classificatin rules based n the matrix variate bservatins with Krnecker delta cvariance structures have been studied in [Ry and Leiva, 2009], and als in [Krzyśk and Skrzybut, 2009]. The mdel in Therem 2.2 the way it is stated is unidentifiable. Hwever, this prblem can easily be reslved by putting restrictins n the cvariance parameters. The apprach we take is t assume that j 1 f the last diagnal elements f matrices A j A j are equal t 1 fr j = 1,2,...,i. The Flip-Flp Algrithm is prven t attain the maximum likelihd estimatrs f the parameters f tw dimensinal array variate nrmal distributin [Srivastava et al., 2008]. The fllwing is similar t the flip flp algrithm. First, assume { X 1, X2,..., X N } is a randm sample frm a N( M,A 1,A 2,...A i ) distributin with j 1 f the last diagnal elements f matrices A j A j equal t 1 fr j = 1,2,...,i. Further, we assume that all A js are square psitive definite matrices f rank at least j. Finally, assume that we have N i j=1 m j > m 2 r fr all r = 1,2,...,i. (4) 8

9 Algrithm fr estimatin: 1. Estimate M by M = 1 N N l=1 X l, and btain the centered array bservatins X c l = X l M fr l = 1,2,...,N. 2. Start with initial estimates f A 2,A 3,...,A i. 3. On the basis f the estimates f A 2,A 3,...,A i calculate an estimate f A 1 by first scaling the array bservatins using Z l = (I) 1 (A 1 2 )2,(A 1 3 )3,...,(A 1 i ) i Xc l, and then calculating the square rt f cvariance alng the 1st dimensin f the arrays Z l, l = 1,2,...,N. 4. On the basis f the mst recent estimates f the mdel parameters, estimate A j j = 2,...,i. by first scaling the array bservatins using Z l = (A 1 1 )1 (A 1 2 )2,...(A 1 j 1 )j 1 I(A 1 j+1 )j+1...(a 1 i ) i Xc l, and then calculating the square rt f cvariance alng the jth dimensin f the arrays Z l s fr j = 2,...,i. Scale the estimate f A j A j s that the last j 1 diagnal elements are equal t Repeat steps 3 and 4 until cnvergence is attained. We will use the estimatin algrithm t illustrate a technique we will call slicing fr btaining an apprximate nnsingular estimate f the cvariance matrix f a p dimensinal vectr variate randm variable when N < p. A vectr x f dimensin p can be sliced int p/m 1 = m 2 pieces and rganized int a matrix f rder p = m 1 m 2 fr sme natural numbers m 1 and m 2. Or, in general, the same vectr can be rganized in an array f dimensin p = m 1 m 2... m i fr sme natural numbers m 1, m 2,..., m i. Once we slicethedataandrerganizeitinarrayfrm, wewillpretendthatthisarraydata was generated frm the mdel in Therem 2.2. We require that the additinal assumptins stated befre the algrithm fr estimatin f the parameters f this mdel hld. An apprximate nnsingular estimate f the cvariance matrix Λ f the p dimensinal vectr variate randm variable can be btained by using the estimatrs frm this algrithm and using ˆΛ = (Â1 i  2 i  i )(Â1 i  2 i  i ) That we d nt have t assume any cvariance cmpnents are zer is the main difference and advantage f this regularizatin methd t the usual shrinkage methds like lass [Friedman et al., 2008]. Example 3.1. Let x N 12 (µ = 0,Λ) where Λ is the identity matrix. We illustrate slicing fr i = 2, m 1 = 3 and m 2 = 4. N = 5,10,20,50 Sets f N = 5,10,20,50 bservatins were generated and Λ was estimated using Λ = Â1 i  2 9

10 Cmpnents Cumulative cntributin f PC s N=5 N=10 N=20 N=50 Figure 2: As N gets larger the estimatrs f the eigenvalues apprach the true values. The true cvariance structure is represented with the black line. assuming the mdel in Therem 2.2. We repeated the whle experiment 5 times. The results are summarized in Figure 2. The eigenvalues cnverge t the true eigenvalues as N gets larger. Example 3.2. The Aln cln data set [Aln et al., 1999] have expressin measurements n 2000 genes and n 1 = 40 tumr tissues and n 2 = 22 nrmal tissue samples. We will cmpare the means f the nrmal and tumr tissue samples. We assume first that nrmal and tumr tissues have the same cvariance Λ, a psitive definite matrix. We slice each f the n = 62 bservatins int a matrix and estimate Λ with Λ = Â1 i  2 assuming the mdel in Therem 2.2 hlds. Fr testing the equality f the means, we calculate the F + statistic prpsed in [Kubkawa and Srivastava, 2008] replacing their estimatr 10

11 f the inverse f cvariance matrix Λ with the inverse f Λ: F + = 2000 (62 1)+1 ( 1 (62 1) ) 1 ( x 1 x 2 ) Λ 1 ( x 1 x 2 ) = Using the sampling distributin F r,n r prpsed in [Kubkawa and Srivastava, 2008] assuming that the rank r f Λ is 62 1, the p-value is calculated as Thus, the hypthesis f equality f the means is rejected. The estimate f the cvariance matrix fr cln data set is available at the url: akdemir. user. cs. ucf. edu/ cvclncancer. txt. Example 3.3. N = 10 i.i.d. bservatins frm a N 12(µ = 0,Λ), distributin are generated fr a randmly generated unstructured nnsingular cvariance matrix Λ. The right plt in Figure 3 cmpares the estimated eigenvalues btained by slicing this data int a array with the rdinary sample cvariance. True and Estimated Eigenvalues % cntributin f eigenvalues True and Estimated Eigenvalues % cntributin f eigenvalues Index Index Figure 3: The left plt in Figure 3 cmpares the estimated eigenvalues t the true eigenvalues (Example 3.4). The right plt in Figure 3 cmpares the estimated eigenvalues btained under different assumptins t the true eigenvalues (Example 3.3). The red s represent the true values, black s are fr estimates under Krnecker delta cvariance assumptin and the blue + s are fr estimates under unrestricted cvariance assumptin. Example 3.4. Let and A 1 = ( ) 1/2, A 2 = A 3 = /2. 1/2, 11

12 Als, let M be the 0 array f dimensins The fllwing are the estimates f A 1, A 2 and A 3 based n a randm sample f size 100 frm the N(A 1,A 2,A 3, M). and  1 = ( ) 1/2,  2 = /2,  3 = The left plt in Figure 3 cmpares the estimated eigenvalues t the true eigenvalues fr this example. Example 3.5. In this example, we will use the heatmap f the true and estimated cvariance matrices under different scenaris t see that slicing gives a reasnable descriptin f the variable variances and cvariances. In Figure 4 the true cvariance matrix is a identity matrix, we estimate this cvariance matrix fr N = 10,50, and 100 independent sets f randm samples by using 15 8 slicing. In Figure 5 the true cvariance matrix is a blck diagnal matrix with krnecker delta structure. Finally, in Figure 6 the true cvariance is a matrix with 4 way krnecker structure. Cnvergence f the estimatrs is bserved even when p >> N. 4 Principal Cmpnents Analysis and Dimensin Reductin Principal cmpnents analysis (PCA) is a useful statistical technique that has fund applicatins in fields such as face recgnitin and image cmpressin, and is a cmmn technique fr finding patterns in data f high dimensin. The end prduct f PCA is a set f new uncrrelated variables rdered in terms f their variances btained frm a linear cmbinatin f the riginal variables. Definitin 4.1. Fr the m 1 m 2... m i dimensinal array variate randm variable X, the principal cmpnents are defined as the principal cmpnents f the d = m 1 m 2...m i -dimensinal randm vectr rvec( X). The main statistical prblem is the estimatin f the cvariance f rvec( X), its eigenvectrs and eigenvalues fr small sample sizes. Let X l, l = 1,2,..,N be a randm sample fr the array variate randm variable X. Let p = m 1 m 2...m i. When N < p, it is well knwn that the usual cvariance estimatr fr rvec( X) will be singular with prbability ne. Therefre, when N < p, there is n cnsistent estimatr f the cvariance f rvec( X) under the unstructured cvariance assumptin. 1/2. 12

13 True Cvariance N=10 N=50 N=100 Figure 4: The true cvariance matrix is a identity matrix with krnecker delta structure. We estimate this cvariance matrix fr N = 10,50, and 100 independent sets f randm samples by using 15 8 slicing. 13

14 True Cvariance N=10 N=50 N=100 Figure 5: The true cvariance matrix is a blck diagnal matrix with krnecker delta structure. We estimate this cvariance matrix fr N = 10, 50, and 100 independent sets f randm samples by using 15 8 slicing. 14

15 True Cvariance N=10 N=50 N=100 Figure 6: The true cvariance matrix is way krnecker delta structured matrix. We estimate this cvariance matrix fr N = 10, 50, and 100 independent sets f randm samples by using 15 8 slicing. 15

16 On the ther hand, if we assume that the cvariance matrix has Krnecker delta structure, we can btain a nnsingular estimate f the cvariance structure with the methds develped in this sectin. The cnditin n the sample size is relaxed cnsiderably. If we have pn > m 2 r fr r = 1,2,...,i and the assumtins stated befre the algrithm fr estimatin f the parameters f this mdel hld, then the estimatr f the cvariance matrix is nnsingular. When the cvariance des nt have Krnecker structure, the estimate btained here culd be used as regularized nnsingular estimate f the cvariance. If {λ(a r ) rj } are the m j eigenvalues f A r A r with the crrespnding eigenvectrs {(x r ) rj } fr r = 1,2,...,i and r j = 1,2,...,m r, then (A 1 i A 2 i A i )(A 1 i A 2 i A i ) will have eigenvalues {λ(a 1 ) r1 λ(a 2 ) r2...λ(a i ) ri } with crrespnding eigenvectrs {(x 1 ) r1 i (x 2 ) r2 i... i (x i ) ri }. By replacing A r by their estimatrs, we estimate the eigenvalues and eigenvectrs f the cvariance f rvec( X) using this relatinship. Since each eigenvectr is a Krnecker prduct f smaller cmpnents the reductin in dimensin btained by this apprach is larger than the ne that culd be btained using the rdinary principle cmpnents n the rdinary sample cvariance matrix. Example 4.1. The 3-way aersl particles data n air quality was first analysed by multiway methds by [Stanimirva and Simenv, 2005]. They wanted t characterize air quality n the basis f particle size, seasnality, and chemical cmpsitin. The 3-way data cnsisting f the cncentratins f 17 chemical cmpunds plus dust fr each f 5 particle sizes in each f the 4 seasns were measured at N = 2 lcatins in Kärnten (Carynthia), Austria. The data is given in [Krnenberg and Crpratin, 2008]. Fr describing the cvariance f the randm variable using 2 bservatins, we culd assume that the data cmes frm the array variate nrmal distributin with unstructured cvariance matrix. Hwever the estimate btained under the unstructured mdel is singular with rank 1. On the ther hand, the array variate mdel with Krnecker delta cvariance structure prvides a nnsingular cvariance estimate. In Figure 7, the cumulative cntributin f the eigenvalues f each cmpnent dimensin are displayed in the first three plts. The eigenvalues f the resulting estimatr f the cvariance f randm variable is shwn in the last plt. In Figure 8, this dimensinal aersl data at tw lcatins is summarized by their prjectin n the first 3 estimated principal cmpnents. 5 Classificatin Suppse the array variable is generated by a mixture f tw densities, i.e., X πn(a 1,A 2,...,A i,m 1 )+(1 π)n(b 1,B 2,...,B i,m 2 ). Based n a set f training bservatins, the classificatin f a new bservatin t the cmpnent density N(A 1,A 2,...,A i,m 1 ) r N(B 1,B 2,...,B i,m 2 ) can be dne using the Bayes rule tgether with the estimatrs btained in Sectin 3. Let the training estimates f parameters be π, Â1, Â2,..., Âi, M 1, B 1, B 2, 16

17 % cntributin f eigenvalues % cntributin f eigenvalues % cntributin f eigenvalues % cntributin f eigenvalues Index Index Index Index Figure 7: The cumulative cntributin f the eigenvalues f each cmpnent dimensin are displayed in the first three plts. The eigenvalues f the resulting estimate f the cvariance f randm variable is shwn in the last plt. First 50 cmpnents ut f the 340 cmpnents explain mre than 90% f the variatin in the aersl data. 17

18 Prjectin f the Aersl Measurements at Tw Lcatins c c2 0 5 c1 Figure 8: In Figure 8, the aersl data at the tw lcatins is summarized by their prjectin n the first 3 estimated principal cmpnents...., B i, M 2. Then, the psterir prbability f an bservatin with value X t cme frm the first cmpnent N(A 1,A 2,...,A i,m 1 ) is given by πφ( X;Â1 P( X N(A 1,A 2,...,A i,m 1 )) =,Â2,...,Âi, M 1 ) πφ( X;Â1,Â2,...,Âi, M 1 ) + (1 π)φ( X; B 1, B 2,..., B i, M. (5) 2 ) Accrding t this we wuld classify the bservatin X t the first cmpnent if the psterir prbability is large (fr example if the prbability is mre than 0.5), therwise we classify it t the secnd cmpnent. The extensin t the case f mre than tw cmpnent densities is straightfrward. Example 5.1. We have used the Fisher s linear discriminant analysis (i.e., π =.5, A j = B j fr j = 1,2,...,i) fr the Aln cln data set [Aln et al., 1999]. The linear discriminant functin was calculated using w = Λ 1 ( x 1 x 2 ) where Λ is the cvariance estimate frm Example 3.2. An bservatin x was classified as nrmal if x w > 0, therwise as tumr. Figure 9 summarizes ur findings. Misclassificatin rate is %

19 Linear Discriminatin fr Aln Cancer Data fr nrmal fr tumr w x Classifed as Nrmal Classified as Tumr Observatin Number Figure 9: Linear discriminant analysis fr the Aln cln data set [Aln et al., 1999]. An bservatin x was classified as nrmal if x w > 0, therwise as tumr. Misclassificatin rate is %

20 References U. Aln, N. Barkai, DA Ntterman, K. Gish, S. Ybarra, D. Mack, and AJ Levine. Brad Patterns f Gene Expressin Revealed by Clustering Analysis f Tumr and Nrmal Cln Tissues Prbed by Olignucletide Arrays. Prceedings f the Natinal Academy f Sciences f the United States f America, 96(12):6745, G. Blaha. A Few Basic Principles and Techniques f Array Algebra. Jurnal f Gedesy, 51(3): , J. Friedman, T. Hastie, and R. Tibshirani. Sparse Inverse Cvariance Estimatin with the Graphical Lass. Bistatistics, 9(3):432, A.K. Gupta and D.K. Nagar. Matrix Variate Distributins. Chapman Hall/CRC Mngraphs and Surveys in Pure and Applied Mathematics. Chapman & Hall, P.M. Krnenberg and Ebks Crpratin. Applied Multiway Data Analysis. Wiley Online Library, M. Krzyśk and M. Skrzybut. Discriminant Analysis f Multivariate Repeated Measures Data with a Krnecker Prduct Structured Cvariance Matrices. Statistical Papers, 50(4): , T. Kubkawa and M.S. Srivastava. Estimatin f the Precisin Matrix f a Singular Wishart Distributin and Its Applicatin in High-Dimensinal Data. Jurnal f Multivariate Analysis, 99(9): , N. Lu and D.L. Zimmerman. The Likelihd Rati Test fr a Separable Cvariance Matrix. Statistics & Prbability Letters, 73(4): , U.A. Rauhala. Array Algebra with Applicatins in Phtgrammetry and Gedesy. Divisin f Phtgrammetry, Ryal Institute f Technlgy, U.A. Rauhala. Intrductin t Array Algebra. Phtgrammetric Engineering and Remte Sensing, 46(2): , A. Ry and R. Khattree. Tests fr Mean and Cvariance Structures Relevant in Repeated Measures Based Discriminant Analysis. Jurnal f Applied Statistical Science, 12(2):91 104, A. Ry and R. Leiva. Likelihd Rati Tests fr Triply Multivariate Data with Structured Crrelatin n Spatial Repeated Measurements. Statistics & Prbability Letters, 78(13): , A. Ry and R. Leiva. Classificatin Rules fr Multivariate Repeated Measures Data with Equicrrelated Crrelatin Structure n bth Time and Spatial Repeated Measurements. UTSA, Cllege f Business,

21 M.S. Srivastava, T. vn Rsen, and D. Vn Rsen. Mdels with a Krnecker Prduct Cvariance Structure: Estimatin and Testing. Mathematical Methds f Statistics, 17(4): , I. Stanimirva and V. Simenv. Mdeling f Envirnmental Fur-Way Data frm Air Quality Cntrl. Chemmetrics and Intelligent Labratry Systems, 77(1-2): ,

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) > Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);

More information

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction T-61.5060 Algrithmic methds fr data mining Slide set 6: dimensinality reductin reading assignment LRU bk: 11.1 11.3 PCA tutrial in mycurses (ptinal) ptinal: An Elementary Prf f a Therem f Jhnsn and Lindenstrauss,

More information

Principal Components

Principal Components Principal Cmpnents Suppse we have N measurements n each f p variables X j, j = 1,..., p. There are several equivalent appraches t principal cmpnents: Given X = (X 1,... X p ), prduce a derived (and small)

More information

Distributions, spatial statistics and a Bayesian perspective

Distributions, spatial statistics and a Bayesian perspective Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics

More information

A Simple Set of Test Matrices for Eigenvalue Programs*

A Simple Set of Test Matrices for Eigenvalue Programs* Simple Set f Test Matrices fr Eigenvalue Prgrams* By C. W. Gear** bstract. Sets f simple matrices f rder N are given, tgether with all f their eigenvalues and right eigenvectrs, and simple rules fr generating

More information

Revision: August 19, E Main Suite D Pullman, WA (509) Voice and Fax

Revision: August 19, E Main Suite D Pullman, WA (509) Voice and Fax .7.4: Direct frequency dmain circuit analysis Revisin: August 9, 00 5 E Main Suite D Pullman, WA 9963 (509) 334 6306 ice and Fax Overview n chapter.7., we determined the steadystate respnse f electrical

More information

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551

More information

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems.

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems. Building t Transfrmatins n Crdinate Axis Grade 5: Gemetry Graph pints n the crdinate plane t slve real-wrld and mathematical prblems. 5.G.1. Use a pair f perpendicular number lines, called axes, t define

More information

A Matrix Representation of Panel Data

A Matrix Representation of Panel Data web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins

More information

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur Cdewrd Distributin fr Frequency Sensitive Cmpetitive Learning with One Dimensinal Input Data Aristides S. Galanpuls and Stanley C. Ahalt Department f Electrical Engineering The Ohi State University Abstract

More information

Admissibility Conditions and Asymptotic Behavior of Strongly Regular Graphs

Admissibility Conditions and Asymptotic Behavior of Strongly Regular Graphs Admissibility Cnditins and Asympttic Behavir f Strngly Regular Graphs VASCO MOÇO MANO Department f Mathematics University f Prt Oprt PORTUGAL vascmcman@gmailcm LUÍS ANTÓNIO DE ALMEIDA VIEIRA Department

More information

Lecture 10, Principal Component Analysis

Lecture 10, Principal Component Analysis Principal Cmpnent Analysis Lecture 10, Principal Cmpnent Analysis Ha Helen Zhang Fall 2017 Ha Helen Zhang Lecture 10, Principal Cmpnent Analysis 1 / 16 Principal Cmpnent Analysis Lecture 10, Principal

More information

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm

More information

, which yields. where z1. and z2

, which yields. where z1. and z2 The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin

More information

NUMBERS, MATHEMATICS AND EQUATIONS

NUMBERS, MATHEMATICS AND EQUATIONS AUSTRALIAN CURRICULUM PHYSICS GETTING STARTED WITH PHYSICS NUMBERS, MATHEMATICS AND EQUATIONS An integral part t the understanding f ur physical wrld is the use f mathematical mdels which can be used t

More information

Module 4: General Formulation of Electric Circuit Theory

Module 4: General Formulation of Electric Circuit Theory Mdule 4: General Frmulatin f Electric Circuit Thery 4. General Frmulatin f Electric Circuit Thery All electrmagnetic phenmena are described at a fundamental level by Maxwell's equatins and the assciated

More information

7 TH GRADE MATH STANDARDS

7 TH GRADE MATH STANDARDS ALGEBRA STANDARDS Gal 1: Students will use the language f algebra t explre, describe, represent, and analyze number expressins and relatins 7 TH GRADE MATH STANDARDS 7.M.1.1: (Cmprehensin) Select, use,

More information

Lyapunov Stability Stability of Equilibrium Points

Lyapunov Stability Stability of Equilibrium Points Lyapunv Stability Stability f Equilibrium Pints 1. Stability f Equilibrium Pints - Definitins In this sectin we cnsider n-th rder nnlinear time varying cntinuus time (C) systems f the frm x = f ( t, x),

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 4: Mdel checing fr ODE mdels In Petre Department f IT, Åb Aademi http://www.users.ab.fi/ipetre/cmpmd/ Cntent Stichimetric matrix Calculating the mass cnservatin relatins

More information

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression 3.3.4 Prstate Cancer Data Example (Cntinued) 3.4 Shrinkage Methds 61 Table 3.3 shws the cefficients frm a number f different selectin and shrinkage methds. They are best-subset selectin using an all-subsets

More information

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression 4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw

More information

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline

More information

Sections 15.1 to 15.12, 16.1 and 16.2 of the textbook (Robbins-Miller) cover the materials required for this topic.

Sections 15.1 to 15.12, 16.1 and 16.2 of the textbook (Robbins-Miller) cover the materials required for this topic. Tpic : AC Fundamentals, Sinusidal Wavefrm, and Phasrs Sectins 5. t 5., 6. and 6. f the textbk (Rbbins-Miller) cver the materials required fr this tpic.. Wavefrms in electrical systems are current r vltage

More information

On Huntsberger Type Shrinkage Estimator for the Mean of Normal Distribution ABSTRACT INTRODUCTION

On Huntsberger Type Shrinkage Estimator for the Mean of Normal Distribution ABSTRACT INTRODUCTION Malaysian Jurnal f Mathematical Sciences 4(): 7-4 () On Huntsberger Type Shrinkage Estimatr fr the Mean f Nrmal Distributin Department f Mathematical and Physical Sciences, University f Nizwa, Sultanate

More information

Chapter 9 Vector Differential Calculus, Grad, Div, Curl

Chapter 9 Vector Differential Calculus, Grad, Div, Curl Chapter 9 Vectr Differential Calculus, Grad, Div, Curl 9.1 Vectrs in 2-Space and 3-Space 9.2 Inner Prduct (Dt Prduct) 9.3 Vectr Prduct (Crss Prduct, Outer Prduct) 9.4 Vectr and Scalar Functins and Fields

More information

MATHEMATICS SYLLABUS SECONDARY 5th YEAR

MATHEMATICS SYLLABUS SECONDARY 5th YEAR Eurpean Schls Office f the Secretary-General Pedaggical Develpment Unit Ref. : 011-01-D-8-en- Orig. : EN MATHEMATICS SYLLABUS SECONDARY 5th YEAR 6 perid/week curse APPROVED BY THE JOINT TEACHING COMMITTEE

More information

Resampling Methods. Chapter 5. Chapter 5 1 / 52

Resampling Methods. Chapter 5. Chapter 5 1 / 52 Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and

More information

Interference is when two (or more) sets of waves meet and combine to produce a new pattern.

Interference is when two (or more) sets of waves meet and combine to produce a new pattern. Interference Interference is when tw (r mre) sets f waves meet and cmbine t prduce a new pattern. This pattern can vary depending n the riginal wave directin, wavelength, amplitude, etc. The tw mst extreme

More information

Tree Structured Classifier

Tree Structured Classifier Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients

More information

A NOTE ON THE EQUIVAImCE OF SOME TEST CRITERIA. v. P. Bhapkar. University of Horth Carolina. and

A NOTE ON THE EQUIVAImCE OF SOME TEST CRITERIA. v. P. Bhapkar. University of Horth Carolina. and ~ A NOTE ON THE EQUVAmCE OF SOME TEST CRTERA by v. P. Bhapkar University f Hrth Carlina University f Pna nstitute f Statistics Mime Series N. 421 February 1965 This research was supprted by the Mathematics

More information

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Céline Ferré The Wrld Bank When can we use matching? What if the assignment t the treatment is nt dne randmly r based n an eligibility index, but n the basis

More information

Determining the Accuracy of Modal Parameter Estimation Methods

Determining the Accuracy of Modal Parameter Estimation Methods Determining the Accuracy f Mdal Parameter Estimatin Methds by Michael Lee Ph.D., P.E. & Mar Richardsn Ph.D. Structural Measurement Systems Milpitas, CA Abstract The mst cmmn type f mdal testing system

More information

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use

More information

Equilibrium of Stress

Equilibrium of Stress Equilibrium f Stress Cnsider tw perpendicular planes passing thrugh a pint p. The stress cmpnents acting n these planes are as shwn in ig. 3.4.1a. These stresses are usuall shwn tgether acting n a small

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Support-Vector Machines

Support-Vector Machines Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material

More information

Math 302 Learning Objectives

Math 302 Learning Objectives Multivariable Calculus (Part I) 13.1 Vectrs in Three-Dimensinal Space Math 302 Learning Objectives Plt pints in three-dimensinal space. Find the distance between tw pints in three-dimensinal space. Write

More information

Emphases in Common Core Standards for Mathematical Content Kindergarten High School

Emphases in Common Core Standards for Mathematical Content Kindergarten High School Emphases in Cmmn Cre Standards fr Mathematical Cntent Kindergarten High Schl Cntent Emphases by Cluster March 12, 2012 Describes cntent emphases in the standards at the cluster level fr each grade. These

More information

ChE 471: LECTURE 4 Fall 2003

ChE 471: LECTURE 4 Fall 2003 ChE 47: LECTURE 4 Fall 003 IDEL RECTORS One f the key gals f chemical reactin engineering is t quantify the relatinship between prductin rate, reactr size, reactin kinetics and selected perating cnditins.

More information

The blessing of dimensionality for kernel methods

The blessing of dimensionality for kernel methods fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

ECE 2100 Circuit Analysis

ECE 2100 Circuit Analysis ECE 2100 Circuit Analysis Lessn 25 Chapter 9 & App B: Passive circuit elements in the phasr representatin Daniel M. Litynski, Ph.D. http://hmepages.wmich.edu/~dlitynsk/ ECE 2100 Circuit Analysis Lessn

More information

AP Statistics Notes Unit Two: The Normal Distributions

AP Statistics Notes Unit Two: The Normal Distributions AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).

More information

Surface and Contact Stress

Surface and Contact Stress Surface and Cntact Stress The cncept f the frce is fundamental t mechanics and many imprtant prblems can be cast in terms f frces nly, fr example the prblems cnsidered in Chapter. Hwever, mre sphisticated

More information

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001)

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001) CN700 Additive Mdels and Trees Chapter 9: Hastie et al. (2001) Madhusudana Shashanka Department f Cgnitive and Neural Systems Bstn University CN700 - Additive Mdels and Trees March 02, 2004 p.1/34 Overview

More information

Chapter 3 Kinematics in Two Dimensions; Vectors

Chapter 3 Kinematics in Two Dimensions; Vectors Chapter 3 Kinematics in Tw Dimensins; Vectrs Vectrs and Scalars Additin f Vectrs Graphical Methds (One and Tw- Dimensin) Multiplicatin f a Vectr b a Scalar Subtractin f Vectrs Graphical Methds Adding Vectrs

More information

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa There are tw parts t this lab. The first is intended t demnstrate hw t request and interpret the spatial diagnstics f a standard OLS regressin mdel using GeDa. The diagnstics prvide infrmatin abut the

More information

5 th grade Common Core Standards

5 th grade Common Core Standards 5 th grade Cmmn Cre Standards In Grade 5, instructinal time shuld fcus n three critical areas: (1) develping fluency with additin and subtractin f fractins, and develping understanding f the multiplicatin

More information

Methods for Determination of Mean Speckle Size in Simulated Speckle Pattern

Methods for Determination of Mean Speckle Size in Simulated Speckle Pattern 0.478/msr-04-004 MEASUREMENT SCENCE REVEW, Vlume 4, N. 3, 04 Methds fr Determinatin f Mean Speckle Size in Simulated Speckle Pattern. Hamarvá, P. Šmíd, P. Hrváth, M. Hrabvský nstitute f Physics f the Academy

More information

Source Coding and Compression

Source Coding and Compression Surce Cding and Cmpressin Heik Schwarz Cntact: Dr.-Ing. Heik Schwarz heik.schwarz@hhi.fraunhfer.de Heik Schwarz Surce Cding and Cmpressin September 22, 2013 1 / 60 PartI: Surce Cding Fundamentals Heik

More information

^YawataR&D Laboratory, Nippon Steel Corporation, Tobata, Kitakyushu, Japan

^YawataR&D Laboratory, Nippon Steel Corporation, Tobata, Kitakyushu, Japan Detectin f fatigue crack initiatin frm a ntch under a randm lad C. Makabe," S. Nishida^C. Urashima,' H. Kaneshir* "Department f Mechanical Systems Engineering, University f the Ryukyus, Nishihara, kinawa,

More information

Dead-beat controller design

Dead-beat controller design J. Hetthéssy, A. Barta, R. Bars: Dead beat cntrller design Nvember, 4 Dead-beat cntrller design In sampled data cntrl systems the cntrller is realised by an intelligent device, typically by a PLC (Prgrammable

More information

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must M.E. Aggune, M.J. Dambrg, M.A. El-Sharkawi, R.J. Marks II and L.E. Atlas, "Dynamic and static security assessment f pwer systems using artificial neural netwrks", Prceedings f the NSF Wrkshp n Applicatins

More information

22.54 Neutron Interactions and Applications (Spring 2004) Chapter 11 (3/11/04) Neutron Diffusion

22.54 Neutron Interactions and Applications (Spring 2004) Chapter 11 (3/11/04) Neutron Diffusion .54 Neutrn Interactins and Applicatins (Spring 004) Chapter (3//04) Neutrn Diffusin References -- J. R. Lamarsh, Intrductin t Nuclear Reactr Thery (Addisn-Wesley, Reading, 966) T study neutrn diffusin

More information

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017 Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with

More information

Least Squares Optimal Filtering with Multirate Observations

Least Squares Optimal Filtering with Multirate Observations Prc. 36th Asilmar Cnf. n Signals, Systems, and Cmputers, Pacific Grve, CA, Nvember 2002 Least Squares Optimal Filtering with Multirate Observatins Charles W. herrien and Anthny H. Hawes Department f Electrical

More information

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards: MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use

More information

February 28, 2013 COMMENTS ON DIFFUSION, DIFFUSIVITY AND DERIVATION OF HYPERBOLIC EQUATIONS DESCRIBING THE DIFFUSION PHENOMENA

February 28, 2013 COMMENTS ON DIFFUSION, DIFFUSIVITY AND DERIVATION OF HYPERBOLIC EQUATIONS DESCRIBING THE DIFFUSION PHENOMENA February 28, 2013 COMMENTS ON DIFFUSION, DIFFUSIVITY AND DERIVATION OF HYPERBOLIC EQUATIONS DESCRIBING THE DIFFUSION PHENOMENA Mental Experiment regarding 1D randm walk Cnsider a cntainer f gas in thermal

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.

More information

Aerodynamic Separability in Tip Speed Ratio and Separability in Wind Speed- a Comparison

Aerodynamic Separability in Tip Speed Ratio and Separability in Wind Speed- a Comparison Jurnal f Physics: Cnference Series OPEN ACCESS Aerdynamic Separability in Tip Speed Rati and Separability in Wind Speed- a Cmparisn T cite this article: M L Gala Sants et al 14 J. Phys.: Cnf. Ser. 555

More information

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal

More information

UNIV1"'RSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION

UNIV1'RSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION UNIV1"'RSITY OF NORTH CAROLINA Department f Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION by N. L. Jlmsn December 1962 Grant N. AFOSR -62..148 Methds f

More information

Eric Klein and Ning Sa

Eric Klein and Ning Sa Week 12. Statistical Appraches t Netwrks: p1 and p* Wasserman and Faust Chapter 15: Statistical Analysis f Single Relatinal Netwrks There are fur tasks in psitinal analysis: 1) Define Equivalence 2) Measure

More information

Introduction: A Generalized approach for computing the trajectories associated with the Newtonian N Body Problem

Introduction: A Generalized approach for computing the trajectories associated with the Newtonian N Body Problem A Generalized apprach fr cmputing the trajectries assciated with the Newtnian N Bdy Prblem AbuBar Mehmd, Syed Umer Abbas Shah and Ghulam Shabbir Faculty f Engineering Sciences, GIK Institute f Engineering

More information

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares

More information

The multivariate skew-slash distribution

The multivariate skew-slash distribution Jurnal f Statistical Planning and Inference 136 (6) 9 wwwelseviercm/lcate/jspi The multivariate skew-slash distributin Jing Wang, Marc G Gentn Department f Statistics, Nrth Carlina State University, Bx

More information

B. Definition of an exponential

B. Definition of an exponential Expnents and Lgarithms Chapter IV - Expnents and Lgarithms A. Intrductin Starting with additin and defining the ntatins fr subtractin, multiplicatin and divisin, we discvered negative numbers and fractins.

More information

Homology groups of disks with holes

Homology groups of disks with holes Hmlgy grups f disks with hles THEOREM. Let p 1,, p k } be a sequence f distinct pints in the interir unit disk D n where n 2, and suppse that fr all j the sets E j Int D n are clsed, pairwise disjint subdisks.

More information

GRAPH EFFECTIVE RESISTANCE AND DISTRIBUTED CONTROL: SPECTRAL PROPERTIES AND APPLICATIONS

GRAPH EFFECTIVE RESISTANCE AND DISTRIBUTED CONTROL: SPECTRAL PROPERTIES AND APPLICATIONS GRAPH EFFECTIVE RESISTANCE AND DISTRIBUTED CONTROL: SPECTRAL PROPERTIES AND APPLICATIONS Prabir Barah Jã P. Hespanha Abstract We intrduce the cncept f matrix-valued effective resistance fr undirected matrix-weighted

More information

Part 3 Introduction to statistical classification techniques

Part 3 Introduction to statistical classification techniques Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms

More information

Curriculum Development Overview Unit Planning for 8 th Grade Mathematics MA10-GR.8-S.1-GLE.1 MA10-GR.8-S.4-GLE.2

Curriculum Development Overview Unit Planning for 8 th Grade Mathematics MA10-GR.8-S.1-GLE.1 MA10-GR.8-S.4-GLE.2 Unit Title It s All Greek t Me Length f Unit 5 weeks Fcusing Lens(es) Cnnectins Standards and Grade Level Expectatins Addressed in this Unit MA10-GR.8-S.1-GLE.1 MA10-GR.8-S.4-GLE.2 Inquiry Questins (Engaging-

More information

SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A.

SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A. SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST Mark C. Ott Statistics Research Divisin, Bureau f the Census Washingtn, D.C. 20233, U.S.A. and Kenneth H. Pllck Department f Statistics, Nrth Carlina State

More information

Perturbation approach applied to the asymptotic study of random operators.

Perturbation approach applied to the asymptotic study of random operators. Perturbatin apprach applied t the asympttic study f rm peratrs. André MAS, udvic MENNETEAU y Abstract We prve that, fr the main mdes f stchastic cnvergence (law f large numbers, CT, deviatins principles,

More information

Pattern Recognition 2014 Support Vector Machines

Pattern Recognition 2014 Support Vector Machines Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft

More information

SAMPLING DYNAMICAL SYSTEMS

SAMPLING DYNAMICAL SYSTEMS SAMPLING DYNAMICAL SYSTEMS Melvin J. Hinich Applied Research Labratries The University f Texas at Austin Austin, TX 78713-8029, USA (512) 835-3278 (Vice) 835-3259 (Fax) hinich@mail.la.utexas.edu ABSTRACT

More information

The standards are taught in the following sequence.

The standards are taught in the following sequence. B L U E V A L L E Y D I S T R I C T C U R R I C U L U M MATHEMATICS Third Grade In grade 3, instructinal time shuld fcus n fur critical areas: (1) develping understanding f multiplicatin and divisin and

More information

1 The limitations of Hartree Fock approximation

1 The limitations of Hartree Fock approximation Chapter: Pst-Hartree Fck Methds - I The limitatins f Hartree Fck apprximatin The n electrn single determinant Hartree Fck wave functin is the variatinal best amng all pssible n electrn single determinants

More information

Kinetic Model Completeness

Kinetic Model Completeness 5.68J/10.652J Spring 2003 Lecture Ntes Tuesday April 15, 2003 Kinetic Mdel Cmpleteness We say a chemical kinetic mdel is cmplete fr a particular reactin cnditin when it cntains all the species and reactins

More information

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation III-l III. A New Evaluatin Measure J. Jiner and L. Werner Abstract The prblems f evaluatin and the needed criteria f evaluatin measures in the SMART system f infrmatin retrieval are reviewed and discussed.

More information

Bayesian nonparametric modeling approaches for quantile regression

Bayesian nonparametric modeling approaches for quantile regression Bayesian nnparametric mdeling appraches fr quantile regressin Athanasis Kttas Department f Applied Mathematics and Statistics University f Califrnia, Santa Cruz Department f Statistics Athens University

More information

MODULAR DECOMPOSITION OF THE NOR-TSUM MULTIPLE-VALUED PLA

MODULAR DECOMPOSITION OF THE NOR-TSUM MULTIPLE-VALUED PLA MODUAR DECOMPOSITION OF THE NOR-TSUM MUTIPE-AUED PA T. KAGANOA, N. IPNITSKAYA, G. HOOWINSKI k Belarusian State University f Infrmatics and Radielectrnics, abratry f Image Prcessing and Pattern Recgnitin.

More information

Marginal Conceptual Predictive Statistic for Mixed Model Selection

Marginal Conceptual Predictive Statistic for Mixed Model Selection Open Jurnal f Statistics, 06, 6, 39-53 Published Online April 06 in Sci http://wwwscirprg/jurnal/js http://dxdirg/0436/js0660 Marginal Cnceptual Predictive Statistic fr Mixed Mdel Selectin Cheng Wenren,

More information

Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart

Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Sandy D. Balkin Dennis K. J. Lin y Pennsylvania State University, University Park, PA 16802 Sandy Balkin is a graduate student

More information

CHAPTER 2 Algebraic Expressions and Fundamental Operations

CHAPTER 2 Algebraic Expressions and Fundamental Operations CHAPTER Algebraic Expressins and Fundamental Operatins OBJECTIVES: 1. Algebraic Expressins. Terms. Degree. Gruping 5. Additin 6. Subtractin 7. Multiplicatin 8. Divisin Algebraic Expressin An algebraic

More information

NOTE ON THE ANALYSIS OF A RANDOMIZED BLOCK DESIGN. Junjiro Ogawa University of North Carolina

NOTE ON THE ANALYSIS OF A RANDOMIZED BLOCK DESIGN. Junjiro Ogawa University of North Carolina NOTE ON THE ANALYSIS OF A RANDOMIZED BLOCK DESIGN by Junjir Ogawa University f Nrth Carlina This research was supprted by the Office f Naval Research under Cntract N. Nnr-855(06) fr research in prbability

More information

Chapter 2 GAUSS LAW Recommended Problems:

Chapter 2 GAUSS LAW Recommended Problems: Chapter GAUSS LAW Recmmended Prblems: 1,4,5,6,7,9,11,13,15,18,19,1,7,9,31,35,37,39,41,43,45,47,49,51,55,57,61,6,69. LCTRIC FLUX lectric flux is a measure f the number f electric filed lines penetrating

More information

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse

More information

Margin Distribution and Learning Algorithms

Margin Distribution and Learning Algorithms ICML 03 Margin Distributin and Learning Algrithms Ashutsh Garg IBM Almaden Research Center, San Jse, CA 9513 USA Dan Rth Department f Cmputer Science, University f Illinis, Urbana, IL 61801 USA ASHUTOSH@US.IBM.COM

More information

Smoothing, penalized least squares and splines

Smoothing, penalized least squares and splines Smthing, penalized least squares and splines Duglas Nychka, www.image.ucar.edu/~nychka Lcally weighted averages Penalized least squares smthers Prperties f smthers Splines and Reprducing Kernels The interplatin

More information

Differentiation Applications 1: Related Rates

Differentiation Applications 1: Related Rates Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm

More information

Particle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD

Particle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD 3 J. Appl. Cryst. (1988). 21,3-8 Particle Size Distributins frm SANS Data Using the Maximum Entrpy Methd By J. A. PTTN, G. J. DANIELL AND B. D. RAINFRD Physics Department, The University, Suthamptn S9

More information

Kinematic transformation of mechanical behavior Neville Hogan

Kinematic transformation of mechanical behavior Neville Hogan inematic transfrmatin f mechanical behavir Neville Hgan Generalized crdinates are fundamental If we assume that a linkage may accurately be described as a cllectin f linked rigid bdies, their generalized

More information

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA Mdelling f Clck Behaviur Dn Percival Applied Physics Labratry University f Washingtn Seattle, Washingtn, USA verheads and paper fr talk available at http://faculty.washingtn.edu/dbp/talks.html 1 Overview

More information

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS 1 Influential bservatins are bservatins whse presence in the data can have a distrting effect n the parameter estimates and pssibly the entire analysis,

More information

Multiple Source Multiple. using Network Coding

Multiple Source Multiple. using Network Coding Multiple Surce Multiple Destinatin Tplgy Inference using Netwrk Cding Pegah Sattari EECS, UC Irvine Jint wrk with Athina Markpulu, at UCI, Christina Fraguli, at EPFL, Lausanne Outline Netwrk Tmgraphy Gal,

More information

MODULE 1. e x + c. [You can t separate a demominator, but you can divide a single denominator into each numerator term] a + b a(a + b)+1 = a + b

MODULE 1. e x + c. [You can t separate a demominator, but you can divide a single denominator into each numerator term] a + b a(a + b)+1 = a + b . REVIEW OF SOME BASIC ALGEBRA MODULE () Slving Equatins Yu shuld be able t slve fr x: a + b = c a d + e x + c and get x = e(ba +) b(c a) d(ba +) c Cmmn mistakes and strategies:. a b + c a b + a c, but

More information

IN a recent article, Geary [1972] discussed the merit of taking first differences

IN a recent article, Geary [1972] discussed the merit of taking first differences The Efficiency f Taking First Differences in Regressin Analysis: A Nte J. A. TILLMAN IN a recent article, Geary [1972] discussed the merit f taking first differences t deal with the prblems that trends

More information

David HORN and Irit OPHER. School of Physics and Astronomy. Raymond and Beverly Sackler Faculty of Exact Sciences

David HORN and Irit OPHER. School of Physics and Astronomy. Raymond and Beverly Sackler Faculty of Exact Sciences Cmplex Dynamics f Neurnal Threshlds David HORN and Irit OPHER Schl f Physics and Astrnmy Raymnd and Beverly Sackler Faculty f Exact Sciences Tel Aviv University, Tel Aviv 69978, Israel hrn@neurn.tau.ac.il

More information

Lim f (x) e. Find the largest possible domain and its discontinuity points. Why is it discontinuous at those points (if any)?

Lim f (x) e. Find the largest possible domain and its discontinuity points. Why is it discontinuous at those points (if any)? THESE ARE SAMPLE QUESTIONS FOR EACH OF THE STUDENT LEARNING OUTCOMES (SLO) SET FOR THIS COURSE. SLO 1: Understand and use the cncept f the limit f a functin i. Use prperties f limits and ther techniques,

More information

LHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers

LHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers LHS Mathematics Department Hnrs Pre-alculus Final Eam nswers Part Shrt Prblems The table at the right gives the ppulatin f Massachusetts ver the past several decades Using an epnential mdel, predict the

More information