On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -1 Bing / 38Li,

Size: px
Start display at page:

Download "On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -1 Bing / 38Li,"

Transcription

1 On an Additive Semigraphoid Model for Statistical Networks With Application to Pathway Analysis - Bing Li, Hyunho Chun & Hongyu Zhao Kim Youngrae SNU Stat. Multivariate Lab Nov 25, 2016 On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -1 Bing / 38Li,

2 Outline 1 Introduction GGM, GCGM 2 Additive conditional independent Graphoid, semigraphoid Conditional independent, additively conditional independent Relation with previous models 3 Estimation ACCO, APO Calculation Tuning parameter 4 Simultation results On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -2 Bing / 38Li,

3 1. Introduction Additive semigraphoid model Gaussian copula graphical model Additively conditional independent Additive conditional covariance Operator Additive precision operator On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -3 Bing / 38Li,

4 1.1 GGM, GCGM Gaussian graphical model Let X = (X 1,..., X p ) T, Suppose X N(µ, Σ), and Σ is postive definite. Let Θ = Σ 1, Θ ij be the (i, j)th elements of Θ. Then X follows a GGM with respect to the graph G = {Γ, E} iff X i X j X (i,j), (i, j) E Multivariate normal assumption. In Gaussian graphical model, X i X j X (i,j) iff Θ ij = 0. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -4 Bing / 38Li,

5 1.1 GGM, GCGM (Cont.) Copula Copula is multivariate probability distribution for which the marginal probability of each variable is uniform. It can deal with the variables with dependency. C [0, 1] d [0, 1], is a d dimensional copula if C is cumulative distribution function of a d dimensional random vector on the unit cube On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -5 Bing / 38Li,

6 1.1 GGM, GCGM (Cont.) Gaussian copula graphical model Let X = (X 1,..., X p ) T, Suppose the existence of unknown injections f 1,..., f p such that (f 1 (X 1 ),..., f p (X p )) T is multivariate gaussian. (f 1 (X 1 ),..., f p (X p )) T N(µ, Σ) Each X i shouldn t be a Gaussian. We can use all of GGM procedures for non-gaussian X i s. But the Gaussian copula assumpotion can be violated even for commonly used interactions. ex. marginally normal doesn t imply jointly normal. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -6 Bing / 38Li,

7 2.1 Graphoid, semigraphoid Graphoid Axioms A three-way relation R is called a graphoid if it satisfies the following conditions. Symmetry: (A, C, B) R (B, C, A) R Decomposition: (A, C, B D) R (A, C, B) R Weak Union: (A, C, B D) R (A, C B, D) R Contraction: (A, C B, D) R, (A, C, B) R (A, C, B D) R Intersection: (A, C D, B) R, (A, C B, D) (A, C, B D) R Conditional indenpendent satisfies this axioms. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -7 Bing / 38Li,

8 2.1 Graphoid, semigraphoid (Cont.) Semigraphoid Axioms A three-way relation R is called a semigraphoid if it satisfies the conditions of Graphoid axioms except Intersection condition. Symmetry: (A, C, B) R (B, C, A) R Decomposition: (A, C, B D) R (A, C, B) R Weak Union: (A, C, B D) R (A, C B, D) R Contraction: (A, C B, D) R, (A, C, B) R (A, C, B D) R Using this semigraphoid relation in place of probabilistic conditional independence can free us from some awkward restrictions. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -8 Bing / 38Li,

9 2.2 Conditional independent, additively conditional independent Notations Let X(X 1,..., X p ) T be a random vector. For subvector U = (U 1,..., U p ) T of X, let L 2 (P U ) be the class of functions of U such that Ef(U) = 0, Ef 2 (U) <. For each U i, let A U i denote a subset of L 2 (P U i). Let A U denote the addtive family A U A U r = {f f r f 1 A U 1,..., f r A U r} In this paper, inner product <.,. > represents the L 2 (P X ) inner product. < f, g >= f(x)g(x)dµ(x) On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -9 Bing / 38Li,

10 2.2 Conditional independent, addtively conditional independent (Cont.) Additively conditional independent We say that U and V are additively conditional independent (ACI) give W iff (A U + A W ) A W (A V + A W ) A W And we write this relation as U A V W. Additively conditional independent is three-way relation. U A V W (A U, A W, A V ) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 10 - Bing / 38Li,

11 2.2 Conditional independent, addtively conditional independent (Cont.) Theorem The additive conditional independece relation is a semigraphoid. Additive semigraphoid model (ASG model) A random vector X follows and ASG model with respect to a graph G = (Γ, E) iff X i A X j X (i,j) (i, j) E. We write this condition as X ASG(G). On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 11 - Bing / 38Li,

12 2.3 Relation with previous models Now we investigate the relation between new-concept, Additive conditional independence and conditional indepdence. Theorem Suppose (a) X has a Gaussian copula distribution with copula functions f 1,..., f p (b) U, V, and W are subvectors of X (c) A X i = span{f i }, i = 1,..., p Then, U A V W iff U V W. Under Gaussian copula assumption with A X i = span{f i }, addtive conditional independence is equivalent to conditional independence. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 12 - Bing / 38Li,

13 2.3 Relation with previous models (Cont.) Theorem Suppose (a) X has a Gaussian copula distribution with copula functions f 1,..., f p (b) U, V, and W are subvectors of X (c) A X i = L 2 (P X i), i = 1,..., p Then, U A V W implies U V W. In this case, they are not equivalent. However, equivalence holds apploximately. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 13 - Bing / 38Li,

14 2.3 Relation with previous models (Cont.) The following proposition suggests implication U V W to U A V W holds approximately. Proposition Suppose Gaussian copula assumption, and WLOG, E[f i (X i )] = 0, V ar[f i (X i )] = 1. Let U = f i (X i ), V = f j (X j ), W = {f k (X k ) k i, j}, and R V W = cov(v, W ), R W W = V ar(w ), R W U = cov(w, U), ρ UV = cor(u, V ). Then, ρ UV W max{ ρ α UV R α V W (R α W W ) 1 R α W U where A α is the α fold Hadamard product. α = 1, 2,...} On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 14 - Bing / 38Li,

15 2.3 Relation with previous models (Cont.) τ UV W is upper bound of ρ UV W in previous proposition. These are calculated under assumption that is in the proposition plus U V W. The small numbers of τ UV W in the table indicate ρ UV W is approximately 0. So, U V W implies U A V W approximately. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 15 - Bing / 38Li,

16 3.1 ACCO, APO In the GGM, X i X j X (i,j) iff (i, j) E, and with precision matrix Θ, we can cut the edge easily. Building the graph in ASG, we introduce two linear operator. Additive conditional covariance operator The ACCO is operator from A V to A U such that Σ UV W = Σ UV Σ UW Σ W V. where Σ UV A V A U such that < f, Σ UV g >= E[f(U)g(V )]. We call the operator the ACCO from V to U given W On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 16 - Bing / 38Li,

17 3.1 ACCO, APO (Cont.) With this operator, there are two important results. Lemma Let P AW A X A W be the projection on to A W. Then P AW A U = Σ W U. Theorem U A V W iff Σ UV W = 0. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 17 - Bing / 38Li,

18 3.1 ACCO, APO (Cont.) Calculating ACCO is difficult because if W is changed, you should calculate again. Similar to GGM, we want to use precision matrix instead of conditional covariance. For defining such operator, we will use different inner product. Let A X be the Hilbert space consisting of the same set of functions as A X, but with the inner product < f f p, g g p > A X = p i=1 < f i, g i > AX i On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 18 - Bing / 38Li,

19 3.1 ACCO, APO (Cont.) Additive precision operator The operator Υ A X A X defined by the matrix of operator {Σ Xi Xj i Γ, j Γ} is called additive covariance operator. If it is invertible then its inverse Θ is called the APO. Υ = {Σ Xi X j i Γ, j Γ} means When Υf i = h h p, Σ Xj X i(f i) = h j. We can easily check for any f, g A X, < f, Υg > A X = cov[f(x), g(x)] =< f, g > On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 19 - Bing / 38Li,

20 3.1 ACCO, APO (Cont.) Question. Is APO well defined? Proposition Suppose Υ is invertible and, for each i j, Υ ij is compact. Then, Υ 1 is bounded. Lemma Suppose that T B(A X, A X ) is a self adjoint, positive definite operator. Then the operators T UU, T V V, T UU T UV TV 1 V T V U, T V V T V U TUU 1 T UV are bounded, self adjoint, and positive definite, and the following identities hold (T 1 ) UU = (T UU T UV T 1 V V T V U ) 1 (1) (T 1 ) V V = (T V V T V U T 1 UU T UV ) 1 (2) (T 1 ) UV = (T 1 ) UU T UV T 1 V V = T 1 UU T UV (T 1 ) V V (3) (T 1 ) V U = (T 1 ) V V T V U T 1 UU = T 1 V V T V U (T 1 ) UU (4) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 20 - Bing / 38Li,

21 3.1 ACCO, APO (Cont.) The above lemma shows that if Υ has bounded inverse, Θ is well defined. And finally we derive this equivalence. Theorem Suppose (U, V, W ) is a partition of X and Υ is invertible. Then, U A V W iff Θ UV = 0. Corollary A random vector X follows an ASG(G) model if and only if Θ Xi X j = 0 whenever (i, j) E. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 21 - Bing / 38Li,

22 3.2 Calculation First, how we build a graph with ACCO. If sample estimate of Σ Xi X j X (i,j) is small enough, we determine there is no edge between X i and X j. We should derive the finite-sample matrix representation of the operator. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 22 - Bing / 38Li,

23 3.2 Calculation (Cont.) Notations For each i = 1,..., p, let κ i Ω X i Ω X i R be a positive kernel. Let A X i = span{κ 1 (., X i 1) E n κ i (X i, X i 1),..., κ i (., X i n) E n κ i (X i, X i n)} where E n κ i (X i, X i k ) = n 1 n l=1 κ i (X i l, Xi k ). Q = I n 1 n 1 T n /n ; Projection matrix Let κ i be the vector valued function x i (κ i (x i, X i 1) E n κ i (X i, X i 1),..., κ i (x i, X i n) E n κ i (X i, X i n)) With this κ i, for any f A X i can be expressed as f = κ i [f]. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 23 - Bing / 38Li,

24 3.2 Calculation (Cont.) With these notations, we can get the followings (f(x i 1),..., f(x i n)) T = QK i [f] (5) < f, g > AX i = n 1 (QK i [f]) T (QK i [g]) = n 1 [f] T K i QK i [g] (6) where K i = {κ i (X i k, Xi l ) k, l = 1,..., n}, f, g A X i. K (i,j) denoted n(p 2) n matrix obtained by removing ith and jth block of (K 1,..., K p ) T. Lemma K i QK i [Σ Xi X j ] = K iqk j (7) K i QK i [Σ Xi X (i,j)] = K iqk T (i,j) (8) K (i,j) QK T ( i,j) [Σ X (i,j) X i] = K (i,j)qk i (9) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 24 - Bing / 38Li,

25 3.2 Calculation (Cont.) K i QK i, K (i,j) QK (i,j) are singular. We can solve it by Tychonoff regularization. [Σ X i X j ] = (K iqk i + ɛ 1 I n ) 1 K i QK j (10) [Σ Xi X (i,j)] = (K iqk i + ɛ 1 I n ) 1 K i QK T (i,j) (11) [Σ X (i,j) X i] = (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) QK i (12) We can represent [Σ Xj X i X (i,j)] = [Σ X j X i] [Σ X j X (i,j)][σ X (i,j) X i] On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 25 - Bing / 38Li,

26 3.2 Calculation (Cont.) So, we can get [Σ Xj X i X (i,j)] = [Σ X j X i] [Σ X j X (i,j)][σ X (i,j) X i] = (K j QK j + ɛ 1 I n ) 1 K j QK i (K j QK j + ɛ 1 I n ) 1 K j QK T (i,j) (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) QK i = (K j QK j + ɛ 1 I n ) 1 K j Q(I n K T (i,j) (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) )QK i On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 26 - Bing / 38Li,

27 3.2 Calculation (Cont.) By definition, Σ Xj X i X (i,j) 2 is the maximum of < Σ Xj X i X (i,j)f, Σ X j X i X (i,j)f > A X j subject to < f, f > AX j = 1. = n 1 [f] T [Σ Xj X i X (i,j)]t K j QK j [Σ Xj X i X (i,j)][f] Let v = (K i QK i ) 1/2 [f], then [f] = (K i QK i + ɛ 1 I n ) 1/2 v. The right hand side of above is v T (K i QK i + ɛ 1 I n ) 1/2 [Σ Xj X i X (i,j)]t K j QK j [Σ Xj X i X (i,j)](k iqk i + ɛ 1 I n ) 1/2 v subject to v T v = 1,which implies that Σ Xj X i X (i,j) 2 is the largest eigenvalue of Λ Xi X j (Cont.) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 27 - Bing / 38Li,

28 3.2 Calculation (Cont.) Λ Xi X j =(K iqk i + ɛ 1 I n ) 1/2 [Σ Xj X i X (i,j)]t K j QK j [Σ Xj X i X (i,j)](k iqk i + ɛ 1 I n ) 1/2 We plug in [Σ Xj X i X (i,j)], then we can get Λ Xi X j = X i X j T X i X j where Xi X j =(K iqk i + ɛ 1 I n ) 1/2 K j Q(I n K T (i,j) (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) )QK i K j (K j QK j + ɛ 1 I n ) 1/2 So, we should get the largest singular value of Xi X j. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 28 - Bing / 38Li,

29 3.2 Calculation (Cont.) We can see the large matrix (K (i,j) QK T ( i,j) + ɛ 2I n(p 2) ) and we have to get inverse of it. But, following proposition shows actual amount computation is not large. Proposition V R s t with s > t, which has SVD where D R u u Then, V = (L 1, L 0 )diag(d, 0)(R 1, R 0 ) T (V V T + ɛi s ) 1 = V R 1 D 1 ((D 2 + ɛi u ) 1 ɛ 1 I u )D 1 R T 1 V T + ɛ 1 I s V T (V V T ɛi s ) 1 V = R 1 D(D 2 + ɛi s ) 1 R T 1 On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 29 - Bing / 38Li,

30 3.2 Calculation (Cont.) Second, we calculate Θ Xi X j. Similarly, if sample estimatre of Θ Xi Xj is small enough, we determine there is no edge between X i and X j. Theorem Let K 1 p = (K 1,..., K p ) T, then matrix representation of Υ satisfies the following equation, So, we can get [Θ] as diag(k 1 QK 1,..., K p QK p )[Υ] = K 1 p QK T 1 p [Θ] = (K 1 p QK T 1 p + ɛ 3 I n p) 1 diag(k 1 QK 1,..., K p QK p ) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 30 - Bing / 38Li,

31 3.2 Calculation (Cont.) To find Θ Xj Xi, we maximize < Θ Xj X if i, Θ Xj X if i > AX j = n 1 [f i ] T [Θ Xj X i]t K j QK j [Θ Xj X i][f i] subject to < f i, f i > AX i = 1. Similarly, Θ Xj X i 2 is the largest eigenvalue of (K i QK i + ɛ 1 I n ) 1/2 [Θ Xj X i]t K j QK j [Θ Xj X i](k iqk i + ɛ 1 I n ) 1/2 And the largest singular value of (K i QK i + ɛ 1 I n ) 1/2 [Θ Xj X i]t (K j QK j + ɛ 1 I n ) 1/2 On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 31 - Bing / 38Li,

32 3.3 Tuning parameter Because of Tychonoff regularization, we used three tuning parameters ɛ 1, ɛ 2, ɛ 3 getting inverse of K i QK i, K (i,j) QK T (i,j), K 1 pqk T 1 p. In this paper, the cross validation procedure is proposed to select these parameters. Notations Let U and V be random vectors and W = (U T, V T ) T. Let F = {f 0, f 1,..., f r } be a set of functions of U. Similarly, G = {g 0, g 1,..., g s } be a set of functions of V. f 0, g 0 represents the constant function 1. For iid sample {W 1,..., W n }, let A = {a 1,..., a m } {1, 2,..., n}, B = A c. Let {W a a A} be a training set, {W b b B} be a testing set. (Cont.) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 32 - Bing / 38Li,

33 3.3 Tuning parameter (Cont.) (Cont.) Suppose Ê(ɛ) G F is a training-set-based operator, such that for each V U g G, Ê (ɛ) g is the best prediction of g(v ) based on functions in F. V U Then, total error is evaluated as s [g µ (V b ) (Ê(ɛ) V U g µ)(u b )] 2 b B µ=0 Let L U be the matrix with ith row is {f i+1 (U a ) a A}, and let L V be with {g i+1 (V a ) a A} And let l U, l V be a vector valued functions (f 0,..., f r ) T, (g 0,..., g s ) T. Then we can express total error as l V (V b ) L V L T U (L U L T U + ɛi r+1 ) 1 l U (U b ) 2 b B On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 33 - Bing / 38Li,

34 3.3 Tuning parameter (Cont.) If you let G U be the bth column is l U (U b ), G V be the bth column is l V (V b ), xthen total error can be further simplified as G V L V L T U (L U L T U + ɛi r+1 ) 1 G U 2 Now, we apply this cross validation procedure for choosing our parameters. p CV ν 1(ɛ 1 ) = CV 1 (ɛ 1 ) = i=1 b B p l i (X i b ) L i L T i (L i L T i + ɛ 1 I m+1 ) 1 l i (X i b) 2 = G i L i L T i (L i L T i + ɛ 1 I m+1 ) 1 G i 2 i=1 k ν=1 CV ν 1 (ɛ 1 ) where ν is k index of k fold CV. Choose ɛ 1 minimizing CV 1 (ɛ 1 ). On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 34 - Bing / 38Li,

35 3.3 Tuning parameter (Cont.) CV ν 2(ɛ 2 ) = G (i,j) L (i,j) L T (i,j) (L (i,j)l T (i,j) + ɛ 2I m(p 2)+1 ) 1 G (i,j) 2 i>j CV 2 (ɛ 2 ) = k ν=1 CV ν 2 (ɛ 2 ) Choose ɛ 2 minimizing CV 2 (ɛ 2 ). And for ɛ 3, in this paper, they recommend to take ɛ 3 to be same as ɛ 2. Because K (i,j) QK T (i,j) and K 1 pqk T 1 p have similar forms and dimensions. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 35 - Bing / 38Li,

36 4. Simulation results They suggest some comparisons to assess their method. We will see some of them. Comparison 1 When the Gaussian copula assumption is violated. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 36 - Bing / 38Li,

37 4. Simulation results (Cont.) Comparison 2 When the Gaussian assumption and Gaussian copula assumption is satisfied. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 37 - Bing / 38Li,

38 4. Simulation results (Cont.) Comparison 3 When the true model is non additive model On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 38 - Bing / 38Li,

Markov properties for undirected graphs

Markov properties for undirected graphs Graphical Models, Lecture 2, Michaelmas Term 2011 October 12, 2011 Formal definition Fundamental properties Random variables X and Y are conditionally independent given the random variable Z if L(X Y,

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

MET Workshop: Exercises

MET Workshop: Exercises MET Workshop: Exercises Alex Blumenthal and Anthony Quas May 7, 206 Notation. R d is endowed with the standard inner product (, ) and Euclidean norm. M d d (R) denotes the space of n n real matrices. When

More information

Conditional Independence and Markov Properties

Conditional Independence and Markov Properties Conditional Independence and Markov Properties Lecture 1 Saint Flour Summerschool, July 5, 2006 Steffen L. Lauritzen, University of Oxford Overview of lectures 1. Conditional independence and Markov properties

More information

Total positivity in Markov structures

Total positivity in Markov structures 1 based on joint work with Shaun Fallat, Kayvan Sadeghi, Caroline Uhler, Nanny Wermuth, and Piotr Zwiernik (arxiv:1510.01290) Faculty of Science Total positivity in Markov structures Steffen Lauritzen

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models

MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models 1/13 MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models Dominique Guillot Departments of Mathematical Sciences University of Delaware May 4, 2016 Recall

More information

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

Functional Analysis Review

Functional Analysis Review Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

Lecture Summaries for Linear Algebra M51A

Lecture Summaries for Linear Algebra M51A These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture

More information

Markov properties for undirected graphs

Markov properties for undirected graphs Graphical Models, Lecture 2, Michaelmas Term 2009 October 15, 2009 Formal definition Fundamental properties Random variables X and Y are conditionally independent given the random variable Z if L(X Y,

More information

E.7 Alaoglu s Theorem

E.7 Alaoglu s Theorem E.7 Alaoglu s Theorem 359 E.7 Alaoglu s Theorem If X is a normed space, then the closed unit ball in X or X is compact if and only if X is finite-dimensional (Problem A.25). Even so, Alaoglu s Theorem

More information

Lecture 19: Isometries, Positive operators, Polar and singular value decompositions; Unitary matrices and classical groups; Previews (1)

Lecture 19: Isometries, Positive operators, Polar and singular value decompositions; Unitary matrices and classical groups; Previews (1) Lecture 19: Isometries, Positive operators, Polar and singular value decompositions; Unitary matrices and classical groups; Previews (1) Travis Schedler Thurs, Nov 18, 2010 (version: Wed, Nov 17, 2:15

More information

Student-t Process as Alternative to Gaussian Processes Discussion

Student-t Process as Alternative to Gaussian Processes Discussion Student-t Process as Alternative to Gaussian Processes Discussion A. Shah, A. G. Wilson and Z. Gharamani Discussion by: R. Henao Duke University June 20, 2014 Contributions The paper is concerned about

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Gaussian vectors and central limit theorem

Gaussian vectors and central limit theorem Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Likelihood Analysis of Gaussian Graphical Models

Likelihood Analysis of Gaussian Graphical Models Faculty of Science Likelihood Analysis of Gaussian Graphical Models Ste en Lauritzen Department of Mathematical Sciences Minikurs TUM 2016 Lecture 2 Slide 1/43 Overview of lectures Lecture 1 Markov Properties

More information

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide

More information

Definition 2.3. We define addition and multiplication of matrices as follows.

Definition 2.3. We define addition and multiplication of matrices as follows. 14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row

More information

Representing Independence Models with Elementary Triplets

Representing Independence Models with Elementary Triplets Representing Independence Models with Elementary Triplets Jose M. Peña ADIT, IDA, Linköping University, Sweden jose.m.pena@liu.se Abstract An elementary triplet in an independence model represents a conditional

More information

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Chapter 26 Semidefinite Programming Zacharias Pitouras 1 Introduction LP place a good lower bound on OPT for NP-hard problems Are there other ways of doing this? Vector programs

More information

Math113: Linear Algebra. Beifang Chen

Math113: Linear Algebra. Beifang Chen Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary

More information

Basic Concepts in Matrix Algebra

Basic Concepts in Matrix Algebra Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1

More information

4 Countability axioms

4 Countability axioms 4 COUNTABILITY AXIOMS 4 Countability axioms Definition 4.1. Let X be a topological space X is said to be first countable if for any x X, there is a countable basis for the neighborhoods of x. X is said

More information

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure. Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,

More information

Basics on Probability. Jingrui He 09/11/2007

Basics on Probability. Jingrui He 09/11/2007 Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Rings and groups. Ya. Sysak

Rings and groups. Ya. Sysak Rings and groups. Ya. Sysak 1 Noetherian rings Let R be a ring. A (right) R -module M is called noetherian if it satisfies the maximum condition for its submodules. In other words, if M 1... M i M i+1...

More information

EECS 598: Statistical Learning Theory, Winter 2014 Topic 11. Kernels

EECS 598: Statistical Learning Theory, Winter 2014 Topic 11. Kernels EECS 598: Statistical Learning Theory, Winter 2014 Topic 11 Kernels Lecturer: Clayton Scott Scribe: Jun Guo, Soumik Chatterjee Disclaimer: These notes have not been subjected to the usual scrutiny reserved

More information

The Multivariate Normal Distribution. In this case according to our theorem

The Multivariate Normal Distribution. In this case according to our theorem The Multivariate Normal Distribution Defn: Z R 1 N(0, 1) iff f Z (z) = 1 2π e z2 /2. Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) T with the Z i independent and each Z i N(0, 1). In this

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

Convergence of Multivariate Quantile Surfaces

Convergence of Multivariate Quantile Surfaces Convergence of Multivariate Quantile Surfaces Adil Ahidar Institut de Mathématiques de Toulouse - CERFACS August 30, 2013 Adil Ahidar (Institut de Mathématiques de Toulouse Convergence - CERFACS) of Multivariate

More information

Introduction. Vectors and Matrices. Vectors [1] Vectors [2]

Introduction. Vectors and Matrices. Vectors [1] Vectors [2] Introduction Vectors and Matrices Dr. TGI Fernando 1 2 Data is frequently arranged in arrays, that is, sets whose elements are indexed by one or more subscripts. Vector - one dimensional array Matrix -

More information

Vector Spaces, Affine Spaces, and Metric Spaces

Vector Spaces, Affine Spaces, and Metric Spaces Vector Spaces, Affine Spaces, and Metric Spaces 2 This chapter is only meant to give a short overview of the most important concepts in linear algebra, affine spaces, and metric spaces and is not intended

More information

Outline of the course

Outline of the course School of Mathematical Sciences PURE MTH 3022 Geometry of Surfaces III, Semester 2, 20 Outline of the course Contents. Review 2. Differentiation in R n. 3 2.. Functions of class C k 4 2.2. Mean Value Theorem

More information

5: MULTIVARATE STATIONARY PROCESSES

5: MULTIVARATE STATIONARY PROCESSES 5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability

More information

Introduction to Probability and Stocastic Processes - Part I

Introduction to Probability and Stocastic Processes - Part I Introduction to Probability and Stocastic Processes - Part I Lecture 2 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

Convex Optimization in Classification Problems

Convex Optimization in Classification Problems New Trends in Optimization and Computational Algorithms December 9 13, 2001 Convex Optimization in Classification Problems Laurent El Ghaoui Department of EECS, UC Berkeley elghaoui@eecs.berkeley.edu 1

More information

MAT-INF4110/MAT-INF9110 Mathematical optimization

MAT-INF4110/MAT-INF9110 Mathematical optimization MAT-INF4110/MAT-INF9110 Mathematical optimization Geir Dahl August 20, 2013 Convexity Part IV Chapter 4 Representation of convex sets different representations of convex sets, boundary polyhedra and polytopes:

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Probability Distribution And Density For Functional Random Variables

Probability Distribution And Density For Functional Random Variables Probability Distribution And Density For Functional Random Variables E. Cuvelier 1 M. Noirhomme-Fraiture 1 1 Institut d Informatique Facultés Universitaires Notre-Dame de la paix Namur CIL Research Contact

More information

Multivariate Gaussian Analysis

Multivariate Gaussian Analysis BS2 Statistical Inference, Lecture 7, Hilary Term 2009 February 13, 2009 Marginal and conditional distributions For a positive definite covariance matrix Σ, the multivariate Gaussian distribution has density

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

Introduction to Machine Learning (67577) Lecture 3

Introduction to Machine Learning (67577) Lecture 3 Introduction to Machine Learning (67577) Lecture 3 Shai Shalev-Shwartz School of CS and Engineering, The Hebrew University of Jerusalem General Learning Model and Bias-Complexity tradeoff Shai Shalev-Shwartz

More information

Gaussians. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

Gaussians. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Gaussians Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Outline Univariate Gaussian Multivariate Gaussian Law of Total Probability Conditioning

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Chapter 4 Euclid Space

Chapter 4 Euclid Space Chapter 4 Euclid Space Inner Product Spaces Definition.. Let V be a real vector space over IR. A real inner product on V is a real valued function on V V, denoted by (, ), which satisfies () (x, y) = (y,

More information

Kernel Principal Component Analysis

Kernel Principal Component Analysis Kernel Principal Component Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

On the inverse matrix of the Laplacian and all ones matrix

On the inverse matrix of the Laplacian and all ones matrix On the inverse matrix of the Laplacian and all ones matrix Sho Suda (Joint work with Michio Seto and Tetsuji Taniguchi) International Christian University JSPS Research Fellow PD November 21, 2012 Sho

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Lecture 15: Multivariate normal distributions

Lecture 15: Multivariate normal distributions Lecture 15: Multivariate normal distributions Normal distributions with singular covariance matrices Consider an n-dimensional X N(µ,Σ) with a positive definite Σ and a fixed k n matrix A that is not of

More information

Consistent Bivariate Distribution

Consistent Bivariate Distribution A Characterization of the Normal Conditional Distributions MATSUNO 79 Therefore, the function ( ) = G( : a/(1 b2)) = N(0, a/(1 b2)) is a solu- tion for the integral equation (10). The constant times of

More information

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of

More information

Linear algebra for computational statistics

Linear algebra for computational statistics University of Seoul May 3, 2018 Vector and Matrix Notation Denote 2-dimensional data array (n p matrix) by X. Denote the element in the ith row and the jth column of X by x ij or (X) ij. Denote by X j

More information

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2 Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution Defn: Z R 1 N(0,1) iff f Z (z) = 1 2π e z2 /2 Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) (a column

More information

Direct Limits. Mathematics 683, Fall 2013

Direct Limits. Mathematics 683, Fall 2013 Direct Limits Mathematics 683, Fall 2013 In this note we define direct limits and prove their basic properties. This notion is important in various places in algebra. In particular, in algebraic geometry

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical

More information

which is a group homomorphism, such that if W V U, then

which is a group homomorphism, such that if W V U, then 4. Sheaves Definition 4.1. Let X be a topological space. A presheaf of groups F on X is a a function which assigns to every open set U X a group F(U) and to every inclusion V U a restriction map, ρ UV

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Stat 206: Linear algebra

Stat 206: Linear algebra Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two

More information

ECON 5111 Mathematical Economics

ECON 5111 Mathematical Economics Test 1 October 1, 2010 1. Construct a truth table for the following statement: [p (p q)] q. 2. A prime number is a natural number that is divisible by 1 and itself only. Let P be the set of all prime numbers

More information

Maths for Signals and Systems Linear Algebra in Engineering

Maths for Signals and Systems Linear Algebra in Engineering Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE

More information

Appendices to the paper "Detecting Big Structural Breaks in Large Factor Models" (2013) by Chen, Dolado and Gonzalo.

Appendices to the paper Detecting Big Structural Breaks in Large Factor Models (2013) by Chen, Dolado and Gonzalo. Appendices to the paper "Detecting Big Structural Breaks in Large Factor Models" 203 by Chen, Dolado and Gonzalo. A.: Proof of Propositions and 2 he proof proceeds by showing that the errors, factors and

More information

MAT 445/ INTRODUCTION TO REPRESENTATION THEORY

MAT 445/ INTRODUCTION TO REPRESENTATION THEORY MAT 445/1196 - INTRODUCTION TO REPRESENTATION THEORY CHAPTER 1 Representation Theory of Groups - Algebraic Foundations 1.1 Basic definitions, Schur s Lemma 1.2 Tensor products 1.3 Unitary representations

More information

Econ 508B: Lecture 5

Econ 508B: Lecture 5 Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline

More information

4.4 Noetherian Rings

4.4 Noetherian Rings 4.4 Noetherian Rings Recall that a ring A is Noetherian if it satisfies the following three equivalent conditions: (1) Every nonempty set of ideals of A has a maximal element (the maximal condition); (2)

More information

Stable Process. 2. Multivariate Stable Distributions. July, 2006

Stable Process. 2. Multivariate Stable Distributions. July, 2006 Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff David Gerard Department of Statistics University of Washington gerard2@uw.edu May 2, 2013 David Gerard (UW)

More information

Mathematical Analysis Outline. William G. Faris

Mathematical Analysis Outline. William G. Faris Mathematical Analysis Outline William G. Faris January 8, 2007 2 Chapter 1 Metric spaces and continuous maps 1.1 Metric spaces A metric space is a set X together with a real distance function (x, x ) d(x,

More information

ELEMENTARY SUBALGEBRAS OF RESTRICTED LIE ALGEBRAS

ELEMENTARY SUBALGEBRAS OF RESTRICTED LIE ALGEBRAS ELEMENTARY SUBALGEBRAS OF RESTRICTED LIE ALGEBRAS J. WARNER SUMMARY OF A PAPER BY J. CARLSON, E. FRIEDLANDER, AND J. PEVTSOVA, AND FURTHER OBSERVATIONS 1. The Nullcone and Restricted Nullcone We will need

More information

Filters in Analysis and Topology

Filters in Analysis and Topology Filters in Analysis and Topology David MacIver July 1, 2004 Abstract The study of filters is a very natural way to talk about convergence in an arbitrary topological space, and carries over nicely into

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

Independence for Full Conditional Measures, Graphoids and Bayesian Networks

Independence for Full Conditional Measures, Graphoids and Bayesian Networks Independence for Full Conditional Measures, Graphoids and Bayesian Networks Fabio G. Cozman Universidade de Sao Paulo Teddy Seidenfeld Carnegie Mellon University February 28, 2007 Abstract This paper examines

More information

A Magiv CV Theory for Large-Margin Classifiers

A Magiv CV Theory for Large-Margin Classifiers A Magiv CV Theory for Large-Margin Classifiers Hui Zou School of Statistics, University of Minnesota June 30, 2018 Joint work with Boxiang Wang Outline 1 Background 2 Magic CV formula 3 Magic support vector

More information

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song Presenter: Jiwei Zhao Department of Statistics University of Wisconsin Madison April

More information

6-1 Study Guide and Intervention Multivariable Linear Systems and Row Operations

6-1 Study Guide and Intervention Multivariable Linear Systems and Row Operations 6-1 Study Guide and Intervention Multivariable Linear Systems and Row Operations Gaussian Elimination You can solve a system of linear equations using matrices. Solving a system by transforming it into

More information

Mean-field dual of cooperative reproduction

Mean-field dual of cooperative reproduction The mean-field dual of systems with cooperative reproduction joint with Tibor Mach (Prague) A. Sturm (Göttingen) Friday, July 6th, 2018 Poisson construction of Markov processes Let (X t ) t 0 be a continuous-time

More information

Slide05 Haykin Chapter 5: Radial-Basis Function Networks

Slide05 Haykin Chapter 5: Radial-Basis Function Networks Slide5 Haykin Chapter 5: Radial-Basis Function Networks CPSC 636-6 Instructor: Yoonsuck Choe Spring Learning in MLP Supervised learning in multilayer perceptrons: Recursive technique of stochastic approximation,

More information

Section 8.1. Vector Notation

Section 8.1. Vector Notation Section 8.1 Vector Notation Definition 8.1 Random Vector A random vector is a column vector X = [ X 1 ]. X n Each Xi is a random variable. Definition 8.2 Vector Sample Value A sample value of a random

More information

Sample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson

Sample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson Sample Geometry Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois Spring

More information

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12 Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following

More information

Stat 710: Mathematical Statistics Lecture 31

Stat 710: Mathematical Statistics Lecture 31 Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:

More information

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond Measure Theory on Topological Spaces Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond May 22, 2011 Contents 1 Introduction 2 1.1 The Riemann Integral........................................ 2 1.2 Measurable..............................................

More information

Nonparameteric Regression:

Nonparameteric Regression: Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,

More information