On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -1 Bing / 38Li,
|
|
- Isaac Dalton
- 5 years ago
- Views:
Transcription
1 On an Additive Semigraphoid Model for Statistical Networks With Application to Pathway Analysis - Bing Li, Hyunho Chun & Hongyu Zhao Kim Youngrae SNU Stat. Multivariate Lab Nov 25, 2016 On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -1 Bing / 38Li,
2 Outline 1 Introduction GGM, GCGM 2 Additive conditional independent Graphoid, semigraphoid Conditional independent, additively conditional independent Relation with previous models 3 Estimation ACCO, APO Calculation Tuning parameter 4 Simultation results On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -2 Bing / 38Li,
3 1. Introduction Additive semigraphoid model Gaussian copula graphical model Additively conditional independent Additive conditional covariance Operator Additive precision operator On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -3 Bing / 38Li,
4 1.1 GGM, GCGM Gaussian graphical model Let X = (X 1,..., X p ) T, Suppose X N(µ, Σ), and Σ is postive definite. Let Θ = Σ 1, Θ ij be the (i, j)th elements of Θ. Then X follows a GGM with respect to the graph G = {Γ, E} iff X i X j X (i,j), (i, j) E Multivariate normal assumption. In Gaussian graphical model, X i X j X (i,j) iff Θ ij = 0. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -4 Bing / 38Li,
5 1.1 GGM, GCGM (Cont.) Copula Copula is multivariate probability distribution for which the marginal probability of each variable is uniform. It can deal with the variables with dependency. C [0, 1] d [0, 1], is a d dimensional copula if C is cumulative distribution function of a d dimensional random vector on the unit cube On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -5 Bing / 38Li,
6 1.1 GGM, GCGM (Cont.) Gaussian copula graphical model Let X = (X 1,..., X p ) T, Suppose the existence of unknown injections f 1,..., f p such that (f 1 (X 1 ),..., f p (X p )) T is multivariate gaussian. (f 1 (X 1 ),..., f p (X p )) T N(µ, Σ) Each X i shouldn t be a Gaussian. We can use all of GGM procedures for non-gaussian X i s. But the Gaussian copula assumpotion can be violated even for commonly used interactions. ex. marginally normal doesn t imply jointly normal. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -6 Bing / 38Li,
7 2.1 Graphoid, semigraphoid Graphoid Axioms A three-way relation R is called a graphoid if it satisfies the following conditions. Symmetry: (A, C, B) R (B, C, A) R Decomposition: (A, C, B D) R (A, C, B) R Weak Union: (A, C, B D) R (A, C B, D) R Contraction: (A, C B, D) R, (A, C, B) R (A, C, B D) R Intersection: (A, C D, B) R, (A, C B, D) (A, C, B D) R Conditional indenpendent satisfies this axioms. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -7 Bing / 38Li,
8 2.1 Graphoid, semigraphoid (Cont.) Semigraphoid Axioms A three-way relation R is called a semigraphoid if it satisfies the conditions of Graphoid axioms except Intersection condition. Symmetry: (A, C, B) R (B, C, A) R Decomposition: (A, C, B D) R (A, C, B) R Weak Union: (A, C, B D) R (A, C B, D) R Contraction: (A, C B, D) R, (A, C, B) R (A, C, B D) R Using this semigraphoid relation in place of probabilistic conditional independence can free us from some awkward restrictions. On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -8 Bing / 38Li,
9 2.2 Conditional independent, additively conditional independent Notations Let X(X 1,..., X p ) T be a random vector. For subvector U = (U 1,..., U p ) T of X, let L 2 (P U ) be the class of functions of U such that Ef(U) = 0, Ef 2 (U) <. For each U i, let A U i denote a subset of L 2 (P U i). Let A U denote the addtive family A U A U r = {f f r f 1 A U 1,..., f r A U r} In this paper, inner product <.,. > represents the L 2 (P X ) inner product. < f, g >= f(x)g(x)dµ(x) On an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -9 Bing / 38Li,
10 2.2 Conditional independent, addtively conditional independent (Cont.) Additively conditional independent We say that U and V are additively conditional independent (ACI) give W iff (A U + A W ) A W (A V + A W ) A W And we write this relation as U A V W. Additively conditional independent is three-way relation. U A V W (A U, A W, A V ) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 10 - Bing / 38Li,
11 2.2 Conditional independent, addtively conditional independent (Cont.) Theorem The additive conditional independece relation is a semigraphoid. Additive semigraphoid model (ASG model) A random vector X follows and ASG model with respect to a graph G = (Γ, E) iff X i A X j X (i,j) (i, j) E. We write this condition as X ASG(G). On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 11 - Bing / 38Li,
12 2.3 Relation with previous models Now we investigate the relation between new-concept, Additive conditional independence and conditional indepdence. Theorem Suppose (a) X has a Gaussian copula distribution with copula functions f 1,..., f p (b) U, V, and W are subvectors of X (c) A X i = span{f i }, i = 1,..., p Then, U A V W iff U V W. Under Gaussian copula assumption with A X i = span{f i }, addtive conditional independence is equivalent to conditional independence. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 12 - Bing / 38Li,
13 2.3 Relation with previous models (Cont.) Theorem Suppose (a) X has a Gaussian copula distribution with copula functions f 1,..., f p (b) U, V, and W are subvectors of X (c) A X i = L 2 (P X i), i = 1,..., p Then, U A V W implies U V W. In this case, they are not equivalent. However, equivalence holds apploximately. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 13 - Bing / 38Li,
14 2.3 Relation with previous models (Cont.) The following proposition suggests implication U V W to U A V W holds approximately. Proposition Suppose Gaussian copula assumption, and WLOG, E[f i (X i )] = 0, V ar[f i (X i )] = 1. Let U = f i (X i ), V = f j (X j ), W = {f k (X k ) k i, j}, and R V W = cov(v, W ), R W W = V ar(w ), R W U = cov(w, U), ρ UV = cor(u, V ). Then, ρ UV W max{ ρ α UV R α V W (R α W W ) 1 R α W U where A α is the α fold Hadamard product. α = 1, 2,...} On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 14 - Bing / 38Li,
15 2.3 Relation with previous models (Cont.) τ UV W is upper bound of ρ UV W in previous proposition. These are calculated under assumption that is in the proposition plus U V W. The small numbers of τ UV W in the table indicate ρ UV W is approximately 0. So, U V W implies U A V W approximately. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 15 - Bing / 38Li,
16 3.1 ACCO, APO In the GGM, X i X j X (i,j) iff (i, j) E, and with precision matrix Θ, we can cut the edge easily. Building the graph in ASG, we introduce two linear operator. Additive conditional covariance operator The ACCO is operator from A V to A U such that Σ UV W = Σ UV Σ UW Σ W V. where Σ UV A V A U such that < f, Σ UV g >= E[f(U)g(V )]. We call the operator the ACCO from V to U given W On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 16 - Bing / 38Li,
17 3.1 ACCO, APO (Cont.) With this operator, there are two important results. Lemma Let P AW A X A W be the projection on to A W. Then P AW A U = Σ W U. Theorem U A V W iff Σ UV W = 0. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 17 - Bing / 38Li,
18 3.1 ACCO, APO (Cont.) Calculating ACCO is difficult because if W is changed, you should calculate again. Similar to GGM, we want to use precision matrix instead of conditional covariance. For defining such operator, we will use different inner product. Let A X be the Hilbert space consisting of the same set of functions as A X, but with the inner product < f f p, g g p > A X = p i=1 < f i, g i > AX i On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 18 - Bing / 38Li,
19 3.1 ACCO, APO (Cont.) Additive precision operator The operator Υ A X A X defined by the matrix of operator {Σ Xi Xj i Γ, j Γ} is called additive covariance operator. If it is invertible then its inverse Θ is called the APO. Υ = {Σ Xi X j i Γ, j Γ} means When Υf i = h h p, Σ Xj X i(f i) = h j. We can easily check for any f, g A X, < f, Υg > A X = cov[f(x), g(x)] =< f, g > On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 19 - Bing / 38Li,
20 3.1 ACCO, APO (Cont.) Question. Is APO well defined? Proposition Suppose Υ is invertible and, for each i j, Υ ij is compact. Then, Υ 1 is bounded. Lemma Suppose that T B(A X, A X ) is a self adjoint, positive definite operator. Then the operators T UU, T V V, T UU T UV TV 1 V T V U, T V V T V U TUU 1 T UV are bounded, self adjoint, and positive definite, and the following identities hold (T 1 ) UU = (T UU T UV T 1 V V T V U ) 1 (1) (T 1 ) V V = (T V V T V U T 1 UU T UV ) 1 (2) (T 1 ) UV = (T 1 ) UU T UV T 1 V V = T 1 UU T UV (T 1 ) V V (3) (T 1 ) V U = (T 1 ) V V T V U T 1 UU = T 1 V V T V U (T 1 ) UU (4) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 20 - Bing / 38Li,
21 3.1 ACCO, APO (Cont.) The above lemma shows that if Υ has bounded inverse, Θ is well defined. And finally we derive this equivalence. Theorem Suppose (U, V, W ) is a partition of X and Υ is invertible. Then, U A V W iff Θ UV = 0. Corollary A random vector X follows an ASG(G) model if and only if Θ Xi X j = 0 whenever (i, j) E. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 21 - Bing / 38Li,
22 3.2 Calculation First, how we build a graph with ACCO. If sample estimate of Σ Xi X j X (i,j) is small enough, we determine there is no edge between X i and X j. We should derive the finite-sample matrix representation of the operator. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 22 - Bing / 38Li,
23 3.2 Calculation (Cont.) Notations For each i = 1,..., p, let κ i Ω X i Ω X i R be a positive kernel. Let A X i = span{κ 1 (., X i 1) E n κ i (X i, X i 1),..., κ i (., X i n) E n κ i (X i, X i n)} where E n κ i (X i, X i k ) = n 1 n l=1 κ i (X i l, Xi k ). Q = I n 1 n 1 T n /n ; Projection matrix Let κ i be the vector valued function x i (κ i (x i, X i 1) E n κ i (X i, X i 1),..., κ i (x i, X i n) E n κ i (X i, X i n)) With this κ i, for any f A X i can be expressed as f = κ i [f]. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 23 - Bing / 38Li,
24 3.2 Calculation (Cont.) With these notations, we can get the followings (f(x i 1),..., f(x i n)) T = QK i [f] (5) < f, g > AX i = n 1 (QK i [f]) T (QK i [g]) = n 1 [f] T K i QK i [g] (6) where K i = {κ i (X i k, Xi l ) k, l = 1,..., n}, f, g A X i. K (i,j) denoted n(p 2) n matrix obtained by removing ith and jth block of (K 1,..., K p ) T. Lemma K i QK i [Σ Xi X j ] = K iqk j (7) K i QK i [Σ Xi X (i,j)] = K iqk T (i,j) (8) K (i,j) QK T ( i,j) [Σ X (i,j) X i] = K (i,j)qk i (9) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 24 - Bing / 38Li,
25 3.2 Calculation (Cont.) K i QK i, K (i,j) QK (i,j) are singular. We can solve it by Tychonoff regularization. [Σ X i X j ] = (K iqk i + ɛ 1 I n ) 1 K i QK j (10) [Σ Xi X (i,j)] = (K iqk i + ɛ 1 I n ) 1 K i QK T (i,j) (11) [Σ X (i,j) X i] = (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) QK i (12) We can represent [Σ Xj X i X (i,j)] = [Σ X j X i] [Σ X j X (i,j)][σ X (i,j) X i] On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 25 - Bing / 38Li,
26 3.2 Calculation (Cont.) So, we can get [Σ Xj X i X (i,j)] = [Σ X j X i] [Σ X j X (i,j)][σ X (i,j) X i] = (K j QK j + ɛ 1 I n ) 1 K j QK i (K j QK j + ɛ 1 I n ) 1 K j QK T (i,j) (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) QK i = (K j QK j + ɛ 1 I n ) 1 K j Q(I n K T (i,j) (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) )QK i On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 26 - Bing / 38Li,
27 3.2 Calculation (Cont.) By definition, Σ Xj X i X (i,j) 2 is the maximum of < Σ Xj X i X (i,j)f, Σ X j X i X (i,j)f > A X j subject to < f, f > AX j = 1. = n 1 [f] T [Σ Xj X i X (i,j)]t K j QK j [Σ Xj X i X (i,j)][f] Let v = (K i QK i ) 1/2 [f], then [f] = (K i QK i + ɛ 1 I n ) 1/2 v. The right hand side of above is v T (K i QK i + ɛ 1 I n ) 1/2 [Σ Xj X i X (i,j)]t K j QK j [Σ Xj X i X (i,j)](k iqk i + ɛ 1 I n ) 1/2 v subject to v T v = 1,which implies that Σ Xj X i X (i,j) 2 is the largest eigenvalue of Λ Xi X j (Cont.) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 27 - Bing / 38Li,
28 3.2 Calculation (Cont.) Λ Xi X j =(K iqk i + ɛ 1 I n ) 1/2 [Σ Xj X i X (i,j)]t K j QK j [Σ Xj X i X (i,j)](k iqk i + ɛ 1 I n ) 1/2 We plug in [Σ Xj X i X (i,j)], then we can get Λ Xi X j = X i X j T X i X j where Xi X j =(K iqk i + ɛ 1 I n ) 1/2 K j Q(I n K T (i,j) (K (i,j)qk T ( i,j) + ɛ 2I n(p 2) ) 1 K (i,j) )QK i K j (K j QK j + ɛ 1 I n ) 1/2 So, we should get the largest singular value of Xi X j. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 28 - Bing / 38Li,
29 3.2 Calculation (Cont.) We can see the large matrix (K (i,j) QK T ( i,j) + ɛ 2I n(p 2) ) and we have to get inverse of it. But, following proposition shows actual amount computation is not large. Proposition V R s t with s > t, which has SVD where D R u u Then, V = (L 1, L 0 )diag(d, 0)(R 1, R 0 ) T (V V T + ɛi s ) 1 = V R 1 D 1 ((D 2 + ɛi u ) 1 ɛ 1 I u )D 1 R T 1 V T + ɛ 1 I s V T (V V T ɛi s ) 1 V = R 1 D(D 2 + ɛi s ) 1 R T 1 On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 29 - Bing / 38Li,
30 3.2 Calculation (Cont.) Second, we calculate Θ Xi X j. Similarly, if sample estimatre of Θ Xi Xj is small enough, we determine there is no edge between X i and X j. Theorem Let K 1 p = (K 1,..., K p ) T, then matrix representation of Υ satisfies the following equation, So, we can get [Θ] as diag(k 1 QK 1,..., K p QK p )[Υ] = K 1 p QK T 1 p [Θ] = (K 1 p QK T 1 p + ɛ 3 I n p) 1 diag(k 1 QK 1,..., K p QK p ) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 30 - Bing / 38Li,
31 3.2 Calculation (Cont.) To find Θ Xj Xi, we maximize < Θ Xj X if i, Θ Xj X if i > AX j = n 1 [f i ] T [Θ Xj X i]t K j QK j [Θ Xj X i][f i] subject to < f i, f i > AX i = 1. Similarly, Θ Xj X i 2 is the largest eigenvalue of (K i QK i + ɛ 1 I n ) 1/2 [Θ Xj X i]t K j QK j [Θ Xj X i](k iqk i + ɛ 1 I n ) 1/2 And the largest singular value of (K i QK i + ɛ 1 I n ) 1/2 [Θ Xj X i]t (K j QK j + ɛ 1 I n ) 1/2 On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 31 - Bing / 38Li,
32 3.3 Tuning parameter Because of Tychonoff regularization, we used three tuning parameters ɛ 1, ɛ 2, ɛ 3 getting inverse of K i QK i, K (i,j) QK T (i,j), K 1 pqk T 1 p. In this paper, the cross validation procedure is proposed to select these parameters. Notations Let U and V be random vectors and W = (U T, V T ) T. Let F = {f 0, f 1,..., f r } be a set of functions of U. Similarly, G = {g 0, g 1,..., g s } be a set of functions of V. f 0, g 0 represents the constant function 1. For iid sample {W 1,..., W n }, let A = {a 1,..., a m } {1, 2,..., n}, B = A c. Let {W a a A} be a training set, {W b b B} be a testing set. (Cont.) On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 32 - Bing / 38Li,
33 3.3 Tuning parameter (Cont.) (Cont.) Suppose Ê(ɛ) G F is a training-set-based operator, such that for each V U g G, Ê (ɛ) g is the best prediction of g(v ) based on functions in F. V U Then, total error is evaluated as s [g µ (V b ) (Ê(ɛ) V U g µ)(u b )] 2 b B µ=0 Let L U be the matrix with ith row is {f i+1 (U a ) a A}, and let L V be with {g i+1 (V a ) a A} And let l U, l V be a vector valued functions (f 0,..., f r ) T, (g 0,..., g s ) T. Then we can express total error as l V (V b ) L V L T U (L U L T U + ɛi r+1 ) 1 l U (U b ) 2 b B On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 33 - Bing / 38Li,
34 3.3 Tuning parameter (Cont.) If you let G U be the bth column is l U (U b ), G V be the bth column is l V (V b ), xthen total error can be further simplified as G V L V L T U (L U L T U + ɛi r+1 ) 1 G U 2 Now, we apply this cross validation procedure for choosing our parameters. p CV ν 1(ɛ 1 ) = CV 1 (ɛ 1 ) = i=1 b B p l i (X i b ) L i L T i (L i L T i + ɛ 1 I m+1 ) 1 l i (X i b) 2 = G i L i L T i (L i L T i + ɛ 1 I m+1 ) 1 G i 2 i=1 k ν=1 CV ν 1 (ɛ 1 ) where ν is k index of k fold CV. Choose ɛ 1 minimizing CV 1 (ɛ 1 ). On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 34 - Bing / 38Li,
35 3.3 Tuning parameter (Cont.) CV ν 2(ɛ 2 ) = G (i,j) L (i,j) L T (i,j) (L (i,j)l T (i,j) + ɛ 2I m(p 2)+1 ) 1 G (i,j) 2 i>j CV 2 (ɛ 2 ) = k ν=1 CV ν 2 (ɛ 2 ) Choose ɛ 2 minimizing CV 2 (ɛ 2 ). And for ɛ 3, in this paper, they recommend to take ɛ 3 to be same as ɛ 2. Because K (i,j) QK T (i,j) and K 1 pqk T 1 p have similar forms and dimensions. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 35 - Bing / 38Li,
36 4. Simulation results They suggest some comparisons to assess their method. We will see some of them. Comparison 1 When the Gaussian copula assumption is violated. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 36 - Bing / 38Li,
37 4. Simulation results (Cont.) Comparison 2 When the Gaussian assumption and Gaussian copula assumption is satisfied. On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 37 - Bing / 38Li,
38 4. Simulation results (Cont.) Comparison 3 When the true model is non additive model On an Additive Semigraphoid Model for Statistical Networks With Application tonov Pathway 25, 2016 Analysis 38 - Bing / 38Li,
Markov properties for undirected graphs
Graphical Models, Lecture 2, Michaelmas Term 2011 October 12, 2011 Formal definition Fundamental properties Random variables X and Y are conditionally independent given the random variable Z if L(X Y,
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationMET Workshop: Exercises
MET Workshop: Exercises Alex Blumenthal and Anthony Quas May 7, 206 Notation. R d is endowed with the standard inner product (, ) and Euclidean norm. M d d (R) denotes the space of n n real matrices. When
More informationConditional Independence and Markov Properties
Conditional Independence and Markov Properties Lecture 1 Saint Flour Summerschool, July 5, 2006 Steffen L. Lauritzen, University of Oxford Overview of lectures 1. Conditional independence and Markov properties
More informationTotal positivity in Markov structures
1 based on joint work with Shaun Fallat, Kayvan Sadeghi, Caroline Uhler, Nanny Wermuth, and Piotr Zwiernik (arxiv:1510.01290) Faculty of Science Total positivity in Markov structures Steffen Lauritzen
More informationEcon 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the
More informationMATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models
1/13 MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models Dominique Guillot Departments of Mathematical Sciences University of Delaware May 4, 2016 Recall
More informationReview (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology
Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationPart II Probability and Measure
Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationLecture Summaries for Linear Algebra M51A
These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture
More informationMarkov properties for undirected graphs
Graphical Models, Lecture 2, Michaelmas Term 2009 October 15, 2009 Formal definition Fundamental properties Random variables X and Y are conditionally independent given the random variable Z if L(X Y,
More informationE.7 Alaoglu s Theorem
E.7 Alaoglu s Theorem 359 E.7 Alaoglu s Theorem If X is a normed space, then the closed unit ball in X or X is compact if and only if X is finite-dimensional (Problem A.25). Even so, Alaoglu s Theorem
More informationLecture 19: Isometries, Positive operators, Polar and singular value decompositions; Unitary matrices and classical groups; Previews (1)
Lecture 19: Isometries, Positive operators, Polar and singular value decompositions; Unitary matrices and classical groups; Previews (1) Travis Schedler Thurs, Nov 18, 2010 (version: Wed, Nov 17, 2:15
More informationStudent-t Process as Alternative to Gaussian Processes Discussion
Student-t Process as Alternative to Gaussian Processes Discussion A. Shah, A. G. Wilson and Z. Gharamani Discussion by: R. Henao Duke University June 20, 2014 Contributions The paper is concerned about
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationGaussian vectors and central limit theorem
Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables
More informationInference For High Dimensional M-estimates. Fixed Design Results
: Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and
More informationLikelihood Analysis of Gaussian Graphical Models
Faculty of Science Likelihood Analysis of Gaussian Graphical Models Ste en Lauritzen Department of Mathematical Sciences Minikurs TUM 2016 Lecture 2 Slide 1/43 Overview of lectures Lecture 1 Markov Properties
More informationEcon 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis
Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide
More informationDefinition 2.3. We define addition and multiplication of matrices as follows.
14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row
More informationRepresenting Independence Models with Elementary Triplets
Representing Independence Models with Elementary Triplets Jose M. Peña ADIT, IDA, Linköping University, Sweden jose.m.pena@liu.se Abstract An elementary triplet in an independence model represents a conditional
More informationHigh Dimensional Inverse Covariate Matrix Estimation via Linear Programming
High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω
More informationApproximation Algorithms
Approximation Algorithms Chapter 26 Semidefinite Programming Zacharias Pitouras 1 Introduction LP place a good lower bound on OPT for NP-hard problems Are there other ways of doing this? Vector programs
More informationMath113: Linear Algebra. Beifang Chen
Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More information4 Countability axioms
4 COUNTABILITY AXIOMS 4 Countability axioms Definition 4.1. Let X be a topological space X is said to be first countable if for any x X, there is a countable basis for the neighborhoods of x. X is said
More information2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.
Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,
More informationBasics on Probability. Jingrui He 09/11/2007
Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationRings and groups. Ya. Sysak
Rings and groups. Ya. Sysak 1 Noetherian rings Let R be a ring. A (right) R -module M is called noetherian if it satisfies the maximum condition for its submodules. In other words, if M 1... M i M i+1...
More informationEECS 598: Statistical Learning Theory, Winter 2014 Topic 11. Kernels
EECS 598: Statistical Learning Theory, Winter 2014 Topic 11 Kernels Lecturer: Clayton Scott Scribe: Jun Guo, Soumik Chatterjee Disclaimer: These notes have not been subjected to the usual scrutiny reserved
More informationThe Multivariate Normal Distribution. In this case according to our theorem
The Multivariate Normal Distribution Defn: Z R 1 N(0, 1) iff f Z (z) = 1 2π e z2 /2. Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) T with the Z i independent and each Z i N(0, 1). In this
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationConvergence of Multivariate Quantile Surfaces
Convergence of Multivariate Quantile Surfaces Adil Ahidar Institut de Mathématiques de Toulouse - CERFACS August 30, 2013 Adil Ahidar (Institut de Mathématiques de Toulouse Convergence - CERFACS) of Multivariate
More informationIntroduction. Vectors and Matrices. Vectors [1] Vectors [2]
Introduction Vectors and Matrices Dr. TGI Fernando 1 2 Data is frequently arranged in arrays, that is, sets whose elements are indexed by one or more subscripts. Vector - one dimensional array Matrix -
More informationVector Spaces, Affine Spaces, and Metric Spaces
Vector Spaces, Affine Spaces, and Metric Spaces 2 This chapter is only meant to give a short overview of the most important concepts in linear algebra, affine spaces, and metric spaces and is not intended
More informationOutline of the course
School of Mathematical Sciences PURE MTH 3022 Geometry of Surfaces III, Semester 2, 20 Outline of the course Contents. Review 2. Differentiation in R n. 3 2.. Functions of class C k 4 2.2. Mean Value Theorem
More information5: MULTIVARATE STATIONARY PROCESSES
5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability
More informationIntroduction to Probability and Stocastic Processes - Part I
Introduction to Probability and Stocastic Processes - Part I Lecture 2 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark
More information8.1 Concentration inequality for Gaussian random matrix (cont d)
MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration
More informationConvex Optimization in Classification Problems
New Trends in Optimization and Computational Algorithms December 9 13, 2001 Convex Optimization in Classification Problems Laurent El Ghaoui Department of EECS, UC Berkeley elghaoui@eecs.berkeley.edu 1
More informationMAT-INF4110/MAT-INF9110 Mathematical optimization
MAT-INF4110/MAT-INF9110 Mathematical optimization Geir Dahl August 20, 2013 Convexity Part IV Chapter 4 Representation of convex sets different representations of convex sets, boundary polyhedra and polytopes:
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationProbability Distribution And Density For Functional Random Variables
Probability Distribution And Density For Functional Random Variables E. Cuvelier 1 M. Noirhomme-Fraiture 1 1 Institut d Informatique Facultés Universitaires Notre-Dame de la paix Namur CIL Research Contact
More informationMultivariate Gaussian Analysis
BS2 Statistical Inference, Lecture 7, Hilary Term 2009 February 13, 2009 Marginal and conditional distributions For a positive definite covariance matrix Σ, the multivariate Gaussian distribution has density
More informationLinear Algebra March 16, 2019
Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R
More informationIntroduction to Machine Learning (67577) Lecture 3
Introduction to Machine Learning (67577) Lecture 3 Shai Shalev-Shwartz School of CS and Engineering, The Hebrew University of Jerusalem General Learning Model and Bias-Complexity tradeoff Shai Shalev-Shwartz
More informationGaussians. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics
Gaussians Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Outline Univariate Gaussian Multivariate Gaussian Law of Total Probability Conditioning
More informationTAMS39 Lecture 2 Multivariate normal distribution
TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationChapter 4 Euclid Space
Chapter 4 Euclid Space Inner Product Spaces Definition.. Let V be a real vector space over IR. A real inner product on V is a real valued function on V V, denoted by (, ), which satisfies () (x, y) = (y,
More informationKernel Principal Component Analysis
Kernel Principal Component Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationOn the inverse matrix of the Laplacian and all ones matrix
On the inverse matrix of the Laplacian and all ones matrix Sho Suda (Joint work with Michio Seto and Tetsuji Taniguchi) International Christian University JSPS Research Fellow PD November 21, 2012 Sho
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationLecture 15: Multivariate normal distributions
Lecture 15: Multivariate normal distributions Normal distributions with singular covariance matrices Consider an n-dimensional X N(µ,Σ) with a positive definite Σ and a fixed k n matrix A that is not of
More informationConsistent Bivariate Distribution
A Characterization of the Normal Conditional Distributions MATSUNO 79 Therefore, the function ( ) = G( : a/(1 b2)) = N(0, a/(1 b2)) is a solu- tion for the integral equation (10). The constant times of
More informationCSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization
CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of
More informationLinear algebra for computational statistics
University of Seoul May 3, 2018 Vector and Matrix Notation Denote 2-dimensional data array (n p matrix) by X. Denote the element in the ith row and the jth column of X by x ij or (X) ij. Denote by X j
More informationNext tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2
Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution Defn: Z R 1 N(0,1) iff f Z (z) = 1 2π e z2 /2 Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) (a column
More informationDirect Limits. Mathematics 683, Fall 2013
Direct Limits Mathematics 683, Fall 2013 In this note we define direct limits and prove their basic properties. This notion is important in various places in algebra. In particular, in algebraic geometry
More informationMultivariate Statistics
Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical
More informationwhich is a group homomorphism, such that if W V U, then
4. Sheaves Definition 4.1. Let X be a topological space. A presheaf of groups F on X is a a function which assigns to every open set U X a group F(U) and to every inclusion V U a restriction map, ρ UV
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationStat 206: Linear algebra
Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two
More informationECON 5111 Mathematical Economics
Test 1 October 1, 2010 1. Construct a truth table for the following statement: [p (p q)] q. 2. A prime number is a natural number that is divisible by 1 and itself only. Let P be the set of all prime numbers
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationAppendices to the paper "Detecting Big Structural Breaks in Large Factor Models" (2013) by Chen, Dolado and Gonzalo.
Appendices to the paper "Detecting Big Structural Breaks in Large Factor Models" 203 by Chen, Dolado and Gonzalo. A.: Proof of Propositions and 2 he proof proceeds by showing that the errors, factors and
More informationMAT 445/ INTRODUCTION TO REPRESENTATION THEORY
MAT 445/1196 - INTRODUCTION TO REPRESENTATION THEORY CHAPTER 1 Representation Theory of Groups - Algebraic Foundations 1.1 Basic definitions, Schur s Lemma 1.2 Tensor products 1.3 Unitary representations
More informationEcon 508B: Lecture 5
Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline
More information4.4 Noetherian Rings
4.4 Noetherian Rings Recall that a ring A is Noetherian if it satisfies the following three equivalent conditions: (1) Every nonempty set of ideals of A has a maximal element (the maximal condition); (2)
More informationStable Process. 2. Multivariate Stable Distributions. July, 2006
Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationSummary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff
Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff David Gerard Department of Statistics University of Washington gerard2@uw.edu May 2, 2013 David Gerard (UW)
More informationMathematical Analysis Outline. William G. Faris
Mathematical Analysis Outline William G. Faris January 8, 2007 2 Chapter 1 Metric spaces and continuous maps 1.1 Metric spaces A metric space is a set X together with a real distance function (x, x ) d(x,
More informationELEMENTARY SUBALGEBRAS OF RESTRICTED LIE ALGEBRAS
ELEMENTARY SUBALGEBRAS OF RESTRICTED LIE ALGEBRAS J. WARNER SUMMARY OF A PAPER BY J. CARLSON, E. FRIEDLANDER, AND J. PEVTSOVA, AND FURTHER OBSERVATIONS 1. The Nullcone and Restricted Nullcone We will need
More informationFilters in Analysis and Topology
Filters in Analysis and Topology David MacIver July 1, 2004 Abstract The study of filters is a very natural way to talk about convergence in an arbitrary topological space, and carries over nicely into
More informationx. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).
.8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics
More informationIndependence for Full Conditional Measures, Graphoids and Bayesian Networks
Independence for Full Conditional Measures, Graphoids and Bayesian Networks Fabio G. Cozman Universidade de Sao Paulo Teddy Seidenfeld Carnegie Mellon University February 28, 2007 Abstract This paper examines
More informationA Magiv CV Theory for Large-Margin Classifiers
A Magiv CV Theory for Large-Margin Classifiers Hui Zou School of Statistics, University of Minnesota June 30, 2018 Joint work with Boxiang Wang Outline 1 Background 2 Magic CV formula 3 Magic support vector
More informationSTAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song
STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song Presenter: Jiwei Zhao Department of Statistics University of Wisconsin Madison April
More information6-1 Study Guide and Intervention Multivariable Linear Systems and Row Operations
6-1 Study Guide and Intervention Multivariable Linear Systems and Row Operations Gaussian Elimination You can solve a system of linear equations using matrices. Solving a system by transforming it into
More informationMean-field dual of cooperative reproduction
The mean-field dual of systems with cooperative reproduction joint with Tibor Mach (Prague) A. Sturm (Göttingen) Friday, July 6th, 2018 Poisson construction of Markov processes Let (X t ) t 0 be a continuous-time
More informationSlide05 Haykin Chapter 5: Radial-Basis Function Networks
Slide5 Haykin Chapter 5: Radial-Basis Function Networks CPSC 636-6 Instructor: Yoonsuck Choe Spring Learning in MLP Supervised learning in multilayer perceptrons: Recursive technique of stochastic approximation,
More informationSection 8.1. Vector Notation
Section 8.1 Vector Notation Definition 8.1 Random Vector A random vector is a column vector X = [ X 1 ]. X n Each Xi is a random variable. Definition 8.2 Vector Sample Value A sample value of a random
More informationSample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson
Sample Geometry Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois Spring
More informationPartitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12
Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following
More informationStat 710: Mathematical Statistics Lecture 31
Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:
More informationMeasure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond
Measure Theory on Topological Spaces Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond May 22, 2011 Contents 1 Introduction 2 1.1 The Riemann Integral........................................ 2 1.2 Measurable..............................................
More informationNonparameteric Regression:
Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,
More information