MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models
|
|
- Letitia Gibbs
- 5 years ago
- Views:
Transcription
1 1/13 MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models Dominique Guillot Departments of Mathematical Sciences University of Delaware May 4, 2016
2 Recall 2/13 An undirected graphical model is a set of random variables {X 1,..., X p } satisfying a Markov property.
3 Recall 2/13 An undirected graphical model is a set of random variables {X 1,..., X p } satisfying a Markov property. Let G = (V, E) be a graph on {1,..., p}.
4 Recall 2/13 An undirected graphical model is a set of random variables {X 1,..., X p } satisfying a Markov property. Let G = (V, E) be a graph on {1,..., p}. The pairwise Markov property: X i X j rest whenever (i, j) E.
5 Recall 2/13 An undirected graphical model is a set of random variables {X 1,..., X p } satisfying a Markov property. Let G = (V, E) be a graph on {1,..., p}. The pairwise Markov property: X i X j rest whenever (i, j) E. If the density of X = (X 1,..., X p ) is continuous and positive, then pairwise local global.
6 Recall 2/13 An undirected graphical model is a set of random variables {X 1,..., X p } satisfying a Markov property. Let G = (V, E) be a graph on {1,..., p}. The pairwise Markov property: X i X j rest whenever (i, j) E. If the density of X = (X 1,..., X p ) is continuous and positive, then pairwise local global. The HammersleyCliord theorem provides a necessary and sucient condition for a random vector to have a Markov random eld structure with respect to a given graph G.
7 Recall 2/13 An undirected graphical model is a set of random variables {X 1,..., X p } satisfying a Markov property. Let G = (V, E) be a graph on {1,..., p}. The pairwise Markov property: X i X j rest whenever (i, j) E. If the density of X = (X 1,..., X p ) is continuous and positive, then pairwise local global. The HammersleyCliord theorem provides a necessary and sucient condition for a random vector to have a Markov random eld structure with respect to a given graph G. We will now turn our attention to the special case of a random vector with a multivariate Gaussian distribution.
8 Recall: Multivariate Gaussian/normal distribution 3/13 Recall: X = (X 1,..., X p ) N(µ, Σ) where µ R p and Σ = (σ ij ) R p p is positive denite if 1 P (X A) = e 1 2 (x µ)t Σ 1 (x µ) dx 1... dx p. (2π)p det Σ A
9 Recall: Multivariate Gaussian/normal distribution 3/13 Recall: X = (X 1,..., X p ) N(µ, Σ) where µ R p and Σ = (σ ij ) R p p is positive denite if 1 P (X A) = e 1 2 (x µ)t Σ 1 (x µ) dx 1... dx p. (2π)p det Σ Bivariate case: A
10 Recall: Multivariate Gaussian/normal distribution 3/13 Recall: X = (X 1,..., X p ) N(µ, Σ) where µ R p and Σ = (σ ij ) R p p is positive denite if 1 P (X A) = e 1 2 (x µ)t Σ 1 (x µ) dx 1... dx p. (2π)p det Σ Bivariate case: A We have E(X) = µ, Cov(X i, X j ) = σ ij.
11 Recall: Multivariate Gaussian/normal distribution 3/13 Recall: X = (X 1,..., X p ) N(µ, Σ) where µ R p and Σ = (σ ij ) R p p is positive denite if 1 P (X A) = e 1 2 (x µ)t Σ 1 (x µ) dx 1... dx p. (2π)p det Σ Bivariate case: A We have E(X) = µ, Cov(X i, X j ) = σ ij. If Y = c + BX, where c R p and B R m p, then Y N(c + Bµ, BΣB T ).
12 Recall: Multivariate Gaussian/normal distribution Recall: X = (X 1,..., X p ) N(µ, Σ) where µ R p and Σ = (σ ij ) R p p is positive denite if 1 P (X A) = e 1 2 (x µ)t Σ 1 (x µ) dx 1... dx p. (2π)p det Σ Bivariate case: A We have E(X) = µ, Cov(X i, X j ) = σ ij. If Y = c + BX, where c R p and B R m p, then Y N(c + Bµ, BΣB T ). Note: Ω := Σ 1 is called the precision matrix or the concentration matrix of the distribution. 3/13
13 The Schur complement 4/13 Let ( ) A B M := C D where A = A m m, B = B m n, C = C n m, and D = D n n.
14 The Schur complement 4/13 Let ( ) A B M := C D where A = A m m, B = B m n, C = C n m, and D = D n n. Assuming D is invertible, the Schur complement of D in M is M/D := A BD 1 C.
15 The Schur complement 4/13 Let ( ) A B M := C D where A = A m m, B = B m n, C = C n m, and D = D n n. Assuming D is invertible, the Schur complement of D in M is Important properties: 1 det M = det D det(m/d). M/D := A BD 1 C.
16 The Schur complement 4/13 Let ( ) A B M := C D where A = A m m, B = B m n, C = C n m, and D = D n n. Assuming D is invertible, the Schur complement of D in M is Important properties: 1 det M = det D det(m/d). M/D := A BD 1 C. 2 M P n+m if and only if D P n and M/D P m. where P k = denotes the cone of k k real symmetric positive semidenite matrices.
17 The Schur complement 4/13 Let ( ) A B M := C D where A = A m m, B = B m n, C = C n m, and D = D n n. Assuming D is invertible, the Schur complement of D in M is Important properties: 1 det M = det D det(m/d). M/D := A BD 1 C. 2 M P n+m if and only if D P n and M/D P m. where P k = denotes the cone of k k real symmetric positive semidenite matrices. Proof: ( Im BD M = 1 ) ( A BD 1 ) ( C 0 Im 0 0 I n 0 D D 1 C I n ).
18 Multivariate Gaussian/normal distribution (cont.) 5/13 Conditional distribution: if A B is a partition of {1,..., p}, then X A X B = x B N(µ A B, Σ A B ),
19 Multivariate Gaussian/normal distribution (cont.) 5/13 Conditional distribution: if A B is a partition of {1,..., p}, then X A X B = x B N(µ A B, Σ A B ), with and µ A B := µ A + Σ AB Σ 1 BB (x B µ B ), Σ A B := Σ AA Σ AB Σ 1 BB Σ BA.
20 Multivariate Gaussian/normal distribution (cont.) 5/13 Conditional distribution: if A B is a partition of {1,..., p}, then X A X B = x B N(µ A B, Σ A B ), with µ A B := µ A + Σ AB Σ 1 BB (x B µ B ), and Σ A B := Σ AA Σ AB Σ 1 BB Σ BA. Marginals: to obtain the joint distribution of (X i, X j ), note that where (X i, X j ) T = B(X 1,..., X p ) T
21 Multivariate Gaussian/normal distribution (cont.) 5/13 Conditional distribution: if A B is a partition of {1,..., p}, then X A X B = x B N(µ A B, Σ A B ), with µ A B := µ A + Σ AB Σ 1 BB (x B µ B ), and Σ A B := Σ AA Σ AB Σ 1 BB Σ BA. Marginals: to obtain the joint distribution of (X i, X j ), note that where (X i, X j ) T = B(X 1,..., X p ) T B = ( I (p 2) ) R 2 p.
22 Multivariate Gaussian/normal distribution (cont.) Conditional distribution: if A B is a partition of {1,..., p}, then X A X B = x B N(µ A B, Σ A B ), with µ A B := µ A + Σ AB Σ 1 BB (x B µ B ), and Σ A B := Σ AA Σ AB Σ 1 BB Σ BA. Marginals: to obtain the joint distribution of (X i, X j ), note that where Therefore and Bµ = (X i, X j ) T = B(X 1,..., X p ) T B = ( I (p 2) ) R 2 p. (X i, X j ) T N(Bµ, BΣB T ), ( µ1 µ 2 ), BΣB T = ( σ11 σ 12 σ 21 σ 22 ). 5/13
23 Multivariate Gaussian/normal distribution (cont.) 6/13 Now, suppose X N(µ, Σ) with µ R p and Σ = (σ ij ) R p p psd.
24 Multivariate Gaussian/normal distribution (cont.) 6/13 Now, suppose X N(µ, Σ) with µ R p and Σ = (σ ij ) R p p psd. Claim: 1 X i X j i σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0.
25 Multivariate Gaussian/normal distribution (cont.) 6/13 Now, suppose X N(µ, Σ) with µ R p and Σ = (σ ij ) R p p psd. Claim: 1 X i X j i σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. Proof of (1): L X i X j X i X j = x j = Xi x j.
26 Multivariate Gaussian/normal distribution (cont.) 6/13 Now, suppose X N(µ, Σ) with µ R p and Σ = (σ ij ) R p p psd. Claim: 1 X i X j i σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. Proof of (1): L X i X j X i X j = x j = Xi x j. Now X i X j = x j N ( µ i + σ ) ii ρ(x j µ j ), (1 ρ 2 )σii 2, σ jj where ρ = σ ij σ ii σ jj is the correlation coecient between X i and X j.
27 6/13 Multivariate Gaussian/normal distribution (cont.) Now, suppose X N(µ, Σ) with µ R p and Σ = (σ ij ) R p p psd. Claim: 1 X i X j i σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. Proof of (1): X i X j X i X j = x j L = Xi x j. Now X i X j = x j N ( µ i + σ ) ii ρ(x j µ j ), (1 ρ 2 )σii 2, σ jj where ρ = σ ij σ ii σ jj is the correlation coecient between X i and X j. Therefore X i X j i ρ = 0 i σ ij = 0.
28 Multivariate Gaussian/normal distribution (cont.) 7/13 Proof of (2): Without loss of generality, assume (i, j) = (1, 2). Write µ, Σ in block form according to the partition A = {1, 2}, B = {3,..., p}: ( ) µ = (µ A, µ B ) T ΣAA Σ, Σ = AB. Σ BA Σ BB
29 Multivariate Gaussian/normal distribution (cont.) 7/13 Proof of (2): Without loss of generality, assume (i, j) = (1, 2). Write µ, Σ in block form according to the partition A = {1, 2}, B = {3,..., p}: ( ) µ = (µ A, µ B ) T ΣAA Σ, Σ = AB. Σ BA Σ BB Now where and (X 1, X 2 ) T rest = x B N(µ A B, Σ A B ), µ A B := µ A + Σ AB Σ 1 BB (x B µ B ), Σ A B := Σ AA Σ AB Σ 1 BB Σ BA
30 Multivariate Gaussian/normal distribution (cont.) 7/13 Proof of (2): Without loss of generality, assume (i, j) = (1, 2). Write µ, Σ in block form according to the partition A = {1, 2}, B = {3,..., p}: ( ) µ = (µ A, µ B ) T ΣAA Σ, Σ = AB. Σ BA Σ BB Now (X 1, X 2 ) T rest = x B N(µ A B, Σ A B ), where and µ A B := µ A + Σ AB Σ 1 BB (x B µ B ), Σ A B := Σ AA Σ AB Σ 1 BB Σ BA By part (1), X 1 X 2 rest i (Σ A B ) 12 = 0.
31 The inverse of a block matrix 8/13 Computing the inverse of a block matrix: Ref.: Petersen and Pedersen, The matrix cookbook.
32 The inverse of a block matrix 8/13 Computing the inverse of a block matrix: Ref.: Petersen and Pedersen, The matrix cookbook. It follows that Σ 1 A B = (Σ 1 ) 1:2,1:2
33 Multivariate Gaussian/normal distribution (cont.) 9/13 We have shown Σ 1 A B = (Σ 1 ) 1:2,1:2.
34 Multivariate Gaussian/normal distribution (cont.) 9/13 We have shown Also, we have ( ) 1 a b = b c Σ 1 A B = (Σ 1 ) 1:2,1:2. 1 ac b 2 ( c b b a ).
35 Multivariate Gaussian/normal distribution (cont.) 9/13 We have shown Also, we have Finally, ( ) 1 a b = b c Σ 1 A B = (Σ 1 ) 1:2,1:2. 1 ac b 2 ( c b b a ). (Σ A B ) 12 = 0 (Σ 1 A B ) 12 = 0 (Σ 1 ) 12 = 0.
36 Multivariate Gaussian/normal distribution (cont.) 9/13 We have shown Also, we have Finally, ( ) 1 a b = b c Σ 1 A B = (Σ 1 ) 1:2,1:2. 1 ac b 2 ( c b b a ). (Σ A B ) 12 = 0 (Σ 1 A B ) 12 = 0 (Σ 1 ) 12 = 0. Therefore, X i X j rest i (Σ 1 ) ij = 0.
37 Estimating the conditional independence structure of a GGM 10/13 We have shown that when X N(µ, Σ), 1 X i X j i Σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0.
38 Estimating the conditional independence structure of a GGM 10/13 We have shown that when X N(µ, Σ), 1 X i X j i Σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1.
39 Estimating the conditional independence structure of a GGM 10/13 We have shown that when X N(µ, Σ), 1 X i X j i Σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1. We will proceed in a way that is similar to the lasso.
40 Estimating the conditional independence structure of a GGM 10/13 We have shown that when X N(µ, Σ), 1 X i X j i Σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1. We will proceed in a way that is similar to the lasso. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1.
41 Estimating the conditional independence structure of a GGM 10/13 We have shown that when X N(µ, Σ), 1 X i X j i Σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1. We will proceed in a way that is similar to the lasso. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1. Suppose x (1),..., x (n) R p are iid observations of X. The associated log-likelihood of (µ, Σ) is given by l(µ, Σ) := n n 2 log det Σ 1 (x (i) µ) T Σ 1 (x (i) µ) np 2 2 log(2π). i=1
42 Estimating the conditional independence structure of a GGM We have shown that when X N(µ, Σ), 1 X i X j i Σ ij = 0. 2 X i X j rest i (Σ 1 ) ij = 0. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1. We will proceed in a way that is similar to the lasso. To discover the conditional structure of X, we need to estimate the structure of zeros of the precision matrix Ω = Σ 1. Suppose x (1),..., x (n) R p are iid observations of X. The associated log-likelihood of (µ, Σ) is given by l(µ, Σ) := n n 2 log det Σ 1 (x (i) µ) T Σ 1 (x (i) µ) np 2 2 log(2π). i=1 Classical result: the MLE of (µ, Σ) is given by ˆµ := 1 n x (i), S := 1 n (x (i) ˆµ)(x (i) ˆµ) T. n n i=1 i=1 10/13
43 Estimating the CI structure of a GGM (cont.) 11/13 Using ˆµ and Σ, we can conveniently rewrite the log-likelihood as: l(µ, Σ) = n 2 log det Σ n 2 Tr(SΣ 1 ) np 2 log(2π) n 2 Tr(Σ 1 (ˆµ µ)(ˆµ µ) T ). (use the identity x T Ax = Tr(Axx T ).
44 Estimating the CI structure of a GGM (cont.) 11/13 Using ˆµ and Σ, we can conveniently rewrite the log-likelihood as: l(µ, Σ) = n 2 log det Σ n 2 Tr(SΣ 1 ) np 2 log(2π) n 2 Tr(Σ 1 (ˆµ µ)(ˆµ µ) T ). (use the identity x T Ax = Tr(Axx T ). Note that the last term is minimized when µ = ˆµ (independently of Σ) since Tr(Σ 1 (ˆµ µ)(ˆµ µ) T ) = (ˆµ µ) T Σ 1 (ˆµ µ) 0. (The last inequality holds since Σ 1 is positive denite.)
45 Estimating the CI structure of a GGM (cont.) 11/13 Using ˆµ and Σ, we can conveniently rewrite the log-likelihood as: l(µ, Σ) = n 2 log det Σ n 2 Tr(SΣ 1 ) np 2 log(2π) n 2 Tr(Σ 1 (ˆµ µ)(ˆµ µ) T ). (use the identity x T Ax = Tr(Axx T ). Note that the last term is minimized when µ = ˆµ (independently of Σ) since Tr(Σ 1 (ˆµ µ)(ˆµ µ) T ) = (ˆµ µ) T Σ 1 (ˆµ µ) 0. (The last inequality holds since Σ 1 is positive denite.) Therefore the log-likelihood of Ω := Σ 1 is l(ω) log det Ω Tr(SΩ) (up to a constant).
46 The Graphical Lasso 12/13 The Graphical Lasso (glasso) algorithm (Friedman, Hastie, Tibshirani, 2007), Banerjee et al. (2007), solves the penalized likelihood problem: ˆΩ ρ = argmax Ω psd log det Ω Tr(SΩ) ρ p Ω 1, i,j=1 where Ω 1 := p i,j=1 Ω ij, and ρ > 0 is a xed regularization parameter.
47 The Graphical Lasso 12/13 The Graphical Lasso (glasso) algorithm (Friedman, Hastie, Tibshirani, 2007), Banerjee et al. (2007), solves the penalized likelihood problem: ˆΩ ρ = argmax Ω psd log det Ω Tr(SΩ) ρ p Ω 1, i,j=1 where Ω 1 := p i,j=1 Ω ij, and ρ > 0 is a xed regularization parameter. Idea: Make a trade-o between maximizing the likelihood and having a sparse Ω.
48 The Graphical Lasso 12/13 The Graphical Lasso (glasso) algorithm (Friedman, Hastie, Tibshirani, 2007), Banerjee et al. (2007), solves the penalized likelihood problem: ˆΩ ρ = argmax Ω psd log det Ω Tr(SΩ) ρ p Ω 1, i,j=1 where Ω 1 := p i,j=1 Ω ij, and ρ > 0 is a xed regularization parameter. Idea: Make a trade-o between maximizing the likelihood and having a sparse Ω. Just like in the lasso problem, using a 1-norm tends to introduce many zeros into Ω.
49 The Graphical Lasso 12/13 The Graphical Lasso (glasso) algorithm (Friedman, Hastie, Tibshirani, 2007), Banerjee et al. (2007), solves the penalized likelihood problem: ˆΩ ρ = argmax Ω psd log det Ω Tr(SΩ) ρ p Ω 1, i,j=1 where Ω 1 := p i,j=1 Ω ij, and ρ > 0 is a xed regularization parameter. Idea: Make a trade-o between maximizing the likelihood and having a sparse Ω. Just like in the lasso problem, using a 1-norm tends to introduce many zeros into Ω. The regularization parameter ρ can be chosen by cross-validation.
50 The Graphical Lasso 12/13 The Graphical Lasso (glasso) algorithm (Friedman, Hastie, Tibshirani, 2007), Banerjee et al. (2007), solves the penalized likelihood problem: ˆΩ ρ = argmax Ω psd log det Ω Tr(SΩ) ρ p Ω 1, i,j=1 where Ω 1 := p i,j=1 Ω ij, and ρ > 0 is a xed regularization parameter. Idea: Make a trade-o between maximizing the likelihood and having a sparse Ω. Just like in the lasso problem, using a 1-norm tends to introduce many zeros into Ω. The regularization parameter ρ can be chosen by cross-validation. The above problem can be eciently solved for problems with up to a few thousand variables (see e.g. ESL, Algorithm 17.2).
51 MLE estimation of a GGM 13/13 From the glasso solution, one infers a conditional independence graph for X = (X 1,..., X p ).
52 MLE estimation of a GGM 13/13 From the glasso solution, one infers a conditional independence graph for X = (X 1,..., X p ). Given a graph G = (V, E) with p vertices, let P G := {A P p : A ij = 0 if (i, j) E}.
53 MLE estimation of a GGM 13/13 From the glasso solution, one infers a conditional independence graph for X = (X 1,..., X p ). Given a graph G = (V, E) with p vertices, let P G := {A P p : A ij = 0 if (i, j) E}. We can now estimate the optimal covariance matrix with the given graph structure by solving: ˆΣ G := argmax l(σ), Σ : Ω=Σ 1 P G where l(σ) denotes the log-likelihood of Σ.
54 MLE estimation of a GGM 13/13 From the glasso solution, one infers a conditional independence graph for X = (X 1,..., X p ). Given a graph G = (V, E) with p vertices, let P G := {A P p : A ij = 0 if (i, j) E}. We can now estimate the optimal covariance matrix with the given graph structure by solving: ˆΣ G := argmax l(σ), Σ : Ω=Σ 1 P G where l(σ) denotes the log-likelihood of Σ. Note: Instead of maximizing the log-likelihood over all possible psd matrices as in the classical case, we restrict ourselves to the matrices having the right conditional independence structure.
55 MLE estimation of a GGM 13/13 From the glasso solution, one infers a conditional independence graph for X = (X 1,..., X p ). Given a graph G = (V, E) with p vertices, let P G := {A P p : A ij = 0 if (i, j) E}. We can now estimate the optimal covariance matrix with the given graph structure by solving: ˆΣ G := argmax l(σ), Σ : Ω=Σ 1 P G where l(σ) denotes the log-likelihood of Σ. Note: Instead of maximizing the log-likelihood over all possible psd matrices as in the classical case, we restrict ourselves to the matrices having the right conditional independence structure. The graphical MLE problem can be solved eciently for up to a few thousand variables (see e.g. ESL, Algorithm 17.1).
MATH 829: Introduction to Data Mining and Analysis Graphical Models III - Gaussian Graphical Models (cont.)
1/12 MATH 829: Introduction to Data Mining and Analysis Graphical Models III - Gaussian Graphical Models (cont.) Dominique Guillot Departments of Mathematical Sciences University of Delaware May 6, 2016
More informationMATH 829: Introduction to Data Mining and Analysis Graphical Models I
MATH 829: Introduction to Data Mining and Analysis Graphical Models I Dominique Guillot Departments of Mathematical Sciences University of Delaware May 2, 2016 1/12 Independence and conditional independence:
More informationMATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression
1/9 MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression Dominique Guillot Deartments of Mathematical Sciences University of Delaware February 15, 2016 Distribution of regression
More informationChapter 17: Undirected Graphical Models
Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More information10708 Graphical Models: Homework 2
10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves
More informationGaussian Graphical Models and Graphical Lasso
ELE 538B: Sparsity, Structure and Inference Gaussian Graphical Models and Graphical Lasso Yuxin Chen Princeton University, Spring 2017 Multivariate Gaussians Consider a random vector x N (0, Σ) with pdf
More informationIntroduction to Normal Distribution
Introduction to Normal Distribution Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 17-Jan-2017 Nathaniel E. Helwig (U of Minnesota) Introduction
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationLearning Gaussian Graphical Models with Unknown Group Sparsity
Learning Gaussian Graphical Models with Unknown Group Sparsity Kevin Murphy Ben Marlin Depts. of Statistics & Computer Science Univ. British Columbia Canada Connections Graphical models Density estimation
More informationMultivariate Normal Models
Case Study 3: fmri Prediction Coping with Large Covariances: Latent Factor Models, Graphical Models, Graphical LASSO Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox February
More informationGraphical Model Selection
May 6, 2013 Trevor Hastie, Stanford Statistics 1 Graphical Model Selection Trevor Hastie Stanford University joint work with Jerome Friedman, Rob Tibshirani, Rahul Mazumder and Jason Lee May 6, 2013 Trevor
More informationRobust and sparse Gaussian graphical modelling under cell-wise contamination
Robust and sparse Gaussian graphical modelling under cell-wise contamination Shota Katayama 1, Hironori Fujisawa 2 and Mathias Drton 3 1 Tokyo Institute of Technology, Japan 2 The Institute of Statistical
More informationJoint Gaussian Graphical Model Review Series I
Joint Gaussian Graphical Model Review Series I Probability Foundations Beilun Wang Advisor: Yanjun Qi 1 Department of Computer Science, University of Virginia http://jointggm.org/ June 23rd, 2017 Beilun
More informationMultivariate Normal Models
Case Study 3: fmri Prediction Graphical LASSO Machine Learning/Statistics for Big Data CSE599C1/STAT592, University of Washington Emily Fox February 26 th, 2013 Emily Fox 2013 1 Multivariate Normal Models
More informationGaussian Models (9/9/13)
STA561: Probabilistic machine learning Gaussian Models (9/9/13) Lecturer: Barbara Engelhardt Scribes: Xi He, Jiangwei Pan, Ali Razeen, Animesh Srivastava 1 Multivariate Normal Distribution The multivariate
More informationInverse Covariance Estimation with Missing Data using the Concave-Convex Procedure
Inverse Covariance Estimation with Missing Data using the Concave-Convex Procedure Jérôme Thai 1 Timothy Hunter 1 Anayo Akametalu 1 Claire Tomlin 1 Alex Bayen 1,2 1 Department of Electrical Engineering
More informationStructure estimation for Gaussian graphical models
Faculty of Science Structure estimation for Gaussian graphical models Steffen Lauritzen, University of Copenhagen Department of Mathematical Sciences Minikurs TUM 2016 Lecture 3 Slide 1/48 Overview of
More informationTAMS39 Lecture 2 Multivariate normal distribution
TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution
More informationHigh Dimensional Inverse Covariate Matrix Estimation via Linear Programming
High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω
More informationHigh-dimensional Covariance Estimation Based On Gaussian Graphical Models
High-dimensional Covariance Estimation Based On Gaussian Graphical Models Shuheng Zhou, Philipp Rutimann, Min Xu and Peter Buhlmann February 3, 2012 Problem definition Want to estimate the covariance matrix
More informationMATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels
1/12 MATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels Dominique Guillot Departments of Mathematical Sciences University of Delaware March 14, 2016 Separating sets:
More informationA Robust Approach to Regularized Discriminant Analysis
A Robust Approach to Regularized Discriminant Analysis Moritz Gschwandtner Department of Statistics and Probability Theory Vienna University of Technology, Austria Österreichische Statistiktage, Graz,
More informationThe Multivariate Normal Distribution. In this case according to our theorem
The Multivariate Normal Distribution Defn: Z R 1 N(0, 1) iff f Z (z) = 1 2π e z2 /2. Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) T with the Z i independent and each Z i N(0, 1). In this
More informationSparse inverse covariance estimation with the lasso
Sparse inverse covariance estimation with the lasso Jerome Friedman Trevor Hastie and Robert Tibshirani November 8, 2007 Abstract We consider the problem of estimating sparse graphs by a lasso penalty
More informationSTAT 730 Chapter 4: Estimation
STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum
More informationPermutation-invariant regularization of large covariance matrices. Liza Levina
Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work
More informationMATH 829: Introduction to Data Mining and Analysis Clustering II
his lecture is based on U. von Luxburg, A Tutorial on Spectral Clustering, Statistics and Computing, 17 (4), 2007. MATH 829: Introduction to Data Mining and Analysis Clustering II Dominique Guillot Departments
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More informationApproximation. Inderjit S. Dhillon Dept of Computer Science UT Austin. SAMSI Massive Datasets Opening Workshop Raleigh, North Carolina.
Using Quadratic Approximation Inderjit S. Dhillon Dept of Computer Science UT Austin SAMSI Massive Datasets Opening Workshop Raleigh, North Carolina Sept 12, 2012 Joint work with C. Hsieh, M. Sustik and
More informationUndirected Graphical Models
Undirected Graphical Models 1 Conditional Independence Graphs Let G = (V, E) be an undirected graph with vertex set V and edge set E, and let A, B, and C be subsets of vertices. We say that C separates
More informationCS 195-5: Machine Learning Problem Set 1
CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:
More informationHigh-dimensional covariance estimation based on Gaussian graphical models
High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,
More informationProperties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation
Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Adam J. Rothman School of Statistics University of Minnesota October 8, 2014, joint work with Liliana
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationGaussian Graphical Models: An Algebraic and Geometric Perspective
Gaussian Graphical Models: An Algebraic and Geometric Perspective Caroline Uhler arxiv:707.04345v [math.st] 3 Jul 07 Abstract Gaussian graphical models are used throughout the natural sciences, social
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationJoint Gaussian Graphical Model Review Series I
Joint Gaussian Graphical Model Review Series I Probability Foundations Beilun Wang Advisor: Yanjun Qi 1 Department of Computer Science, University of Virginia http://jointggm.org/ June 23rd, 2017 Beilun
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,
More informationAn efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss
An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss arxiv:1811.04545v1 [stat.co] 12 Nov 2018 Cheng Wang School of Mathematical Sciences, Shanghai Jiao
More informationLearning Multiple Tasks with a Sparse Matrix-Normal Penalty
Learning Multiple Tasks with a Sparse Matrix-Normal Penalty Yi Zhang and Jeff Schneider NIPS 2010 Presented by Esther Salazar Duke University March 25, 2011 E. Salazar (Reading group) March 25, 2011 1
More informationSparse Gaussian conditional random fields
Sparse Gaussian conditional random fields Matt Wytock, J. ico Kolter School of Computer Science Carnegie Mellon University Pittsburgh, PA 53 {mwytock, zkolter}@cs.cmu.edu Abstract We propose sparse Gaussian
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationDecomposable and Directed Graphical Gaussian Models
Decomposable Decomposable and Directed Graphical Gaussian Models Graphical Models and Inference, Lecture 13, Michaelmas Term 2009 November 26, 2009 Decomposable Definition Basic properties Wishart density
More informationNotes on the Multivariate Normal and Related Topics
Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions
More informationProbabilistic Models of Past Climate Change
Probabilistic Models of Past Climate Change Julien Emile-Geay USC College, Department of Earth Sciences SIAM International Conference on Data Mining Anaheim, California April 27, 2012 Probabilistic Models
More informationGaussian Mixture Models
Gaussian Mixture Models David Rosenberg, Brett Bernstein New York University April 26, 2017 David Rosenberg, Brett Bernstein (New York University) DS-GA 1003 April 26, 2017 1 / 42 Intro Question Intro
More informationSparse Permutation Invariant Covariance Estimation: Motivation, Background and Key Results
Sparse Permutation Invariant Covariance Estimation: Motivation, Background and Key Results David Prince Biostat 572 dprince3@uw.edu April 19, 2012 David Prince (UW) SPICE April 19, 2012 1 / 11 Electronic
More informationLearning Markov Network Structure using Brownian Distance Covariance
arxiv:.v [stat.ml] Jun 0 Learning Markov Network Structure using Brownian Distance Covariance Ehsan Khoshgnauz May, 0 Abstract In this paper, we present a simple non-parametric method for learning the
More informationLinear Regression and Discrimination
Linear Regression and Discrimination Kernel-based Learning Methods Christian Igel Institut für Neuroinformatik Ruhr-Universität Bochum, Germany http://www.neuroinformatik.rub.de July 16, 2009 Christian
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Gaussian graphical models and Ising models: modeling networks Eric Xing Lecture 0, February 7, 04 Reading: See class website Eric Xing @ CMU, 005-04
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Recitation 3 1 Gaussian Graphical Models: Schur s Complement Consider
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationCovariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance
Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture
More informationLikelihood Analysis of Gaussian Graphical Models
Faculty of Science Likelihood Analysis of Gaussian Graphical Models Ste en Lauritzen Department of Mathematical Sciences Minikurs TUM 2016 Lecture 2 Slide 1/43 Overview of lectures Lecture 1 Markov Properties
More informationCMPE 58K Bayesian Statistics and Machine Learning Lecture 5
CMPE 58K Bayesian Statistics and Machine Learning Lecture 5 Multivariate distributions: Gaussian, Bernoulli, Probability tables Department of Computer Engineering, Boğaziçi University, Istanbul, Turkey
More informationAn Introduction to Graphical Lasso
An Introduction to Graphical Lasso Bo Chang Graphical Models Reading Group May 15, 2015 Bo Chang (UBC) Graphical Lasso May 15, 2015 1 / 16 Undirected Graphical Models An undirected graph, each vertex represents
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationProximity-Based Anomaly Detection using Sparse Structure Learning
Proximity-Based Anomaly Detection using Sparse Structure Learning Tsuyoshi Idé (IBM Tokyo Research Lab) Aurelie C. Lozano, Naoki Abe, and Yan Liu (IBM T. J. Watson Research Center) 2009/04/ SDM 2009 /
More informationLecture 1 October 9, 2013
Probabilistic Graphical Models Fall 2013 Lecture 1 October 9, 2013 Lecturer: Guillaume Obozinski Scribe: Huu Dien Khue Le, Robin Bénesse The web page of the course: http://www.di.ens.fr/~fbach/courses/fall2013/
More informationMATH 829: Introduction to Data Mining and Analysis Support vector machines
1/10 MATH 829: Introduction to Data Mining and Analysis Support vector machines Dominique Guillot Departments of Mathematical Sciences University of Delaware March 11, 2016 Hyperplanes 2/10 Recall: A hyperplane
More information11 : Gaussian Graphic Models and Ising Models
10-708: Probabilistic Graphical Models 10-708, Spring 2017 11 : Gaussian Graphic Models and Ising Models Lecturer: Bryon Aragam Scribes: Chao-Ming Yen 1 Introduction Different from previous maximum likelihood
More informationRandom Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30
Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationThe Multivariate Gaussian Distribution
The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationBig & Quic: Sparse Inverse Covariance Estimation for a Million Variables
for a Million Variables Cho-Jui Hsieh The University of Texas at Austin NIPS Lake Tahoe, Nevada Dec 8, 2013 Joint work with M. Sustik, I. Dhillon, P. Ravikumar and R. Poldrack FMRI Brain Analysis Goal:
More informationChapter 5. The multivariate normal distribution. Probability Theory. Linear transformations. The mean vector and the covariance matrix
Probability Theory Linear transformations A transformation is said to be linear if every single function in the transformation is a linear combination. Chapter 5 The multivariate normal distribution When
More informationMultivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013
Multivariate Gaussian Distribution Auxiliary notes for Time Series Analysis SF2943 Spring 203 Timo Koski Department of Mathematics KTH Royal Institute of Technology, Stockholm 2 Chapter Gaussian Vectors.
More informationChapter 4: Asymptotic Properties of the MLE (Part 2)
Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response
More informationFinite Singular Multivariate Gaussian Mixture
21/06/2016 Plan 1 Basic definitions Singular Multivariate Normal Distribution 2 3 Plan Singular Multivariate Normal Distribution 1 Basic definitions Singular Multivariate Normal Distribution 2 3 Multivariate
More informationOn an Additive Semigraphoid Model for Statistical Networks With Application to Nov Pathway 25, 2016 Analysis -1 Bing / 38Li,
On an Additive Semigraphoid Model for Statistical Networks With Application to Pathway Analysis - Bing Li, Hyunho Chun & Hongyu Zhao Kim Youngrae SNU Stat. Multivariate Lab Nov 25, 2016 On an Additive
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationLecture 2 Part 1 Optimization
Lecture 2 Part 1 Optimization (January 16, 2015) Mu Zhu University of Waterloo Need for Optimization E(y x), P(y x) want to go after them first, model some examples last week then, estimate didn t discuss
More informationBIOS 2083 Linear Models Abdus S. Wahed. Chapter 2 84
Chapter 2 84 Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random variables. For instance, X = X 1 X 2.
More informationx. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).
.8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics
More informationThe lasso: some novel algorithms and applications
1 The lasso: some novel algorithms and applications Newton Institute, June 25, 2008 Robert Tibshirani Stanford University Collaborations with Trevor Hastie, Jerome Friedman, Holger Hoefling, Gen Nowak,
More informationRegulatory Inferece from Gene Expression. CMSC858P Spring 2012 Hector Corrada Bravo
Regulatory Inferece from Gene Expression CMSC858P Spring 2012 Hector Corrada Bravo 2 Graphical Model Let y be a vector- valued random variable Suppose some condi8onal independence proper8es hold for some
More informationMATH 567: Mathematical Techniques in Data Science Logistic regression and Discriminant Analysis
Logistic regression MATH 567: Mathematical Techniques in Data Science Logistic regression and Discriminant Analysis Dominique Guillot Departments of Mathematical Sciences University of Delaware March 6,
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Gaussian graphical models and Ising models: modeling networks Eric Xing Lecture 0, February 5, 06 Reading: See class website Eric Xing @ CMU, 005-06
More informationThe Multivariate Normal Distribution 1
The Multivariate Normal Distribution 1 STA 302 Fall 2014 1 See last slide for copyright information. 1 / 37 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should
More informationHigh Dimensional Covariance and Precision Matrix Estimation
High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance
More informationProbability Distributions
Probability Distributions Seungjin Choi Department of Computer Science Pohang University of Science and Technology, Korea seungjin@postech.ac.kr 1 / 25 Outline Summarize the main properties of some of
More informationMAS223 Statistical Modelling and Inference Examples
Chapter MAS3 Statistical Modelling and Inference Examples Example : Sample spaces and random variables. Let S be the sample space for the experiment of tossing two coins; i.e. Define the random variables
More informationEstimation of Graphical Models with Shape Restriction
Estimation of Graphical Models with Shape Restriction BY KHAI X. CHIONG USC Dornsife INE, Department of Economics, University of Southern California, Los Angeles, California 989, U.S.A. kchiong@usc.edu
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More informationMultivariate Gaussians. Sargur Srihari
Multivariate Gaussians Sargur srihari@cedar.buffalo.edu 1 Topics 1. Multivariate Gaussian: Basic Parameterization 2. Covariance and Information Form 3. Operations on Gaussians 4. Independencies in Gaussians
More information2.3. The Gaussian Distribution
78 2. PROBABILITY DISTRIBUTIONS Figure 2.5 Plots of the Dirichlet distribution over three variables, where the two horizontal axes are coordinates in the plane of the simplex and the vertical axis corresponds
More informationMath 152. Rumbos Fall Solutions to Assignment #12
Math 52. umbos Fall 2009 Solutions to Assignment #2. Suppose that you observe n iid Bernoulli(p) random variables, denoted by X, X 2,..., X n. Find the LT rejection region for the test of H o : p p o versus
More informationGenetic Networks. Korbinian Strimmer. Seminar: Statistical Analysis of RNA-Seq Data 19 June IMISE, Universität Leipzig
Genetic Networks Korbinian Strimmer IMISE, Universität Leipzig Seminar: Statistical Analysis of RNA-Seq Data 19 June 2012 Korbinian Strimmer, RNA-Seq Networks, 19/6/2012 1 Paper G. I. Allen and Z. Liu.
More informationSparse Inverse Covariance Estimation for a Million Variables
Sparse Inverse Covariance Estimation for a Million Variables Inderjit S. Dhillon Depts of Computer Science & Mathematics The University of Texas at Austin SAMSI LDHD Opening Workshop Raleigh, North Carolina
More informationGRAPHICAL MODELS FOR HIGH DIMENSIONAL GENOMIC DATA. Min Jin Ha
GRAPHICAL MODELS FOR HIGH DIMENSIONAL GENOMIC DATA Min Jin Ha A dissertation submitted to the faculty of the University of North Carolina at Chapel Hill in partial fulfillment of the requirements for the
More informationLecture 25: Review. Statistics 104. April 23, Colin Rundel
Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April
More information