On the Conditional Distribution of a Multivariate Normal given a Transformation - the Linear Case BY RAJESHWARI MAJUMDAR* AND SUMAN MAJUMDAR

Size: px
Start display at page:

Download "On the Conditional Distribution of a Multivariate Normal given a Transformation - the Linear Case BY RAJESHWARI MAJUMDAR* AND SUMAN MAJUMDAR"

Transcription

1 On the Conditional Distribution of a Multivariate Normal given a Transformation - the Linear Case BY RAJESHWARI MAJUMDAR* AND SUMAN MAJUMDAR We show that the orthogonal projection operator onto the range of the adjoint of a linear operator X can be represented as YX, where Y is an invertible linear operator. Given a Normal random vector ] and a linear operator X, we use this representation to obtain a linear operator Xs such that X] s is independent of X] and ] X] s is an affine function of X]. We then use this decomposition to prove that the conditional distribution of a Normal random vector ] given g ], where g is a linear transformation, is again a multivariate Normal distribution. This result is equivalent to the well-known result that given a 5-dimensional component of a -dimensional Normal random vector, where 5, the conditional distribution of the remaining 5-dimensional component is a 5-dimensional multivariate Normal distribution, and sets the stage for approximating the conditional distribution of ] given 1 ], where 1 is a continuously differentiable vector field. 1. Introduction. What can we ascertain about the conditional distribution of a multivariate Normal random vector ] d given 1], where 1Àd Èd 7 is a measurable function? Clearly, given a particular functional form of 1, the problem is a very specific one, and depending on the functional form, may or may not have a closed form solution. Our objective is to derive an approximation to the conditional distribution in question based on some regularity properties of 1. Specifically, in this paper we find the conditional distribution when 1 is a linear transformation, and in a companion paper use that to derive the desired approximation when 1 is a continuously differentiable ( G " ) vector field by exploiting the local linearity of 1. Before proceeding further, we present a brief review of what is known about the conditional distribution when 1 is a linear transformation, to be denoted by g in what follows. Casella and Berger (2002) define the bivariate Normal distribution by specifying the joint density in terms of the five parameters of the distribution - the means." and., the variances 5" and 5, and the correlation 3 [Definition ]. They calculate the marginal density of ] " and note that (using the joint and marginal densities) it is easy to verify that the conditional distribution of ] given ] œ C is Normal with " " 5 mean œ. 3 C". " and variance œ 5 "3. 5" Anderson (194, Section 2.5.1) and Flury (1997, Theorem 3.3.1) generalize this result to MSC 2010 subject classifications: 15A1, 46B2, 47A05, 60E99, 62B05, 62E15. Keywords and phrases: Positive operator; Orthogonal projection operator; Spectral Theorem; Multivariate Normal distribution; Independence; Sufficiency; Analysis of Variance.

2 the multivariate Normal distribution. While Anderson (194) deals with the Lebesgue density of the multivariate Normal distribution (which exists only when the covariance matrix is of full rank), Flury (1997) avoids dealing with the density by defining the multivariate Normal distribution in terms of linear functionals, but requires the covariance matrix of the conditioning component to be of full rank. Though their approaches to defining the multivariate Normal distribution differ, both Muirhead (192, Theorem ) and Eaton (193, Proposition 3.13) obtain, as described below, the conditional distribution without any restriction on the rank of the covariance matrix. Let ] " have the multivariate Normal distribution with mean vector. " and covariance matrix D. Partition ],., and Das ]"." D"" D" ]œ ß. œ ß D œ ]. D" D where ]" and." are 5 " and D"" is 5 5. Then ] D" D "" ]" and ]" are jointly Normal and uncorrelated, hence independent, where D"" is a generalized inverse of D"". Consequently, the conditional distribution of ] D" D "" ]" given ]" equals the unconditional distribution of ] D" D "" ]", which is multivariate Normal in 5 dimensions with mean. D" D ""." and covariance D D" D"" D". Thus, the conditional distribution of ] given ]" is multivariate Normal in 5dimensions with mean. D D ]. and covariance D D D D. " "" " " " "" " Given two topological vector spaces Z and [, let _Zß[ denote the linear space of continuous linear transformations from Z to [, and let _ ZßZ œ _ Z. Now note 7 w 7 that, if g _ d ßd, the joint distribution of g ]ß] d is multivariate Normal with mean œ E. E E E and covariance œ D D w. DE D [ Muirhead (192, Theorem 1.2.6)], where the 7 matrix E represents the 7 transformation g with respect to the standard orthonormal bases of d and d. By the results of Muirhead (192) and Eaton (193) cited in the previous paragraph, we obtain that the conditional distribution of ] given g ] is multivariate Normal in d with mean w w w w. DE ED E E ]. and covariance DDE EDE ED. Unfortunately, this derivation of the conditional distribution of ] given g ], because of its dependence on manipulative matrix algebra, is not of much help when it comes to " approximating the conditional distribution of ] given 1 ] for a 1 in G by exploiting the local linearity of 1. In what follows, we present an alternative derivation of the conditional distribution of ] given g ]. Given X _d, we find in Theorem 3 X s _d, depending on X and the covariance of ], such that X] s is independent of X] ] X] s 7 and is an affine function of X]. In Theorem 4, given g _d ßd, we find X _d such that the conditional distribution of ] given X] equals that given g ], and use the decomposition obtained in Theorem 3 to obtain the conditional distribution of ] given X], hence that given g ]. Our derivation facilitates the w

3 approximation of the conditional distribution when the transformation 1 is nonlinear but continuously differentiable; see Remark 4 for a brief outline. We define a multivariate Normal distribution in terms of an affine transformation of the random vector with coordinates independent and identically distributed (iid, hereinafter) standard Normals, as in Muirhead (192), but work with the covariance operator (instead of matrix) and characteristic function. This coordinate-free approach allows us to seamlessly subsume the possibility that the multivariate Normal distribution is supported on a proper subspace of d which is not spanned by a subset of the standard orthonormal basis /ßâß/ ". In the spirit of Axler (2015), references to which in the sequel are by omission only, this relegates the manipulative matrix algebra that so dominates the multivariate Normal literature to the back burner. The paper is organized as follows. In Section 2, we introduce all the notations, definitions, and results from linear algebra used in Section 3. The main result of this section is Theorem 1, where we show, using the Spectral Theorem, that the orthogonal projection operator onto the range of the adjoint of a linear operator X is YX, where Y is an invertible linear operator. Theorem 1 is used in our proof of Theorem 3. In section 3, we first summarize the basic facts about the multivariate Normal distribution in our coordinate-free setup. Theorem 2 shows that if ] is d -valued multivariate 7 : 7 Normal, then f]ß g ] is d d -valued multivariate Normal, where f _ d ßd : and g _ d ßd. Corollary 1, which is used in our proof of Theorem 3, then formulates a necessary and sufficient condition in terms of the operators H, f, and g for the independence of f] and g ], where H is the covariance of ]. Note that, since a component of a vector is a linear transformation of the vector and a linear transformation of a multivariate Normal random variable is another multivariate Normal random variable [Lemma 5], Theorem 4 allows us to deduce Theorem (b) of Muirhead (192) (and Proposition 3.13 of Eaton (193)) as an immediate corollary. We also present alternative derivations of the independence of the sample mean and the sample variance of a random sample from a Normal distribution [Remark 1], the "partial out" formula for population regression in the Normal model [Corollary 2], and the sufficiency of the sample mean in the Normal location model [Remark 3]. We simplify the expressions for the conditional mean and covariance obtained in Theorem 4 in Remark 2, leading to a direct verification of the iterated expectation and the analysis of variance formulae. We outline a direction in which our method can possibly be extended in Remark 5. The Appendix contains two technical lemmas that are used in the proofs of Theorem 4 and Corollary 2. The following notational conventions and their consequences are used throughout the rest of the paper. The equality of two random variables, unless otherwise mentioned, implies equality almost surely. For any Polish space Ë, let UË denote the Borel 5-algebra of Ë.

4 Let 2 be a map from an arbitrary set Ì into a measurable space Íßm. For an arbitrary subset F of Í, let 2 " F denote C Ì À 2C F, whereas for an arbitrary subset E " of Ì, let 2E denote 2C ÀC E. Note that 2 2E œefor every subset Eof Ì. Let 52 denote the smallest 5-algebra of subsets of Ì that makes 2 measurable. Since 2 " F À F m is a 5 -algebra of subsets of Ì [Dudley (199, page 9)], we " obtain 52 œ 2 F À F m. 2. The results from Linear Algebra. Let Zß[ be finite-dimensional vector spaces. For any f _ [ßZ, let ef Z denote the range of f and af [ denote the null space of f. The result of Lemma 1 is used in the proof of Theorem 1. It is mentioned in Exercise 3.D.3; we present a proof for the sake of completeness. Lemma 1. Let Z be a finite-dimensional vector space, [ a subspace of Z, and h _ [ßZ. There exists an invertible operator Y _ Z such that YA œ ha for every A [ if and only if h is injective. Proof of Lemma 1. Let Y _Z be invertible such that YAœhA for every A [. Let A [ be such that haœ!, implying YAœ!. Since Y is invertible, hence injective by Theorem 3.69, we conclude that Aœ!, showing that h is injective. To show the converse, let U be a direct sum complement of [ and \ a direct sum complement of eh, where the existence of U and \ are guaranteed by Theorem Let ;ßâß; " 5 be a basis for Uand BßâB " 7 a basis for \. By the fundamental theorem of linear maps [Theorem 3.22], dim eh Ÿ dim [, implying, by Theorem 2.43, 5Ÿ7. Recall that Z can be uniquely decomposed as A;, where A [ and ; U. Define Y _Z as 5 5 Y@œh AB, where Bœ-B \ and ;œ-; œ" 4œ" Since \ is a direct sum complement of eh, Y@œ! implies haœbœ!. Since h is injective and BßâB " 5 is linearly independent, Yis injective, hence invertible (again by Theorem 3.69). In the sequel we assume that [ and Z are real inner product spaces. For g _Zß[, let g _ [ßZ denote the adjoint of g [Definition 7.2]. For any subspace U of Z, let ¼ U denote the orthogonal complement of U [Definition 6.45] and C U _Z denote the orthogonal projection operator onto U [Definition 6.53]. Theorem 1. Given X _ Z, there exists Y _ Z, invertible and depending on X, such that œyx. C e X

5 Proof of Theorem 1. We first observe that X X is a positive operator [Definition 7.31]. By Theorems 7.6(e) and 7.6(c), X X is self-adjoint; the observation follows, since, by the definition of the adjoint operator, for Z, X X@ß@ œ X@!. 1 By the Real Spectral Theorem [Theorem 7.29(b)], Z has an orthonormal basis 0ßâß0 " consisting of eigenvectors of X X with corresponding eigenvalues -" ßâß -. By Theorem 7.35(b), -4! for all "Ÿ4Ÿ. Since, by 1, -4 œx04, we obtain - - œ!íx0 œ!!íx0 Á! If œ! for every 4œ"ßâß, then X Xis the zero operator, implying, by 1, that X is the zero operator. By Theorem 7.7, X is the zero operator as well, and the theorem trivially holds with Y œm. Hence, à œ4à"ÿ4ÿß-4! is non-empty without - loss of generality; let à œ4à"ÿ4ÿß- œ!. - 4 X For 4 Ã, since X œ0, we have 0 ex, equivalently, - 4 C ex œ 0. 3 For 4 Ã, 0 ax by 2 ; since ax œex [Theorem 7.7(c)], By 3 and 4, for any B Z, C e X C ex 4 ¼ 0 œ!. 4 B œ ØBß 0 Ù Ã By 2 and the definition of à -, for any B Z, 4 4 XB œ ØBß0 ÙX à 4 4 By definition of 04 and -4, "Ÿ4Ÿ, the list of vectors X04X04 À4 à in Z is orthonormal, and consequently, by Theorem 6.26, linearly independent; the same conclusion holds for the list 04 À 4 Ã. For [ œ span X04 À 4 Ã, h _ [ßZ defined by hx04 œ 04 is clearly injective. By Lemma 1 there exists an invertible operator Y _Z such that Now note that, for any YA œ h Afor every A [. 7 B Z, by 6, 7, the definition of h, and 5, in that order,

6 Y X B œ ØBß 0 ÙY X 0 œ ØBß 0 ÙhX 0 œ ØBß 0 Ù0 œ C B, e X 4 à 4 à 4 à completing the proof. Lemma 2. Given H _ Z positive, there exists H _ Zpositive such that H H œ H H œ C eh and where H denotes the unique positive square root of H. Proof of Lemma 2. By Theorem 7.36, HH œh œh H, 9 H exists and is defined by H 0 œ - 0, 10 where 0ß0ßâß0 " is an orthonormal basis of Z consisting of eigenvectors of Hwith - corresponding eigenvalues -" ß- ßâß-, and à and à are as in the proof of Theorem 1. Let H _Z be defined by H if 4 à 0 œ 11 -! if 4 Ã. Clearly, H BßC œ - ØBß0 ÙØCß0 Ùœ BßH C, showing that H is self- 4 à adjoint and implying H BßB!, that is, H _Z is positive. For any C Z, by 10 and 11, H H C œ ØCß 0 Ù0 œ H H C à Since H is self-adjoint, by Theorem 7.7(d), eh œ ah ¼. 13 Since 04 À 4 à is contained in ah and 04 À 4 à in eh, it follows that 04 À 4 à is a basis of eh, and follows from 12. Note that 9 follows from the observation that both HH C and H HC equal ØCß 0 Ù- 0 œ H C. 4 à The multivariate Normal distribution results. Let ^ßâß^ " be iid standard Normal random variables. The distribution of

7 ^œ /^ is defined to be the standard multivariate Normal distribution Á is the identity operator. 5œ" !ß M, where M _ d Lemma 3. The characteristic function G!ßM of the Á!ß M distribution is given by G!ßM > œ I exp3ø>ß^ù œ exp > Î. Proof of Lemma 3. Follows from Proposition 9.4.2(a) of Dudley (199). Lemma 4. Given ^µ Á!ßM,. d, and X _ d, let Then, for any =ß > d, ]œ. X^. 15 " I 3Ø>ß ] Ù œ 3Ø>ß Ù exp exp. >ß X X > IØ>ß ] Ù œ Ø>ß. Ù Ø>ß ] Ùß Ø=ß ] Ù œ Cov >ß X X =. Proof of Lemma 4. Straightforward algebra using the bilinearity of the inner product and the definitions of X, ^ in 14, and ] in 15, along with the fact that ^" ßâß^ are iid standard Normal random variables, proves the lemma. Definition 1. The distribution of ] in 15 is defined to be the multivariate Normal distribution with mean. and covariance XX. Recall that a distribution on d is uniquely determined by its characteristic function [Dudley (199, Theorem 9.5.1)]; since the characteristic function of ] is determined by its mean. and covariance XX, the multivariate Normal distribution Á. ßH with mean. and covariance H (where H is any positive operator on d ) is uniquely defined in terms of the characteristic function " G.ßH> œ exp 3Ø>ß. Ù Ø>ß H>Ù. 16 Lemma 5. Given ] µ Á. ßH and f _ d ßd, 7 f]µ Á f. ßfHf. 7 Proof of Lemma 5. The proof is a straightforward consequence of Definition 1. Given =ß> ß =ß> d d œd, the inner product on d d is given by " " 7 : 7: 7 : =ß> ß =ß> œø=ß=ùø>ß>ù " " " ". Theorem 2. Given ] µ Á. ßH, f _ d ßd, and g _ d ßd, the random 7 :

8 7 : vector f]ß g ] µ Á7: with mean œf. ßg. d d and covariance operator 7 : ^ _ d d given by ^=ß> œ fh f =fh g >ßg H f =g H g >. 7 : 7 : Proof of Theorem 2. We first verify that ^ Àd d Èd d is a positive operator. 7 : The verification of ^ _ d d being linear and self-adjoint is routine. Since fh f =ß= fh g >ß= g H f =ß> g H g >ß> œ H f = H g >ßH f = H g >, the non-negativity of 7 : ^ =ß > ß =ß > for every =ß > d d follows. By the 7 : definition of the inner product in d d and 16, the characteristic function G of the 7 : random vector f]ß g ] taking values in d d is given by G " f g. =ß> œ exp 3 = >ß f = g >ßH f = g >. 17 Since f = g >ß. œ =ß> ßf. ßg. and f = g >ßH f = g > f f f g œ =ß H = H > >ßg H f =g H g > œ =ß> ßfH f =fh g >ßg H f =g H g >, 1 the proof follows. 7 : Corollary 1. Given ] µ Á. ßH, f _ d ßd, and g _ d ßd, f] and g ] : 7 7 : are independent if and only if fhg _ d ßd, equivalently g Hf _ d ßd, is the zero operator. Proof of Corollary 1. From 17, " G f. g. =ß > œ exp 3 =ß exp 3 >ß exp f = g >ß H f = g >. Now by 1, f = g >ßH f = g > œ =ßfH f = =ßfH g > >ßg H g >. Thus, by 16 and Lemma 5, G f g =ß > œ I exp 3 =ß ] I exp 3 >ß ] exp =ß fh g >. That is, the characteristic function of f]ß g ] is the product of two factors; one factor is the characteristic function of the product measure of the distributions induced by f] 7 : on U and g on U, whereas the other factor is d ] d exp =ßfH g >. Therefore, f] and g ] are independent if and only if exp =ßfH g > œ " for every

9 7 : = d and > d, equivalently, =ßfH g > œ! for every = and >. Since g Hf œ fhg by Theorems 7.6(e) and 7.6(c), the proof follows. Remark 1. For \ßâß\ iid Normal random variables with mean ) and variance 5, " " " \œ \ 3 and W œ \ 3 \ " 3œ" 3œ" are independent [ Casella and Berger (2002, Theorem 5.3.1(a) )]. Most textbooks prove this result by working with the (joint) density of the sample and using the Jacobian formula for finding the density of the transformation that maps the sample to the sample mean and the sample variance. Some textbooks use Basu's (1955) Theorem on an ancillary statistic (the sample variance) being independent of a complete sufficient statistic (the sample mean) to prove this result. We are going to show that this result is a straightforward consequence of Corollary 1. Let N be the sum of the standard orthonormal basis vectors, and Nthe span of N. we have BœN ØBßNÙN, C N " \œn C \ßNand W œn " M C \, N N Since where \œ \/ µ Á ) Nß5 M. By Corollary 1, using the fact that an orthogonal 5œ" 5 5 projection operator is self-adjoint, we obtain CN] and M CN] are independent if and only if MC NH CNis the zero operator, where ] µ Á. ßH. Clearly, MCNHCNBœ! for all B d if and only if MCNHNœ!, equivalently, N is an eigenvector of H. Since N is an eigenvector of 5 M, the independence of \ and W follows from that of C \ and M C \. N N The class of positive operators with N as an eigenvector does not reduce to the singleton set M. For œ, define H _ d by H? " ß? œ? "? ß? "? ; clearly, Nœ "ß" is an eigenvector for H. Thus, for the sample mean and the sample variance to be independent, it is not necessary for the sample to be iid. As long as the joint distribution of the sample is multivariate Normal such that N is an eigenvector of the covariance, the independence of the sample mean and the sample variance holds. However, for Normal random variables that are dependent, whether the joint distribution is multivariate Normal becomes a modeling question, as the joint distribution of even pairwise uncorrelated Normal random variables may not be multivariate Normal. // Theorem 3. Given ]µ Á. ßH and X _ d, define WœXH ; 19

10 then H C a H ] is independent of X] 20 W and ]H C a H ] is an affine function of X]. 21 W Proof of Theorem 3. For any B d and J _d, we obtain from 9, aj aj aj, XH H H C B œ XHH C H B œ XH C H B implying, by 19, that XH H C a W H is the zero operator, whence 20 follows from Corollary 1. To prove 21 we first observe that, by 13 and in that order, eh ah ah ]œ C ] C ]œh H ] C ]Þ 22 Now we are going to show that C ] œ C., 23 ah ah implying, by 22, that ]œh H ] C H. a Let 04, -4, Ã, and à be as in the proof of Lemma 2. Recall that 04 À 4 à is an orthonormal basis for ah and 04 À 4 à is an orthonormal basis for eh. For 4 à - and? d, by 16, " Iexp3? ] ß 04 œ exp3?04ß.?04ß H?04 œ exp3?. ß 04, implying ]ß0 - œ. ß0for all 4 Ã, that is, By Theorems 6.47 and 7.7(c), MœC C, 25 ew aw implying, by 24, aw ew ah ]H C H ] œh C H ]C.. 26 By Theorem 1, there exists Y _d, invertible, such that C ew œyw; 27 applying 27, 19, and 24 in that order, we obtain that RHS26 equals

11 H YWH ]C. œh YX] MH YX C., ah ah thereby establishing Theorem 4. Given ]µ Á. ßH and g _ d, d, the conditional distribution of ] given g ] is multivariate Normal on Ud. Proof of Theorem 4. The idea is to first construct X _ d such that 5g ] œ5x] and then use Theorem 3 to find the conditional distribution of ] given X]. Let Ç _ d, eg be defined by ÇB œ g B for every B d. Clearly, Ç is surjective, hence injective and invertible [Theorem 3.69]. By Theorem 7.7, Ç _ g d e, is invertible as well. Let X _ d be defined by XBœÇ ÇB; note that X, being the composition of two invertible transformations, is invertible. Let HY ß ßT denote the probability space underlying ]. Since g is continuous, hence measurable, g ]ÀHßY Èd 7 ßUd 7 is measurable, implying g ] " I Y for every I Ud 7. Note that, g ] " I œ Ç] " I eg. Further, eg, by 7 7 virtue of being closed, is an element of Ud, and Ueg is the trace of Ud on 7 eg ; that is, Ueg œ I eg À I Ud, implying 5g ] œ 5Ç ]. Since Ç and X are injective (and measurable), by Lemma A.1, 5Ç œ Ud œ 5X; that is, Ç " K À K Ue g œ Ud œ X " J À J Ud. Since, for any K Ue g and J Ud, we obtain Ç] " " " K œ ] Ç K X] " " J œ ] " and X J, 5Ç " Ç " ] œ ] K À K Ue g œ X] J À J Ud œ 5X]. That completes the proof that 5g ] œ 5X]. By the pull-out property of conditional expectation, 20, and 21, Iexp3 >ß] X] œ exp3 >ß] H H ] Iexp3>ßH C a C a H ], where W is as in 19. By Lemma 5 and 16, W Iexp œ exp " 3>ßH Ca W H ] 3>ßH C a W H. >ßK>, where the operator KœH CaWH HH CaW H equals, by 9 and, H CaWCeHCaW H. Therefore, the conditional distribution of ] given X], hence that given g ], is Normal with mean ] œ ]H C a WH ]. and covariance K. W

12 Remark 2. The expressions for the mean and the covariance K of the conditional distribution of ] given g ] can be considerably simplified, rendering the verification of the iterated expectation formula and the analysis of variance formula immediate. Since ]. œ C eh]. by 23, applying 25 and in that order, ] œ. H C H ].. 2 ew To simplify the expression of the covariance operator K (which does not depend on ]), first note that, by and 9, By 27, 19, and 29, in that order, Consequently, K equals H H C e œ H H H œ HH œ H. 29 ew eh eh ew C C œyxh C œyxh œ C. 30 H CeHCaWH H CeW CeHCaWH by 25 œ H CaWH H CeW CeHCaWH " Î œ H CaWH H CeW CaWH by 29 by 30 œh C a W H by Recall that if Z and [ are random vectors in d such that Z is an affine function of [, i.e., Z œ +V[, where + d and V _d, then IZ œ+vi[ and CovZ œv Cov [V, implying II ] X] œ., verifying the iterated expectation formula. Defining the expected value of an operator-valued random element E as the operator IE such that I E= œ I E = for every = d, provided the expected value on the right hand side is well defined for every =, we conclude from 31 that By 2 and Lemma 5, I ] X] Cov œ H H. C a W Cov I ] X] œh C H HH H œh H C C, ew ew ew where the second equality follows by 9,, and 30. The analysis of variance formula, I Cov] X] CovI ] X] œ Cov ], is verified by 25. // An immediate corollary of Theorem 4 is the "partial out" formula for population regression in the Normal model.

13 Corollary 2. If \ß] ß^ is multivariate Normal such that ^ œ ^", âß^ d and \ß] d, then Cov\ß] l^ I ]l\ß^i ]l^ œ Var \l^! \ I \l^. 32 Var\l^ " Proof of Corollary 2. Let c "ß$ _ d ßd and c $ _ d ßd be defined by c AœBßD and c AœD, where AœBßCßD. By 2, "ß$ $ e W W LHS32 œ H C C e H [. ß/ 33 "ß$ $, where [œ \ß]ß^,. d is the mean of [ and H _ d is the covariance of [, W"ß$ œt"ß$ H with T"ß$ _d defined by T"ß$ AœBß!ßD, and W$ œt$ H with T _d defined by T A œ!ß!ß D. $ $ To show RHS32 œ RHS33, we first observe, using 31 and 2, Cov \ß] l^ œ H C H / ß/ a W " $ Var\l^ œ H CaW H /" ß/ " œ CaW H /" 34 $ $ \I \l^ œmh C ew H [. ß/ $ ". Since T "ß$/ 4 œ/ 4 if 4Á, T "ß$/ œ!, T/ $ 4 œ/ 4 if 4 $, and T/ $ 4 œ! if 4œ"ß, a W"ß$ œ spanh /" ßH / $ ßâßH / a W œ ¼ span H / ßâßH /. ¼ $ $ If H /" span H / $ ßâßH / œ ew $, then RHS33 œ!, in which case Var \l^, a constant function of ^, is also equal to!, thereby vacuously satisfying 32. If H /  span H / ßâßH /, by Lemma A.2, " $ C C H [. ew ew "ß$ $ œ H / C C H [. ßC H / C H / aw " aw aw " aw " Therefore, RHS 33 equals C $ $ $ $. H [. ßCa W H / " H Ca W H / $ " ß/, C H / " aw $ $ a W $ which, by the first two rows in 34, equals, since C a W$ is self-adjoint and idempotent,

14 Cov\ß] l^ H C a H [. ß/ $ Var\l^ W " Since H C a W$ H œ MH CeW H C $ ah by 25,, and 13, the proof follows by the third row in 34 and 23.. Remark 3. Given a random sample \" ßâß\ from the Normal distribution with mean ) and (known) variance 5, \ is a sufficient statistic for ) [Example 6.2.4, Casella and Berger (2002)]. While the typical proof uses the powerful factorization theorem [Theorem 6.2.6, Casella and Berger (2002)], we are going to show that the conditional distribution of the sample \œ \/ 5 5 µ Á ) Nß5 M given \, that is, " C N \, 5œ" does not depend on ), thereby proving the sufficiency of \ directly. The said conditional distribution, by Theorem 2, 2, and 31, is multivariate Normal with mean ) NH C ew H \) N and covariance H CaWH, where Hœ5 M and " " Wœ CNH, implying W œ5 CNand consequently, Ce W œcn, further implying that the mean equals C \ and the covariance equals 5 M C. // N Remark 4. In a companion paper we are currently working on, we use the local linearity " 7 of a G vector field 1 ( on d that takes values in d ) and the decomposition of ] in Theorem 3 to approximate the conditional distribution of ] given 1 ]. If 1 is constant, then the 5-algebra generated by 1] is the trivial 5-algebra consisting of the empty set and the entire sample space, making ] independent of 1 ], so that the conditional distribution of ] given 1] is the unconditional distribution of ]. If 1is injective, 51 equals Ud by Lemma A.1; consequently, the conditional distribution of ] given 1 ] is the point mass at ]. Thus, the interesting problem unfolds when 1 is a non-constant, non-injective, G " vector field. The conditional distribution of ] given 1 ] will not be multivariate Normal if 1 is not linear. Let Àd È_ d ßd 7 be defined to be the continuous function that maps + d to the total derivative of 1 at +. Given %!, there exists a compact subset O% of d such that T ] O% "%. The function, restricted to O%, is uniformly continuous, and admits a uniformly continuous modulus of continuity ' %. We are working to show that, for a suitably chosen metric for weak convergence of probability measures and $! (that depends on the given % through the supremum of the modulus of continuity ' % on the closed interval!ß F %, where F % is the diameter of O % ), the conditional distribution of ] given 1 ] is within $ -neighborhood of the family of multivariate Normal distributions. // Remark 5. Proposition 3.13 of Eaton (193) holds at a much greater level of generality; see Bogachev (199, Theorem ). Extending our approach, in the absence of an inner product, to finding the conditional distribution of a Gaussian random element given a (non-injective, non-trivial) linear transformation appears to be an interesting area of future research. In particular, if ] is the Brownian motion, i.e., the random element in N

15 V!ß " distributed according to the Wiener measure, /" ß âß / 5 are measures on 5 U!ß ", and g À V!ß " È d is given by g 0 œ 0. /" ß âß 0. / 5, what is the conditional distribution of ] given g ]? // Appendix. We present a lemma on measurability [Lemma A.1] that was used in the proof of Theorem 4 and another lemma on orthogonal projection operators in a Hilbert space [Lemma A.2] that was used in the proof of Corollary 2. Lemma A.1 Let Ë and Ì be Polish spaces. If 2ÀË ÈÌ is Borel measurable and injective, then 52 œ UË. Proof of Lemma A.1. The inclusion 52 UË follows from the pertinent definitions. For Q UË, define Fœ2Q. Since 2 " F œqand F UÌ by Theorem I.3.9 of Parthasarathy (1967), the reverse inclusion follows. Lemma A.2 For a proper and closed subspace Z of a Hilbert space L and B LZ, let Z B denote the closure of the Zß- d. Then, for any C L, C Z B CC CœC ¼B C ¼CßC ¼BC ¼BÞ 35 Z Z Z Z Z Proof of Lemma A.2. By the definition of Z B there exists a À " Z and a sequence - À " d such that C Z Cœ B Ä - B. 36 Since Z Z B, CZCœCZCZ B C; since CZ is continuous, by 36, C Cœ - C B. 37 Z Z Ä Since BC Z BœC Z ¼B, subtracting 37 from 36, LHS35 œ lim - C ¼ B. 3 Ä Since LHS35 œ CZ ¼CC Z C, taking the inner product of both sides of 3 with B ¼ C Z ¼B and using the linearity and homogeneity of inner product, we obtain C ¼Cß C ¼B C Cß C ¼B œ lim - C ¼B. 39 Z Z ZB ¼ Z Z Ä Since B ZB and C Z B Z Z B, CZ ¼B ZB, implying C Z CßCZ ¼B œ! ; since B ¼ BÂZ, that is, C Z ¼B Á!, the lemma follows from 39 and 3. Z

16 References nd Anderson, T. W. (194). An Introduction to Multivariate Statistical Analysis. 2 Ed., John Wiley & Sons, New York. Axler, Sheldon (2015). rd Linear Algebra Done Right. 3 Ed., Springer, New York. Basu, D. (1955). On Statistics Independent of a Complete Sufficient Statistic. Sankhya, Vol 15, Bogachev, Vladimir I. (199). Gaussian Measures. 1 Ed., American Mathematical Society, Providence, Rhode Island. nd Casella, George and Roger L. Berger (2002). Statistical Inference. 2 Ed., Brooks/Cole, California. st Dudley, R. M. (199). Brooks/Cole, California. st Real Analysis and Probability. 1 Ed., Wadsworth and Eaton, Morris L. (193). Multivariate Statistics: A Vector Space Approach. 1 Ed., John Wiley & Sons, New York. Flury, Bernard (1997). York. st A First Course in Multivariate Statistics. 1 Ed., Springer, New st Muirhead, Robb J. (192). Aspects of Multivariate Statistical Theory. 1 Ed., John Wiley & Sons, New York. Parthasarathy, K. R. (1967). New York, NY. Probability Measures on Metric Spaces, Academic Press, st RAJESHWARI MAJUMDAR SUMAN MAJUMDAR rajeshwari.majumdar@uconn.edu suman.majumdar@uconn.edu PO Box 47 1 University Place Coventry, CT 0623 Stamford, CT 06901

On Least Squares Linear Regression Without Second Moment

On Least Squares Linear Regression Without Second Moment On Least Squares Linear Regression Without Second Moment BY RAJESHWARI MAJUMDAR University of Connecticut If \ and ] are real valued random variables such that the first moments of \, ], and \] exist and

More information

NECESSARY AND SUFFICIENT CONDITION FOR ASYMPTOTIC NORMALITY OF STANDARDIZED SAMPLE MEANS

NECESSARY AND SUFFICIENT CONDITION FOR ASYMPTOTIC NORMALITY OF STANDARDIZED SAMPLE MEANS NECESSARY AND SUFFICIENT CONDITION FOR ASYMPTOTIC NORMALITY OF STANDARDIZED SAMPLE MEANS BY RAJESHWARI MAJUMDAR* AND SUMAN MAJUMDAR University of Connecticut and University of Connecticut The double sequence

More information

MATH 8253 ALGEBRAIC GEOMETRY WEEK 12

MATH 8253 ALGEBRAIC GEOMETRY WEEK 12 MATH 8253 ALGEBRAIC GEOMETRY WEEK 2 CİHAN BAHRAN 3.2.. Let Y be a Noetherian scheme. Show that any Y -scheme X of finite type is Noetherian. Moreover, if Y is of finite dimension, then so is X. Write f

More information

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016 Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016 1. Let V be a vector space. A linear transformation P : V V is called a projection if it is idempotent. That

More information

Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution

Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution Relevant

More information

9 Brownian Motion: Construction

9 Brownian Motion: Construction 9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of

More information

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,

More information

1 Math 241A-B Homework Problem List for F2015 and W2016

1 Math 241A-B Homework Problem List for F2015 and W2016 1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let

More information

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory Part V 7 Introduction: What are measures and why measurable sets Lebesgue Integration Theory Definition 7. (Preliminary). A measure on a set is a function :2 [ ] such that. () = 2. If { } = is a finite

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

The best generalised inverse of the linear operator in normed linear space

The best generalised inverse of the linear operator in normed linear space Linear Algebra and its Applications 420 (2007) 9 19 www.elsevier.com/locate/laa The best generalised inverse of the linear operator in normed linear space Ping Liu, Yu-wen Wang School of Mathematics and

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

Multivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013

Multivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013 Multivariate Gaussian Distribution Auxiliary notes for Time Series Analysis SF2943 Spring 203 Timo Koski Department of Mathematics KTH Royal Institute of Technology, Stockholm 2 Chapter Gaussian Vectors.

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Review and problem list for Applied Math I

Review and problem list for Applied Math I Review and problem list for Applied Math I (This is a first version of a serious review sheet; it may contain errors and it certainly omits a number of topic which were covered in the course. Let me know

More information

Normed & Inner Product Vector Spaces

Normed & Inner Product Vector Spaces Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed

More information

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations Linear Algebra in Computer Vision CSED441:Introduction to Computer Vision (2017F Lecture2: Basic Linear Algebra & Probability Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Mathematics in vector space Linear

More information

Stat 643 Review of Probability Results (Cressie)

Stat 643 Review of Probability Results (Cressie) Stat 643 Review of Probability Results (Cressie) Probability Space: ( HTT,, ) H is the set of outcomes T is a 5-algebra; subsets of H T is a probability measure mapping from T onto [0,] Measurable Space:

More information

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where: VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector

More information

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna

More information

Spectral theory for compact operators on Banach spaces

Spectral theory for compact operators on Banach spaces 68 Chapter 9 Spectral theory for compact operators on Banach spaces Recall that a subset S of a metric space X is precompact if its closure is compact, or equivalently every sequence contains a Cauchy

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2 Contents Preface for the Instructor xi Preface for the Student xv Acknowledgments xvii 1 Vector Spaces 1 1.A R n and C n 2 Complex Numbers 2 Lists 5 F n 6 Digression on Fields 10 Exercises 1.A 11 1.B Definition

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)

More information

Boolean Inner-Product Spaces and Boolean Matrices

Boolean Inner-Product Spaces and Boolean Matrices Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver

More information

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t ) LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data

More information

CAAM 336 DIFFERENTIAL EQUATIONS IN SCI AND ENG Examination 1

CAAM 336 DIFFERENTIAL EQUATIONS IN SCI AND ENG Examination 1 CAAM 6 DIFFERENTIAL EQUATIONS IN SCI AND ENG Examination Instructions: Time limit: uninterrupted hours There are four questions worth a total of 5 points Please do not look at the questions until you begin

More information

Math 413/513 Chapter 6 (from Friedberg, Insel, & Spence)

Math 413/513 Chapter 6 (from Friedberg, Insel, & Spence) Math 413/513 Chapter 6 (from Friedberg, Insel, & Spence) David Glickenstein December 7, 2015 1 Inner product spaces In this chapter, we will only consider the elds R and C. De nition 1 Let V be a vector

More information

ECE 275A Homework #3 Solutions

ECE 275A Homework #3 Solutions ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =

More information

Gaussian random variables inr n

Gaussian random variables inr n Gaussian vectors Lecture 5 Gaussian random variables inr n One-dimensional case One-dimensional Gaussian density with mean and standard deviation (called N, ): fx x exp. Proposition If X N,, then ax b

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

Karhunen-Loève decomposition of Gaussian measures on Banach spaces

Karhunen-Loève decomposition of Gaussian measures on Banach spaces Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix jean-charles.croix@emse.fr Génie Mathématique et Industriel (GMI) First workshop on Gaussian processes at Saint-Etienne

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND

More information

The Dirichlet-to-Neumann operator

The Dirichlet-to-Neumann operator Lecture 8 The Dirichlet-to-Neumann operator The Dirichlet-to-Neumann operator plays an important role in the theory of inverse problems. In fact, from measurements of electrical currents at the surface

More information

DS-GA 1002 Lecture notes 10 November 23, Linear models

DS-GA 1002 Lecture notes 10 November 23, Linear models DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.

More information

COMMON COMPLEMENTS OF TWO SUBSPACES OF A HILBERT SPACE

COMMON COMPLEMENTS OF TWO SUBSPACES OF A HILBERT SPACE COMMON COMPLEMENTS OF TWO SUBSPACES OF A HILBERT SPACE MICHAEL LAUZON AND SERGEI TREIL Abstract. In this paper we find a necessary and sufficient condition for two closed subspaces, X and Y, of a Hilbert

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

REAL ANALYSIS II HOMEWORK 3. Conway, Page 49

REAL ANALYSIS II HOMEWORK 3. Conway, Page 49 REAL ANALYSIS II HOMEWORK 3 CİHAN BAHRAN Conway, Page 49 3. Let K and k be as in Proposition 4.7 and suppose that k(x, y) k(y, x). Show that K is self-adjoint and if {µ n } are the eigenvalues of K, each

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information

1 Invariant subspaces

1 Invariant subspaces MATH 2040 Linear Algebra II Lecture Notes by Martin Li Lecture 8 Eigenvalues, eigenvectors and invariant subspaces 1 In previous lectures we have studied linear maps T : V W from a vector space V to another

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments

Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments Austrian Journal of Statistics April 27, Volume 46, 67 78. AJS http://www.ajs.or.at/ doi:.773/ajs.v46i3-4.672 Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments Yuliya

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Five Mini-Courses on Analysis

Five Mini-Courses on Analysis Christopher Heil Five Mini-Courses on Analysis Metrics, Norms, Inner Products, and Topology Lebesgue Measure and Integral Operator Theory and Functional Analysis Borel and Radon Measures Topological Vector

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

2 Garrett: `A Good Spectral Theorem' 1. von Neumann algebras, density theorem The commutant of a subring S of a ring R is S 0 = fr 2 R : rs = sr; 8s 2

2 Garrett: `A Good Spectral Theorem' 1. von Neumann algebras, density theorem The commutant of a subring S of a ring R is S 0 = fr 2 R : rs = sr; 8s 2 1 A Good Spectral Theorem c1996, Paul Garrett, garrett@math.umn.edu version February 12, 1996 1 Measurable Hilbert bundles Measurable Banach bundles Direct integrals of Hilbert spaces Trivializing Hilbert

More information

Linear Algebra 2 Spectral Notes

Linear Algebra 2 Spectral Notes Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex

More information

Empirical Processes: General Weak Convergence Theory

Empirical Processes: General Weak Convergence Theory Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

Math Solutions to homework 5

Math Solutions to homework 5 Math 75 - Solutions to homework 5 Cédric De Groote November 9, 207 Problem (7. in the book): Let {e n } be a complete orthonormal sequence in a Hilbert space H and let λ n C for n N. Show that there is

More information

VARIETIES OF ABELIAN TOPOLOGICAL GROUPS AND SCATTERED SPACES

VARIETIES OF ABELIAN TOPOLOGICAL GROUPS AND SCATTERED SPACES Bull. Austral. Math. Soc. 78 (2008), 487 495 doi:10.1017/s0004972708000877 VARIETIES OF ABELIAN TOPOLOGICAL GROUPS AND SCATTERED SPACES CAROLYN E. MCPHAIL and SIDNEY A. MORRIS (Received 3 March 2008) Abstract

More information

Linear algebra 2. Yoav Zemel. March 1, 2012

Linear algebra 2. Yoav Zemel. March 1, 2012 Linear algebra 2 Yoav Zemel March 1, 2012 These notes were written by Yoav Zemel. The lecturer, Shmuel Berger, should not be held responsible for any mistake. Any comments are welcome at zamsh7@gmail.com.

More information

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms (February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26 Estimation: Regression and Least Squares This note explains how to use observations to estimate unobserved random variables.

More information

Measurable functions are approximately nice, even if look terrible.

Measurable functions are approximately nice, even if look terrible. Tel Aviv University, 2015 Functions of real variables 74 7 Approximation 7a A terrible integrable function........... 74 7b Approximation of sets................ 76 7c Approximation of functions............

More information

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October Finding normalized and modularity cuts by spectral clustering Marianna Bolla Institute of Mathematics Budapest University of Technology and Economics marib@math.bme.hu Ljubjana 2010, October Outline Find

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

SPECTRAL THEOREM FOR SYMMETRIC OPERATORS WITH COMPACT RESOLVENT

SPECTRAL THEOREM FOR SYMMETRIC OPERATORS WITH COMPACT RESOLVENT SPECTRAL THEOREM FOR SYMMETRIC OPERATORS WITH COMPACT RESOLVENT Abstract. These are the letcure notes prepared for the workshop on Functional Analysis and Operator Algebras to be held at NIT-Karnataka,

More information

Systems of Linear Equations

Systems of Linear Equations Systems of Linear Equations Math 108A: August 21, 2008 John Douglas Moore Our goal in these notes is to explain a few facts regarding linear systems of equations not included in the first few chapters

More information

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality. 88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal

More information

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work. Assignment 1 Math 5341 Linear Algebra Review Give complete answers to each of the following questions Show all of your work Note: You might struggle with some of these questions, either because it has

More information

Two remarks on normality preserving Borel automorphisms of R n

Two remarks on normality preserving Borel automorphisms of R n Proc. Indian Acad. Sci. (Math. Sci.) Vol. 3, No., February 3, pp. 75 84. c Indian Academy of Sciences Two remarks on normality preserving Borel automorphisms of R n K R PARTHASARATHY Theoretical Statistics

More information

Spectral Theory, with an Introduction to Operator Means. William L. Green

Spectral Theory, with an Introduction to Operator Means. William L. Green Spectral Theory, with an Introduction to Operator Means William L. Green January 30, 2008 Contents Introduction............................... 1 Hilbert Space.............................. 4 Linear Maps

More information

The Essential Equivalence of Pairwise and Mutual Conditional Independence

The Essential Equivalence of Pairwise and Mutual Conditional Independence The Essential Equivalence of Pairwise and Mutual Conditional Independence Peter J. Hammond and Yeneng Sun Probability Theory and Related Fields, forthcoming Abstract For a large collection of random variables,

More information

LINEAR ALGEBRA SUMMARY SHEET.

LINEAR ALGEBRA SUMMARY SHEET. LINEAR ALGEBRA SUMMARY SHEET RADON ROSBOROUGH https://intuitiveexplanationscom/linear-algebra-summary-sheet/ This document is a concise collection of many of the important theorems of linear algebra, organized

More information

CLOSURE OF INVERTIBLE OPERATORS ON A HILBERT SPACE

CLOSURE OF INVERTIBLE OPERATORS ON A HILBERT SPACE proceedings of the american mathematical society Volume 108, Number 3, March 1990 CLOSURE OF INVERTIBLE OPERATORS ON A HILBERT SPACE RICHARD BOULDIN (Communicated by Palle E. T. Jorgensen) Abstract. Although

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

MET Workshop: Exercises

MET Workshop: Exercises MET Workshop: Exercises Alex Blumenthal and Anthony Quas May 7, 206 Notation. R d is endowed with the standard inner product (, ) and Euclidean norm. M d d (R) denotes the space of n n real matrices. When

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Linear Algebra Review

Linear Algebra Review Linear Algebra Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Linear Algebra Review 1 / 45 Definition of Matrix Rectangular array of elements arranged in rows and

More information

Karhunen-Loève decomposition of Gaussian measures on Banach spaces

Karhunen-Loève decomposition of Gaussian measures on Banach spaces Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix GT APSSE - April 2017, the 13th joint work with Xavier Bay. 1 / 29 Sommaire 1 Preliminaries on Gaussian processes 2

More information

CHAPTER II HILBERT SPACES

CHAPTER II HILBERT SPACES CHAPTER II HILBERT SPACES 2.1 Geometry of Hilbert Spaces Definition 2.1.1. Let X be a complex linear space. An inner product on X is a function, : X X C which satisfies the following axioms : 1. y, x =

More information

First we introduce the sets that are going to serve as the generalizations of the scalars.

First we introduce the sets that are going to serve as the generalizations of the scalars. Contents 1 Fields...................................... 2 2 Vector spaces.................................. 4 3 Matrices..................................... 7 4 Linear systems and matrices..........................

More information

REPRESENTATION THEORY OF S n

REPRESENTATION THEORY OF S n REPRESENTATION THEORY OF S n EVAN JENKINS Abstract. These are notes from three lectures given in MATH 26700, Introduction to Representation Theory of Finite Groups, at the University of Chicago in November

More information

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra D. R. Wilkins Contents 3 Topics in Commutative Algebra 2 3.1 Rings and Fields......................... 2 3.2 Ideals...............................

More information

µ X (A) = P ( X 1 (A) )

µ X (A) = P ( X 1 (A) ) 1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

Ð"Ñ + Ð"Ñ, Ð"Ñ +, +, + +, +,,

ÐÑ + ÐÑ, ÐÑ +, +, + +, +,, Handout #11 Confounding: Complete factorial experiments in incomplete blocks Blocking is one of the important principles in experimental design. In this handout we address the issue of designing complete

More information

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability... Functional Analysis Franck Sueur 2018-2019 Contents 1 Metric spaces 1 1.1 Definitions........................................ 1 1.2 Completeness...................................... 3 1.3 Compactness......................................

More information

Gaussian Hilbert spaces

Gaussian Hilbert spaces Gaussian Hilbert spaces Jordan Bell jordan.bell@gmail.com Department of Mathematics, University of Toronto July 11, 015 1 Gaussian measures Let γ be a Borel probability measure on. For a, if γ = δ a then

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

Chapter 2 Linear Transformations

Chapter 2 Linear Transformations Chapter 2 Linear Transformations Linear Transformations Loosely speaking, a linear transformation is a function from one vector space to another that preserves the vector space operations. Let us be more

More information

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 59 CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 4. INTRODUCTION Weighted average-based fusion algorithms are one of the widely used fusion methods for multi-sensor data integration. These methods

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Analysis Preliminary Exam Workshop: Hilbert Spaces

Analysis Preliminary Exam Workshop: Hilbert Spaces Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H

More information

von Neumann algebras, II 1 factors, and their subfactors V.S. Sunder (IMSc, Chennai)

von Neumann algebras, II 1 factors, and their subfactors V.S. Sunder (IMSc, Chennai) von Neumann algebras, II 1 factors, and their subfactors V.S. Sunder (IMSc, Chennai) Lecture 3 at IIT Mumbai, April 24th, 2007 Finite-dimensional C -algebras: Recall: Definition: A linear functional tr

More information

The Multivariate Gaussian Distribution [DRAFT]

The Multivariate Gaussian Distribution [DRAFT] The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,

More information

LECTURE 25-26: CARTAN S THEOREM OF MAXIMAL TORI. 1. Maximal Tori

LECTURE 25-26: CARTAN S THEOREM OF MAXIMAL TORI. 1. Maximal Tori LECTURE 25-26: CARTAN S THEOREM OF MAXIMAL TORI 1. Maximal Tori By a torus we mean a compact connected abelian Lie group, so a torus is a Lie group that is isomorphic to T n = R n /Z n. Definition 1.1.

More information

On the simplest expression of the perturbed Moore Penrose metric generalized inverse

On the simplest expression of the perturbed Moore Penrose metric generalized inverse Annals of the University of Bucharest (mathematical series) 4 (LXII) (2013), 433 446 On the simplest expression of the perturbed Moore Penrose metric generalized inverse Jianbing Cao and Yifeng Xue Communicated

More information