Analysis of the Balanced Incomplete Block Design. Each of t treatments is applied a total of r times, distributed across b blocks, each made up

Size: px
Start display at page:

Download "Analysis of the Balanced Incomplete Block Design. Each of t treatments is applied a total of r times, distributed across b blocks, each made up"

Transcription

1 1 Analysis of the Balanced Incomplete Bloc Design Each of t treatments is applied a total of r times, distributed across b blocs, each made up of plots. For = 1,...,t and = 1,...,b write n for the number of times treatment appears in bloc. Each n is either 0 or 1, and so n = and n = t. The word balanced comes from the fact that each pair of treatments appears together in λ blocs. Constraints on the parameters The total number of plots equals n = tr = b. The parameters are also connected by a relationship obtained by counting the total number of plots in all r blocs where treatment 1 appears. Each of those r blocs consists of plots, giving a total of r plots. Moreover, each of the r blocs contains treatment 1 together with the 1 other treatments that appear together with 1 somewhere. Thus <1> r 1 = 1 The linear model Define indicators B 1,...,B b to identify which bloc a plot belongs to, and indicators T 1,...,T t to identify the treatment applied to the plot. Notice that T B = n. The model asserts that the vector of yields Y has variance matrix σ 2 I n and vector of expected values in the subspace X spanned by B 1,...,B b,t 1,...,T t. That is, EY = β B + τ T, for unnown parameters β 1,...,τ t. The representation is overparametrized, because the spanning vectors are linearly dependent. For example, 1 = B = T. In particular, the τ are not identifiable unless they are subected to a constraint such as τ = 0, as explained later. Bloc and treatment subspaces The vectors {B / } form an orthonormal basis for a b-dimensional subspace X B of X. Denote by X T the part of X that is orthogonal to X B. In general, X T is not a subspce of spant 1,...,T t. Write H T for the linear map that is, a matrix, if you are thining of the elements of R n as column vectors and not as b t tables with some cells empty that proects R n onto X T. Write T for H T T, the component of T that is orthogonal to X B : T T = T component of T in X B = T Note that B B = T 1 n B <2> T T = T T because T T X T = T T 1 n 2 = r r and, for, <3> T T = T T = T T 1 n n = λ

2 2 Dimension of the treatment subspace <4> The subspace X T is spanned by the vectors H T B 1,...,H T B b,h T T 1,...,H T T t, that is, by T 1,..., T t because H T B = 0. The { T } are not linearly independent, because T = H T T = H T 1 = H T B = 0. However, the vectors T 2,..., T t are linearly independent. For suppose 0 = 2 θ T. Dot with T 1 to get 0 = θ λ That is, θ = 0. Dot with T, for 2toget 0=θ r r λ θ = θ r r + λ because θ = θ It follows that θ = 0 because, by <1>, <5> r r + λ = 0. Thus X T has dimension t 1. Proection matrix The proection H T onto X T equals M = T T To establish this fact, first notice that Mx = 0ifxis orthogonal to each T. That is, Mx = 0if x X T. To show that Mz = z for each z in X T it suffices to chec that M T = T for each, because each vector in X T is a linear combination of the spanning set { T }. T T T M T = = r r T λ = T from <4> and <5> That is, M T = T. It follows that H T = M, as asserted. T from <2> and <3> Least squares estimates The vector of fitted values must be a linear combination of the vectors spanning X, that is, Ŷ = β B + τ T Of course the coefficients are not uniquely determined by this equality, because the vectors are linearly dependent. We could replace the T by T plus a linear combination of the B s, then rewrite the representation as <6> Ŷ = β B + τ T where β is a linear combination of the β and the τ s. This representation has the advantage that the first sum gives the component of Ŷ in X B and the second sum gives the component in

3 3 the orthogonal subspace X T. In particular, β = 1 Y B = Y + = average yield in bloc. The τ will be uniquely determined if we add the constraint that τ = 0. Write Y + for the total yield from plots receiving treatment. Then, Y + = Y T = Ŷ T because Y Ŷ X T = β B T + τ T T = 1 = 1 Y + n + r r τ λ n Y + + τ λ τ τ by <5> Subect to τ = 0, these equations have the unique solution τ = Y + 1 n Y + That is, τ = Y T 1 n B Y = T Y Notice that n B Y is ust the sum of all yields over blocs in which treatment appears. Estimates of treatment differences The individual τ in the representation for EY are not uniquely determined, and hence they are not estimable, but we will see that their differences are functions of the expected values. With the nonunique choice of τ as above, E τ = EY T = τ T T because B T = 0 = r r τ λ τ = τ λ τ by <5> Consequently, E τ τ = τ τ. The differences are estimable because they equal the expectations of linear functions of the yields. Notice that the functions τ 1 t τ are estimable. The τ are uniquely determined only up to an additive constant. If they are constrained by τ = 0 then the τ are uniquely determined. Variances var τ τ = From <2> and <3>, and then <5>, 2 var Y T 2 T = σ 2 T T 2. T T 2 = T 2 + T 2 2 T T =2 r r +2 λ =2

4 It follows that var τ τ = 2σ 2 /. Notice the symmetry: all treatment differences can be estimated with the same accuracy because of the balance in the design. 4 A 2 15 factorial design in blocs of size 16. The following explanation is based on the discussion in Fisher s boo Design of Experiments, and on ideas in the solutions of Feng Liang and Andreas Schulz. The problem ased for a design such that no main effects or interactions are confounded. The solution amounts to finding a subgroup of size 2 11 of the group of all products of factors, such that no singletons or doublets are in the group. The group operation corresponds to pointwise multiplication of signed indicator vectors, that is, of vectors { +1 if factor i appears at high level F i = 1 if factor i appears at low level Notice that F i F i = 1, the constant vector of 1 s, for every i. As suggested in Section 45.1 of Fisher s boo, the subgroup G of factor combinations that are confounded with the blocs should be generated by S = S 0 S 1 S 2, the following collection of 11 generators, each consisting of a product of three factors: S 0 S 1 S 2 F 1 F 2 F 3 F 2 F 4 F 6 F 4 F 8 F 12 F 1 F 4 F 5 F 2 F 8 F 10 F 1 F 6 F 7 F 2 F 12 F 14 F 1 F 8 F 9 F 1 F 10 F 11 F 1 F 12 F 13 F 1 F 14 F 15 Notice the pattern. The 8 members of S 0 are obtained by writing F 1 followed by successive pairs, F 2 F 4,...,F 14 F 15. The set S 1 is obtained from S 0 by multiplying each subscript by 2 then discarding any products involving a subscript greater than 15. The set S 2 is obtained in similar fashion from S 1. To simplify the exposition I will refer to a factor whose subscript is an odd number as odd, to a factor whose subscript is an even number as even, to a factor whose subscript is twice an odd number as 2 odd, and so on. Notice that: A Products of one or two generators from S 0 contain two odd factors and at least one even factor. B Products of three or more generators from S 0 contain at least three odd factors. There are several facts that need to be established. i All different products of generators or 2 11 if we include the empty product as giving 1 give distinct factor combinations. For example, if F 1 F 3 F 4 F 6 = F 1 F 2 F 3 F 2 F 4 F 6 were in the list of generators then many of the products would correspond to the same factor combination. ii No singleton a product of the form F i belongs to G. iii No doublet a product of the form F i F, with i belongs to G. Proof of i It is sufficient to show that no generator can be expressed as a product of generators appearing earlier in the listing of S.

5 5 For generators in S 0 the property is obvious, because each new generator introduces a new odd factor. No generator in S 0 can appear as a product for a generator from S 1 S 2, because those generators contain no odd factors. No generator in S 1 is a product of earlier generators from S 1 because each generator introduces a new 2 odd factor. And so on. I have written the argument to suggest the extension to the more general case of n = 2 1 factors in blocs of size 2. Proof of ii From A, any product containing a generator from S 0 must contain at least two odd factors, which could not be cancelled out by generators in S 1 S 2. Thus no generators from S 0 appear in the product. A similar argument, with odd replaced by 2 odd, can be applied to any product of generators from S 1. Thus no generators from S 1 appear in the product. And so on. Proof of iii Suppose a product of generators gave a doublet. Consider how many generators could come from S 0. From B, three or more generators from S 0 would give three odd factors, which could not be cancelled out by generators from S 1 S 2. A single generator from S 0 would contribute two odd factors together with a single even factor that would have to be cancelled out by a product from S 1 S 2 that reduces a singleton, contradicting ii. A pair of generators from S 0 would contribute two odd factors, together with a pair of 2 odd factors that would have to be cancelled out by a product from S 1 S 2. That is, it would have to be possible to get a doublet of 2 odd factors using products from S 1 S 2. With no generators from S 0 we would be left with the tas of finding a product of generators from S 1 S 2 that reduced to a doublet. The last two possibilities both require construction of a doublet as a product from S 1 S 2. But that problem is essentially a copy of the problem of constructing a doublet from a product of generators from S 0 S 1 S 2, except that all subscripts get multiplied by 2. Repeat the argument to reduce the last problem to the impossible tas of getting a doublet as a product of generators from S 1 alone. Again I have written the argument to suggest the extension to the general case.

is Use at most six elementary row operations. (Partial

is Use at most six elementary row operations. (Partial MATH 235 SPRING 2 EXAM SOLUTIONS () (6 points) a) Show that the reduced row echelon form of the augmented matrix of the system x + + 2x 4 + x 5 = 3 x x 3 + x 4 + x 5 = 2 2x + 2x 3 2x 4 x 5 = 3 is. Use

More information

SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1. I. Topics from linear algebra

SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1. I. Topics from linear algebra SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1 Winter 2009 I. Topics from linear algebra I.0 : Background 1. Suppose that {x, y} is linearly dependent. Then there are scalars a, b which are not both

More information

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016 Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016 1. Let V be a vector space. A linear transformation P : V V is called a projection if it is idempotent. That

More information

Lecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018

Lecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018 Lecture 13: Orthogonal projections and least squares (Section 3.2-3.3) Thang Huynh, UC San Diego 2/9/2018 Orthogonal projection onto subspaces Theorem. Let W be a subspace of R n. Then, each x in R n can

More information

Algebraic Methods in Combinatorics

Algebraic Methods in Combinatorics Algebraic Methods in Combinatorics Po-Shen Loh 27 June 2008 1 Warm-up 1. (A result of Bourbaki on finite geometries, from Răzvan) Let X be a finite set, and let F be a family of distinct proper subsets

More information

MTH 2310, FALL Introduction

MTH 2310, FALL Introduction MTH 2310, FALL 2011 SECTION 6.2: ORTHOGONAL SETS Homework Problems: 1, 5, 9, 13, 17, 21, 23 1, 27, 29, 35 1. Introduction We have discussed previously the benefits of having a set of vectors that is linearly

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

1 Principal Components Analysis

1 Principal Components Analysis Lecture 3 and 4 Sept. 18 and Sept.20-2006 Data Visualization STAT 442 / 890, CM 462 Lecture: Ali Ghodsi 1 Principal Components Analysis Principal components analysis (PCA) is a very popular technique for

More information

Solutions to Review Problems for Chapter 6 ( ), 7.1

Solutions to Review Problems for Chapter 6 ( ), 7.1 Solutions to Review Problems for Chapter (-, 7 The Final Exam is on Thursday, June,, : AM : AM at NESBITT Final Exam Breakdown Sections % -,7-9,- - % -9,-,7,-,-7 - % -, 7 - % Let u u and v Let x x x x,

More information

There are two things that are particularly nice about the first basis

There are two things that are particularly nice about the first basis Orthogonality and the Gram-Schmidt Process In Chapter 4, we spent a great deal of time studying the problem of finding a basis for a vector space We know that a basis for a vector space can potentially

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

REFLECTIONS IN A EUCLIDEAN SPACE

REFLECTIONS IN A EUCLIDEAN SPACE REFLECTIONS IN A EUCLIDEAN SPACE PHILIP BROCOUM Let V be a finite dimensional real linear space. Definition 1. A function, : V V R is a bilinear form in V if for all x 1, x, x, y 1, y, y V and all k R,

More information

Chapter 6. Orthogonality

Chapter 6. Orthogonality 6.4 The Projection Matrix 1 Chapter 6. Orthogonality 6.4 The Projection Matrix Note. In Section 6.1 (Projections), we projected a vector b R n onto a subspace W of R n. We did so by finding a basis for

More information

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections Math 6 Lecture Notes: Sections 6., 6., 6. and 6. Orthogonal Sets and Projections We will not cover general inner product spaces. We will, however, focus on a particular inner product space the inner product

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

The Gram Schmidt Process

The Gram Schmidt Process u 2 u The Gram Schmidt Process Now we will present a procedure, based on orthogonal projection, that converts any linearly independent set of vectors into an orthogonal set. Let us begin with the simple

More information

The Gram Schmidt Process

The Gram Schmidt Process The Gram Schmidt Process Now we will present a procedure, based on orthogonal projection, that converts any linearly independent set of vectors into an orthogonal set. Let us begin with the simple case

More information

Math 61CM - Solutions to homework 6

Math 61CM - Solutions to homework 6 Math 61CM - Solutions to homework 6 Cédric De Groote November 5 th, 2018 Problem 1: (i) Give an example of a metric space X such that not all Cauchy sequences in X are convergent. (ii) Let X be a metric

More information

Answer Key for Exam #2

Answer Key for Exam #2 . Use elimination on an augmented matrix: Answer Key for Exam # 4 4 8 4 4 4 The fourth column has no pivot, so x 4 is a free variable. The corresponding system is x + x 4 =, x =, x x 4 = which we solve

More information

Dot product. The dot product is an inner product on a coordinate vector space (Definition 1, Theorem

Dot product. The dot product is an inner product on a coordinate vector space (Definition 1, Theorem Dot product The dot product is an inner product on a coordinate vector space (Definition 1, Theorem 1). Definition 1 Given vectors v and u in n-dimensional space, the dot product is defined as, n v u v

More information

Further Mathematical Methods (Linear Algebra) 2002

Further Mathematical Methods (Linear Algebra) 2002 Further Mathematical Methods (Linear Algebra) Solutions For Problem Sheet 9 In this problem sheet, we derived a new result about orthogonal projections and used them to find least squares approximations

More information

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent. Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u

More information

Vectors in Function Spaces

Vectors in Function Spaces Jim Lambers MAT 66 Spring Semester 15-16 Lecture 18 Notes These notes correspond to Section 6.3 in the text. Vectors in Function Spaces We begin with some necessary terminology. A vector space V, also

More information

On Least Squares Linear Regression Without Second Moment

On Least Squares Linear Regression Without Second Moment On Least Squares Linear Regression Without Second Moment BY RAJESHWARI MAJUMDAR University of Connecticut If \ and ] are real valued random variables such that the first moments of \, ], and \] exist and

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r Practice final solutions. I did not include definitions which you can find in Axler or in the course notes. These solutions are on the terse side, but would be acceptable in the final. However, if you

More information

Orthogonality. Orthonormal Bases, Orthogonal Matrices. Orthogonality

Orthogonality. Orthonormal Bases, Orthogonal Matrices. Orthogonality Orthonormal Bases, Orthogonal Matrices The Major Ideas from Last Lecture Vector Span Subspace Basis Vectors Coordinates in different bases Matrix Factorization (Basics) The Major Ideas from Last Lecture

More information

CS 322 Homework 4 Solutions

CS 322 Homework 4 Solutions CS 322 Homework 4 Solutions out: Thursday 1 March 2007 due: Wednesday 7 March 2007 Problem 1: QR mechanics 1 Prove that a Householder reflection is symmetric and orthogonal. Answer: Symmetry: H T = (I

More information

(1) is an invertible sheaf on X, which is generated by the global sections

(1) is an invertible sheaf on X, which is generated by the global sections 7. Linear systems First a word about the base scheme. We would lie to wor in enough generality to cover the general case. On the other hand, it taes some wor to state properly the general results if one

More information

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition 6 Vector Spaces with Inned Product Basis and Dimension Section Objective(s): Vector Spaces and Subspaces Linear (In)dependence Basis and Dimension Inner Product 6 Vector Spaces and Subspaces Definition

More information

Math 3A Winter 2016 Midterm

Math 3A Winter 2016 Midterm Math 3A Winter 016 Midterm Name Signature UCI ID # E-mail address There are 7 problems for a total of 115 points. Present your work as clearly as possible. Partial credit will be awarded, and you must

More information

Winter 2017 Ma 1b Analytical Problem Set 2 Solutions

Winter 2017 Ma 1b Analytical Problem Set 2 Solutions 1. (5 pts) From Ch. 1.10 in Apostol: Problems 1,3,5,7,9. Also, when appropriate exhibit a basis for S. Solution. (1.10.1) Yes, S is a subspace of V 3 with basis {(0, 0, 1), (0, 1, 0)} and dimension 2.

More information

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. Orthogonal sets Let V be a vector space with an inner product. Definition. Nonzero vectors v 1,v

More information

Inner Product and Orthogonality

Inner Product and Orthogonality Inner Product and Orthogonality P. Sam Johnson October 3, 2014 P. Sam Johnson (NITK) Inner Product and Orthogonality October 3, 2014 1 / 37 Overview In the Euclidean space R 2 and R 3 there are two concepts,

More information

Section 6.2, 6.3 Orthogonal Sets, Orthogonal Projections

Section 6.2, 6.3 Orthogonal Sets, Orthogonal Projections Section 6. 6. Orthogonal Sets Orthogonal Projections Main Ideas in these sections: Orthogonal set = A set of mutually orthogonal vectors. OG LI. Orthogonal Projection of y onto u or onto an OG set {u u

More information

Vectors. Vectors and the scalar multiplication and vector addition operations:

Vectors. Vectors and the scalar multiplication and vector addition operations: Vectors Vectors and the scalar multiplication and vector addition operations: x 1 x 1 y 1 2x 1 + 3y 1 x x n 1 = 2 x R n, 2 2 y + 3 2 2x = 2 + 3y 2............ x n x n y n 2x n + 3y n I ll use the two terms

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

Orthogonal Complements

Orthogonal Complements Orthogonal Complements Definition Let W be a subspace of R n. If a vector z is orthogonal to every vector in W, then z is said to be orthogonal to W. The set of all such vectors z is called the orthogonal

More information

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 : Length, Angle and the Inner Product The length (or norm) of a vector v R 2 (viewed as connecting the origin to a point (v 1,v 2 )) is easily determined by the Pythagorean Theorem and is denoted v : v =

More information

MATH SOLUTIONS TO PRACTICE PROBLEMS - MIDTERM I. 1. We carry out row reduction. We begin with the row operations

MATH SOLUTIONS TO PRACTICE PROBLEMS - MIDTERM I. 1. We carry out row reduction. We begin with the row operations MATH 2 - SOLUTIONS TO PRACTICE PROBLEMS - MIDTERM I. We carry out row reduction. We begin with the row operations yielding the matrix This is already upper triangular hence The lower triangular matrix

More information

LINEAR ALGEBRA W W L CHEN

LINEAR ALGEBRA W W L CHEN LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008. This chapter is available free to all individuals, on the understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,

More information

22 Approximations - the method of least squares (1)

22 Approximations - the method of least squares (1) 22 Approximations - the method of least squares () Suppose that for some y, the equation Ax = y has no solutions It may happpen that this is an important problem and we can t just forget about it If we

More information

6.1. Inner Product, Length and Orthogonality

6.1. Inner Product, Length and Orthogonality These are brief notes for the lecture on Friday November 13, and Monday November 1, 2009: they are not complete, but they are a guide to what I want to say on those days. They are guaranteed to be incorrect..1.

More information

Chapter 6: Orthogonality

Chapter 6: Orthogonality Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products

More information

GEOMETRY OF MATRICES x 1

GEOMETRY OF MATRICES x 1 GEOMETRY OF MATRICES. SPACES OF VECTORS.. Definition of R n. The space R n consists of all column vectors with n components. The components are real numbers... Representation of Vectors in R n.... R. The

More information

Lecture 10: Vector Algebra: Orthogonal Basis

Lecture 10: Vector Algebra: Orthogonal Basis Lecture 0: Vector Algebra: Orthogonal Basis Orthogonal Basis of a subspace Computing an orthogonal basis for a subspace using Gram-Schmidt Orthogonalization Process Orthogonal Set Any set of vectors that

More information

Math 2331 Linear Algebra

Math 2331 Linear Algebra 6.2 Orthogonal Sets Math 233 Linear Algebra 6.2 Orthogonal Sets Jiwen He Department of Mathematics, University of Houston jiwenhe@math.uh.edu math.uh.edu/ jiwenhe/math233 Jiwen He, University of Houston

More information

(Ref: Schensted Part II) If we have an arbitrary tensor with k indices W i 1,,i k. we can act on it 1 2 k with a permutation P = = w ia,i b,,i l

(Ref: Schensted Part II) If we have an arbitrary tensor with k indices W i 1,,i k. we can act on it 1 2 k with a permutation P = = w ia,i b,,i l Chapter S k and Tensor Representations Ref: Schensted art II) If we have an arbitrary tensor with k indices W i,,i k ) we can act on it k with a permutation = so a b l w) i,i,,i k = w ia,i b,,i l. Consider

More information

Linear Algebra Final Exam Study Guide Solutions Fall 2012

Linear Algebra Final Exam Study Guide Solutions Fall 2012 . Let A = Given that v = 7 7 67 5 75 78 Linear Algebra Final Exam Study Guide Solutions Fall 5 explain why it is not possible to diagonalize A. is an eigenvector for A and λ = is an eigenvalue for A diagonalize

More information

LINEAR ALGEBRA REVIEW

LINEAR ALGEBRA REVIEW LINEAR ALGEBRA REVIEW When we define a term, we put it in boldface. This is a very compressed review; please read it very carefully and be sure to ask questions on parts you aren t sure of. x 1 WedenotethesetofrealnumbersbyR.

More information

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar

More information

MIDTERM I LINEAR ALGEBRA. Friday February 16, Name PRACTICE EXAM SOLUTIONS

MIDTERM I LINEAR ALGEBRA. Friday February 16, Name PRACTICE EXAM SOLUTIONS MIDTERM I LIEAR ALGEBRA MATH 215 Friday February 16, 2018. ame PRACTICE EXAM SOLUTIOS Please answer the all of the questions, and show your work. You must explain your answers to get credit. You will be

More information

P = A(A T A) 1 A T. A Om (m n)

P = A(A T A) 1 A T. A Om (m n) Chapter 4: Orthogonality 4.. Projections Proposition. Let A be a matrix. Then N(A T A) N(A). Proof. If Ax, then of course A T Ax. Conversely, if A T Ax, then so Ax also. x (A T Ax) x T A T Ax (Ax) T Ax

More information

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6 Chapter 6 Orthogonality 6.1 Orthogonal Vectors and Subspaces Recall that if nonzero vectors x, y R n are linearly independent then the subspace of all vectors αx + βy, α, β R (the space spanned by x and

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

LINEAR ALGEBRA KNOWLEDGE SURVEY

LINEAR ALGEBRA KNOWLEDGE SURVEY LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.

More information

Algebraic Methods in Combinatorics

Algebraic Methods in Combinatorics Algebraic Methods in Combinatorics Po-Shen Loh June 2009 1 Linear independence These problems both appeared in a course of Benny Sudakov at Princeton, but the links to Olympiad problems are due to Yufei

More information

CMPSCI 250: Introduction to Computation. Lecture #15: The Fundamental Theorem of Arithmetic David Mix Barrington 24 February 2014

CMPSCI 250: Introduction to Computation. Lecture #15: The Fundamental Theorem of Arithmetic David Mix Barrington 24 February 2014 CMPSCI 250: Introduction to Computation Lecture #15: The Fundamental Theorem of Arithmetic David Mix Barrington 24 February 2014 The Fundamental Theorem Statement of the Theorem Existence of a Factorization

More information

Approximations - the method of least squares (1)

Approximations - the method of least squares (1) Approximations - the method of least squares () In many applications, we have to consider the following problem: Suppose that for some y, the equation Ax = y has no solutions It could be that this is an

More information

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................

More information

Mathematical Foundations: Intro

Mathematical Foundations: Intro Mathematical Foundations: Intro Graphics relies on 3 basic objects: 1. Scalars 2. Vectors 3. Points Mathematically defined in terms of spaces: 1. Vector space 2. Affine space 3. Euclidean space Math required:

More information

Proofs for Quizzes. Proof. Suppose T is a linear transformation, and let A be a matrix such that T (x) = Ax for all x R m. Then

Proofs for Quizzes. Proof. Suppose T is a linear transformation, and let A be a matrix such that T (x) = Ax for all x R m. Then Proofs for Quizzes 1 Linear Equations 2 Linear Transformations Theorem 1 (2.1.3, linearity criterion). A function T : R m R n is a linear transformation if and only if a) T (v + w) = T (v) + T (w), for

More information

Section 6.4. The Gram Schmidt Process

Section 6.4. The Gram Schmidt Process Section 6.4 The Gram Schmidt Process Motivation The procedures in 6 start with an orthogonal basis {u, u,..., u m}. Find the B-coordinates of a vector x using dot products: x = m i= x u i u i u i u i Find

More information

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold: Inner products Definition: An inner product on a real vector space V is an operation (function) that assigns to each pair of vectors ( u, v) in V a scalar u, v satisfying the following axioms: 1. u, v

More information

Systems of Identical Particles

Systems of Identical Particles qmc161.tex Systems of Identical Particles Robert B. Griffiths Version of 21 March 2011 Contents 1 States 1 1.1 Introduction.............................................. 1 1.2 Orbitals................................................

More information

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam January 23, 2015

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam January 23, 2015 University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra PhD Preliminary Exam January 23, 2015 Name: Exam Rules: This exam lasts 4 hours and consists of

More information

Math Linear Algebra

Math Linear Algebra Math 220 - Linear Algebra (Summer 208) Solutions to Homework #7 Exercise 6..20 (a) TRUE. u v v u = 0 is equivalent to u v = v u. The latter identity is true due to the commutative property of the inner

More information

Consistent Histories. Chapter Chain Operators and Weights

Consistent Histories. Chapter Chain Operators and Weights Chapter 10 Consistent Histories 10.1 Chain Operators and Weights The previous chapter showed how the Born rule can be used to assign probabilities to a sample space of histories based upon an initial state

More information

Functional Analysis HW #5

Functional Analysis HW #5 Functional Analysis HW #5 Sangchul Lee October 29, 2015 Contents 1 Solutions........................................ 1 1 Solutions Exercise 3.4. Show that C([0, 1]) is not a Hilbert space, that is, there

More information

Domesticity in projective spaces

Domesticity in projective spaces Innovations in Incidence Geometry Volume 12 (2011), Pages 141 149 ISSN 1781-6475 ACADEMIA PRESS Domesticity in projective spaces Beukje Temmermans Joseph A. Thas Hendrik Van Maldeghem Abstract Let J be

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra Throughout these notes, F denotes a field (often called the scalars in this context). 1 Definition of a vector space Definition 1.1. A F -vector space or simply a vector space

More information

Footnotes to Linear Algebra (MA 540 fall 2013), T. Goodwillie, Bases

Footnotes to Linear Algebra (MA 540 fall 2013), T. Goodwillie, Bases Footnotes to Linear Algebra (MA 540 fall 2013), T. Goodwillie, Bases November 18, 2013 1 Spanning and linear independence I will outline a slightly different approach to the material in Chapter 2 of Axler

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

THERE IS NO Sz(8) IN THE MONSTER

THERE IS NO Sz(8) IN THE MONSTER THERE IS NO Sz(8) IN THE MONSTER ROBERT A. WILSON Abstract. As a contribution to an eventual solution of the problem of the determination of the maximal subgroups of the Monster we show that there is no

More information

Some notes on Coxeter groups

Some notes on Coxeter groups Some notes on Coxeter groups Brooks Roberts November 28, 2017 CONTENTS 1 Contents 1 Sources 2 2 Reflections 3 3 The orthogonal group 7 4 Finite subgroups in two dimensions 9 5 Finite subgroups in three

More information

CAAM 336 DIFFERENTIAL EQUATIONS IN SCI AND ENG Examination 1

CAAM 336 DIFFERENTIAL EQUATIONS IN SCI AND ENG Examination 1 CAAM 6 DIFFERENTIAL EQUATIONS IN SCI AND ENG Examination Instructions: Time limit: uninterrupted hours There are four questions worth a total of 5 points Please do not look at the questions until you begin

More information

Chapter 6 - Orthogonality

Chapter 6 - Orthogonality Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/

More information

1 Basic Combinatorics

1 Basic Combinatorics 1 Basic Combinatorics 1.1 Sets and sequences Sets. A set is an unordered collection of distinct objects. The objects are called elements of the set. We use braces to denote a set, for example, the set

More information

More chapter 3...linear dependence and independence... vectors

More chapter 3...linear dependence and independence... vectors More chapter 3...linear dependence and independence... vectors It is important to determine if a set of vectors is linearly dependent or independent Consider a set of vectors A, B, and C. If we can find

More information

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix Definition: Let L : V 1 V 2 be a linear operator. The null space N (L) of L is the subspace of V 1 defined by N (L) = {x

More information

The Gram-Schmidt Process 1

The Gram-Schmidt Process 1 The Gram-Schmidt Process In this section all vector spaces will be subspaces of some R m. Definition.. Let S = {v...v n } R m. The set S is said to be orthogonal if v v j = whenever i j. If in addition

More information

ISOMETRIES OF R n KEITH CONRAD

ISOMETRIES OF R n KEITH CONRAD ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x

More information

MATH 12 CLASS 4 NOTES, SEP

MATH 12 CLASS 4 NOTES, SEP MATH 12 CLASS 4 NOTES, SEP 28 2011 Contents 1. Lines in R 3 1 2. Intersections of lines in R 3 2 3. The equation of a plane 4 4. Various problems with planes 5 4.1. Intersection of planes with planes or

More information

Inner Product Spaces

Inner Product Spaces Inner Product Spaces Introduction Recall in the lecture on vector spaces that geometric vectors (i.e. vectors in two and three-dimensional Cartesian space have the properties of addition, subtraction,

More information

The Symmetric Space for SL n (R)

The Symmetric Space for SL n (R) The Symmetric Space for SL n (R) Rich Schwartz November 27, 2013 The purpose of these notes is to discuss the symmetric space X on which SL n (R) acts. Here, as usual, SL n (R) denotes the group of n n

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

VECTOR AND TENSOR ALGEBRA

VECTOR AND TENSOR ALGEBRA PART I VECTOR AND TENSOR ALGEBRA Throughout this book: (i) Lightface Latin and Greek letters generally denote scalars. (ii) Boldface lowercase Latin and Greek letters generally denote vectors, but the

More information

Criteria for existence of semigroup homomorphisms and projective rank functions. George M. Bergman

Criteria for existence of semigroup homomorphisms and projective rank functions. George M. Bergman Criteria for existence of semigroup homomorphisms and projective rank functions George M. Bergman Suppose A, S, and T are semigroups, e: A S and f: A T semigroup homomorphisms, and X a generating set for

More information

0.1. Linear transformations

0.1. Linear transformations Suggestions for midterm review #3 The repetitoria are usually not complete; I am merely bringing up the points that many people didn t now on the recitations Linear transformations The following mostly

More information

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,

More information

Lecture 14: The Rank-Nullity Theorem

Lecture 14: The Rank-Nullity Theorem Math 108a Professor: Padraic Bartlett Lecture 14: The Rank-Nullity Theorem Week 6 UCSB 2013 In today s talk, the last before we introduce the concept of matrices, we prove what is arguably the strongest

More information

The 'linear algebra way' of talking about "angle" and "similarity" between two vectors is called "inner product". We'll define this next.

The 'linear algebra way' of talking about angle and similarity between two vectors is called inner product. We'll define this next. Orthogonality and QR The 'linear algebra way' of talking about "angle" and "similarity" between two vectors is called "inner product". We'll define this next. So, what is an inner product? An inner product

More information

Let p 2 ( t), (2 t k), we have the scaling relation,

Let p 2 ( t), (2 t k), we have the scaling relation, Multiresolution Analysis and Daubechies N Wavelet We have discussed decomposing a signal into its Haar wavelet components of varying frequencies. The Haar wavelet scheme relied on two functions: the Haar

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

No Solution Equations Let s look at the following equation: 2 +3=2 +7

No Solution Equations Let s look at the following equation: 2 +3=2 +7 5.4 Solving Equations with Infinite or No Solutions So far we have looked at equations where there is exactly one solution. It is possible to have more than solution in other types of equations that are

More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures

More information

Metric Space Topology (Spring 2016) Selected Homework Solutions. HW1 Q1.2. Suppose that d is a metric on a set X. Prove that the inequality d(x, y)

Metric Space Topology (Spring 2016) Selected Homework Solutions. HW1 Q1.2. Suppose that d is a metric on a set X. Prove that the inequality d(x, y) Metric Space Topology (Spring 2016) Selected Homework Solutions HW1 Q1.2. Suppose that d is a metric on a set X. Prove that the inequality d(x, y) d(z, w) d(x, z) + d(y, w) holds for all w, x, y, z X.

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

11 Block Designs. Linear Spaces. Designs. By convention, we shall

11 Block Designs. Linear Spaces. Designs. By convention, we shall 11 Block Designs Linear Spaces In this section we consider incidence structures I = (V, B, ). always let v = V and b = B. By convention, we shall Linear Space: We say that an incidence structure (V, B,

More information

Math 250B Midterm II Information Spring 2019 SOLUTIONS TO PRACTICE PROBLEMS

Math 250B Midterm II Information Spring 2019 SOLUTIONS TO PRACTICE PROBLEMS Math 50B Midterm II Information Spring 019 SOLUTIONS TO PRACTICE PROBLEMS Problem 1. Determine whether each set S below forms a subspace of the given vector space V. Show carefully that your answer is

More information