Average mixing of continuous quantum walks
|
|
- Melina Preston
- 5 years ago
- Views:
Transcription
1 Average mixing of continuous quantum walks Juergen Kritschgau Department of Mathematics Iowa State University Ames, IA April 7, 2017
2 Juergen Kritschgau (ISU) 2 of 17 This is a presentation of: Average mixing of continuous quantum walks, Journal of Combinatorial Theory, Series A June 2013, by Chris Godsil
3 Juergen Kritschgau (ISU) 3 of 17 Outline 1) Background and Definitions, particularly, the average mixing matrix (AMM) 2) How to work with the AMM 3) The AMM of paths 4) The AMM and cospectral vertices
4 Juergen Kritschgau (ISU) 4 of 17 Continuous Quantum Walk Let X be a graph on V (X ) = {1,..., n} vertices with adjacency matrix A. We will be considering the operator H(t) = exp(ıta) = k=0 1 k! (ıta)k where t R. The idea is that H(t) determines a continuous quantum walk by associating the standard basis vector e i C n with the vertex i. Doing so, yields that e T u H(t)e v 2 is the probability that the quantum walk starting at vertex v is at vertex u at time t.
5 Juergen Kritschgau (ISU) 5 of 17 Average Mixing Matrix of X The Schur Product of two n m matrices A, B, denoted by A B, is the entry-wise product of A and B.Then define the following: M X (t) := H(t) H( t) 1 T ˆM X := lim M X (t)dt T T 0 The matrix ˆM X is the average mixing matrix of X. The goal is to relate properties of ˆMX (t) to X.
6 Juergen Kritschgau (ISU) 6 of 17 Properties of the Matrix Exponential There are a few properties of the matrix exponential that are useful for us. Let B, C be complex square matrices. 1) If B is skew hermitian, then exp(b) is unitary. 2) If BC = CB, then exp(b + C) = exp(b)exp(c). 3) If B is idempotent, then expb = k=0 1 k! Bk = I + ( k=1 1 k!) B = I + (e 1)P Property 1) gives us that H(t) = exp(ıta) is unitary in virtue of (ıta) = ıta
7 Juergen Kritschgau (ISU) 7 of 17 Working with ˆM X (t) Recall that the spectral decomposition of A is given by A = r θ r E r where E r is idempotent. Then we have: H(t) = r exp(ıtθ r )E r M X (t) = r E r E r + 2 r s cos((θ r θ s )t)e r E s Lemma If A = r θ r E r, then ˆM X = r E r E r.
8 Juergen Kritschgau (ISU) 8 of 17 Connectivity Lemma If X is connected, then all entries of ˆM X are positive. Proof. Notice that E r E r is nonnegative. Suppose ( ˆM X ) u,v = 0. This implies (E r E r ) u,v = 0 for all r. Furthermore, this implies that (E r ) u,v = 0 for all r. Recall that A is a linear combination of E r and E i E j = 0 for i j. Therefore, (A k ) u,v = 0 for all k, and X is disconnected.
9 Average Mixing on Paths Lemma The idempotents of E 1,..., E n in the spectral decomposition of P n are given by (E r ) j,k = 2 ( ) ( ) jrπ krπ n + 1 sin sin. n + 1 n + 1 Proof. Recall that the adjacency eigenvalues of P n are 2 cos(β) with sin(β) sin(2β) eigenvectors z(β) = where β ranges over. sin(nβ) β = πr, r [n]. To be continued... n + 1 Juergen Kritschgau (ISU) 9 of 17
10 Juergen Kritschgau (ISU) 10 of 17 Average Mixing on Paths [FUN FACT!] If λ is a simple eigenvalue with vector x, then 1 x T x xx T is the projection onto the eigenspace of λ. proof continued. 1 Using the fun fact, we get that z(β) T z(β) = 2 n+1 and ( ) ( ) jrπ krπ (z(β)z(β) T ) j,k = sin(jβ) sin(kβ) = sin sin. n + 1 n + 1
11 Juergen Kritschgau (ISU) 11 of 17 Average Mixing on Paths Lemma If E 1,..., E n are idempotents for P n, then ˆM Pn = r E r E r = 1 (2J + I + J). 2n + 2 Proof. By the previous Lemma we have (E r E r ) j,k = 4 (n + 1) 2 sin2 Using some trigonometry we get... ( ) ( ) jrπ krπ sin 2. n + 1 n + 1
12 Juergen Kritschgau (ISU) 12 of 17 Average Mixing on Paths proof continued. (n + 1) 2 (E r ) E r ) j,k = 1 cos ( 2jrπ n+1 cos ( 2krπ n+1 ) ( ) cos 2(j+k)rπ n+1 ). ( cos 2(j k)rπ n+1 ) = 1, Summing across r and using the fact that ( n r=1 cos 2lrπ n+1 the right hand side (RHS) evaluates to 3(n + 1)/2 j = k; 3(n + 1)/2 j + k = n + 1; RHS = 2n + 2 j = k, j + k = n + 1; n + 1 otherwise.
13 Juergen Kritschgau (ISU) 13 of 17 An Example: P 3 Recall that the eigenvalues of P 3 are 0, 2, 2 with corresponding vectors [1, 0, 1] T, [1, 2, 1] T, [1, 2, 1] T. These are all simple eigenvalues, so the corresponding projections are E 1 = E 2 = E 3 = Then 2 1 3/8 2/8 3/8 ˆM P3 = E 1 /2 + E 2 /4 + E 3 /4 = 2/8 4/8 2/8 3/8 2/8 3/8
14 Juergen Kritschgau (ISU) 14 of 17 Cospectral Vertices Theorem Let X be a graph with vertices u and v, and let E 1,..., E m be the idempotents in the spectral decomposition of A(X ). TFAE: 1) (A k ) u,u = (A k ) v,v for all k 0 2) E r e u 2 = E r e v 2 3) The graphs X u and X v. If any of the conditions above hold for vertices u and v, we say that u and b are cospectral. Furthermore, if E r e u = ±E r e v for all r, then we say that u and v are strongly cospectral. Lemma If u and V are vertices in X and the eigenvalues of X are all simple, then U and v are strongly cospecral if and only if they are cospectral.
15 Juergen Kritschgau (ISU) 15 of 17 Cospectral Vertices Theorem Let ˆM X be the average mixing matrix of the graph X. Then vertices u andv are strongly cospectral if and only if ˆM x (e v e u ) = 0. Proof. Some algebra shows that if N is PSD and N(e u e v ) = 0, then N[v, u](e u e v ) = 0. ˆM X (e u e v ) = 0 implies 0 = (e u e v ) T ˆM X (e u e v ) = r (e u e v ) T (E r E r )(e u e v ). Notice that each summand E r E r is PSD, so we have (E r E r )(e u e v ) = 0 for all r. This implies ((E r ) u,u ) 2 = ((E r ) u,v ) 2 = ((E r ) v,v ) 2. Applying Cauchy-Schwarz gives that E r e u = ±E r e v.
16 Juergen Kritschgau (ISU) 16 of 17 Returning to our Example Recall that that the average mixing matrix of P 3 is 3/8 2/8 3/8 ˆM P3 = 2/8 4/8 2/8. 3/8 2/8 3/8 The first and last row are equal, and P 3 has all simple eigenvalues. Therefore, P 3 v 1 and P 3 v 3 are cospectral.
17 Juergen Kritschgau (ISU) 17 of 17 The End Thank you!
Periodicity & State Transfer Some Results Some Questions. Periodic Graphs. Chris Godsil. St John s, June 7, Chris Godsil Periodic Graphs
St John s, June 7, 2009 Outline 1 Periodicity & State Transfer 2 Some Results 3 Some Questions Unitary Operators Suppose X is a graph with adjacency matrix A. Definition We define the operator H X (t)
More informationAverage Mixing Matrix of Trees
Electronic Journal of Linear Algebra Volume 34 Volume 34 08 Article 9 08 Average Mixing Matrix of Trees Chris Godsil University of Waterloo, cgodsil@uwaterloo.ca Krystal Guo Université libre de Bruxelles,
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationWhich Quantum Walk on Graphs?
Which Quantum Walk on Graphs? Christino Tamon Department of Computer Science Clarkson University August 13, 2015 Outline I Walks on graphs (random and quantum) I State transfer on graphs (motivation) I
More informationWhen can perfect state transfer occur?
Electronic Journal of Linear Algebra Volume 23 Volume 23 (2012) Article 62 2012 When can perfect state transfer occur? Chris Godsil cgodsil@uwaterloo.ca Follow this and additional works at: http://repository.uwyo.edu/ela
More informationDiagonalizing Matrices
Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B
More informationEigenvalues and Eigenvectors
/88 Chia-Ping Chen Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Eigenvalue Problem /88 Eigenvalue Equation By definition, the eigenvalue equation for matrix
More informationLinear algebra and applications to graphs Part 1
Linear algebra and applications to graphs Part 1 Written up by Mikhail Belkin and Moon Duchin Instructor: Laszlo Babai June 17, 2001 1 Basic Linear Algebra Exercise 1.1 Let V and W be linear subspaces
More informationCompound matrices and some classical inequalities
Compound matrices and some classical inequalities Tin-Yau Tam Mathematics & Statistics Auburn University Dec. 3, 04 We discuss some elegant proofs of several classical inequalities of matrices by using
More informationEigenvalues of 2-edge-coverings
Eigenvalues of -edge-coverings Steve Butler October 3, 007 Abstract A -edge-covering between G and H is an onto homomorphism from the vertices of G to the vertices of H so that each edge is covered twice
More informationData Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs
Data Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture
More informationPeter J. Dukes. 22 August, 2012
22 August, 22 Graph decomposition Let G and H be graphs on m n vertices. A decompostion of G into copies of H is a collection {H i } of subgraphs of G such that each H i = H, and every edge of G belongs
More informationNo Laplacian Perfect State Transfer in Trees
No Laplacian Perfect State Transfer in Trees arxiv:1408.2935v1 [math.co] 13 Aug 2014 Gabriel Coutinho Henry Liu October 15, 2018 Abstract We consider a system of qubits coupled via nearest-neighbour interaction
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationLecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?
KTH ROYAL INSTITUTE OF TECHNOLOGY Norms for vectors and matrices Why? Lecture 5 Ch. 5, Norms for vectors and matrices Emil Björnson/Magnus Jansson/Mats Bengtsson April 27, 2016 Problem: Measure size of
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationInner products and Norms. Inner product of 2 vectors. Inner product of 2 vectors x and y in R n : x 1 y 1 + x 2 y x n y n in R n
Inner products and Norms Inner product of 2 vectors Inner product of 2 vectors x and y in R n : x 1 y 1 + x 2 y 2 + + x n y n in R n Notation: (x, y) or y T x For complex vectors (x, y) = x 1 ȳ 1 + x 2
More informationMa/CS 6b Class 20: Spectral Graph Theory
Ma/CS 6b Class 20: Spectral Graph Theory By Adam Sheffer Recall: Parity of a Permutation S n the set of permutations of 1,2,, n. A permutation σ S n is even if it can be written as a composition of an
More informationGraph fundamentals. Matrices associated with a graph
Graph fundamentals Matrices associated with a graph Drawing a picture of a graph is one way to represent it. Another type of representation is via a matrix. Let G be a graph with V (G) ={v 1,v,...,v n
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationMa/CS 6b Class 20: Spectral Graph Theory
Ma/CS 6b Class 20: Spectral Graph Theory By Adam Sheffer Eigenvalues and Eigenvectors A an n n matrix of real numbers. The eigenvalues of A are the numbers λ such that Ax = λx for some nonzero vector x
More informationIr O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )
Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O
More information1 Adjacency matrix and eigenvalues
CSC 5170: Theory of Computational Complexity Lecture 7 The Chinese University of Hong Kong 1 March 2010 Our objective of study today is the random walk algorithm for deciding if two vertices in an undirected
More informationLecture 13: Spectral Graph Theory
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 13: Spectral Graph Theory Lecturer: Shayan Oveis Gharan 11/14/18 Disclaimer: These notes have not been subjected to the usual scrutiny reserved
More informationProperties of Linear Transformations from R n to R m
Properties of Linear Transformations from R n to R m MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Topic Overview Relationship between the properties of a matrix transformation
More informationNotes on Linear Algebra and Matrix Theory
Massimo Franceschet featuring Enrico Bozzo Scalar product The scalar product (a.k.a. dot product or inner product) of two real vectors x = (x 1,..., x n ) and y = (y 1,..., y n ) is not a vector but a
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationIntroduction to Matrix Algebra
Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p
More informationLinear Algebra and Dirac Notation, Pt. 2
Linear Algebra and Dirac Notation, Pt. 2 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 2 February 1, 2017 1 / 14
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationAn Introduction to Spectral Graph Theory
An Introduction to Spectral Graph Theory Mackenzie Wheeler Supervisor: Dr. Gary MacGillivray University of Victoria WheelerM@uvic.ca Outline Outline 1. How many walks are there from vertices v i to v j
More informationLINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM
LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator
More informationSingular Value Decomposition (SVD)
School of Computing National University of Singapore CS CS524 Theoretical Foundations of Multimedia More Linear Algebra Singular Value Decomposition (SVD) The highpoint of linear algebra Gilbert Strang
More informationLecture 1 and 2: Random Spanning Trees
Recent Advances in Approximation Algorithms Spring 2015 Lecture 1 and 2: Random Spanning Trees Lecturer: Shayan Oveis Gharan March 31st Disclaimer: These notes have not been subjected to the usual scrutiny
More information. The following is a 3 3 orthogonal matrix: 2/3 1/3 2/3 2/3 2/3 1/3 1/3 2/3 2/3
Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. An n n matrix
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationIntroduction to quantum information processing
Introduction to quantum information processing Measurements and quantum probability Brad Lackey 25 October 2016 MEASUREMENTS AND QUANTUM PROBABILITY 1 of 22 OUTLINE 1 Probability 2 Density Operators 3
More informationSpectral Graph Theory
Spectral Graph Theory Aaron Mishtal April 27, 2016 1 / 36 Outline Overview Linear Algebra Primer History Theory Applications Open Problems Homework Problems References 2 / 36 Outline Overview Linear Algebra
More informationI. Multiple Choice Questions (Answer any eight)
Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY
More informationDifferential equations
Differential equations Math 7 Spring Practice problems for April Exam Problem Use the method of elimination to find the x-component of the general solution of x y = 6x 9x + y = x 6y 9y Soln: The system
More information22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices
m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix
More informationSymmetric and anti symmetric matrices
Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal
More informationTHIS paper is an extended version of the paper [1],
INTERNATIONAL JOURNAL OF MATHEMATICAL MODELS AND METHODS IN APPLIED SCIENCES Volume 8 04 Alternating Schur Series and Necessary Conditions for the Existence of Strongly Regular Graphs Vasco Moço Mano and
More informationLecture Notes: Eigenvalues and Eigenvectors. 1 Definitions. 2 Finding All Eigenvalues
Lecture Notes: Eigenvalues and Eigenvectors Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk 1 Definitions Let A be an n n matrix. If there
More informationFiedler s Theorems on Nodal Domains
Spectral Graph Theory Lecture 7 Fiedler s Theorems on Nodal Domains Daniel A Spielman September 9, 202 7 About these notes These notes are not necessarily an accurate representation of what happened in
More informationAN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES
AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim
More informationA Glimpse of Quantum Computation
A Glimpse of Quantum Computation Zhengfeng Ji (UTS:QSI) QCSS 2018, UTS 1. 1 Introduction What is quantum computation? Where does the power come from? Superposition Incompatible states can coexist Transformation
More informationEigenvectors and Reconstruction
Eigenvectors and Reconstruction Hongyu He Department of Mathematics Louisiana State University, Baton Rouge, USA hongyu@mathlsuedu Submitted: Jul 6, 2006; Accepted: Jun 14, 2007; Published: Jul 5, 2007
More informationExercise Set 7.2. Skills
Orthogonally diagonalizable matrix Spectral decomposition (or eigenvalue decomposition) Schur decomposition Subdiagonal Upper Hessenburg form Upper Hessenburg decomposition Skills Be able to recognize
More informationSpan and Linear Independence
Span and Linear Independence It is common to confuse span and linear independence, because although they are different concepts, they are related. To see their relationship, let s revisit the previous
More informationALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA
ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND
More informationComputational math: Assignment 1
Computational math: Assignment 1 Thanks Ting Gao for her Latex file 11 Let B be a 4 4 matrix to which we apply the following operations: 1double column 1, halve row 3, 3add row 3 to row 1, 4interchange
More informationCLASSIFICATION OF COMPLETELY POSITIVE MAPS 1. INTRODUCTION
CLASSIFICATION OF COMPLETELY POSITIVE MAPS STEPHAN HOYER ABSTRACT. We define a completely positive map and classify all completely positive linear maps. We further classify all such maps that are trace-preserving
More informationThe Matrix-Tree Theorem
The Matrix-Tree Theorem Christopher Eur March 22, 2015 Abstract: We give a brief introduction to graph theory in light of linear algebra. Our results culminates in the proof of Matrix-Tree Theorem. 1 Preliminaries
More informationQuantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002
Quantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002 1 QMA - the quantum analog to MA (and NP). Definition 1 QMA. The complexity class QMA is the class of all languages
More informationLecture Note 1: Background
ECE5463: Introduction to Robotics Lecture Note 1: Background Prof. Wei Zhang Department of Electrical and Computer Engineering Ohio State University Columbus, Ohio, USA Spring 2018 Lecture 1 (ECE5463 Sp18)
More informationLinear Algebra. Workbook
Linear Algebra Workbook Paul Yiu Department of Mathematics Florida Atlantic University Last Update: November 21 Student: Fall 2011 Checklist Name: A B C D E F F G H I J 1 2 3 4 5 6 7 8 9 10 xxx xxx xxx
More information8. Diagonalization.
8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard
More informationDef. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as
MAHALANOBIS DISTANCE Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as d E (x, y) = (x 1 y 1 ) 2 + +(x p y p ) 2
More informationThe Spectral Theorem for normal linear maps
MAT067 University of California, Davis Winter 2007 The Spectral Theorem for normal linear maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 14, 2007) In this section we come back to the question
More informationLinGloss. A glossary of linear algebra
LinGloss A glossary of linear algebra Contents: Decompositions Types of Matrices Theorems Other objects? Quasi-triangular A matrix A is quasi-triangular iff it is a triangular matrix except its diagonal
More informationNumerical Linear Algebra Homework Assignment - Week 2
Numerical Linear Algebra Homework Assignment - Week 2 Đoàn Trần Nguyên Tùng Student ID: 1411352 8th October 2016 Exercise 2.1: Show that if a matrix A is both triangular and unitary, then it is diagonal.
More informationMatching Polynomials of Graphs
Spectral Graph Theory Lecture 25 Matching Polynomials of Graphs Daniel A Spielman December 7, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened in class The notes
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationSpectral Graph Theory Lecture 3. Fundamental Graphs. Daniel A. Spielman September 5, 2018
Spectral Graph Theory Lecture 3 Fundamental Graphs Daniel A. Spielman September 5, 2018 3.1 Overview We will bound and derive the eigenvalues of the Laplacian matrices of some fundamental graphs, including
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationNew feasibility conditions for directed strongly regular graphs
New feasibility conditions for directed strongly regular graphs Sylvia A. Hobart Jason Williford Department of Mathematics University of Wyoming Laramie, Wyoming, U.S.A sahobart@uwyo.edu, jwillif1@uwyo.edu
More informationLinear Algebra using Dirac Notation: Pt. 2
Linear Algebra using Dirac Notation: Pt. 2 PHYS 476Q - Southern Illinois University February 6, 2018 PHYS 476Q - Southern Illinois University Linear Algebra using Dirac Notation: Pt. 2 February 6, 2018
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 12 1 / 18 Overview
More informationDot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.
Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................
More informationEigenvalues, random walks and Ramanujan graphs
Eigenvalues, random walks and Ramanujan graphs David Ellis 1 The Expander Mixing lemma We have seen that a bounded-degree graph is a good edge-expander if and only if if has large spectral gap If G = (V,
More informationIBM Research Report. Conditions for Separability in Generalized Laplacian Matrices and Diagonally Dominant Matrices as Density Matrices
RC23758 (W58-118) October 18, 25 Physics IBM Research Report Conditions for Separability in Generalized Laplacian Matrices and Diagonally Dominant Matrices as Density Matrices Chai Wah Wu IBM Research
More informationA = 3 B = A 1 1 matrix is the same as a number or scalar, 3 = [3].
Appendix : A Very Brief Linear ALgebra Review Introduction Linear Algebra, also known as matrix theory, is an important element of all branches of mathematics Very often in this course we study the shapes
More informationLecture 5: Random Walks and Markov Chain
Spectral Graph Theory and Applications WS 20/202 Lecture 5: Random Walks and Markov Chain Lecturer: Thomas Sauerwald & He Sun Introduction to Markov Chains Definition 5.. A sequence of random variables
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More information1 Matrices and vector spaces
Matrices and vector spaces. Which of the following statements about linear vector spaces are true? Where a statement is false, give a counter-example to demonstrate this. (a) Non-singular N N matrices
More informationSymmetric and self-adjoint matrices
Symmetric and self-adjoint matrices A matrix A in M n (F) is called symmetric if A T = A, ie A ij = A ji for each i, j; and self-adjoint if A = A, ie A ij = A ji or each i, j Note for A in M n (R) that
More information2.2. OPERATOR ALGEBRA 19. If S is a subset of E, then the set
2.2. OPERATOR ALGEBRA 19 2.2 Operator Algebra 2.2.1 Algebra of Operators on a Vector Space A linear operator on a vector space E is a mapping L : E E satisfying the condition u, v E, a R, L(u + v) = L(u)
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationChapter 2 Spectra of Finite Graphs
Chapter 2 Spectra of Finite Graphs 2.1 Characteristic Polynomials Let G = (V, E) be a finite graph on n = V vertices. Numbering the vertices, we write down its adjacency matrix in an explicit form of n
More informationVectors Summary. can slide along the line of action. not restricted, defined by magnitude & direction but can be anywhere.
Vectors Summary A vector includes magnitude (size) and direction. Academic Skills Advice Types of vectors: Line vector: Free vector: Position vector: Unit vector (n ): can slide along the line of action.
More informationThe Local Spectra of Regular Line Graphs
The Local Spectra of Regular Line Graphs M. A. Fiol a, M. Mitjana b, a Departament de Matemàtica Aplicada IV Universitat Politècnica de Catalunya Barcelona, Spain b Departament de Matemàtica Aplicada I
More informationTriply Regular Graphs
Triply Regular Graphs by Krystal J. Guo Simon Fraser University Burnaby, British Columbia, Canada, 2011 c Krystal J. Guo 2011 Abstract This project studies a regularity condition on graphs. A graph X
More informationMatrix Solutions to Linear Systems of ODEs
Matrix Solutions to Linear Systems of ODEs James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 3, 216 Outline 1 Symmetric Systems of
More informationHermitian adjacency matrix of digraphs and mixed graphs
Hermitian adjacency matrix of digraphs and mixed graphs arxiv:1505.01321v1 [math.co] 6 May 2015 Krystal Guo Department of Mathematics Simon Fraser University Burnaby, B.C. V5A 1S6 krystalg@sfu.ca May 7,
More informationSimple eigenvalues of graphs and digraphs
Simple eigenvalues of graphs and digraphs by Krystal Guo M.Maths., University of Waterloo, 2010 B.Maths., University of Waterloo, 2009 Dissertation Submitted in Partial Fulfillment of the Requirements
More informationU.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018
U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 Lecture 3 In which we show how to find a planted clique in a random graph. 1 Finding a Planted Clique We will analyze
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationj=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.
Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u
More informationFiedler s Theorems on Nodal Domains
Spectral Graph Theory Lecture 7 Fiedler s Theorems on Nodal Domains Daniel A. Spielman September 19, 2018 7.1 Overview In today s lecture we will justify some of the behavior we observed when using eigenvectors
More informationLecture 3: QR-Factorization
Lecture 3: QR-Factorization This lecture introduces the Gram Schmidt orthonormalization process and the associated QR-factorization of matrices It also outlines some applications of this factorization
More informationData Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings
Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline
More informationCSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming
CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming
More informationPart IA. Vectors and Matrices. Year
Part IA Vectors and Matrices Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2018 Paper 1, Section I 1C Vectors and Matrices For z, w C define the principal value of z w. State de Moivre s
More informationLINEAR ALGEBRA SUMMARY SHEET.
LINEAR ALGEBRA SUMMARY SHEET RADON ROSBOROUGH https://intuitiveexplanationscom/linear-algebra-summary-sheet/ This document is a concise collection of many of the important theorems of linear algebra, organized
More informationMath 408 Advanced Linear Algebra
Math 408 Advanced Linear Algebra Chi-Kwong Li Chapter 4 Hermitian and symmetric matrices Basic properties Theorem Let A M n. The following are equivalent. Remark (a) A is Hermitian, i.e., A = A. (b) x
More information