COMPUTATIONAL METHODS IN MRI: MATHEMATICS
|
|
- Dylan Gervais Booth
- 5 years ago
- Views:
Transcription
1 COMPUTATIONAL METHODS IN MATHEMATICS Imaging Sciences-KCL November 20, 2008
2 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD
3 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD 2 LINEAR SYSTEMS: INVERSE PROBLEMS
4 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD 2 LINEAR SYSTEMS: INVERSE PROBLEMS 3 JACOBIANS
5 MATRICES Matrix: table of numbers; array; used to store data.
6 MATRICES Matrix: table of numbers; array; used to store data. Matrix size is M (rows) times N columns
7 MATRICES Matrix: table of numbers; array; used to store data. Matrix size is M (rows) times N columns If M = 1 or N = 1: vector; used to store coordinates.
8 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ 2.3 ]
9 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ 2.3 ]
10 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis
11 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis 2 [ 2.3 ]
12 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis 2 [ 2.3 ]
13 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ ] b11 b B = 12 b 21 b 22 A
14 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ ] b11 b B = 12 b 21 b 22 A
15 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ ] a11 +b A+B = 11 a 12 +b 12 a 21 +b 21 a 22 +b 22
16 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)
17 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)
18 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)
19 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)
20 LINEAR OPERATIONS: MATRIX PRODUCTS [ ] cos(θ) sin(θ) A = sin(θ) cos(θ) Operation f is linear iff f(αa+βb) = αf(a)+βf(b) Transformation of coordinates by a matrix represents linear transforms: write Ax
21 MATRIX PRODUCTS =
22 MATRIX PRODUCTS P N N P M M =
23 MATRIX PRODUCTS A(m,:) B(:,n) = (AB)(m,n)
24 MATRIX PRODUCTS A mk B kn k A mk B kn =
25 MATRIX PRODUCTS =
26 MATRIX PRODUCTS =
27 MATRIX PRODUCTS = + +
28 LINEAR OPERATIONS: DIMENSIONS :
29 LINEAR OPERATIONS: DIMENSIONS :
30 LINEAR OPERATIONS: DIMENSIONS :
31 LINEAR OPERATIONS: DIMENSIONS?! :
32 LINEAR OPERATIONS: DIMENSIONS
33 LINEAR OPERATIONS: DIMENSIONS Entire line maps to a point. 2 1 [ ]
34 EXAMPLE I: FOURIER TRANSFORM S(k) S(k j ) = n e 2πix nk j s(x n )
35 EXAMPLE I: FOURIER TRANSFORM S(k) S(k j ) = n e 2πix nk j s(x n )
36 EXAMPLE I: FOURIER TRANSFORM S(k) S(k j ) = n F jn s(x n )
37 EXAMPLE II: CONVOLUTION c(x m ) Convolution is another example of linear transform (s c)(x m ) = n c(x m x n )s(x n )
38 EXAMPLE II: CONVOLUTION c(x m ) Convolution is another example of linear transform (s c)(x m ) = n c(x m x n )s(x n )
39 EXAMPLE II: CONVOLUTION c(x m x n ) Convolution is another example of linear transform (s c)(x m ) = n c(x m x n )s(x n ) Matrix is C nm :=c(x m x n ). (Circulant matrix)
40 COMPUTATIONAL ASPECTS Matlab commands: *, kron,.*, +, - BLAS: Level 1 (vec-vec): x αx+y Level 2 (mat-vec): x αax+βy Level 3 (mat-mat): A αab+βc (available as: netlib, GSL, ATLAS, Goto, CUDA, Sunperf, Intel MKL, AMD ACML) Example:
41 BLAS DGEMV SUBROUTINE DGEMV(TRANS,M,N,ALPHA,A,LDA,X,INCX,BETA,Y,INCY) x αax+βy * On entry, TRANS specifies the operation to be performed as * follows: * TRANS = N or n y := alpha*a*x + beta*y. * TRANS = T or t y := alpha*a *x + beta*y. * TRANS = C or c y := alpha*a *x + beta*y. * Unchanged on exit.
42 BLAS DGEMV SUBROUTINE DGEMV(TRANS,M,N,ALPHA,A,LDA,X,INCX,BETA,Y,INCY) x αax+βy * On entry, M specifies the number of rows of the matrix A. * M must be at least zero. * Unchanged on exit. * * N - INTEGER. * On entry, N specifies the number of columns of the matrix A. * N must be at least zero. * Unchanged on exit.
43 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation).
44 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation). problem: compute x such that y = Ax: image reconstruction.
45 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation). problem: compute x such that y = Ax: image reconstruction.
46 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation). problem: compute x such that y = Ax: image reconstruction. Now list of some possible strategies. Possible only if M N.
47 I: DIRECT INVERSION Find B such that BA =I. Then we know that BAx =By. B :=A 1. Example: θ (x +iy)e iθ x +iy [ ] cosθ sinθ R = sinθ cosθ yxr
48 I: DIRECT INVERSION Find B such that BA =I. Then we know that BAx =By. B :=A 1. Example: θ (x +iy)e iθ x +iy [ ] cosθ sinθ R = sinθ cosθ Here D = 2. High dimensional example: F F =I for Fourier. yxr
49 I: DIRECT INVERSION Find B such that BA =I. Then we know that BAx =By. B :=A 1. Example: θ (x +iy)e iθ x +iy [ ] cosθ sinθ R = sinθ cosθ yxr Here D = 2. High dimensional example: F F =I for Fourier. Cost: in general high (mem and speed), unless B =A t (rotation) etc Matlab: inv
50 II: DIAGONALISATION Find U,D =diag(λ) such that U 1 AU =D,A =UDU 1. NB: A 1 =Udiag(1./λ)U 1 ex = exp(-(x.*x + Y.*Y)/128; ; Cost: in general very high, unless we can find U 1 =U t Matlab: diag.
51 II: DIAGONALISATION Find U,D =diag(λ) such that U 1 AU =D,A =UDU 1. NB: A 1 =Udiag(1./λ)U 1 ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); ; Cost: in general very high, unless we can find U 1 =U t Matlab: diag.
52 II: DIAGONALISATION Find U,D =diag(λ) such that U 1 AU =D,A =UDU 1. NB: A 1 =Udiag(1./λ)U 1 ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex)); Cost: in general very high, unless we can find U 1 =U t Matlab: diag.
53 III: TRIANGULAR DECOMPOSITION: LU Find L,U such that A =LU Solve Ax =y by solving Lz =y and Ux =z 0 0 x = y
54 III: TRIANGULAR DECOMPOSITION: LU Find L,U such that A =LU Solve Ax =y by solving Lz =y and Ux =z 0 0 x = y Matlab default \ (lu) Cost moderate. (Default for M such that M
55 IV: ITERATIVE METHODS Generate a sequence of x k which converge to x.
56 IV: ITERATIVE METHODS Gradient descent: Ellipses are level lines (iso-contours) for x t Ax x t y
57 IV: ITERATIVE METHODS Gradient descent: Ellipses are level lines (iso-contours) for x t Ax x t y
58 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A
59 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A
60 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A
61 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A low
62 BENCHMARK Test your own machinebench in Matlab. NB: LINPACK test used forhttp://
63 BENCHMARK Test your own machinebench in Matlab. NB: LINPACK test used 80 0 forhttp:// for k=100:100:10000; A=randn(k,k); b = randn(k,1); tic; A\b; t(k) = toc; end
64 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
65 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
66 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/064; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
67 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/064; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
68 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/032; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
69 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/032; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
70 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/032; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));
71 SVD Problem with deconvolution: division by zero! SVD generalises this statement to all matrices! A =USV t where U,V are orthogonal, S rectangular-diagonal. Condition number is ratio of non-zero diagonal elements. 1 is good, quality of numerical solutions degrades with condition going up. 0
72 NUMBER OF EQS. NUMBER OF UNKNOWNS If matrices are not square, they can t be inverted. If M >N (more eqs than unknown). In general no exact solution. It makes sense to pick the closest: ˆx == argmin (Ax) n y n 2 =: argmin Ax y 2 2. If M <N, infinity of solutions. We need a way to pick one. Usually: pick x which minimises some condition.
73 V: PSEUDO-INVERSE (M >N) Equivalent to min Ax y normal equation A t Ax =A t y Pseudo-inverse (A t A) 1 A t y y
74 V: PSEUDO-INVERSE (M >N) Equivalent to min Ax y normal equation A t Ax =A t y Pseudo-inverse (A t A) 1 A t y y
75 V: PSEUDO-INVERSE (M >N) A\y Equivalent to min Ax y normal equation A t Ax =A t y Pseudo-inverse (A t A) 1 A t y y
76 EXAMPLE: SENSE y = subsample(fx)
77 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax
78 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y
79 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y x 3 :Ax 3 =y
80 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y x 3 :Ax 3 =y x 4 :Ax 4 =y
81 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y A(tx 0 +(1 t)x) x 3 :Ax 3 =y x 4 :Ax 4 =y
82 UNDERSAMPLING Undersampling in k-space means aliasing. For example, factor 2: s(y) = 1 (s(x)+s(x +F/2)). 2 Write x 1 :=x x 2 :=x+f/2.
83 UNDERSAMPLING Undersampling in k-space means aliasing. For example, factor 2: s(y) = 1 (s(x)+s(x +F/2)). 2 Write x 1 :=x x 2 :=x+f/2. In other words, we have 2 unknowns, but only 1 equation.
84 Parallel MR: multiple coils measure weighted signals: s i (x) =c i (x)s(x).
85 SENSE Combined with undersampling: s 1 = 1 2 ( s(x 1) + s(x 2 )) s 2 = 1 2 ( s(x 1) + s(x 2 ))
86 SENSE Combined with undersampling: s 1 = 1 2 (c 1(x 1 )s(x 1 ) +c 1 (x 2 )s(x 2 )) s 2 = 1 2 (c 2(x 1 )s(x 1 ) +c 2 (x 2 )s(x 2 )) We can solve for s(x 1 ),s(x 2 )
87 SENSE Combined with undersampling: s 1 = 1 2 (c 1(x 1 )s(x 1 ) +c 1 (x 2 )s(x 2 )) s 2 = 1 2 (c 2(x 1 )s(x 1 ) +c 2 (x 2 )s(x 2 )) We can solve for s(x 1 ),s(x 2 )if the matrix with c nm =c n (x m ) is well conditi oned
88 NORMS x 2 Euclidean norm, L 2, SOS, relevant in physical problems (energy), etc x 1 sum of absolute values, useful for robust estimation in statistics x p := ( n x n p ) 1/p x 0 number of nonzeros x =max x n
89 NORMS x 2 Euclidean norm, L 2, SOS, relevant in physical problems (energy), etc x 1 sum of absolute values, useful for robust estimation in statistics x p := ( n x n p ) 1/p x 0 number of nonzeros x =max x n P-norms unit circles for p=0.5,1,1.5,2,2.5 (from inside to outside)
90 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x
91 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x Minimum norm min x given Ax =y (see below).
92 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x Minimum norm min x given Ax =y (see below). SVD truncation: pinv(a) =Vdiag(1./σ i )U t.
93 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x Minimum norm min x given Ax =y (see below). SVD truncation: pinv(a) =Vdiag(1./σ i )U t. Prior information: e.g. min Ax y +α Bx. B 1 : object autocorrelation. Tikhonov: Vdiag(σ i./(σ 2 i + α 2 ))U t
94 EXAMPLE: DECONVOLUTION Blurry - Noisy deconvwnr: using SNR with prior info (A t A+αB t B) 1 A t y
95 NORMAL EQUATION-ITERATIVE METHODS For M >N: min Ax y 2 implies A Ax =A y the normal equation CGNE: Apply CG to A Ax =A y: LSQR Matlab: lsqr
96 LAPACK operations: BLAS library, inverse (and others): LAPACK Example SUBROUTINE SGGLSE( M, N, P, A, LDA, B, LDB, C, D, X, WORK, LWORK,INFO ) * Purpose * ======= * SGGLSE solves the linear equality-constrained least squares (LSE) * problem: * minimize c - A*x 2 subject to B*x = d * where A is an M-by-N matrix, B is a P-by-N matrix, c is a given * M-vector, and d is a given P-vector. It is assumed that * P <= N <= M+P, and rank(b) = P and rank( (A) ) = N.
97 COMPRESSED SENSING Total undersampling: M << N L 2. Solution A t (AA t ) 1 y. min x 2 such that Ax =y
98 COMPRESSED SENSING Total undersampling: M << N L 2. Solution A t (AA t ) 1 y. min x 2 such that Ax =y
99 COMPRESSED SENSING Total undersampling: M << N L 0. Sparsest solution: A\y. min x 0 such that Ax =y
100 COMPRESSED SENSING Total undersampling: M << N L 0. Sparsest solution: A\y. min Ex 0 such that Ax =y
101 COMPRESSED SENSING Total undersampling: M << N min x 1 such that Ax =y L 1. Under some conditions, equivalent to L 0. Acquisition-reconstruction is Compressed Sensing OMP algorithms: remnove strongest alias(es), re-project, iterate.
102 SUMMARY SENSE: deconvolution with undersampling compensated by overdetermination. Deconvolution: regularised. Undersampling: prior assumption.
103 DETERMINANT Determinant of a matrix: product of eigenvalues
104 DETERMINANT Determinant of a matrix: product of eigenvalues Area /Volume of shape generated by columns
105 DETERMINANT Determinant of a matrix: product of eigenvalues Area /Volume of shape generated by columns det=0 iff matrix is singular
106 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration)
107 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration) Multivariate derivative at x: best linear (affine) transform approximating Φ at x. Denoted DΦ(x) dφ(x)/dx. Thus in dimension D it is a D D matrix.
108 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration) Multivariate derivative at x: best linear (affine) transform approximating Φ at x. Denoted DΦ(x) dφ(x)/dx. Thus in dimension D it is a D D matrix. Imaging: integration or sums. Change of coordinates implies need change of are/volume element.
109 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration) Multivariate derivative at x: best linear (affine) transform approximating Φ at x. Denoted DΦ(x) dφ(x)/dx. Thus in dimension D it is a D D matrix. Imaging: integration or sums. Change of coordinates implies need change of are/volume element. Jacobian determinant: determinant of this matrix
110 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k)
111 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k) : R s(ax)e 2πiktx dx = R s(y)e 2πikt (A 1 y) dx/dydy = R s(y)e 2πiA t k t (y) 1 det(a) dy = 1 F det(a) s(a t (k))
112 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k) : R s(ax)e 2πiktx dx = R s(y)e 2πikt (A 1 y) dx/dydy = R s(y)e 2πiA t k t (y) 1 det(a) dy = 1 F det(a) s(a t (k)) Rotation: special case where det(a) = 1
113 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k) : R s(ax)e 2πiktx dx = R s(y)e 2πikt (A 1 y) dx/dydy = R s(y)e 2πiA t k t (y) 1 det(a) dy = 1 F det(a) s(a t (k)) Rotation: special case where det(a) = 1 Affine: linear with translation.
Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??
Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationStatistical Geometry Processing Winter Semester 2011/2012
Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian
More information2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian
FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More information6 Linear Systems of Equations
6 Linear Systems of Equations Read sections 2.1 2.3, 2.4.1 2.4.5, 2.4.7, 2.7 Review questions 2.1 2.37, 2.43 2.67 6.1 Introduction When numerically solving two-point boundary value problems, the differential
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationGeometric Modeling Summer Semester 2010 Mathematical Tools (1)
Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationLinear Algebra for Machine Learning. Sargur N. Srihari
Linear Algebra for Machine Learning Sargur N. srihari@cedar.buffalo.edu 1 Overview Linear Algebra is based on continuous math rather than discrete math Computer scientists have little experience with it
More informationSingular Value Decomposition
Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =
More informationTHE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR
THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},
More informationMobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti
Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes
More informationPreconditioning. Noisy, Ill-Conditioned Linear Systems
Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationTBP MATH33A Review Sheet. November 24, 2018
TBP MATH33A Review Sheet November 24, 2018 General Transformation Matrices: Function Scaling by k Orthogonal projection onto line L Implementation If we want to scale I 2 by k, we use the following: [
More informationGeometric Modeling Summer Semester 2012 Linear Algebra & Function Spaces
Geometric Modeling Summer Semester 2012 Linear Algebra & Function Spaces (Recap) Announcement Room change: On Thursday, April 26th, room 024 is occupied. The lecture will be moved to room 021, E1 4 (the
More informationLinear Algebra - Part II
Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T
More informationMultivariable Calculus
2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationApplied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization
More informationPreconditioning. Noisy, Ill-Conditioned Linear Systems
Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationSpring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier
Spring 0 CIS 55 Fundamentals of Linear Algebra and Optimization Jean Gallier Homework 5 & 6 + Project 3 & 4 Note: Problems B and B6 are for extra credit April 7 0; Due May 7 0 Problem B (0 pts) Let A be
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationMatrices A brief introduction
Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 44 Definitions Definition A matrix is a set of N real or complex
More informationSingular Value Decomposition
Singular Value Decomposition (Com S 477/577 Notes Yan-Bin Jia Sep, 7 Introduction Now comes a highlight of linear algebra. Any real m n matrix can be factored as A = UΣV T where U is an m m orthogonal
More informationReview Questions REVIEW QUESTIONS 71
REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of
More informationReview of some mathematical tools
MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationReview of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationMatrices and systems of linear equations
Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationLecture 6. Numerical methods. Approximation of functions
Lecture 6 Numerical methods Approximation of functions Lecture 6 OUTLINE 1. Approximation and interpolation 2. Least-square method basis functions design matrix residual weighted least squares normal equation
More informationLeast-Squares Fitting of Model Parameters to Experimental Data
Least-Squares Fitting of Model Parameters to Experimental Data Div. of Mathematical Sciences, Dept of Engineering Sciences and Mathematics, LTU, room E193 Outline of talk What about Science and Scientific
More informationLecture Notes in Linear Algebra
Lecture Notes in Linear Algebra Dr. Abdullah Al-Azemi Mathematics Department Kuwait University February 4, 2017 Contents 1 Linear Equations and Matrices 1 1.2 Matrices............................................
More informationLecture notes on Quantum Computing. Chapter 1 Mathematical Background
Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For
More informationIntroduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello
Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Vectors Arrays of numbers Vectors represent a point
More informationMath Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p
Math Bootcamp 2012 1 Review of matrix algebra 1.1 Vectors and rules of operations An p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the
More informationA Quick Tour of Linear Algebra and Optimization for Machine Learning
A Quick Tour of Linear Algebra and Optimization for Machine Learning Masoud Farivar January 8, 2015 1 / 28 Outline of Part I: Review of Basic Linear Algebra Matrices and Vectors Matrix Multiplication Operators
More informationLinear Algebra and Matrices
Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.
More informationFundamentals of Engineering Analysis (650163)
Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is
More informationIntroduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz
Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Vectors Arrays of numbers Vectors represent a point in a n dimensional space
More informationYORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions
YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 222 3. M Test # July, 23 Solutions. For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For
More informationOR MSc Maths Revision Course
OR MSc Maths Revision Course Tom Byrne School of Mathematics University of Edinburgh t.m.byrne@sms.ed.ac.uk 15 September 2017 General Information Today JCMB Lecture Theatre A, 09:30-12:30 Mathematics revision
More informationScientific Computing: Dense Linear Systems
Scientific Computing: Dense Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 February 9th, 2012 A. Donev (Courant Institute)
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More information(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =
. (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)
More informationMatrices A brief introduction
Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino September 2013 Basilio Bona (DAUIN) Matrices September 2013 1 / 74 Definitions Definition A matrix is a set of N real or complex numbers
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More informationEE263 homework 3 solutions
EE263 Prof. S. Boyd EE263 homework 3 solutions 2.17 Gradient of some common functions. Recall that the gradient of a differentiable function f : R n R, at a point x R n, is defined as the vector f(x) =
More informationComputer Assignment 2
Computer Assignment 2 Chapter 2, section, problem 3 Apes = [7 5;3 6] Diet = [25 4 2;5 3 25] Result = Apes * Diet Apes = 7 5 3 6 Diet = 25 4 2 5 3 25 Result = 25 43 265 65 3 2 The meaning of the calculation:
More informationEIGENVALUES AND SINGULAR VALUE DECOMPOSITION
APPENDIX B EIGENVALUES AND SINGULAR VALUE DECOMPOSITION B.1 LINEAR EQUATIONS AND INVERSES Problems of linear estimation can be written in terms of a linear matrix equation whose solution provides the required
More information5 Linear Algebra and Inverse Problem
5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More informationJordan Normal Form and Singular Decomposition
University of Debrecen Diagonalization and eigenvalues Diagonalization We have seen that if A is an n n square matrix, then A is diagonalizable if and only if for all λ eigenvalues of A we have dim(u λ
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More informationMath 515 Fall, 2008 Homework 2, due Friday, September 26.
Math 515 Fall, 2008 Homework 2, due Friday, September 26 In this assignment you will write efficient MATLAB codes to solve least squares problems involving block structured matrices known as Kronecker
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More informationProblem # Max points possible Actual score Total 120
FINAL EXAMINATION - MATH 2121, FALL 2017. Name: ID#: Email: Lecture & Tutorial: Problem # Max points possible Actual score 1 15 2 15 3 10 4 15 5 15 6 15 7 10 8 10 9 15 Total 120 You have 180 minutes to
More informationIntroduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Bastian Steder
Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Bastian Steder Reference Book Thrun, Burgard, and Fox: Probabilistic Robotics Vectors Arrays of numbers Vectors represent
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More information11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.
C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a
More informationInverses. Stephen Boyd. EE103 Stanford University. October 28, 2017
Inverses Stephen Boyd EE103 Stanford University October 28, 2017 Outline Left and right inverses Inverse Solving linear equations Examples Pseudo-inverse Left and right inverses 2 Left inverses a number
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,
More information5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors
EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationMATH.2720 Introduction to Programming with MATLAB Vector and Matrix Algebra
MATH.2720 Introduction to Programming with MATLAB Vector and Matrix Algebra A. Vectors A vector is a quantity that has both magnitude and direction, like velocity. The location of a vector is irrelevant;
More informationPose estimation from point and line correspondences
Pose estimation from point and line correspondences Giorgio Panin October 17, 008 1 Problem formulation Estimate (in a LSE sense) the pose of an object from N correspondences between known object points
More informationEvaluation and Benchmarking of Highly Scalable Parallel Numerical Libraries
Evaluation and Benchmarking of Highly Scalable Parallel Numerical Libraries Christos Theodosiou (ctheodos@grid.auth.gr) User and Application Support Scientific Computing Centre @ AUTH Presentation Outline
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationComputational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science
Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Lecture 5: Numerical Linear Algebra Cho-Jui Hsieh UC Davis April 20, 2017 Linear Algebra Background Vectors A vector has a direction and a magnitude
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationSingular value decomposition
Singular value decomposition The eigenvalue decomposition (EVD) for a square matrix A gives AU = UD. Let A be rectangular (m n, m > n). A singular value σ and corresponding pair of singular vectors u (m
More informationMath Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT
Math Camp II Basic Linear Algebra Yiqing Xu MIT Aug 26, 2014 1 Solving Systems of Linear Equations 2 Vectors and Vector Spaces 3 Matrices 4 Least Squares Systems of Linear Equations Definition A linear
More informationPractical Linear Algebra: A Geometry Toolbox
Practical Linear Algebra: A Geometry Toolbox Third edition Chapter 12: Gauss for Linear Systems Gerald Farin & Dianne Hansford CRC Press, Taylor & Francis Group, An A K Peters Book www.farinhansford.com/books/pla
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationA VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 2010
A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 00 Introduction Linear Algebra, also known as matrix theory, is an important element of all branches of mathematics
More informationMATH 106 LINEAR ALGEBRA LECTURE NOTES
MATH 6 LINEAR ALGEBRA LECTURE NOTES FALL - These Lecture Notes are not in a final form being still subject of improvement Contents Systems of linear equations and matrices 5 Introduction to systems of
More informationThe Singular Value Decomposition and Least Squares Problems
The Singular Value Decomposition and Least Squares Problems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 27, 2009 Applications of SVD solving
More informationVector Spaces, Orthogonality, and Linear Least Squares
Week Vector Spaces, Orthogonality, and Linear Least Squares. Opening Remarks.. Visualizing Planes, Lines, and Solutions Consider the following system of linear equations from the opener for Week 9: χ χ
More informationEigenvectors and SVD 1
Eigenvectors and SVD 1 Definition Eigenvectors of a square matrix Ax=λx, x=0. Intuition: x is unchanged by A (except for scaling) Examples: axis of rotation, stationary distribution of a Markov chain 2
More informationMath for ML: review. ML and knowledge of other fields
ath for L: review ilos Hauskrecht milos@cs.pitt.edu Sennott Square x- people.cs.pitt.edu/~milos/ L and knowledge of other fields L solutions and algorithms rely on knowledge of many other disciplines:
More informationLinear Inverse Problems
Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationVector and Matrix Norms. Vector and Matrix Norms
Vector and Matrix Norms Vector Space Algebra Matrix Algebra: We let x x and A A, where, if x is an element of an abstract vector space n, and A = A: n m, then x is a complex column vector of length n whose
More informationMTH 464: Computational Linear Algebra
MTH 464: Computational Linear Algebra Lecture Outlines Exam 2 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University February 6, 2018 Linear Algebra (MTH
More informationMATH2210 Notebook 2 Spring 2018
MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................
More informationCSL361 Problem set 4: Basic linear algebra
CSL361 Problem set 4: Basic linear algebra February 21, 2017 [Note:] If the numerical matrix computations turn out to be tedious, you may use the function rref in Matlab. 1 Row-reduced echelon matrices
More information