COMPUTATIONAL METHODS IN MRI: MATHEMATICS

Size: px
Start display at page:

Download "COMPUTATIONAL METHODS IN MRI: MATHEMATICS"

Transcription

1 COMPUTATIONAL METHODS IN MATHEMATICS Imaging Sciences-KCL November 20, 2008

2 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD

3 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD 2 LINEAR SYSTEMS: INVERSE PROBLEMS

4 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD 2 LINEAR SYSTEMS: INVERSE PROBLEMS 3 JACOBIANS

5 MATRICES Matrix: table of numbers; array; used to store data.

6 MATRICES Matrix: table of numbers; array; used to store data. Matrix size is M (rows) times N columns

7 MATRICES Matrix: table of numbers; array; used to store data. Matrix size is M (rows) times N columns If M = 1 or N = 1: vector; used to store coordinates.

8 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ 2.3 ]

9 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ 2.3 ]

10 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis

11 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis 2 [ 2.3 ]

12 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis 2 [ 2.3 ]

13 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ ] b11 b B = 12 b 21 b 22 A

14 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ ] b11 b B = 12 b 21 b 22 A

15 MATRIX OPERATIONS: SUM AND SCALAR PRODUCTS geometric basis [ ] a11 +b A+B = 11 a 12 +b 12 a 21 +b 21 a 22 +b 22

16 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)

17 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)

18 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)

19 LINEAR OPERATIONS: MATRIX PRODUCTS Operation f is linear iff f(αa+βb) = αf(a)+βf(b)

20 LINEAR OPERATIONS: MATRIX PRODUCTS [ ] cos(θ) sin(θ) A = sin(θ) cos(θ) Operation f is linear iff f(αa+βb) = αf(a)+βf(b) Transformation of coordinates by a matrix represents linear transforms: write Ax

21 MATRIX PRODUCTS =

22 MATRIX PRODUCTS P N N P M M =

23 MATRIX PRODUCTS A(m,:) B(:,n) = (AB)(m,n)

24 MATRIX PRODUCTS A mk B kn k A mk B kn =

25 MATRIX PRODUCTS =

26 MATRIX PRODUCTS =

27 MATRIX PRODUCTS = + +

28 LINEAR OPERATIONS: DIMENSIONS :

29 LINEAR OPERATIONS: DIMENSIONS :

30 LINEAR OPERATIONS: DIMENSIONS :

31 LINEAR OPERATIONS: DIMENSIONS?! :

32 LINEAR OPERATIONS: DIMENSIONS

33 LINEAR OPERATIONS: DIMENSIONS Entire line maps to a point. 2 1 [ ]

34 EXAMPLE I: FOURIER TRANSFORM S(k) S(k j ) = n e 2πix nk j s(x n )

35 EXAMPLE I: FOURIER TRANSFORM S(k) S(k j ) = n e 2πix nk j s(x n )

36 EXAMPLE I: FOURIER TRANSFORM S(k) S(k j ) = n F jn s(x n )

37 EXAMPLE II: CONVOLUTION c(x m ) Convolution is another example of linear transform (s c)(x m ) = n c(x m x n )s(x n )

38 EXAMPLE II: CONVOLUTION c(x m ) Convolution is another example of linear transform (s c)(x m ) = n c(x m x n )s(x n )

39 EXAMPLE II: CONVOLUTION c(x m x n ) Convolution is another example of linear transform (s c)(x m ) = n c(x m x n )s(x n ) Matrix is C nm :=c(x m x n ). (Circulant matrix)

40 COMPUTATIONAL ASPECTS Matlab commands: *, kron,.*, +, - BLAS: Level 1 (vec-vec): x αx+y Level 2 (mat-vec): x αax+βy Level 3 (mat-mat): A αab+βc (available as: netlib, GSL, ATLAS, Goto, CUDA, Sunperf, Intel MKL, AMD ACML) Example:

41 BLAS DGEMV SUBROUTINE DGEMV(TRANS,M,N,ALPHA,A,LDA,X,INCX,BETA,Y,INCY) x αax+βy * On entry, TRANS specifies the operation to be performed as * follows: * TRANS = N or n y := alpha*a*x + beta*y. * TRANS = T or t y := alpha*a *x + beta*y. * TRANS = C or c y := alpha*a *x + beta*y. * Unchanged on exit.

42 BLAS DGEMV SUBROUTINE DGEMV(TRANS,M,N,ALPHA,A,LDA,X,INCX,BETA,Y,INCY) x αax+βy * On entry, M specifies the number of rows of the matrix A. * M must be at least zero. * Unchanged on exit. * * N - INTEGER. * On entry, N specifies the number of columns of the matrix A. * N must be at least zero. * Unchanged on exit.

43 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation).

44 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation). problem: compute x such that y = Ax: image reconstruction.

45 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation). problem: compute x such that y = Ax: image reconstruction.

46 INVERSE PROBLEMS problem: compute y = Ax given A, x. Physically: image acquisition (or simulation). problem: compute x such that y = Ax: image reconstruction. Now list of some possible strategies. Possible only if M N.

47 I: DIRECT INVERSION Find B such that BA =I. Then we know that BAx =By. B :=A 1. Example: θ (x +iy)e iθ x +iy [ ] cosθ sinθ R = sinθ cosθ yxr

48 I: DIRECT INVERSION Find B such that BA =I. Then we know that BAx =By. B :=A 1. Example: θ (x +iy)e iθ x +iy [ ] cosθ sinθ R = sinθ cosθ Here D = 2. High dimensional example: F F =I for Fourier. yxr

49 I: DIRECT INVERSION Find B such that BA =I. Then we know that BAx =By. B :=A 1. Example: θ (x +iy)e iθ x +iy [ ] cosθ sinθ R = sinθ cosθ yxr Here D = 2. High dimensional example: F F =I for Fourier. Cost: in general high (mem and speed), unless B =A t (rotation) etc Matlab: inv

50 II: DIAGONALISATION Find U,D =diag(λ) such that U 1 AU =D,A =UDU 1. NB: A 1 =Udiag(1./λ)U 1 ex = exp(-(x.*x + Y.*Y)/128; ; Cost: in general very high, unless we can find U 1 =U t Matlab: diag.

51 II: DIAGONALISATION Find U,D =diag(λ) such that U 1 AU =D,A =UDU 1. NB: A 1 =Udiag(1./λ)U 1 ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); ; Cost: in general very high, unless we can find U 1 =U t Matlab: diag.

52 II: DIAGONALISATION Find U,D =diag(λ) such that U 1 AU =D,A =UDU 1. NB: A 1 =Udiag(1./λ)U 1 ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex)); Cost: in general very high, unless we can find U 1 =U t Matlab: diag.

53 III: TRIANGULAR DECOMPOSITION: LU Find L,U such that A =LU Solve Ax =y by solving Lz =y and Ux =z 0 0 x = y

54 III: TRIANGULAR DECOMPOSITION: LU Find L,U such that A =LU Solve Ax =y by solving Lz =y and Ux =z 0 0 x = y Matlab default \ (lu) Cost moderate. (Default for M such that M

55 IV: ITERATIVE METHODS Generate a sequence of x k which converge to x.

56 IV: ITERATIVE METHODS Gradient descent: Ellipses are level lines (iso-contours) for x t Ax x t y

57 IV: ITERATIVE METHODS Gradient descent: Ellipses are level lines (iso-contours) for x t Ax x t y

58 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A

59 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A

60 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A

61 IV: ITERATIVE METHODS Conjugate gradient: correct for distortion by A low

62 BENCHMARK Test your own machinebench in Matlab. NB: LINPACK test used forhttp://

63 BENCHMARK Test your own machinebench in Matlab. NB: LINPACK test used 80 0 forhttp:// for k=100:100:10000; A=randn(k,k); b = randn(k,1); tic; A\b; t(k) = toc; end

64 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

65 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/128; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

66 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/064; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

67 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/064; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

68 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/032; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

69 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/032; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

70 NON-REGULAR PROBLEMS Deconvolution: Operator C diagonalised by FT: F CF =diag(c) But this usually implies division by zero (or very small number) (coil sensitivities are zero in some places) Makes the problem ill-posed. ex = exp(-(x.*x + Y.*Y)/032; cr = ifft2(fft2(r).*fft2(ex)); dcr = ifft2(fft2(cr)./fft2(ex));

71 SVD Problem with deconvolution: division by zero! SVD generalises this statement to all matrices! A =USV t where U,V are orthogonal, S rectangular-diagonal. Condition number is ratio of non-zero diagonal elements. 1 is good, quality of numerical solutions degrades with condition going up. 0

72 NUMBER OF EQS. NUMBER OF UNKNOWNS If matrices are not square, they can t be inverted. If M >N (more eqs than unknown). In general no exact solution. It makes sense to pick the closest: ˆx == argmin (Ax) n y n 2 =: argmin Ax y 2 2. If M <N, infinity of solutions. We need a way to pick one. Usually: pick x which minimises some condition.

73 V: PSEUDO-INVERSE (M >N) Equivalent to min Ax y normal equation A t Ax =A t y Pseudo-inverse (A t A) 1 A t y y

74 V: PSEUDO-INVERSE (M >N) Equivalent to min Ax y normal equation A t Ax =A t y Pseudo-inverse (A t A) 1 A t y y

75 V: PSEUDO-INVERSE (M >N) A\y Equivalent to min Ax y normal equation A t Ax =A t y Pseudo-inverse (A t A) 1 A t y y

76 EXAMPLE: SENSE y = subsample(fx)

77 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax

78 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y

79 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y x 3 :Ax 3 =y

80 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y x 3 :Ax 3 =y x 4 :Ax 4 =y

81 EXAMPLE: SENSE x 1 :Ax 1 =y y =Ax x 2 :Ax 2 =y A(tx 0 +(1 t)x) x 3 :Ax 3 =y x 4 :Ax 4 =y

82 UNDERSAMPLING Undersampling in k-space means aliasing. For example, factor 2: s(y) = 1 (s(x)+s(x +F/2)). 2 Write x 1 :=x x 2 :=x+f/2.

83 UNDERSAMPLING Undersampling in k-space means aliasing. For example, factor 2: s(y) = 1 (s(x)+s(x +F/2)). 2 Write x 1 :=x x 2 :=x+f/2. In other words, we have 2 unknowns, but only 1 equation.

84 Parallel MR: multiple coils measure weighted signals: s i (x) =c i (x)s(x).

85 SENSE Combined with undersampling: s 1 = 1 2 ( s(x 1) + s(x 2 )) s 2 = 1 2 ( s(x 1) + s(x 2 ))

86 SENSE Combined with undersampling: s 1 = 1 2 (c 1(x 1 )s(x 1 ) +c 1 (x 2 )s(x 2 )) s 2 = 1 2 (c 2(x 1 )s(x 1 ) +c 2 (x 2 )s(x 2 )) We can solve for s(x 1 ),s(x 2 )

87 SENSE Combined with undersampling: s 1 = 1 2 (c 1(x 1 )s(x 1 ) +c 1 (x 2 )s(x 2 )) s 2 = 1 2 (c 2(x 1 )s(x 1 ) +c 2 (x 2 )s(x 2 )) We can solve for s(x 1 ),s(x 2 )if the matrix with c nm =c n (x m ) is well conditi oned

88 NORMS x 2 Euclidean norm, L 2, SOS, relevant in physical problems (energy), etc x 1 sum of absolute values, useful for robust estimation in statistics x p := ( n x n p ) 1/p x 0 number of nonzeros x =max x n

89 NORMS x 2 Euclidean norm, L 2, SOS, relevant in physical problems (energy), etc x 1 sum of absolute values, useful for robust estimation in statistics x p := ( n x n p ) 1/p x 0 number of nonzeros x =max x n P-norms unit circles for p=0.5,1,1.5,2,2.5 (from inside to outside)

90 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x

91 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x Minimum norm min x given Ax =y (see below).

92 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x Minimum norm min x given Ax =y (see below). SVD truncation: pinv(a) =Vdiag(1./σ i )U t.

93 REGULARISATION Problem: small (or even 0!) singular values. Regularisation:= do something about small singular values-choose a solution among the many possibles Tikhonov: min Ax y +α x Minimum norm min x given Ax =y (see below). SVD truncation: pinv(a) =Vdiag(1./σ i )U t. Prior information: e.g. min Ax y +α Bx. B 1 : object autocorrelation. Tikhonov: Vdiag(σ i./(σ 2 i + α 2 ))U t

94 EXAMPLE: DECONVOLUTION Blurry - Noisy deconvwnr: using SNR with prior info (A t A+αB t B) 1 A t y

95 NORMAL EQUATION-ITERATIVE METHODS For M >N: min Ax y 2 implies A Ax =A y the normal equation CGNE: Apply CG to A Ax =A y: LSQR Matlab: lsqr

96 LAPACK operations: BLAS library, inverse (and others): LAPACK Example SUBROUTINE SGGLSE( M, N, P, A, LDA, B, LDB, C, D, X, WORK, LWORK,INFO ) * Purpose * ======= * SGGLSE solves the linear equality-constrained least squares (LSE) * problem: * minimize c - A*x 2 subject to B*x = d * where A is an M-by-N matrix, B is a P-by-N matrix, c is a given * M-vector, and d is a given P-vector. It is assumed that * P <= N <= M+P, and rank(b) = P and rank( (A) ) = N.

97 COMPRESSED SENSING Total undersampling: M << N L 2. Solution A t (AA t ) 1 y. min x 2 such that Ax =y

98 COMPRESSED SENSING Total undersampling: M << N L 2. Solution A t (AA t ) 1 y. min x 2 such that Ax =y

99 COMPRESSED SENSING Total undersampling: M << N L 0. Sparsest solution: A\y. min x 0 such that Ax =y

100 COMPRESSED SENSING Total undersampling: M << N L 0. Sparsest solution: A\y. min Ex 0 such that Ax =y

101 COMPRESSED SENSING Total undersampling: M << N min x 1 such that Ax =y L 1. Under some conditions, equivalent to L 0. Acquisition-reconstruction is Compressed Sensing OMP algorithms: remnove strongest alias(es), re-project, iterate.

102 SUMMARY SENSE: deconvolution with undersampling compensated by overdetermination. Deconvolution: regularised. Undersampling: prior assumption.

103 DETERMINANT Determinant of a matrix: product of eigenvalues

104 DETERMINANT Determinant of a matrix: product of eigenvalues Area /Volume of shape generated by columns

105 DETERMINANT Determinant of a matrix: product of eigenvalues Area /Volume of shape generated by columns det=0 iff matrix is singular

106 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration)

107 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration) Multivariate derivative at x: best linear (affine) transform approximating Φ at x. Denoted DΦ(x) dφ(x)/dx. Thus in dimension D it is a D D matrix.

108 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration) Multivariate derivative at x: best linear (affine) transform approximating Φ at x. Denoted DΦ(x) dφ(x)/dx. Thus in dimension D it is a D D matrix. Imaging: integration or sums. Change of coordinates implies need change of are/volume element.

109 CHANGE OF COORDINATES Non-rigid transform: Φ(x) (e.e. in registration) Multivariate derivative at x: best linear (affine) transform approximating Φ at x. Denoted DΦ(x) dφ(x)/dx. Thus in dimension D it is a D D matrix. Imaging: integration or sums. Change of coordinates implies need change of are/volume element. Jacobian determinant: determinant of this matrix

110 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k)

111 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k) : R s(ax)e 2πiktx dx = R s(y)e 2πikt (A 1 y) dx/dydy = R s(y)e 2πiA t k t (y) 1 det(a) dy = 1 F det(a) s(a t (k))

112 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k) : R s(ax)e 2πiktx dx = R s(y)e 2πikt (A 1 y) dx/dydy = R s(y)e 2πiA t k t (y) 1 det(a) dy = 1 F det(a) s(a t (k)) Rotation: special case where det(a) = 1

113 INTEGRATION R s(x)dx = R s(y)(dx/dy)dy valid in any dimension. Thus it involves the Jacobian determinant. Can be used to derive formulas for behaviour of Fourier Transform f Ff Translation R s(x+u)e 2πik tx dx = R s(y)e 2πikt (y u) dx =e 2πikt u F s(k) : R s(ax)e 2πiktx dx = R s(y)e 2πikt (A 1 y) dx/dydy = R s(y)e 2πiA t k t (y) 1 det(a) dy = 1 F det(a) s(a t (k)) Rotation: special case where det(a) = 1 Affine: linear with translation.

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Numerical Methods. Elena loli Piccolomini. Civil Engeneering.  piccolom. Metodi Numerici M p. 1/?? Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

Statistical Geometry Processing Winter Semester 2011/2012

Statistical Geometry Processing Winter Semester 2011/2012 Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

6 Linear Systems of Equations

6 Linear Systems of Equations 6 Linear Systems of Equations Read sections 2.1 2.3, 2.4.1 2.4.5, 2.4.7, 2.7 Review questions 2.1 2.37, 2.43 2.67 6.1 Introduction When numerically solving two-point boundary value problems, the differential

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Linear Algebra for Machine Learning. Sargur N. Srihari

Linear Algebra for Machine Learning. Sargur N. Srihari Linear Algebra for Machine Learning Sargur N. srihari@cedar.buffalo.edu 1 Overview Linear Algebra is based on continuous math rather than discrete math Computer scientists have little experience with it

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =

More information

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},

More information

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

TBP MATH33A Review Sheet. November 24, 2018

TBP MATH33A Review Sheet. November 24, 2018 TBP MATH33A Review Sheet November 24, 2018 General Transformation Matrices: Function Scaling by k Orthogonal projection onto line L Implementation If we want to scale I 2 by k, we use the following: [

More information

Geometric Modeling Summer Semester 2012 Linear Algebra & Function Spaces

Geometric Modeling Summer Semester 2012 Linear Algebra & Function Spaces Geometric Modeling Summer Semester 2012 Linear Algebra & Function Spaces (Recap) Announcement Room change: On Thursday, April 26th, room 024 is occupied. The lecture will be moved to room 021, E1 4 (the

More information

Linear Algebra - Part II

Linear Algebra - Part II Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T

More information

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Spring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier

Spring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier Spring 0 CIS 55 Fundamentals of Linear Algebra and Optimization Jean Gallier Homework 5 & 6 + Project 3 & 4 Note: Problems B and B6 are for extra credit April 7 0; Due May 7 0 Problem B (0 pts) Let A be

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Matrices A brief introduction

Matrices A brief introduction Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 44 Definitions Definition A matrix is a set of N real or complex

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition (Com S 477/577 Notes Yan-Bin Jia Sep, 7 Introduction Now comes a highlight of linear algebra. Any real m n matrix can be factored as A = UΣV T where U is an m m orthogonal

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

A matrix over a field F is a rectangular array of elements from F. The symbol

A matrix over a field F is a rectangular array of elements from F. The symbol Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=

More information

Matrices and systems of linear equations

Matrices and systems of linear equations Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Lecture 6. Numerical methods. Approximation of functions

Lecture 6. Numerical methods. Approximation of functions Lecture 6 Numerical methods Approximation of functions Lecture 6 OUTLINE 1. Approximation and interpolation 2. Least-square method basis functions design matrix residual weighted least squares normal equation

More information

Least-Squares Fitting of Model Parameters to Experimental Data

Least-Squares Fitting of Model Parameters to Experimental Data Least-Squares Fitting of Model Parameters to Experimental Data Div. of Mathematical Sciences, Dept of Engineering Sciences and Mathematics, LTU, room E193 Outline of talk What about Science and Scientific

More information

Lecture Notes in Linear Algebra

Lecture Notes in Linear Algebra Lecture Notes in Linear Algebra Dr. Abdullah Al-Azemi Mathematics Department Kuwait University February 4, 2017 Contents 1 Linear Equations and Matrices 1 1.2 Matrices............................................

More information

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For

More information

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Vectors Arrays of numbers Vectors represent a point

More information

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p Math Bootcamp 2012 1 Review of matrix algebra 1.1 Vectors and rules of operations An p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the

More information

A Quick Tour of Linear Algebra and Optimization for Machine Learning

A Quick Tour of Linear Algebra and Optimization for Machine Learning A Quick Tour of Linear Algebra and Optimization for Machine Learning Masoud Farivar January 8, 2015 1 / 28 Outline of Part I: Review of Basic Linear Algebra Matrices and Vectors Matrix Multiplication Operators

More information

Linear Algebra and Matrices

Linear Algebra and Matrices Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Vectors Arrays of numbers Vectors represent a point in a n dimensional space

More information

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 222 3. M Test # July, 23 Solutions. For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For

More information

OR MSc Maths Revision Course

OR MSc Maths Revision Course OR MSc Maths Revision Course Tom Byrne School of Mathematics University of Edinburgh t.m.byrne@sms.ed.ac.uk 15 September 2017 General Information Today JCMB Lecture Theatre A, 09:30-12:30 Mathematics revision

More information

Scientific Computing: Dense Linear Systems

Scientific Computing: Dense Linear Systems Scientific Computing: Dense Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 February 9th, 2012 A. Donev (Courant Institute)

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax = . (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)

More information

Matrices A brief introduction

Matrices A brief introduction Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino September 2013 Basilio Bona (DAUIN) Matrices September 2013 1 / 74 Definitions Definition A matrix is a set of N real or complex numbers

More information

Maths for Signals and Systems Linear Algebra in Engineering

Maths for Signals and Systems Linear Algebra in Engineering Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Linear Algebra- Final Exam Review

Linear Algebra- Final Exam Review Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.

More information

EE263 homework 3 solutions

EE263 homework 3 solutions EE263 Prof. S. Boyd EE263 homework 3 solutions 2.17 Gradient of some common functions. Recall that the gradient of a differentiable function f : R n R, at a point x R n, is defined as the vector f(x) =

More information

Computer Assignment 2

Computer Assignment 2 Computer Assignment 2 Chapter 2, section, problem 3 Apes = [7 5;3 6] Diet = [25 4 2;5 3 25] Result = Apes * Diet Apes = 7 5 3 6 Diet = 25 4 2 5 3 25 Result = 25 43 265 65 3 2 The meaning of the calculation:

More information

EIGENVALUES AND SINGULAR VALUE DECOMPOSITION

EIGENVALUES AND SINGULAR VALUE DECOMPOSITION APPENDIX B EIGENVALUES AND SINGULAR VALUE DECOMPOSITION B.1 LINEAR EQUATIONS AND INVERSES Problems of linear estimation can be written in terms of a linear matrix equation whose solution provides the required

More information

5 Linear Algebra and Inverse Problem

5 Linear Algebra and Inverse Problem 5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude

More information

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.

More information

Jordan Normal Form and Singular Decomposition

Jordan Normal Form and Singular Decomposition University of Debrecen Diagonalization and eigenvalues Diagonalization We have seen that if A is an n n square matrix, then A is diagonalizable if and only if for all λ eigenvalues of A we have dim(u λ

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

Math 515 Fall, 2008 Homework 2, due Friday, September 26.

Math 515 Fall, 2008 Homework 2, due Friday, September 26. Math 515 Fall, 2008 Homework 2, due Friday, September 26 In this assignment you will write efficient MATLAB codes to solve least squares problems involving block structured matrices known as Kronecker

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.

More information

Problem # Max points possible Actual score Total 120

Problem # Max points possible Actual score Total 120 FINAL EXAMINATION - MATH 2121, FALL 2017. Name: ID#: Email: Lecture & Tutorial: Problem # Max points possible Actual score 1 15 2 15 3 10 4 15 5 15 6 15 7 10 8 10 9 15 Total 120 You have 180 minutes to

More information

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Bastian Steder

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Bastian Steder Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Bastian Steder Reference Book Thrun, Burgard, and Fox: Probabilistic Robotics Vectors Arrays of numbers Vectors represent

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly. C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a

More information

Inverses. Stephen Boyd. EE103 Stanford University. October 28, 2017

Inverses. Stephen Boyd. EE103 Stanford University. October 28, 2017 Inverses Stephen Boyd EE103 Stanford University October 28, 2017 Outline Left and right inverses Inverse Solving linear equations Examples Pseudo-inverse Left and right inverses 2 Left inverses a number

More information

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,

More information

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

MATH.2720 Introduction to Programming with MATLAB Vector and Matrix Algebra

MATH.2720 Introduction to Programming with MATLAB Vector and Matrix Algebra MATH.2720 Introduction to Programming with MATLAB Vector and Matrix Algebra A. Vectors A vector is a quantity that has both magnitude and direction, like velocity. The location of a vector is irrelevant;

More information

Pose estimation from point and line correspondences

Pose estimation from point and line correspondences Pose estimation from point and line correspondences Giorgio Panin October 17, 008 1 Problem formulation Estimate (in a LSE sense) the pose of an object from N correspondences between known object points

More information

Evaluation and Benchmarking of Highly Scalable Parallel Numerical Libraries

Evaluation and Benchmarking of Highly Scalable Parallel Numerical Libraries Evaluation and Benchmarking of Highly Scalable Parallel Numerical Libraries Christos Theodosiou (ctheodos@grid.auth.gr) User and Application Support Scientific Computing Centre @ AUTH Presentation Outline

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Lecture 5: Numerical Linear Algebra Cho-Jui Hsieh UC Davis April 20, 2017 Linear Algebra Background Vectors A vector has a direction and a magnitude

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Singular value decomposition

Singular value decomposition Singular value decomposition The eigenvalue decomposition (EVD) for a square matrix A gives AU = UD. Let A be rectangular (m n, m > n). A singular value σ and corresponding pair of singular vectors u (m

More information

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT Math Camp II Basic Linear Algebra Yiqing Xu MIT Aug 26, 2014 1 Solving Systems of Linear Equations 2 Vectors and Vector Spaces 3 Matrices 4 Least Squares Systems of Linear Equations Definition A linear

More information

Practical Linear Algebra: A Geometry Toolbox

Practical Linear Algebra: A Geometry Toolbox Practical Linear Algebra: A Geometry Toolbox Third edition Chapter 12: Gauss for Linear Systems Gerald Farin & Dianne Hansford CRC Press, Taylor & Francis Group, An A K Peters Book www.farinhansford.com/books/pla

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 2010

A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 2010 A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 00 Introduction Linear Algebra, also known as matrix theory, is an important element of all branches of mathematics

More information

MATH 106 LINEAR ALGEBRA LECTURE NOTES

MATH 106 LINEAR ALGEBRA LECTURE NOTES MATH 6 LINEAR ALGEBRA LECTURE NOTES FALL - These Lecture Notes are not in a final form being still subject of improvement Contents Systems of linear equations and matrices 5 Introduction to systems of

More information

The Singular Value Decomposition and Least Squares Problems

The Singular Value Decomposition and Least Squares Problems The Singular Value Decomposition and Least Squares Problems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 27, 2009 Applications of SVD solving

More information

Vector Spaces, Orthogonality, and Linear Least Squares

Vector Spaces, Orthogonality, and Linear Least Squares Week Vector Spaces, Orthogonality, and Linear Least Squares. Opening Remarks.. Visualizing Planes, Lines, and Solutions Consider the following system of linear equations from the opener for Week 9: χ χ

More information

Eigenvectors and SVD 1

Eigenvectors and SVD 1 Eigenvectors and SVD 1 Definition Eigenvectors of a square matrix Ax=λx, x=0. Intuition: x is unchanged by A (except for scaling) Examples: axis of rotation, stationary distribution of a Markov chain 2

More information

Math for ML: review. ML and knowledge of other fields

Math for ML: review. ML and knowledge of other fields ath for L: review ilos Hauskrecht milos@cs.pitt.edu Sennott Square x- people.cs.pitt.edu/~milos/ L and knowledge of other fields L solutions and algorithms rely on knowledge of many other disciplines:

More information

Linear Inverse Problems

Linear Inverse Problems Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Vector and Matrix Norms. Vector and Matrix Norms

Vector and Matrix Norms. Vector and Matrix Norms Vector and Matrix Norms Vector Space Algebra Matrix Algebra: We let x x and A A, where, if x is an element of an abstract vector space n, and A = A: n m, then x is a complex column vector of length n whose

More information

MTH 464: Computational Linear Algebra

MTH 464: Computational Linear Algebra MTH 464: Computational Linear Algebra Lecture Outlines Exam 2 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University February 6, 2018 Linear Algebra (MTH

More information

MATH2210 Notebook 2 Spring 2018

MATH2210 Notebook 2 Spring 2018 MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................

More information

CSL361 Problem set 4: Basic linear algebra

CSL361 Problem set 4: Basic linear algebra CSL361 Problem set 4: Basic linear algebra February 21, 2017 [Note:] If the numerical matrix computations turn out to be tedious, you may use the function rref in Matlab. 1 Row-reduced echelon matrices

More information