Physics 202 Laboratory 5. Linear Algebra 1. Laboratory 5. Physics 202 Laboratory
|
|
- Clarissa Stokes
- 5 years ago
- Views:
Transcription
1 Physics 202 Laboratory 5 Linear Algebra Laboratory 5 Physics 202 Laboratory We close our whirlwind tour of numerical methods by advertising some elements of (numerical) linear algebra. There are three major problems one needs to solve numerically the eigenvalue problem, the matrix inverse problem, and least squares. We ll focus on the physical targets for the first two, and outline the solution methods relevant there. Least squares, since we will ignore it, deserves additional mention. The least squares problem is defined as follows given a matrix A IR n m, with n rows and m columns 2, and a vector b IR n, find a vector x IR m such that: A x b (5.) is minimized. This problem is similar in spirit to matrix inversion (to which it reduces when m = n), and can be used to solve for best fit coefficients in data analysis, for example. 5. Physical Motivation We ll think about problems that reduce to matrix inversion first, then consider physics that leads to the eigenvalue problem. You have seen multiple examples of each, and my job here is to remind you of them, and set up the matrix form of their solution. The matrix inversion problem can be expressed as: Given a matrix A IR n n, and a vector b IR n, find x IR n such that: A x = b. (5.2) Some of this material appears in Computational Methods for Physics by Joel Franklin, Cambridge University Press, 203 and is used with the author s permission. 2 It is customary to let n > m, so we think of long matrices.
2 5.. PHYSICAL MOTIVATION Laboratory 5 When the solution exists, it is generally written: x = A b. The eigenvalue problem is defined as follows: Given a matrix A IR n n, find the vectors {v i } n i= and numbers {λ i} n i= such that A v i = λ i v i for all i = n. 5.. ODEs and PDEs by Inversion A linear, first order differential equation with initial condition, like: f (x) + p(x) f(x) = q(x) f(0) = f 0 (5.3) for given p(x) and q(x) can be solved on a grid with uniform spacing let x j j x for constant x, and write the projections of p, q and (the approximation to) f(x) on the grid as p j p(x j ), q j q(x j ) and f j f(x j ) respectively. Then the ODE is approximated by the difference equation: f j+ f j + p j f j = q j for all j =... N. (5.4) Suppose we think of the elements of the projection f j as entries in a vector of length N: f f 2 f =., (5.5) f N then we can construct a matrix that represents the left-hand side of the approximation (5.4), and a vector approximating the right-hand side: p f q + f 0 p f 2 q 2 p f 3 q p f 4 = q p 5... f 5 q p N f N q N (5.6) Now the discretization is relatively straightforward the first row is a little different than the others because, for j =, our ODE approximation reads: f N+. f 2 f 0 + p f = q, (5.7) 2
3 5.. PHYSICAL MOTIVATION Laboratory 5 and since f 0 is given, it goes over to the right-hand side of the equation and acts as part of the known vector. The same thing happens at j = N: f N+ f N + p N f N = q N, (5.8) but we are not justified in counting f N+ as a known value. Our matrix method appears to require two boundary conditions, and artificially so usually, something in the problem will help us approximate a value at N + (beyond the grid, and hence outside of our interest in the solution). For example, take f (x) + x f(x) = 0, so that p(x) = x and q(x) = 0. We know that for x very large, we have x f(x) 0, so that f(x) must be small, and we might use that observation to define a large value for x N+ and set f(x N+ ) = 0, at least numerically. However we manage to set the final value, it is clear that we have achieved matrix inverse form, since (5.6) clearly defines a known matrix A IR N N, and right-hand-side vector b IR N, our solution to this, f = A b is precisely our numerical approximation to the function f(x) evaluated on the grid. Many times, we are interested in second order ODEs, as in Newton s second law, for example, and in these cases, a boundary condition like f N+ = f f, provided, is reasonable, so our numerical approximation matches the physical boundaries. Think of approximating the most general second order linear ODE: f (x) + s(x) f (x) + p(x) f(x) = q(x) f(0) = f 0 f(x) = f X (5.9) (for final position X, where the value of f is given: f X )this equation would have second derivative approximated using: f (x j ) f j+ 2 f j + f j x 2 (5.0) and would also rely on known values for f 0 and f N+, but these are provided in the problem specification. We might let x N = X x = N x, giving us the definition of x in terms of the endpoint and N: x = X N +. (5.) Then x N+ = X, and we are given the value f N+ = f X. 3
4 5.. PHYSICAL MOTIVATION Laboratory 5 Example Suppose we take a damped harmonic oscillator that will give us a good check of the numerical method. Start with the usual form: where ω k m ẍ(t) = ω 2 x(t) 2 γ ω ẋ(t) (5.2) is the usual frequency of oscillation, and the damping term is governed by the parameter γ (dimensionless, in this case). We ll start by nondimensionalizing the equation, let t t 0 s and x x 0 q, then the above becomes d 2 q(s) ds 2 = q(s) ω 2 t γ dq(s) ds ω t 0, (5.3) and it is clear that we should take t 0 /ω, defining a natural timescale. The only parameter left is γ, and our choice of scaling for x 0 (typically the initial extension): d 2 q(s) ds 2 = q(s) 2 γ dq(s) ds. (5.4) We ll start, and end, with the ansatz: through (5.4) gives q(s) = e α s, running this α γ α + α = 2 γ ± 4 γ = γ ± γ 2 (5.5) from which we extract the two independent solutions (for γ ): ( ) ( ) γ+ γ q(s) = A e 2 s γ γ + B e 2 s. (5.6) Let s put all the constants back in, just to see how that goes: x(t) = x 0 e γ ω t [ A e γ 2 ω t + B e γ 2 ω t ]. (5.7) Now the particular behavior of the system is governed by the factor γ 2 if this is less than zero, there is oscillation (in addition to the damping), while if it is greater than zero, the system has only the decaying (and growing) exponentials as solutions. 4
5 5.. PHYSICAL MOTIVATION Laboratory 5 Partial differential equations can also be handled using a discretization, and reduction to matrix inverse. In E&M, the electrostatic potential V is sourced by charge density (charge per unit volume) ρ via: 2 V = ρ ɛ 0, where ɛ 0 sets the units for V given charge in Coulombs. This is an example of Poisson s equation, and it is the PDE version of the ODE f (x) = q(x). The analytic solution is more complicated, but numerically, all we do is take a grid (in two dimensions, just to show the pattern): x j j x, y k k y, and then approximate the Laplacian on the grid via: ( 2 2 ) V V (x j, y k ) x V y 2 x=x j,y=y k V (x j+, y k ) 2 V (x j, y k ) + V (x j, y k ) x 2 + V (x j, y k+ ) 2 V (x j, y k ) + V (x j, y k ) y 2. (5.8) The unknown values V (x j, y k ) must be put in a column vector somehow, and the matrix approximating the derivative values must be adapted to that embedding. But, when the dust has settled, we recover a matrix inverse problem Eigenvalue Problem The eigenvalue problem is already interesting in its pure matrix instantiation here we ll look at how discretization can render continuous versions of this problem into a finite form. Schrödinger s equation governing the allowed energies, E, of a particle of mass m moving under the influence of a potential V (x) reads: 2 2 m ψ (x) + V (x) ψ(x) = E ψ(x). (5.9) There are two targets to a solution of this equation, the first is ψ(x) itself, giving us a tool to calculate experimental outcomes (remember that ψ(x) ψ(x) dx represents the probability of finding the particle in the dx vicinity of x). The second is E, interpreted as the energy of the particle that is not an input, and must come from the solution. So if you squint at (5.9), you see that [ 2 2 m d 2 ] dx 2 + V (x) ψ(x) = E ψ(x) (5.20) 5
6 5.. PHYSICAL MOTIVATION Laboratory 5 defines a sort of continuous eigenvalue problem, with ψ(x) playing the role of the eigenvectors, and E playing the role of the eigenvalues. Remember that, as a second order ODE, the solutions will have two undetermined constants that can be set by provided boundary conditions. For simplicity, we ll try to find solutions that have well-defined zeroes for the wavefunction, so let ψ(0) = ψ(a) = 0. These are reasonable, and could come from the form of the potential, or our demand that at some local infinity, the wavefunction must vanish. We nondimensionalize the above in the usual way, letting x = a q and then defining Ẽ 2 m a2 2 m a2 E and Ṽ (q) V (a q), so that: 2 2 d2 ψ(q) dq 2 + Ṽ (q) ψ(q) = Ẽ ψ(q). (5.2) Suppose we put this problem onto a uniform grid with q j = j q for fixed q. We have q 0 = 0, and we ll let q N+ = q f, giving q = q f N+. Now the discrete approximation to (5.2), using the usual notation ψ j ψ(q j ), is: ψ j+ 2 ψ j + ψ j q 2 + Ṽj ψ j = Ẽ ψ j j =,..., N. (5.22) Once again, we can put the elements ψ j into a vector, ψ = ψ ψ 2. ψ N, (5.23) and define the matrix Q as: Ṽ q 2 q 2 Ṽ q q 2 q 2 0 Ṽ q q 2 q Ṽ q q 2 q Ṽ q q 2 q Ṽ q 2 N + 2 q 2, (5.24) 6
7 5.. PHYSICAL MOTIVATION Laboratory 5 and then the discretized problem (5.22) reads: Q ψ = Ẽ ψ. (5.25) There will be N vectors ψ and N values of Ẽ, so this equation defines an eigenvalue problem. Note that the matrix Q is symmetric and real, so it is guaranteed to have orthogonal eigenvectors with real eigenvalues (which is good if we are to understand Ẽ as an energy). Notice that our boundary condition is enforced: ψ 0 = ψ N+ = 0, since the j = version of (5.22) reads: ψ 2 2 ψ q 2 + Ṽ ψ = Ẽ ψ (5.26) which is correctly encoded in the first row of the matrix Q similar considerations apply to the j = N case. Example the hydrogen spectrum Hydrogen, from a physics point of view, consists of an electron (carrying charge e) moving in the central potential set up by a proton (with charge e). The potential is just the usual Coulomb one: e2 V (x) = 4 π ɛ 0 x, (5.27) and we can put this in our non-dimensionalized form: Ṽ (q) = 2 m a2 e 2 [ ] m e 2 a 4 π ɛ 0 2 a q = 2 4 π ɛ 0 2 q. (5.28) The factor in brackets on the right defines a length, called the Bohr radius, and usually denoted a 4 π ɛ 0 2 (lucky coincidence) let this m e 2 be our fundamental scaling length a, and then the final form of the potential is: Ṽ (q) = 2 q. (5.29) The nondimensionalized Schrödinger equation for this specific potential reads d2 ψ(q) dq 2 2 ψ(q) = Ẽ ψ(q). (5.30) q 7
8 5.2. METHODS Laboratory 5 If we solve this equation by discretizing it, we can get accurate approximations to the bound state energies of hydrogen. There are two fundamentally different behaviors we can study in both classical and quantum mechanics bound states are, here, states of negative energy, and are something like the classical orbits associated with central potentials. There are also scattering states. In the case of hydrogen, these have positive energy, and are a quantum mechanical analogue of unbound classical motion (shoot an electron passed a central charge). 5.2 Methods Here, I will just sketch the ideas, since they are fundamentally good ideas, and form the basis for much of the actual algorithms that are implemented to solve the matrix inversion and eigenvalue problems. But, in the end, this is just to give you a taste it is interesting to note that what a numerical linear algebra package does to solve these problems is very different than what you have learned to do Inverse Problem Most methods for solving for x in A x = b proceed from a particular factorization of the matrix A. We ll consider the special case of the so-called QR-factorization. For an invertible matrix A IR n n, it is possible to decompose A into an orthogonal matrix Q (orthogonal matrices, remember, have the property that Q = Q T ), and an upper-triangular matrix R (upper triangular matrices have non-zero entries only on and above the diagonal) such that: A = Q R. (5.3) The advantage of this observation is that we have reduced the inversion of A to the inversion of two matrices, each with simple inverses: and then we just need to solve x in R x = c. Q R x = b R x = Q T b c (5.32) 8
9 5.2. METHODS Laboratory 5 Let s think about that process the equation R x = c is really n equalities. The last of these reads: R nn x n = c n (5.33) (since R is upper triangular, its last row only has one entry, the last) from which we can easily solve for x n : x n = c n /R nn. One row up, we have, again using the upper triangularity of R: R (n )(n ) x n + R (n )n x n = c n (5.34) but since we know x n, we can again solve directly for x n : x n = c n R (n )n x n R (n )(n ). (5.35) We can continue to work upwards until we reach x, which will depend on the now known x 2... x n. This inversion scheme is easy to implement, and that was the whole point of the factorization the Q matrix is trivial to invert, and R isn t much worse. How do we know that this factorization, A = Q R, exists? The proof is typical of linear algebra factorization proofs we can sketch it by constructing the desired decomposition. Recall the Gram-Schmidt procedure: Given a collection of vectors a, a 2,..., a n all in IR n, we can form a spanning set (orthogonal and normalized in this case) by working sequentially. Start with q = a, and let R = q q, then take Q = q /R, a vector that points in the same direction as a but is normalized. Next let q 2 = a 2 (Q a 2 ) }{{} Q, (5.36) R 2 i.e. a vector that points in the a 2 direction but with the component in the Q direction projected out. Again define R 22 = q 2 q 2 and let Q 2 = q 2 /R 22, a normalized vector. Moving along, we start with a vector that points in the a 3 direction, with components in the Q and Q 2 directions projected out: q 3 = a 3 (Q a 3 ) Q }{{} (Q 2 a 3 ) Q }{{} 2, (5.37) R 3 R 23 and as before, set R 33 = q 3 q 3, then Q 3 = q 3 /R 33. So far, we have a collection of three orthonormal vectors {Q i } 3 i=, and a = R Q a 2 = R 2 Q + R 22 Q 2 a 3 = R 3 Q + R 23 Q 2 + R 33 Q 3, (5.38) 9
10 5.2. METHODS Laboratory 5 a situation we can summarize as follows: [ ] [ ] a a 2 a 3 = Q Q 2 Q 3 R R 2 R 3 0 R 22 R R 33, (5.39) where the square brackets on the left and right indicate matrices whose columns are the {a i } 3 i= (on the left) and {Q i} 3 i= (on the right). If we continue the process, we will eventually get the matrix A on the left (with columns that are precisely the provided {a i } n i= ), the first matrix on the left is an orthogonal matrix, call it Q, and the second matrix on the right is upper triangular, the R in A = Q R. The Power Method Finally, we close with a clever idea for computing the eigenvalues and eigenvectors of a matrix. For our purposes, let A IR n n be a symmetric matrix, so that its eigenvalues are real and its eigenvectors are orthogonal and complete (they span IR n, and can be taken as a basis set) take the eigenvectors to be normalized, so that vj T v j = for all j =... n. We ll further assume that the eigenvalues are non-degenerate, and can be ordered as follows: λ > λ 2 > λ 3 >... > λ n >. Take a generic vector x IR n. We know that it can be decomposed into the eigenvectors, even though we don t know what the eigenvectors are: n x = a i v i. (5.40) i= Now multiply both sides of this equation by A by assumption, we have A v i = λ i v i, so: n n A x = a i A v i = a i λ i v i. (5.4) i= Because the set {v i } n i= are eigenvectors, subsequent multiplication will just introduce additional factors of λ i : n A p x = a i λ p i v i. (5.42) i= Since λ > {λ i } n i=2 >, we know that the factor λp will grow faster than all the others, so for p large: i= A p x a λ p v, (5.43) 0
11 5.2. METHODS Laboratory 5 and if we set y A p x, it is clear that v = y y T y. (5.44) We have found the first eigenvector, and can even find the associated eigenvalue. But what about the rest? Suppose we now want to find v 2? Well, with v in hand, we can iterate the following to find v 2 this is another example of the Gram-Schmidt procedure: w 0 = x for j = p w j = A w j w j = w j ( w T j v ) v end v 2 = w p w T p w p. (5.45) Here, at each stage of the multiplication, we project out any v component, leaving us with the second-fastest-growing eigenvalue, the one associated with v 2. The process continues we now find the third eigenvector by multiplying by A and projecting out the v and v 2 directions. Finally, with all the eigenvectors in place, we can collect the eigenvalues: A v i = λ i (5.46) i.e. we just take the norm of the product A v i.
12 5.2. METHODS Laboratory 5 Lab Problem 5. Take the matrix: and the vector A =5 w = (5.47). (5.48) Use the power method to find the first eigenvector and eigenvalue of A, record the normalized eigenvector, the eigenvalue, and the number of multiplications by A you used (how did you settle on this number?) below: 2
13 5.2. METHODS Laboratory 5 Problem 5.2 Find the negative energy eigenstates of hydrogen using the built-in function Eigenvalues (start by writing a function that generates the appropriate matrix). Use q f = 5, 20, 40, and 80 and N = 000. a. Write the negative energies you get for each value of q f below (you can use the built-in function Sort to order the eigenvalues): b. Take the q f = 80 case using the first four (negative) eigenvalues, find a function f(n) for n =, 2, 3, 4 that (roughly) gives you the numerical result. Write your function below: Finally, calculate the pre-factor to get E(n) for hydrogen (what you have in your numerical routine is Ẽ(n)) use the electron-volt as your unit of energy. 3
1.5 Eigenvalue Problems
5 Eigenvalue Problems The eigenvalue problem, for matrices, reads: Given a matrix A 2 IR n n,find some/all of the set of vectors {v i } n i= and numbers { i} n i= such that: A v i = i v i (45) In general,
More informationPhysics 200 Lecture 4. Integration. Lecture 4. Physics 200 Laboratory
Physics 2 Lecture 4 Integration Lecture 4 Physics 2 Laboratory Monday, February 21st, 211 Integration is the flip-side of differentiation in fact, it is often possible to write a differential equation
More informationHarmonic Oscillator I
Physics 34 Lecture 7 Harmonic Oscillator I Lecture 7 Physics 34 Quantum Mechanics I Monday, February th, 008 We can manipulate operators, to a certain extent, as we would algebraic expressions. By considering
More informationSection 6.4. The Gram Schmidt Process
Section 6.4 The Gram Schmidt Process Motivation The procedures in 6 start with an orthogonal basis {u, u,..., u m}. Find the B-coordinates of a vector x using dot products: x = m i= x u i u i u i u i Find
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to
More informationLINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS
LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,
More informationTime-Independent Perturbation Theory
4 Phys46.nb Time-Independent Perturbation Theory.. Overview... General question Assuming that we have a Hamiltonian, H = H + λ H (.) where λ is a very small real number. The eigenstates of the Hamiltonian
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationPHY 407 QUANTUM MECHANICS Fall 05 Problem set 1 Due Sep
Problem set 1 Due Sep 15 2005 1. Let V be the set of all complex valued functions of a real variable θ, that are periodic with period 2π. That is u(θ + 2π) = u(θ), for all u V. (1) (i) Show that this V
More informationPhysics 342 Lecture 23. Radial Separation. Lecture 23. Physics 342 Quantum Mechanics I
Physics 342 Lecture 23 Radial Separation Lecture 23 Physics 342 Quantum Mechanics I Friday, March 26th, 2010 We begin our spherical solutions with the simplest possible case zero potential. Aside from
More informationQuantum Mechanics in Three Dimensions
Physics 342 Lecture 21 Quantum Mechanics in Three Dimensions Lecture 21 Physics 342 Quantum Mechanics I Monday, March 22nd, 21 We are used to the temporal separation that gives, for example, the timeindependent
More informationPage 404. Lecture 22: Simple Harmonic Oscillator: Energy Basis Date Given: 2008/11/19 Date Revised: 2008/11/19
Page 404 Lecture : Simple Harmonic Oscillator: Energy Basis Date Given: 008/11/19 Date Revised: 008/11/19 Coordinate Basis Section 6. The One-Dimensional Simple Harmonic Oscillator: Coordinate Basis Page
More informationSeparation of Variables in Linear PDE: One-Dimensional Problems
Separation of Variables in Linear PDE: One-Dimensional Problems Now we apply the theory of Hilbert spaces to linear differential equations with partial derivatives (PDE). We start with a particular example,
More informationMATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION
MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether
More informationREVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION
REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationLab 1: Iterative Methods for Solving Linear Systems
Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as
More informationPhysics 221A Fall 2017 Notes 27 The Variational Method
Copyright c 2018 by Robert G. Littlejohn Physics 221A Fall 2017 Notes 27 The Variational Method 1. Introduction Very few realistic problems in quantum mechanics are exactly solvable, so approximation methods
More information1 GSW Sets of Systems
1 Often, we have to solve a whole series of sets of simultaneous equations of the form y Ax, all of which have the same matrix A, but each of which has a different known vector y, and a different unknown
More informationThe quantum state as a vector
The quantum state as a vector February 6, 27 Wave mechanics In our review of the development of wave mechanics, we have established several basic properties of the quantum description of nature:. A particle
More informationTP computing lab - Integrating the 1D stationary Schrödinger eq
TP computing lab - Integrating the 1D stationary Schrödinger equation September 21, 2010 The stationary 1D Schrödinger equation The time-independent (stationary) Schrödinger equation is given by Eψ(x)
More informationEcon Slides from Lecture 7
Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for
More information18.06 Professor Johnson Quiz 1 October 3, 2007
18.6 Professor Johnson Quiz 1 October 3, 7 SOLUTIONS 1 3 pts.) A given circuit network directed graph) which has an m n incidence matrix A rows = edges, columns = nodes) and a conductance matrix C [diagonal
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationHarmonic Oscillator with raising and lowering operators. We write the Schrödinger equation for the harmonic oscillator in one dimension as follows:
We write the Schrödinger equation for the harmonic oscillator in one dimension as follows: H ˆ! = "!2 d 2! + 1 2µ dx 2 2 kx 2! = E! T ˆ = "! 2 2µ d 2 dx 2 V ˆ = 1 2 kx 2 H ˆ = ˆ T + ˆ V (1) where µ is
More information18.06 Quiz 2 April 7, 2010 Professor Strang
18.06 Quiz 2 April 7, 2010 Professor Strang Your PRINTED name is: 1. Your recitation number or instructor is 2. 3. 1. (33 points) (a) Find the matrix P that projects every vector b in R 3 onto the line
More informationLecture 10: Powers of Matrices, Difference Equations
Lecture 10: Powers of Matrices, Difference Equations Difference Equations A difference equation, also sometimes called a recurrence equation is an equation that defines a sequence recursively, i.e. each
More information[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]
Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the
More information1. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal
. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal 3 9 matrix D such that A = P DP, for A =. 3 4 3 (a) P = 4, D =. 3 (b) P = 4, D =. (c) P = 4 8 4, D =. 3 (d) P
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More informationPhysics 202 Laboratory 3. Root-Finding 1. Laboratory 3. Physics 202 Laboratory
Physics 202 Laboratory 3 Root-Finding 1 Laboratory 3 Physics 202 Laboratory The fundamental question answered by this week s lab work will be: Given a function F (x), find some/all of the values {x i }
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More information8.04: Quantum Mechanics Professor Allan Adams. Problem Set 7. Due Tuesday April 9 at 11.00AM
8.04: Quantum Mechanics Professor Allan Adams Massachusetts Institute of Technology Thursday April 4 Problem Set 7 Due Tuesday April 9 at 11.00AM Assigned Reading: E&R 6 all Li. 7 1 9, 8 1 Ga. 4 all, 6
More informationc 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0
LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =
More informationBound and Scattering Solutions for a Delta Potential
Physics 342 Lecture 11 Bound and Scattering Solutions for a Delta Potential Lecture 11 Physics 342 Quantum Mechanics I Wednesday, February 20th, 2008 We understand that free particle solutions are meant
More informationTotal Angular Momentum for Hydrogen
Physics 4 Lecture 7 Total Angular Momentum for Hydrogen Lecture 7 Physics 4 Quantum Mechanics I Friday, April th, 008 We have the Hydrogen Hamiltonian for central potential φ(r), we can write: H r = p
More informationThe Schrödinger Equation
Chapter 13 The Schrödinger Equation 13.1 Where we are so far We have focused primarily on electron spin so far because it s a simple quantum system (there are only two basis states!), and yet it still
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More informationLinear Algebra in Hilbert Space
Physics 342 Lecture 16 Linear Algebra in Hilbert Space Lecture 16 Physics 342 Quantum Mechanics I Monday, March 1st, 2010 We have seen the importance of the plane wave solutions to the potentialfree Schrödinger
More informationEXAMPLES OF PROOFS BY INDUCTION
EXAMPLES OF PROOFS BY INDUCTION KEITH CONRAD 1. Introduction In this handout we illustrate proofs by induction from several areas of mathematics: linear algebra, polynomial algebra, and calculus. Becoming
More informationYORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions
YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 3. M Test # Solutions. (8 pts) For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For this
More informationQuantum Mechanics Solutions
Quantum Mechanics Solutions (a (i f A and B are Hermitian, since (AB = B A = BA, operator AB is Hermitian if and only if A and B commute So, we know that [A,B] = 0, which means that the Hilbert space H
More informationQuantum Theory and Group Representations
Quantum Theory and Group Representations Peter Woit Columbia University LaGuardia Community College, November 1, 2017 Queensborough Community College, November 15, 2017 Peter Woit (Columbia University)
More informationStabilization and Acceleration of Algebraic Multigrid Method
Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration
More informationLINEAR ALGEBRA KNOWLEDGE SURVEY
LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.
More informationThe Particle in a Box
Page 324 Lecture 17: Relation of Particle in a Box Eigenstates to Position and Momentum Eigenstates General Considerations on Bound States and Quantization Continuity Equation for Probability Date Given:
More informationNotes for CS542G (Iterative Solvers for Linear Systems)
Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,
More informationMath 115 Spring 11 Written Homework 10 Solutions
Math 5 Spring Written Homework 0 Solutions. For following its, state what indeterminate form the its are in and evaluate the its. (a) 3x 4x 4 x x 8 Solution: This is in indeterminate form 0. Algebraically,
More informationPDEs in Spherical and Circular Coordinates
Introduction to Partial Differential Equations part of EM, Scalar and Vector Fields module (PHY2064) This lecture Laplacian in spherical & circular polar coordinates Laplace s PDE in electrostatics Schrödinger
More informationECE 501b Homework #4 Due: 10/22/2012
1. Game Strategy: Consider a multiplayer boardgame that takes place on the following board and with the following rules: 7 8 9 10 6 11 5 12 4 3 2 1 The board contains six squares that are property (the
More informationMathematical Physics Homework 10
Georgia Institute of Technology Mathematical Physics Homework Conner Herndon November, 5 Several types of orthogonal polynomials frequently occur in various physics problems. For instance, Hermite polynomials
More informationLinear Least-Squares Data Fitting
CHAPTER 6 Linear Least-Squares Data Fitting 61 Introduction Recall that in chapter 3 we were discussing linear systems of equations, written in shorthand in the form Ax = b In chapter 3, we just considered
More informationMidterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015
Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015 The test lasts 1 hour and 15 minutes. No documents are allowed. The use of a calculator, cell phone or other equivalent electronic
More informationPhysics 106a, Caltech 13 November, Lecture 13: Action, Hamilton-Jacobi Theory. Action-Angle Variables
Physics 06a, Caltech 3 November, 08 Lecture 3: Action, Hamilton-Jacobi Theory Starred sections are advanced topics for interest and future reference. The unstarred material will not be tested on the final
More information1 Mathematical preliminaries
1 Mathematical preliminaries The mathematical language of quantum mechanics is that of vector spaces and linear algebra. In this preliminary section, we will collect the various definitions and mathematical
More informationPractice problems for Exam 3 A =
Practice problems for Exam 3. Let A = 2 (a) Determine whether A is diagonalizable. If so, find a matrix S such that S AS is diagonal. If not, explain why not. (b) What are the eigenvalues of A? Is A diagonalizable?
More informationSpectral Processing. Misha Kazhdan
Spectral Processing Misha Kazhdan [Taubin, 1995] A Signal Processing Approach to Fair Surface Design [Desbrun, et al., 1999] Implicit Fairing of Arbitrary Meshes [Vallet and Levy, 2008] Spectral Geometry
More informationMATH 22A: LINEAR ALGEBRA Chapter 4
MATH 22A: LINEAR ALGEBRA Chapter 4 Jesús De Loera, UC Davis November 30, 2012 Orthogonality and Least Squares Approximation QUESTION: Suppose Ax = b has no solution!! Then what to do? Can we find an Approximate
More informationBrief review of Quantum Mechanics (QM)
Brief review of Quantum Mechanics (QM) Note: This is a collection of several formulae and facts that we will use throughout the course. It is by no means a complete discussion of QM, nor will I attempt
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationCheat Sheet for MATH461
Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationRepresentation theory and quantum mechanics tutorial Spin and the hydrogen atom
Representation theory and quantum mechanics tutorial Spin and the hydrogen atom Justin Campbell August 3, 2017 1 Representations of SU 2 and SO 3 (R) 1.1 The following observation is long overdue. Proposition
More informationSolution of Linear Equations
Solution of Linear Equations (Com S 477/577 Notes) Yan-Bin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass
More informationLinear algebra for MATH2601: Theory
Linear algebra for MATH2601: Theory László Erdős August 12, 2000 Contents 1 Introduction 4 1.1 List of crucial problems............................... 5 1.2 Importance of linear algebra............................
More informationName Solutions Linear Algebra; Test 3. Throughout the test simplify all answers except where stated otherwise.
Name Solutions Linear Algebra; Test 3 Throughout the test simplify all answers except where stated otherwise. 1) Find the following: (10 points) ( ) Or note that so the rows are linearly independent, so
More informationExtreme Values and Positive/ Negative Definite Matrix Conditions
Extreme Values and Positive/ Negative Definite Matrix Conditions James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 8, 016 Outline 1
More informationCS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works
CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The
More information20 The Hydrogen Atom. Ze2 r R (20.1) H( r, R) = h2 2m 2 r h2 2M 2 R
20 The Hydrogen Atom 1. We want to solve the time independent Schrödinger Equation for the hydrogen atom. 2. There are two particles in the system, an electron and a nucleus, and so we can write the Hamiltonian
More informationLecture 4: Applications of Orthogonality: QR Decompositions
Math 08B Professor: Padraic Bartlett Lecture 4: Applications of Orthogonality: QR Decompositions Week 4 UCSB 204 In our last class, we described the following method for creating orthonormal bases, known
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationlecture 2 and 3: algorithms for linear algebra
lecture 2 and 3: algorithms for linear algebra STAT 545: Introduction to computational statistics Vinayak Rao Department of Statistics, Purdue University August 27, 2018 Solving a system of linear equations
More informationMATH 167: APPLIED LINEAR ALGEBRA Least-Squares
MATH 167: APPLIED LINEAR ALGEBRA Least-Squares October 30, 2014 Least Squares We do a series of experiments, collecting data. We wish to see patterns!! We expect the output b to be a linear function of
More informationChasing the Bulge. Sebastian Gant 5/19/ The Reduction to Hessenberg Form 3
Chasing the Bulge Sebastian Gant 5/9/207 Contents Precursers and Motivation 2 The Reduction to Hessenberg Form 3 3 The Algorithm 5 4 Concluding Remarks 8 5 References 0 ntroduction n the early days of
More informationQuantum Mechanics for Mathematicians: Energy, Momentum, and the Quantum Free Particle
Quantum Mechanics for Mathematicians: Energy, Momentum, and the Quantum Free Particle Peter Woit Department of Mathematics, Columbia University woit@math.columbia.edu November 28, 2012 We ll now turn to
More informationEigenvectors and Hermitian Operators
7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding
More informationDegenerate Perturbation Theory. 1 General framework and strategy
Physics G6037 Professor Christ 12/22/2015 Degenerate Perturbation Theory The treatment of degenerate perturbation theory presented in class is written out here in detail. The appendix presents the underlying
More information1.6 and 5.3. Curve Fitting One of the broadest applications of linear algebra is to curve fitting, especially in determining unknown coefficients in
16 and 53 Curve Fitting One of the broadest applications of linear algebra is to curve fitting, especially in determining unknown coefficients in functions You should know that, given two points in the
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More informationFinite Difference Methods for Boundary Value Problems
Finite Difference Methods for Boundary Value Problems October 2, 2013 () Finite Differences October 2, 2013 1 / 52 Goals Learn steps to approximate BVPs using the Finite Difference Method Start with two-point
More informationPhysics 342 Lecture 17. Midterm I Recap. Lecture 17. Physics 342 Quantum Mechanics I
Physics 342 Lecture 17 Midterm I Recap Lecture 17 Physics 342 Quantum Mechanics I Monday, March 1th, 28 17.1 Introduction In the context of the first midterm, there are a few points I d like to make about
More informationPhysics 115C Homework 2
Physics 5C Homework Problem Our full Hamiltonian is H = p m + mω x +βx 4 = H +H where the unperturbed Hamiltonian is our usual and the perturbation is H = p m + mω x H = βx 4 Assuming β is small, the perturbation
More informationLINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12,
LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12, 2000 74 6 Summary Here we summarize the most important information about theoretical and numerical linear algebra. MORALS OF THE STORY: I. Theoretically
More informationECE 487 Lecture 5 : Foundations of Quantum Mechanics IV Class Outline:
ECE 487 Lecture 5 : Foundations of Quantum Mechanics IV Class Outline: Linearly Varying Potential Triangular Potential Well Time-Dependent Schrödinger Equation Things you should know when you leave Key
More informationStatistical Interpretation
Physics 342 Lecture 15 Statistical Interpretation Lecture 15 Physics 342 Quantum Mechanics I Friday, February 29th, 2008 Quantum mechanics is a theory of probability densities given that we now have an
More informationLecture 4 Eigenvalue problems
Lecture 4 Eigenvalue problems Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn
More informationChapter 7. Bound Systems are perhaps the most interesting cases for us to consider. We see much of the interesting features of quantum mechanics.
Chapter 7 In chapter 6 we learned about a set of rules for quantum mechanics. Now we want to apply them to various cases and see what they predict for the behavior of quanta under different conditions.
More informationMIT Final Exam Solutions, Spring 2017
MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of
More informationMTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education
MTH 3 Linear Algebra Study Guide Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education June 3, ii Contents Table of Contents iii Matrix Algebra. Real Life
More informationMITOCW watch?v=fxlzy2l1-4w
MITOCW watch?v=fxlzy2l1-4w PROFESSOR: We spoke about the hydrogen atom. And in the hydrogen atom, we drew the spectrum, so the table, the data of spectrum of a quantum system. So this is a question that
More informationQuantum Mechanics- I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras
Quantum Mechanics- I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras Lecture - 6 Postulates of Quantum Mechanics II (Refer Slide Time: 00:07) In my last lecture,
More informationInverses. Stephen Boyd. EE103 Stanford University. October 28, 2017
Inverses Stephen Boyd EE103 Stanford University October 28, 2017 Outline Left and right inverses Inverse Solving linear equations Examples Pseudo-inverse Left and right inverses 2 Left inverses a number
More informationQuantum Mechanics-I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras. Lecture - 21 Square-Integrable Functions
Quantum Mechanics-I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras Lecture - 21 Square-Integrable Functions (Refer Slide Time: 00:06) (Refer Slide Time: 00:14) We
More informationQuantum Mechanics C (130C) Winter 2014 Final exam
University of California at San Diego Department of Physics Prof. John McGreevy Quantum Mechanics C (130C Winter 014 Final exam Please remember to put your name on your exam booklet. This is a closed-book
More informationMath 471 (Numerical methods) Chapter 3 (second half). System of equations
Math 47 (Numerical methods) Chapter 3 (second half). System of equations Overlap 3.5 3.8 of Bradie 3.5 LU factorization w/o pivoting. Motivation: ( ) A I Gaussian Elimination (U L ) where U is upper triangular
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More information8.3 Partial Fraction Decomposition
8.3 partial fraction decomposition 575 8.3 Partial Fraction Decomposition Rational functions (polynomials divided by polynomials) and their integrals play important roles in mathematics and applications,
More information18.303: Introduction to Green s functions and operator inverses
8.33: Introduction to Green s functions and operator inverses S. G. Johnson October 9, 2 Abstract In analogy with the inverse A of a matri A, we try to construct an analogous inverse  of differential
More informationThe 3 dimensional Schrödinger Equation
Chapter 6 The 3 dimensional Schrödinger Equation 6.1 Angular Momentum To study how angular momentum is represented in quantum mechanics we start by reviewing the classical vector of orbital angular momentum
More informationNotes on wavefunctions IV: the Schrödinger equation in a potential and energy eigenstates.
Notes on wavefunctions IV: the Schrödinger equation in a potential and energy eigenstates. We have now seen that the wavefunction for a free electron changes with time according to the Schrödinger Equation
More information