Multi-Frame Factorization Techniques

Similar documents
Camera Models and Affine Multiple Views Geometry

Trinocular Geometry Revisited

6.801/866. Affine Structure from Motion. T. Darrell

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Affine Structure From Motion

M3: Multiple View Geometry

A Factorization Method for 3D Multi-body Motion Estimation and Segmentation

EE Camera & Image Formation

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Multiple View Geometry in Computer Vision

Linear Algebra Review. Vectors

A Study of Kruppa s Equation for Camera Self-calibration

Robert Collins CSE486, Penn State. Lecture 25: Structure from Motion

On the Absolute Quadratic Complex and its Application to Autocalibration

A Practical Method for Decomposition of the Essential Matrix

Basic Math for

Lecture 8: Coordinate Frames. CITS3003 Graphics & Animation

Uncertainty Models in Quasiconvex Optimization for Geometric Reconstruction

Robust Motion Segmentation by Spectral Clustering

Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix

Pose estimation from point and line correspondences

Properties of Matrices and Operations on Matrices

Augmented Reality VU Camera Registration. Prof. Vincent Lepetit

Multiple View Geometry in Computer Vision

CSE 252B: Computer Vision II

Algorithms for Computing a Planar Homography from Conics in Correspondence

Sparse Levenberg-Marquardt algorithm.

Linear Algebra Review. Fei-Fei Li

Camera Calibration The purpose of camera calibration is to determine the intrinsic camera parameters (c 0,r 0 ), f, s x, s y, skew parameter (s =

Visual SLAM Tutorial: Bundle Adjustment

Multi-frame Infinitesimal Motion Model for the Reconstruction of (Dynamic) Scenes with Multiple Linearly Moving Objects

Linear Algebra for Machine Learning. Sargur N. Srihari

Introduction to pinhole cameras

Degeneracies, Dependencies and their Implications in Multi-body and Multi-Sequence Factorizations

Two-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix

Kinematics. Chapter Multi-Body Systems

1 Linearity and Linear Systems

CS123 INTRODUCTION TO COMPUTER GRAPHICS. Linear Algebra /34

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Determining the Translational Speed of a Camera from Time-Varying Optical Flow

A geometric interpretation of the homogeneous coordinates is given in the following Figure.

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti

Linear Differential Algorithm for Motion Recovery: A Geometric Approach

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

Linear Algebra: Matrix Eigenvalue Problems

Mathematical foundations - linear algebra

Homographies and Estimating Extrinsics

Induced Planar Homologies in Epipolar Geometry

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Lecture 5. Epipolar Geometry. Professor Silvio Savarese Computational Vision and Geometry Lab. 21-Jan-15. Lecture 5 - Silvio Savarese

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Notes on Eigenvalues, Singular Values and QR

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Introduction to Matrix Algebra

Linear Algebra - Part II

Linear Algebra. Min Yan

Pseudoinverse & Moore-Penrose Conditions

Lecture 7. Econ August 18

CS123 INTRODUCTION TO COMPUTER GRAPHICS. Linear Algebra 1/33

Singular Value Decomposition

Maths for Signals and Systems Linear Algebra in Engineering

On Single-Sequence and Multi-Sequence Factorizations

CS 231A Computer Vision (Fall 2011) Problem Set 2

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang

CS 231A Section 1: Linear Algebra & Probability Review

Lecture 5 Singular value decomposition

B553 Lecture 5: Matrix Algebra Review

Homogeneous Coordinates

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

STA141C: Big Data & High Performance Statistical Computing

The Singular Value Decomposition

Principal Components Theory Notes

Linear Algebra Review. Fei-Fei Li

Deep Learning Book Notes Chapter 2: Linear Algebra

Review of similarity transformation and Singular Value Decomposition

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Motion Recovery from Image Sequences: Discrete Viewpoint vs. Differential Viewpoint

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 2: Rigid Motions and Homogeneous Transformations

STA141C: Big Data & High Performance Statistical Computing

Reconstruction from projections using Grassmann tensors

Rigid Geometric Transformations

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Determining Constant Optical Flow

Principal Component Analysis

Optimisation on Manifolds

Linear System Theory

3D from Photographs: Camera Calibration. Dr Francesco Banterle

Lecture: Face Recognition and Feature Reduction

University oftoronto, Toronto, Ontario M5S 3H5. we eliminate the need to estimate the scale of the covariance

Camera calibration Triangulation

Linear Subspace Models

Screw Theory and its Applications in Robotics

Multiple View Geometry in Computer Vision

EE731 Lecture Notes: Matrix Computations for Signal Processing

2. Review of Linear Algebra

Chapter 3 Transformations

Mathematical foundations - linear algebra

Computational Methods. Eigenvalues and Singular Values

Linear Algebra V = T = ( 4 3 ).

ECE 275A Homework #3 Solutions

Transcription:

Multi-Frame Factorization Techniques Suppose { x j,n } J,N j=1,n=1 is a set of corresponding image coordinates, where the index n = 1,...,N refers to the n th scene point and j = 1,..., J refers to the j th image. 15 Image 1 15 Image 2 1 1 5 5 5 5 1 1 15 15 1 5 5 1 15 15 Image 1 15 15 1 5 5 1 15 1 5 5 1 15 15 1 5 5 1 15 Such corresponding points may be obtained from local feature points, for example. Problem. Estimate the 3D point positions, { X n } N n=1, along with the placement and calibration parameters for the J cameras. 253: Multi-Frame Factorization c Allan Jepson, 27 Page: 1

Perspective Projection The image points and the scene points are related by perspective projection, p j,n = 1 z j,n M j Pn. (1) Here p j,n = (x j,n, y j,n, 1) T is in homogeneous pixel coordinates, and the scene point P n = (P n,1, P n,2, P n,3, 1) T is in homogeneous 3D coordinates. Also M j = M in,j M ex,j is the 3 4 camera matrix formed from the product of the intrinsic and extrinsic calibration matrices. Finally, z j,n is the projective depth, z j,n = e T 3 M j Pn, where e T 3 = (,, 1) (i.e. e 3 is the third standard unit vector). For convenience we assume the intrinsic matrices have the form M in,j = f j f j 1 The extrinsic calibration matrices are in general given by. (2) M ex,j = ( R j, R j dj ), (3) where R j is the rotation from the world to the j th -camera s coordinates, and d j is the position, in world coordinates, of the nodal point for the j th camera. 253: Multi-Frame Factorization Page: 2

Bundle Adjustment We wish to solve for the point positions P n for n = 1,..., N and the camera matrices M j for j = 1,..., J by minimizing O({M j } J j=1, { P n } N n=1 ) j,n x j,n y j,n 1 ( I2, ) M e 3 T M jpn jpn Here the camera matrices M j must be of the form M in,j M ex,j where M in,j and M ex,j are as given in equations (2) and (3). This nonlinear optimization problem is called bundle adjustment. In these notes we discuss two approximations to bundle adjustment: 1. Approximate perspective projection by scaled orthographic projection. 2. Rescale each term in the bundle adjustment objective function (4) and solve a bilinear problem. 2. (4) 253: Multi-Frame Factorization Page: 3

Scaled-Orthographic Projection Scaled-orthographic projection provides an approximation of perspective projection (1) for the case of narrow fields of view, max{ x j,n, y j,n } << f j, and relatively shallow depth variations, z j,n 1/s. For scaled-orthographic projection, the image points and the scene points are related by ( I2, ) p j,n = s ( I 2, ) M j Pn. (5) Here p j,n, Pn and M j are as above, and s is a constant scale factor. This is bilinear in the scaled camera matrix sm j and the 3D point P n. 253: Multi-Frame Factorization Page: 4

Differences from Mean Image Points Let p j = 1 N N n=1 p j,n be the average image point, and P = 1 N be the average scene point. Then, by equation (5), we can show where N n=1 Pn d j,n = M j Dn, (6) d j,n = ( I 2, ) ( p j,n p j ), D n = (I 3, )( P n P), M j = s ( I 2, ) M j ( I3, ) T. Moreover, from equations (2) and (3) it follows that the scaled-orthographic projection matrix M j has the form M j = s f j f j R ( j = sf j I2, ) R j, (7) where R j is the rotation matrix for the j th camera, as above. 253: Multi-Frame Factorization Page: 5

Derivation: Difference from Mean Let p j = 1 Nn=1 p N j,n be the average image point in the j th image, and P = 1 Nn=1 Pn be the N average scene point. Then, by equation (5), we have ( I2, ) p j = s ( I 2, ) M j P Subtracting this from (5) we find ( I2, ) ( p j,n p j ) = s ( I 2, ) M j ( P n P). Note the 4 th component of P n P is equal to 1 1 =. Therefore we can drop this 4 th component, and obtain d j,n = M jdn, where d j,n = ( I 2, ) ( p j,n p j ), D n = (I 3, )( P n P), M j = s ( I 2, ) ( M j I3, ) T. Which is what we set out to show. Notice we can use the definitions of M in,j and M ex,j to simplify M j above. We find This gives equation (7) above. M j = s ( I 2, ) M j ( I3, ) T, = s ( I 2, ) ( M in,j M ex,j I3, ) T, = s f j R j f j 253: Multi-Frame Factorization Notes: 6

Scaled-Orthographic Factorization Let C = ( ) dj,n be the 2J N data matrix formed by stacking dj,n, for j = 1,...,J in columns, and combining these columns for n = 1,..., N (here j is the camera index and n the feature point index). From above, d j,n = x j,n x j, where x j,n is the observed pixel position of the j th point in the n th frame, and x j is the average of these over all n. In particular, the data matrix can be built from the observed corresponding points. From equation (6) we then have C = MD, (8) where M is the 2J 3 matrix formed by stacking the M j matrices, and D is the 3 N shape matrix having columns given by D n. This equation states that the data matrix has at most rank 3 (without considering noise). 253: Multi-Frame Factorization Page: 7

Factorization via SVD Performing an SVD on the data matrix C, for a case with J = 3 images, provides C = WΣV T with the singular values shown below: 25 Singular Values of Data Matrix 2 Singlur Value 15 1 5 1 2 3 4 5 6 Singular Value Index See the 3dRecon Matlab tutorialorthomassagedino.m (σ n = 1 pixel). 253: Multi-Frame Factorization Page: 8

Affine Shape What does the factorization C = WΣV T tell us about the shape of the objects being imaged? For notational convenience, we assume that all but the first 3 singular values of Σ have been set to zero or, equivalently, Σ is 3 3, W is 2J 3 and V T is 3 N. We now have two rank 3 factorizations of C, namely MD and WΣV T. But this factorization is only unique up to a nonsingular matrix A, as follows: C = MD = (WA)(A 1 ΣV T ) = WΣV T. (9) That is, for some 3 3 matrix A, the 3D point positions and the camera matrices must be given by D = A 1 ΣV T M = WA. (1) Equivalently, we could place AA 1 between the Σ and the V T in equation (9). Therefore we know the shape D up to the 9 parameters in A, namely AD = ΣV T. This is known as an affine reconstruction of the shape D. 253: Multi-Frame Factorization Page: 9

Affine Shape (Cont.) What can A (or, equivalenty, A 1 ) do to a shape? For example, consider a configuration of 3D points as specified by the 3 N matrix D above. Suppose we have a nonsingular matrix A. What does the configuration AD look like? Use SVD to decompose A into U a Σ a Va T. So AD = U a (Σ a (Va T D)) is obtained by rotating/reflecting D using Va T, then stretching/shrinking the result along the axes according to Σ a, and finally rotating/reflecting this result using U a. (Imagine applying such transforms to your lecturer s head.) The equivalence class of all configurations that can be obtained with transformations of this form is called affine shape. It can be shown that affine shape preserves parallel lines and intersecting lines, but not angles and lengths. See Tomasi and Kanade, IJCV, Vol. 9, 1992, pp.137-154, for the original factorization method. 253: Multi-Frame Factorization Page: 1

Euclidean Reconstructions We can determine many of the parameters in A from knowledge about the cameras. In particular, suppose we know the projection matrix M j satisfies M j = sf j ( I2, ) R j, for some value of sf j. From (1) we have M j = W j A where W j is the j th 2 3 block in W corresponding to the same two rows as Mj occupies in M. Since R j R T j = I 3 it then follows that M j MT j = s 2 f 2 j I 2 = W j AA T W T j. (11) Here the scale factor for the j th image sf j and the 3 3 symmetric positive definite matrix Q = AA T are the only unknowns. For each j, equation (11) provides 2 linear homogeneous equations for the coefficients of Q. Then for J 3 we have 2J 6 homogeneous linear equations which we can solve for Q, up to a scalar multiple r 2 q. Finally, given Q we can factor it (assuming the eigenvalues are all nonnegative) by computing the eigenvalues, Q = U q Λ q U T q, and then recognizing A = 1 r q U q Λq 1/2 Rq T. Here r q is the unknown scale factor in Q, and R q is an arbitrary orthogonal 3 3 matrix. 253: Multi-Frame Factorization Page: 11

Euclidean Reconstruction (Cont.) Therefore we have recovered A = 1 r q K q Rq T where K q = U q Λq 1/2 is known. As a consequence we have recovered the shape matrix D r and the camera matrix M r where D = r q R q D r, for D r = K 1 q ΣV T, M = 1 r q M r R T q, for M r = WK q. (12) This is called a Euclidean reconstruction, since we have recovered the shape up to a 3D scale r q, and a rotation/reflection R q. Equivalently, this is referred to as a metric shape recovery. The ambiguity of the overall rotation R q reflects the fact that we cannot recover the orientation of the original world coordinate frame. This unknown rotation R q affects both the shape, via D = R q D r, and all of the camera matrices, via M = M r Rq T. That is, R q rotates the both the scene and the cameras together. Similarly, the ambiguity of the overall scale r q reflects the fact that we do not know the scale of the world coordinate frame. We could be imaging a tiny scene with large scale factors sf j, and we could not tell from the images alone. (Think about making the movie Titanic.) Here r q rescales the shape via D = r q D r, and also rescales all the scale parameters f j in the cameras, via M = 1 r q M r. 253: Multi-Frame Factorization Page: 12

Remaining Ambiguities The remaining ambiguity in R q is the Necker ambiguity, that is, R q could be a reflection (say R q = diag(1, 1, 1)). Effectively, with orthographic projection we cannot tell the difference between a concave-in shape viewed from the left, and the reflected concave-out shape viewed from the right. Unlike the previous two ambiguities, this ambiguity does not persist (mathematically) when we switch to perspective projection. For J = 2 orthographic views there is an additional ambiguity, known as the bas-relief ambiguity. For this ambiguity, there is an additional unknown parameter (in K q above), which ties the overall depth variation of the shape to the amount of rotation between the two cameras. See orthomassagedino.m. Refs: See the classic paper by Koenderink and van Doorn, 1991. 253: Multi-Frame Factorization Page: 13

Dino Example, Orthographic Case Dino Model Euclidean Reconstruction 1 5 1 5 Y Y 5 5 2 1 1 5 5 1 X Recovered Coord (b), True(r) Recovered Coord (b), True(r) 1 1 1 X Recovered Coord (b), True(r) 2 3D coord X 15 1 5 5 1 15 3D coord Y 5 5 1 3D coord Z 2 4 6 8 2 1 2 3 4 Point index.5 Error in estimated projection directions 15 1 2 3 4 Point index 1.4 1 1 2 3 4 Point index Scale estimation (Est(b) True(r) %Error(k)) Error in Degress.4.3.2.1 1 1.5 2 2.5 3 Image Number Scale or % Error 1.2 1.8.6.4.2 1 1.5 2 2.5 3 Image Number 253: Multi-Frame Factorization Page: 14

Introduction to Projective Reconstruction Returning to perspective projection, it is tempting to modify the bundle adjustment objective function (4) by multiplying each term in the sum by the projective depths z j,n = e T 3 M j Pn, providing a reweighted version of (4): O = j,n z j,n x j,n y j,n 1 M j Pn 2 (13) where z j,n, M j and P n are all unknowns for j = 1,...,J and n = 1,..., N. The form of (13) suggests the following factorization approach. 253: Multi-Frame Factorization Page: 15

Projective Factorization Suppose we know the projective depths z j,n, and form the data matrix C = (z j,n p j,n ). This is a 3J N matrix formed by stacking the 3- vectors z j,n p j,n in columns for the same point n, and then arranging these columns side by side for n = 1,...,N. By equation (1) we have z j,n p j,n = M j P n. Therefore C (for the correct z j,n s) must have the rank 4 factorization C = MP, (14) where M is the 3J 4 matrix formed by stacking up the camera matrices M j, and P = ( P 1,..., P N ) is the 4 N shape matrix. 253: Multi-Frame Factorization Page: 16

Iterative Projective Factorization Suppose we normalize C n = CL so that the columns have unit length (using a diagonal matrix L). Then we factor C n using SVD to form C n = WΣV T, (15) where we set all but the first 4 singular values to zero. Equivalently, we have W is 3J 4, Σ is 4 4, and V T is 4 N. We can rewrite the n th column of C n as Z n z n, where Z n is a 3J N matrix obtained from the image points p j,n and the n th weight L n,n. Here z n = (z 1,n,...,z J,n ) T, which are the projective depths for the n th point in each of the J frames. We then update z n to better match the current factorization. That is, we wish to minimize Z n z n WΣV T e n (16) wrt z n, subject to the constraint that the updated column of C n still has unit length, i.e., Z n z n = 1. Here e n is the n th standard unit vector, e n,i = δ n,i. (In projectivemassagedino.m this update of z n is done with one step along the gradient direction for this constrained optimization problem.) Once all the projective depths have been updated, we reform the normalized data matrix C n, and redo the factorization (15). This process is iterated until convergence. 253: Multi-Frame Factorization Page: 17

Projective Reconstruction Upon convergence we have a projective factorization C n = WΣV T. As in the orthographic case, this factorization is only unique up to a nonsingular matrix H. In this case, H is a 4 4, 3D homography matrix. In particular, we have the factorization, C = C n L 1 = MP with P = H 1 ΣV T L 1 (17) M = WH. Since the shape matrix P is known up to a 3D homography H, this is called a projective reconstruction. This projective reconstruction can be upgraded to a metric reconstruction by using information about the camera matrices M j to constrain the 3D homography matrix H. In order to understand this, we must first introduce the absolute dual quadric from projective geometry. 253: Multi-Frame Factorization Page: 18

Absolute Dual Quadric (Canonical Coords) The equation of a plane in 3D is m T P =, where P = (X, Y, Z, 1) T is a 3D point written in homogeneous coordinates. Imagine two planes, with coefficient vectors m 1 and m 2. Then the angle between these two planes is θ where cos(θ) = m T 1 ˆQ m 2 ( m T 1 ˆQ m 1 )( m T 2 ˆQ m 2 ). (18) Here ˆQ is the absolute dual quadric in canonical coordinates, ˆQ = I 3 T. (19) 253: Multi-Frame Factorization Page: 19

Absolute Dual Quadric (General Coords) Suppose H is any nonsingular 3D homography matrix, and consider the projective coordinates P = H P. The planes m k P = can be expressed in these new coordinates as m k P =, where m k = H T m k. We can measure the same angle between these two planes using the absolute dual quadric in general projective coordinates, namely In fact, it follows that cos(θ) = Q = H ˆQ H T. (2) ( m 1 ) T Q m 2 (( m 1 ) T Q m 1)(( m 2) T Q m 2). The general idea behind upgrading a projective reconstruction to a metric one is to use the absolute dual quadric to express known properties of the camera coordinates, such as the fact that the planes perpendicular to the X, Y, and Z axes are mutually perpendicular (i.e. cos(θ) = ). 253: Multi-Frame Factorization Page: 2

Upgrading to a Metric Reconstruction In particular, from (2) and (3) it follows that M j ˆQ M T j = fj 2 fj 2 1. (21) This is the analogue of equation (11) in the orthographic case. From equation (17) we also have M = WH, where W is known from the projective factorization. So M j ˆQ Mj T = W j H ˆQ H T Wj T = W j Q Wj T. (22) We see that, for the j th camera, (22) and (21) provide 5 linear equations for Q (6 linear equations if f j is known). So we have (at least) 5J linear constraints on Q. Since we know Q is a symmetric 4 4 matrix there are only 1 degrees of freedom to determine, and J = 2 frames are enough. (Note we also know Q has rank 3 and has non-negative eigenvalues.) 253: Multi-Frame Factorization Page: 21

Solving for H Given Q (which is symmetric, positive semi-definite, rank 3) we can compute its eigenvalue decompositon Q = U q Λ q U T q, Λ q = diag[λ 1, λ 2, λ 3, ], with λ i >. It then follows from (2) that H = U q diag[λ 1/2 1, λ 1/2 2, λ 1/2 3, 1]A, (23) where A is the matrix for a general 3D similarity transform A = R d T s Here R is a unitary matrix. The reason A remains unknown is that the absolute dual quadric ˆQ is invariant to similarity transformations. A ˆQ A T = ˆQ. This is easy to verify from the forms of A and ˆQ. Finally, equations (2, 3, 17) can be used to remove the reflection ambiguity. The only remaining ambiguities are the overall orientation, origin and scale of the world coordinate frame. 253: Multi-Frame Factorization Page: 22

Dino Example, Projective Case 12 1 Singular Values of Data Matrix Iteration 1 Iteration 5 Singular Value 8 6 4 2 5 1 15 2 25 3 Singular Value Index 5 5 1 Equivalent Projective Reconstruction 15 1 5 5 Equivalent Projective Reconstruction 15 2 4 6 8 1 12 1 Equivalent 5 Projective Reconstruction 5 1 1 3 Equivalent 2 Projective 1 Reconstruction 5 1 15 2 25 14 15 Equivalent 1 Projective 5 Reconstruction 15 1 5 5 1 15 2 3 4 2 2 4 2 Equivalent 1 Projective Reconstruction 1 25 1 1 2 3 12 1 8 6 4 2 253: Multi-Frame Factorization Page: 23

Dino Example, Projective Case Euclidean Reconstruction 1 5 Dino Model 3 2 1 Y 5 1 2 3 1 5 5 1 X 4 2 2 15 Recovered Euclidean Model (b), Ground Truth (r) 6 2 X(1) 1 5 5 1 15 1 2 3 4 Point index X(2) 4 2 2 4 6 8 1 1 2 3 4 Point index X(3) 2 4 6 8 1 2 3 4 Point index 253: Multi-Frame Factorization Page: 24

Structure from Motion The use of the theoretical rank for a set of observations provides a key insight into the structure from motion problems (see Jepson and Heeger, 1991). Consider a camera travelling through a stationary environment. Then the scene appears to move with translational velocity T and angular velocity Ω. In the camera s coordinates, the motion of any scene point X is dx dt = T + Ω X. Suppose we observe the motion field u( x k ) at K image points, { x k } K k=1, in this camera s image. Let X( x k ) be the 3D scene point associated with the k th image point x k. Then it can be shown that U T ( u T 1,..., u T K ) satisfies U = C( T) z Ω = A( T) z + B Ω. (24) Here z is a K-vector, with elements z k = 1/ X( x k ), A( T) is a 2K K matrix that depends linearly on T, and B is a 2K 3 matrix that depends only on the image points x k. Notice, for T = we have A( T) =, and (24) states that the flow field U must be in the rank 3 subspace formed by the range of the matrix B. Similarly, for nonzero T, equation (24) states that the 2K-dimensional flow field U must be in the K +3-dimensional subspace formed by the range of C( T). This range condition can be used to identify T (up to a speed ambiguity, i.e., T remains unknown) and Ω given the motion field U. Moreover, given T/ T and Ω, equation (24) can be used to solve for the inverse depths z (up to an overall scale ambiguity). 253: Multi-Frame Factorization Notes: 25

Further Readings Hartley and Zisserman, Multiple View Geometry in Computer Vision, Cambridge Univ. Press, 2. Jepson and Heeger, A fast subspace algorithm for recovering rigid motion, Proc. IEEE Workshop on Visual Motion, Princeton, 1991, pp. 124-131. Koenderink and van Doorn, Affine structure from motion, Journal of the Optical Society of America, 8(2), 1991, pp. 377-385. S. Mahamud, M. Hebert, Y. Omori and J. Ponce, Provably-Convergent Iterative Methods for Projective Structure from Motion, CVPR 21. Polleyfeys, Koch and van Gool, Self-calibration and metric reconstruction in spite of varying and unknown intrinsic camera parameters, IJCV 1999. Tomasi and Kanade, Shape and motion from image streams under orthography: A factorization approach, IJCV, Vol. 9 (No. 2), 1992, pp.137-154. 253: Multi-Frame Factorization Notes: 26