Multiview Geometry and Bundle Adjustment. CSE P576 David M. Rosen
|
|
- Jodie Spencer
- 5 years ago
- Views:
Transcription
1 Multiview Geometry and Bundle Adjustment CSE P576 David M. Rosen 1
2 Recap Previously: Image formation Feature extraction + matching Two-view (epipolar geometry) Today: Add some geometry, statistics, optimization Turn it up to 11 N! 2
3 Motivating example: Photogrammetry The science of measurement using cameras 3
4 Application: Remote Sensing Mars Reconnaissance Orbiter Launched 12 Aug 2005 Entered orbit 10 Mar 2006 ~ 112 minute orbital period ~ 12.8 orbits / (Earth) day Sensors: High Resolution Imaging Science Experiment (HiRISE) Context Camera Mars Color Imager Image credit: NASA/JPL 4
5 Image credit: NASA/JPL-Caltech/MSSS 5
6 Image credit: NASA/JPL-Caltech/MSSS 6
7 Image credit: NASA/JPL/University of Arizona/USGS 7
8 Image credit: NASA / JPL-Caltech / UA / Kevin M. Gill 8
9 Image credit: NASA/JPL/University of Arizona/USGS 9
10 Image credit: NASA / JPL-Caltech / UA / Kevin M. Gill 10
11 Video credit: NASA / JPL / AU / Doug Ellison 11
12 Application: 3D Reconstruction Goal: Build a 3D model of a scene from a collection of images [S. Agarwal et al., Building Rome in a Day, Communications of the ACM, 2011] 12
13 13
14 14
15 15
16 Application: Robotics Video credit: MIT DRC team 16
17 Video credit: MIT DRC team 17
18 Video credit: MIT DRC team 18
19 19
20 In this lecture Photogrammetry: The problem of measurement using imagery Maximum-likelihood estimation and bundle adjustment: Solving the photogrammetry problem Practicalities Problem scale Robust estimation Representation of rotations 20
21 Photogrammetry: The Problem Given: A collection of images Estimate: 3D positions of imaged points Poses of the imaging cameras Intrinsic parameters of the imaging cameras 21
22 Photogrammetry: Generative model Q: How are variables of estimation: 3D point positions p j Camera poses x i = t i, R i Camera intrinsics K i related to images? p 1 u 11 u 21 u 31 u ij = f(x i, K i, p j ) x 1,K 1 x 3,K 3 where f is the camera projection function (from Lecture 1) 22 x 2,K 2
23 Photogrammetry: Estimation procedure Main idea: Given a set of images 1. Extract features 2. Match features (identify the set of 3D points) 3. Estimate parameters so that: u ij = f(x i, K i, p j ) for all point projections u ij x 1,K 1 x 2,K 2 x 3,K 3 23
24 The problem of measurement noise We want to find x i, K i, p j so that u ij = f(x i, K i, p j ) (i.e. our model matches the data) for all u ij. But: All real-world measurements have errors What we actually measure is: u ij = f x i, K i, p j + ε ij We cannot find parameters x i, K i, p j that fit the measured projections u ij exactly 24
25 Example: Linear regression 25
26 Maximum likelihood estimation MLE is a method for fitting parameters to noisy data y = y 1,, y n, given a sampling model y~p( θ). Basic idea: Choose the that best fits the data y. But: How can we measure goodness of fit? Likelihood function: L θ p y θ. Measures how likely the data y is for each choice of. MLE principle: Best fit Maximum likelihood Pick the that maximizes the likelihood of data y: መθ = max θ L θ 26
27 Example: Regression under Gaussian noise Consider fitting a function f ; θ to data D = y i = f x i ; θ + ε i, (x i, y i ) N i=1 ε i ~N 0, Σ i under the model: For each choice of, for each (x i, y i ): The pdf for ε i is: ε i = y i f x i ; θ p ε i = det(2πσ i ) 1 2exp 1 2 ε i T Σ i 1 ε i The likelihood of the ith data point (x i, y i ) is: p x i, y i θ = det(2πσ i ) 1 2exp 1 2 (y i f x i ; θ ) T Σ 1 i (y i f(x i ; θ)) The likelihood for the entire dataset D is: N L(D θ) exp 1 2 y i f(x i ; θ) T Σ 1 i y i f(x i ; θ) i=1 ε i 27
28 Example: Regression under Gaussian noise Consider fitting a function f ; θ to data D = y i = f x i ; θ + ε i, (x i, y i ) N i=1 ε i ~N 0, Σ i under the model: The likelihood for the entire dataset D is: Taking the logarithm: N L(D θ) exp 1 2 y i f(x i ; θ) T Σ 1 i y i f(x i ; θ) i=1 N log L(D θ) = c 1 2 i=1 y i f x i ; θ MLE under additive Gaussian noise is a nonlinear least-squares problem: 2 Σi θ = min θ N i=1 2 y i f(x i ; θ) Σi ε i 28
29 Exercise: Linear regression Consider fitting a linear function to the data: x y under the model: y i = ax i + b + ε i, ε i ~N(0,.35 2 ) 29
30 Exercise: Linear regression x y Model: y i = ax i + b Estimated: a = 1.73 b = 1.06 True: a = 1.70 b =
31 Bundle adjustment Recall: Given a set of point projections u ij, we want to estimate: 3D point positions p j camera poses x i camera intrinsics K i p j Assuming the measurement model: u ij = f x i, K i, p j + ε ij, ε i ~N 0, Σ i Maximum-likelihood estimation is then: x i, K i, pƹ j = min x i,k i,p j i,j u ij f(x i, K i, p j ) Σij 2 x 1,K 1 x 2,K 2 x 3,K 3 Minimize (weighted) reprojection error 31
32 Photogrammetry and Bundle Adjustment Given: A set of images 1. Extract features 2. Match features (identify 3D points) 3. Bundle adjust (minimize reprojection error): p j x i, K i, pƹ j = min x i,k i,p j i,j u ij f(x i, K i, p j ) Σij 2 x 1,K 1 x 2,K 2 x 3,K 3 32
33 Special Case: Perspective-n-Point (PnP) Given: Known point positions p j Known camera intrinsics K Estimate: Camera pose x = (R,t) x = min x N j=1 2 u j f(x, K, p j ) Σj Image credit: OpenCV 33
34 Special Case: Camera calibration Given: Known point positions p j (on calibration object) Estimate: Camera poses x i Intrinsic parameters K x i, K = min x i,k i,j u ij f(x i, K, p j ) Σij 2 Image credit: OpenCV 34
35 Given: A pair of cameras with: Known poses x 1, x 2 Known intrinsics K 1, K 2 Special Case: Stereovision Estimate: 3D point positions p j 2 pƹ j = min σ p j u 1j f(x 1, K 1, p j ) j Σ1j + u 2j f(x 2, K 2, p j ) Σ2j 2 Image credit: MIT Space Systems Laboratory pƹ j = min p j independently for all j u 1j f(x 1, K 1, p j ) Σ1j 2 + u 2j f(x 2, K 2, p j ) Σ2j 2 35
36 Practicality: Problem Scale Piazza San Marco reconstruction: ~14,000 images ~4.5 million points Assuming each point is observed 20x, what is the size of the BA problem? 36
37 Practicality: Problem Scale Camera variables (0-skew, equal pixel scaling): 14, = 126,000 Point variables: 4,500,000 3 = 13,500,000 Point observations: 4,500, = 180,000,000 Totals: 13,626,000-dimensional state vector 180,000,000-dimensional residual vector This is a huge optimization problem! 37
38 Practicality: Feature mismatches Recall: We construct the bundle adjustment problem: x i, K i, pƹ j = min x i,k i,p j i,j u ij f(x i, K i, p j ) Σij 2 Image credit: Tian Zhou using estimated feature matches. But: What happens if these are mis-estimated? 38
39 Example: Contaminated linear regression Consider fitting the linear model: y i = ax i + ε i, ε i ~N 0,.35 2 to a contaminated data set x y
40 Exercise: Contaminated linear regression x y Model: y i = ax i + b Estimated: a = 1.37 b = 1.58 True: a = 1.70 b =
41 The problem of outliers Recall: We assumed additive Gaussian image noise: u ij = f x i, K i, p j + ε ij, ε ij ~N(0, Σ ij ) and obtained a nonlinear least-squares problem: x i, K i, pƹ j = min x i,k i,p j i,j u ij f(x i, K i, p j ) Σij 2 ε i NB: This loss weights extreme errors very heavily. This estimator is not robust! ε i 41
42 Robust loss functions One solution: Replace quadratic loss with a function that attenuates gross errors Tradeoff: More robust to outliers (Slightly) less statistical power Quadratic Huber Cauchy Geman-McClure 42
43 Example: Contaminated linear regression Least-squares loss Huber loss 43
44 Photogrammetry and Bundle Adjustment: Summary Given: A set of images 1. Extract features 2. Match features (identify 3D points) 3. Bundle adjust using a robust loss function : p j x i, K i, pƹ j = min x i,k i,p j i,j ρ u ij f(x i, K i, p j ) Σij x 1,K 1 x 2,K 2 x 3,K 3 44
45 Practicality: Representing rotations So far, we ve represented rotations using rotation matrices: R = r 11 r 12 r 13 r 21 r 31 r 22 r 32 r 23, r 33 R T R = I 3, det R = +1 Pro: Trivial point operations Con: Over-parameterized Not super convenient for optimization (requires constraints) 45
46 Euler s Rotation Theorem Theorem: Every rotation of 3D space has a fixed axis e. We can describe a rotation using: An axis (unit vector e) (Right-handed) rotation angle This is the axis-angle parameterization of rotations. Can also combine these into a single axis-angle vector: θ = θe 46
47 Rodrigues Formula How does a rotation parameterized as θ = θe act on points? Rodrigues formula: Given a vector v, v rot = v cos θ + sin θ e v + 1 cos θ e v e NB: Any 3D vector θ determinates a valid rotation Rodrigues formula is differentiable in θ Axis-angle is much more convenient for use in optimization! 47
48 Rodrigues Formula How does a rotation parameterized as θ = θe act on points? Rodrigues formula: Given a vector v, v rot = v cos θ + sin θ e v + 1 cos θ e v e Matrix form: where R θ = I + (sin θ) E + 1 cos θ E 2, E = 0 e 3 e 2 e 3 0 e 1 e 2 e
49 Photogrammetry and Bundle Adjustment: Summary Given: A set of images 1. Extract features 2. Match features (identify 3D points) 3. Bundle adjust using a robust loss function : p j x i, K i, pƹ j = min x i,k i,p j i,j ρ u ij f(x i, K i, p j ) Σij x 1,K 1 x 2,K 2 x 3,K 3 49
Augmented Reality VU numerical optimization algorithms. Prof. Vincent Lepetit
Augmented Reality VU numerical optimization algorithms Prof. Vincent Lepetit P3P: can exploit only 3 correspondences; DLT: difficult to exploit the knowledge of the internal parameters correctly, and the
More informationMachine Learning
Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 1, 2011 Today: Generative discriminative classifiers Linear regression Decomposition of error into
More informationCamera calibration Triangulation
Triangulation Perspective projection in homogenous coordinates ~x img I 0 apple R t 0 T 1 ~x w ~x img R t ~x w Matrix transformations in 2D ~x img K R t ~x w K = 2 3 1 0 t u 40 1 t v 5 0 0 1 Translation
More informationSimultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors
Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,
More informationNumerical computation II. Reprojection error Bundle adjustment Family of Newtonʼs methods Statistical background Maximum likelihood estimation
Numerical computation II Reprojection error Bundle adjustment Family of Newtonʼs methods Statistical background Maximum likelihood estimation Reprojection error Reprojection error = Distance between the
More informationParameterizing the Trifocal Tensor
Parameterizing the Trifocal Tensor May 11, 2017 Based on: Klas Nordberg. A Minimal Parameterization of the Trifocal Tensor. In Computer society conference on computer vision and pattern recognition (CVPR).
More informationVision par ordinateur
Vision par ordinateur Géométrie épipolaire Frédéric Devernay Avec des transparents de Marc Pollefeys Epipolar geometry π Underlying structure in set of matches for rigid scenes C1 m1 l1 M L2 L1 l T 1 l
More informationPose Tracking II! Gordon Wetzstein! Stanford University! EE 267 Virtual Reality! Lecture 12! stanford.edu/class/ee267/!
Pose Tracking II! Gordon Wetzstein! Stanford University! EE 267 Virtual Reality! Lecture 12! stanford.edu/class/ee267/!! WARNING! this class will be dense! will learn how to use nonlinear optimization
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 4, 2015 Today: Generative discriminative classifiers Linear regression Decomposition of error into
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationHomework 2 Solutions Kernel SVM and Perceptron
Homework 2 Solutions Kernel SVM and Perceptron CMU 1-71: Machine Learning (Fall 21) https://piazza.com/cmu/fall21/17115781/home OUT: Sept 25, 21 DUE: Oct 8, 11:59 PM Problem 1: SVM decision boundaries
More informationGAUSSIAN PROCESS REGRESSION
GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The
More informationNonlinear Programming Models
Nonlinear Programming Models Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Nonlinear Programming Models p. Introduction Nonlinear Programming Models p. NLP problems minf(x) x S R n Standard form:
More informationRigid Structure from Motion from a Blind Source Separation Perspective
Noname manuscript No. (will be inserted by the editor) Rigid Structure from Motion from a Blind Source Separation Perspective Jeff Fortuna Aleix M. Martinez Received: date / Accepted: date Abstract We
More informationStructure from Motion. CS4670/CS Kevin Matzen - April 15, 2016
Structure from Motion CS4670/CS5670 - Kevin Matzen - April 15, 2016 Video credit: Agarwal, et. al. Building Rome in a Day, ICCV 2009 Roadmap What we ve seen so far Single view modeling (1 camera) Stereo
More informationECE521 lecture 4: 19 January Optimization, MLE, regularization
ECE521 lecture 4: 19 January 2017 Optimization, MLE, regularization First four lectures Lectures 1 and 2: Intro to ML Probability review Types of loss functions and algorithms Lecture 3: KNN Convexity
More informationConvex Optimization: Applications
Convex Optimization: Applications Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh Convex Optimization 1-75/36-75 Based on material from Boyd, Vandenberghe Norm Approximation minimize Ax b (A R m
More informationBayesian Linear Regression [DRAFT - In Progress]
Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory
More informationCSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1
CSCI5654 (Linear Programming, Fall 2013) Lectures 10-12 Lectures 10,11 Slide# 1 Today s Lecture 1. Introduction to norms: L 1,L 2,L. 2. Casting absolute value and max operators. 3. Norm minimization problems.
More informationWeek 5: Logistic Regression & Neural Networks
Week 5: Logistic Regression & Neural Networks Instructor: Sergey Levine 1 Summary: Logistic Regression In the previous lecture, we covered logistic regression. To recap, logistic regression models and
More informationLearning from Data: Regression
November 3, 2005 http://www.anc.ed.ac.uk/ amos/lfd/ Classification or Regression? Classification: want to learn a discrete target variable. Regression: want to learn a continuous target variable. Linear
More informationModeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop
Modeling Data with Linear Combinations of Basis Functions Read Chapter 3 in the text by Bishop A Type of Supervised Learning Problem We want to model data (x 1, t 1 ),..., (x N, t N ), where x i is a vector
More informationA Practical Method for Decomposition of the Essential Matrix
Applied Mathematical Sciences, Vol. 8, 2014, no. 176, 8755-8770 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.410877 A Practical Method for Decomposition of the Essential Matrix Georgi
More informationStatistics 203: Introduction to Regression and Analysis of Variance Course review
Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More informationMassachusetts Institute of Technology
Massachusetts Institute of Technology 6.867 Machine Learning, Fall 2006 Problem Set 5 Due Date: Thursday, Nov 30, 12:00 noon You may submit your solutions in class or in the box. 1. Wilhelm and Klaus are
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationy Xw 2 2 y Xw λ w 2 2
CS 189 Introduction to Machine Learning Spring 2018 Note 4 1 MLE and MAP for Regression (Part I) So far, we ve explored two approaches of the regression framework, Ordinary Least Squares and Ridge Regression:
More informationLoss Functions and Optimization. Lecture 3-1
Lecture 3: Loss Functions and Optimization Lecture 3-1 Administrative Assignment 1 is released: http://cs231n.github.io/assignments2017/assignment1/ Due Thursday April 20, 11:59pm on Canvas (Extending
More informationGradient Descent. Ryan Tibshirani Convex Optimization /36-725
Gradient Descent Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: canonical convex programs Linear program (LP): takes the form min x subject to c T x Gx h Ax = b Quadratic program (QP): like
More informationSupport Vector Machines and Bayes Regression
Statistical Techniques in Robotics (16-831, F11) Lecture #14 (Monday ctober 31th) Support Vector Machines and Bayes Regression Lecturer: Drew Bagnell Scribe: Carl Doersch 1 1 Linear SVMs We begin by considering
More informationUnsupervised Learning: Dimensionality Reduction
Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationOptimisation on Manifolds
Optimisation on Manifolds K. Hüper MPI Tübingen & Univ. Würzburg K. Hüper (MPI Tübingen & Univ. Würzburg) Applications in Computer Vision Grenoble 18/9/08 1 / 29 Contents 2 Examples Essential matrix estimation
More informationLinear Regression. Machine Learning CSE546 Kevin Jamieson University of Washington. Oct 5, Kevin Jamieson 1
Linear Regression Machine Learning CSE546 Kevin Jamieson University of Washington Oct 5, 2017 1 The regression problem Given past sales data on zillow.com, predict: y = House sale price from x = {# sq.
More informationMasters Comprehensive Examination Department of Statistics, University of Florida
Masters Comprehensive Examination Department of Statistics, University of Florida May 6, 003, 8:00 am - :00 noon Instructions: You have four hours to answer questions in this examination You must show
More informationLecture 8: Signal Detection and Noise Assumption
ECE 830 Fall 0 Statistical Signal Processing instructor: R. Nowak Lecture 8: Signal Detection and Noise Assumption Signal Detection : X = W H : X = S + W where W N(0, σ I n n and S = [s, s,..., s n ] T
More informationTracking for VR and AR
Tracking for VR and AR Hakan Bilen November 17, 2017 Computer Graphics University of Edinburgh Slide credits: Gordon Wetzstein and Steven M. La Valle 1 Overview VR and AR Inertial Sensors Gyroscopes Accelerometers
More informationLinear Regression. Aarti Singh. Machine Learning / Sept 27, 2010
Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X
More informationSpring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier
Spring 0 CIS 55 Fundamentals of Linear Algebra and Optimization Jean Gallier Homework 5 & 6 + Project 3 & 4 Note: Problems B and B6 are for extra credit April 7 0; Due May 7 0 Problem B (0 pts) Let A be
More informationIntroduction to Machine Learning
1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer
More informationCPSC 340: Machine Learning and Data Mining
CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released
More informationCSC321 Lecture 20: Reversible and Autoregressive Models
CSC321 Lecture 20: Reversible and Autoregressive Models Roger Grosse Roger Grosse CSC321 Lecture 20: Reversible and Autoregressive Models 1 / 23 Overview Four modern approaches to generative modeling:
More informationLinear classifiers: Overfitting and regularization
Linear classifiers: Overfitting and regularization Emily Fox University of Washington January 25, 2017 Logistic regression recap 1 . Thus far, we focused on decision boundaries Score(x i ) = w 0 h 0 (x
More informationGaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008
Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:
More informationMethods in Computer Vision: Introduction to Matrix Lie Groups
Methods in Computer Vision: Introduction to Matrix Lie Groups Oren Freifeld Computer Science, Ben-Gurion University June 14, 2017 June 14, 2017 1 / 46 Definition and Basic Properties Definition (Matrix
More informationRobust Camera Location Estimation by Convex Programming
Robust Camera Location Estimation by Convex Programming Onur Özyeşil and Amit Singer INTECH Investment Management LLC 1 PACM and Department of Mathematics, Princeton University SIAM IS 2016 05/24/2016,
More informationLinear Regression (9/11/13)
STA561: Probabilistic machine learning Linear Regression (9/11/13) Lecturer: Barbara Engelhardt Scribes: Zachary Abzug, Mike Gloudemans, Zhuosheng Gu, Zhao Song 1 Why use linear regression? Figure 1: Scatter
More informationMachine Learning! in just a few minutes. Jan Peters Gerhard Neumann
Machine Learning! in just a few minutes Jan Peters Gerhard Neumann 1 Purpose of this Lecture Foundations of machine learning tools for robotics We focus on regression methods and general principles Often
More informationIntroduction to Machine Learning
1, DATA11002 Introduction to Machine Learning Lecturer: Antti Ukkonen TAs: Saska Dönges and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer,
More information1 Kalman Filter Introduction
1 Kalman Filter Introduction You should first read Chapter 1 of Stochastic models, estimation, and control: Volume 1 by Peter S. Maybec (available here). 1.1 Explanation of Equations (1-3) and (1-4) Equation
More informationCPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017
CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class
More informationMultiple View Geometry in Computer Vision
Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Scene Planes & Homographies Lecture 19 March 24, 2005 2 In our last lecture, we examined various
More informationWHEN IS A MAXIMAL INVARIANT HYPOTHESIS TEST BETTER THAN THE GLRT? Hyung Soo Kim and Alfred O. Hero
WHEN IS A MAXIMAL INVARIANT HYPTHESIS TEST BETTER THAN THE GLRT? Hyung Soo Kim and Alfred. Hero Department of Electrical Engineering and Computer Science University of Michigan, Ann Arbor, MI 489-222 ABSTRACT
More informationConvex Optimization M2
Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial
More informationRobust and accurate inference via a mixture of Gaussian and terrors
Robust and accurate inference via a mixture of Gaussian and terrors Hyungsuk Tak ICTS/SAMSI Time Series Analysis for Synoptic Surveys and Gravitational Wave Astronomy 22 March 2017 Joint work with Justin
More informationM3: Multiple View Geometry
M3: Multiple View Geometry L18: Projective Structure from Motion: Iterative Algorithm based on Factorization Based on Sections 13.4 C. V. Jawahar jawahar-at-iiit.net Mar 2005: 1 Review: Reconstruction
More informationMidterm. Introduction to Machine Learning. CS 189 Spring Please do not open the exam before you are instructed to do so.
CS 89 Spring 07 Introduction to Machine Learning Midterm Please do not open the exam before you are instructed to do so. The exam is closed book, closed notes except your one-page cheat sheet. Electronic
More informationWeek 3: Linear Regression
Week 3: Linear Regression Instructor: Sergey Levine Recap In the previous lecture we saw how linear regression can solve the following problem: given a dataset D = {(x, y ),..., (x N, y N )}, learn to
More information2D Image Processing. Bayes filter implementation: Kalman filter
2D Image Processing Bayes filter implementation: Kalman filter Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationConvex Optimization in Computer Vision:
Convex Optimization in Computer Vision: Segmentation and Multiview 3D Reconstruction Yiyong Feng and Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) ELEC 5470 - Convex Optimization
More informationThe Multibody Trifocal Tensor: Motion Segmentation from 3 Perspective Views
The Multibody Trifocal Tensor: Motion Segmentation from 3 Perspective Views Richard Hartley 1,2 and RenéVidal 2,3 1 Dept. of Systems Engineering 3 Center for Imaging Science Australian National University
More informationCOMS 4771 Lecture Course overview 2. Maximum likelihood estimation (review of some statistics)
COMS 4771 Lecture 1 1. Course overview 2. Maximum likelihood estimation (review of some statistics) 1 / 24 Administrivia This course Topics http://www.satyenkale.com/coms4771/ 1. Supervised learning Core
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationASTRONOMY. Chapter 7 OTHER WORLDS: AN INTRODUCTION TO THE SOLAR SYSTEM PowerPoint Image Slideshow
ASTRONOMY Chapter 7 OTHER WORLDS: AN INTRODUCTION TO THE SOLAR SYSTEM PowerPoint Image Slideshow FIGURE 7.1 Self-Portrait of Mars. This picture was taken by the Curiosity Rover on Mars in 2012. The image
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationCamera Models and Affine Multiple Views Geometry
Camera Models and Affine Multiple Views Geometry Subhashis Banerjee Dept. Computer Science and Engineering IIT Delhi email: suban@cse.iitd.ac.in May 29, 2001 1 1 Camera Models A Camera transforms a 3D
More informationMachine learning - HT Maximum Likelihood
Machine learning - HT 2016 3. Maximum Likelihood Varun Kanade University of Oxford January 27, 2016 Outline Probabilistic Framework Formulate linear regression in the language of probability Introduce
More informationCSC 411: Lecture 09: Naive Bayes
CSC 411: Lecture 09: Naive Bayes Class based on Raquel Urtasun & Rich Zemel s lectures Sanja Fidler University of Toronto Feb 8, 2015 Urtasun, Zemel, Fidler (UofT) CSC 411: 09-Naive Bayes Feb 8, 2015 1
More informationReview of Maximum Likelihood Estimators
Libby MacKinnon CSE 527 notes Lecture 7, October 7, 2007 MLE and EM Review of Maximum Likelihood Estimators MLE is one of many approaches to parameter estimation. The likelihood of independent observations
More informationMachine Learning and Computational Statistics, Spring 2017 Homework 2: Lasso Regression
Machine Learning and Computational Statistics, Spring 2017 Homework 2: Lasso Regression Due: Monday, February 13, 2017, at 10pm (Submit via Gradescope) Instructions: Your answers to the questions below,
More informationMaximum Likelihood, Logistic Regression, and Stochastic Gradient Training
Maximum Likelihood, Logistic Regression, and Stochastic Gradient Training Charles Elkan elkan@cs.ucsd.edu January 17, 2013 1 Principle of maximum likelihood Consider a family of probability distributions
More informationModern Methods of Data Analysis - WS 07/08
Modern Methods of Data Analysis Lecture VIc (19.11.07) Contents: Maximum Likelihood Fit Maximum Likelihood (I) Assume N measurements of a random variable Assume them to be independent and distributed according
More informationNeural Network Training
Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationLecture 6: Linear Regression
Lecture 6: Linear Regression Reading: Sections 3.1-3 STATS 202: Data mining and analysis Jonathan Taylor, 10/5 Slide credits: Sergio Bacallado 1 / 30 Simple linear regression Model: y i = β 0 + β 1 x i
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationLocalization of Radioactive Sources Zhifei Zhang
Localization of Radioactive Sources Zhifei Zhang 4/13/2016 1 Outline Background and motivation Our goal and scenario Preliminary knowledge Related work Our approach and results 4/13/2016 2 Background and
More informationMotion Models (cont) 1 2/15/2012
Motion Models (cont 1 Odometry Motion Model the key to computing p( xt ut, xt 1 for the odometry motion model is to remember that the robot has an internal estimate of its pose θ x t 1 x y θ θ true poses
More informationMysteries of Parameterizing Camera Motion - Part 2
Mysteries of Parameterizing Camera Motion - Part 2 Instructor - Simon Lucey 16-623 - Advanced Computer Vision Apps Today Gauss-Newton and SO(3) Alternative Representations of SO(3) Exponential Maps SL(3)
More informationCovariance Matrix Simplification For Efficient Uncertainty Management
PASEO MaxEnt 2007 Covariance Matrix Simplification For Efficient Uncertainty Management André Jalobeanu, Jorge A. Gutiérrez PASEO Research Group LSIIT (CNRS/ Univ. Strasbourg) - Illkirch, France *part
More informationOverfitting, Bias / Variance Analysis
Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic
More informationIntroduction to Machine Learning
Introduction to Machine Learning Linear Regression Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1
More informationReliability Monitoring Using Log Gaussian Process Regression
COPYRIGHT 013, M. Modarres Reliability Monitoring Using Log Gaussian Process Regression Martin Wayne Mohammad Modarres PSA 013 Center for Risk and Reliability University of Maryland Department of Mechanical
More informationDiscriminative Learning and Big Data
AIMS-CDT Michaelmas 2016 Discriminative Learning and Big Data Lecture 2: Other loss functions and ANN Andrew Zisserman Visual Geometry Group University of Oxford http://www.robots.ox.ac.uk/~vgg Lecture
More informationConstrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.
Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization
More informationPartitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12
Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following
More information3D from Photographs: Camera Calibration. Dr Francesco Banterle
3D from Photographs: Camera Calibration Dr Francesco Banterle francesco.banterle@isti.cnr.it 3D from Photographs Automatic Matching of Images Camera Calibration Photographs Surface Reconstruction Dense
More informationAugmented Reality VU Camera Registration. Prof. Vincent Lepetit
Augmented Reality VU Camera Registration Prof. Vincent Lepetit Different Approaches to Vision-based 3D Tracking [From D. Wagner] [From Drummond PAMI02] [From Davison ICCV01] Consider natural features Consider
More informationL11. EKF SLAM: PART I. NA568 Mobile Robotics: Methods & Algorithms
L11. EKF SLAM: PART I NA568 Mobile Robotics: Methods & Algorithms Today s Topic EKF Feature-Based SLAM State Representation Process / Observation Models Landmark Initialization Robot-Landmark Correlation
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 2, 2015 Today: Logistic regression Generative/Discriminative classifiers Readings: (see class website)
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationInf2b Learning and Data
Inf2b Learning and Data Lecture : Single layer Neural Networks () (Credit: Hiroshi Shimodaira Iain Murray and Steve Renals) Centre for Speech Technology Research (CSTR) School of Informatics University
More informationUCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011
UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and
More information