Motion Estimation (I) Ce Liu Microsoft Research New England
|
|
- Ann Wilcox
- 6 years ago
- Views:
Transcription
1 Motion Estimation (I) Ce Liu Microsoft Research New England
2 We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives
3 Motion estimation: a core problem of Related topics: computer vision Image correspondence, image registration, image matching, image alignment, Applications Video enhancement: stabilization, denoising, super resolution 3D reconstruction: structure from motion (SFM) Video segmentation Tracking/recognition Advanced video editing (label propagation)
4 Contents (today) Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
5 Contents (next time) Discrete optical flow Layer motion analysis Contour motion analysis Obtaining motion ground truth SIFT flow: generalized optical flow Applications (2)
6 Readings Rick s book: Chapter 8 Ce Liu s PhD thesis (appendix A & B) S. Baker and I. Matthews. Lucas-Kanade 20 years on: a unifying framework. IJCV 2004 Horn-Schunck (wikipedia) A. Bruhn, J. Weickert, C. Schnorr. Lucas/Kanade meets Horn/Schunk: combining local and global optical flow methods. IJCV 2005
7 Contents Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
8 Seeing motion from a static picture?
9 More examples
10 How is this possible? The true mechanism is to be revealed FMRI data suggest that illusion is related to some component of eye movements We don t expect computer vision to see motion from these stimuli, yet
11 What do you see?
12 In fact,
13 We still don t touch these areas
14 Motion analysis: human vs. computer Computers can only analyze motion for opaque and solid objects Challenges: Shapeless or transparent scenes Key: motion representation
15 Contents Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
16 Motion forms Mapping: x 1, y 1 (x 2, y 2 ) Global parametric motion: x 2, y 2 Motion types Translation: x 2 y 2 Similarity: x 2 y 2 Affine: x 2 y 2 = x 1 + a y 1 + b = s Homography: x 2 y 2 cos α sin α = ax 1 + by 1 + c dx 1 + ey 1 + f = 1 z sin α cos α = f(x 1, y 1 ; θ) x 1 + a y 1 + b ax 1 + by 1 + c dx 1 + ey 1 + f, z = gx 1 + y 1 + i
17 Illustration of motion types Translation
18 Optical flow field Parametric motion is limited and cannot describe the motion of arbitrary videos Optical flow field: assign a flow vector u x, y, v x, y to each pixel (x, y) Projection from 3D world to 2D
19 Optical flow field visualization Too messy to plot flow vector for every pixel Map flow vector to color Magnitude: saturation Orientation: hue Input Ground-truth flow field Visualization code [Baker et al. 2007]
20 Matching criterion Brightness constancy assumption I 1 x, y = I 2 x + u, y + v + r + g r N 0, σ 2, g U 1,1 Noise r, outlier g (occlusion, lighting change) Matching criteria What s invariant between two images? Brightness, gradients, phase, other features Distance metric (L2, L1, truncated L1, Lorentzian) E u, v = x,y ρ I 1 x, y I 2 x + u, y + v Correlation, normalized cross correlation (NCC)
21 Error functions L2 norm ρ z = z L1 norm ρ z = z Truncated L1 norm ρ z = min ( z, η) Lorentzian ρ z = log (1 + γz 2 )
22 Robust statistics Traditional L2 norm: only noise, no outlier Example: estimate the average of 0.95, 1.04, 0.91, 1.02, 1.10, Estimate with minimum error z = arg min z L2 norm: z = L1 norm: z = i ρ z z i Truncated L1: z = Lorentzian: z = L2 norm ρ z = z Truncated L1 norm ρ z = min ( z, η) L1 norm ρ z = z Lorentzian ρ z = log (1 + γz 2 )
23 Contents Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
24 Lucas-Kanade: problem setup Given two images I 1 (x, y) and I 2 (x, y), estimate a parametric motion that transforms I 1 to I 2 Let x = x, y T be a column vector indexing pixel coordinate Two typical transforms Translation: W x; p = x + p 1 y + p 2 Affine: W x; p = p 1x + p 2 y + p 3 p 4 x + p 5 y + p = p 1 p 2 p x 3 6 p 4 p 5 p y 6 1 Goal of the Lucas-Kanade algorithm p = arg min p I 2 W x; p I 1 x 2 x
25 An incremental algorithm Difficult to directly optimize the objective function p = arg min p I 2 W x; p I 1 x 2 x Instead, we try to optimize each step Δp = arg min Δp The transform parameter is updated: x I 2 W x; p + Δp I 1 x 2 p p + Δp
26 Taylor expansion The term I 2 W x; p + Δp is highly nonlinear Taylor expansion: I 2 W x; p + Δp I 2 W x; p + I 2 W p Δp W : Jacobian of the warp p If W x; p = W x x; p, W y x; p T, then W p = W x W x p 1 p n W y W y p 1 p n
27 Jacobian matrix For affine transform: W x; p = p 1 p 2 p x 3 p 4 p 5 p y 6 1 The Jacobian is W = x y p x 0 0 y 1 For translation : W x; p = x + p 1 y + p 2 The Jacobian is W p =
28 Taylor expansion I 2 = I x I y is the gradient of image I 2 evaluated at W(x; p): compute the gradients in the coordinate of I 2 and warp back to the coordinate of I 1 For affine transform W = x y p x 0 0 y 1 W I 2 p = I x x I x y I x I y x I y y I y Let matrix B = [I x X I x Y I x I y X I y Y I y ] R n 6, I x and X are both column vectors. I x X is element-wise vector multiplication.
29 Gauss-Newton With Taylor expansion, the objective function becomes Δp = arg min Δp x Or in a vector form: I 2 W x; p + I 2 W p Δp I 1 x 2 Δp = arg min Δp I t + BΔp T (I t + BΔp) Where B = [I x X I x Y I x I y X I y Y I y ] R n 6 I t = I 2 W p I 1 Solution: Δp = B T B 1 B T I t
30 Jacobian: W p = I 2 W p = I x I y B = [I x I y ] R n 2 Solution: Translation Δp = B T B 1 B T I t = I x T I x I T x I y I T x I y I T y I y 1 Ix T I t I y T I t
31 How it works
32 Coarse-to-fine refinement Lucas-Kanade is a greedy algorithm that converges to local minimum Initialization is crucial: if initialized with zero, then the underlying motion must be small If underlying transform is significant, then coarse-to-fine is a must (u 2, v 2 ) 2 Smooth & downsampling (u 1, v 1 ) 2 (u, v)
33 Variations Variations of Lucas Kanade: Additive algorithm [Lucas-Kanade, 81] Compositional algorithm [Shum & Szeliski, 98] Inverse compositional algorithm [Baker & Matthews, 01] Inverse additive algorithm [Hager & Belhumeur, 98] Although inverse algorithms run faster (avoiding recomputing Hessian), they have the same complexity for robust error functions!
34 From parametric motion to flow field Incremental flow update (du, dv) for pixel x, y I 2 x + u + du, y + v + dv I 1 x, y = I 2 x + u, y + v + I x x + u, y + v du + I y x + u, y + v dv I 1 x, y I x du + I y dv + I t = 0 We obtain the following function within a patch T I x du dv = I x I T x I y I x T I y I y T I y 1 Ix T I t I y T I t The flow vector of each pixel is updated independently Median filtering can be applied for spatial smoothness
35 Example Input two frames Coarse-to-fine LK Flow visualization Coarse-to-fine LK with median filtering
36 Contents Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
37 Motion ambiguities When will the Lucas-Kanade algorithm fail? T I x du dv = I x I T x I y I x T I y I y T I y 1 Ix T I t I y T I t The inverse may not exist!!! How? All the derivatives are zero: flat regions X- and y- derivatives are linearly correlated: lines
38 The aperture problem Corners Lines Flat regions
39 Dense optical flow with spatial regularity Local motion is inherently ambiguous Corners: definite, no ambiguity Lines: definite along the normal, ambiguous along the tangent Flat regions: totally ambiguous Solution: imposing spatial smoothness to the flow field Adjacent pixels should move together as much as possible Horn & Schunck equation u, v = arg min I x u + I y v + I t 2 + α u 2 + v 2 dxdy u 2 = u 2 u 2 + = ux 2 + u2 x y y α: smoothness coefficient Data term Smoothness term
40 2D Euler Lagrange 2D Euler Lagrange: the functional S = L x, y, f, f x, f y dxdy Ω is minimized only if f satisfies the partial differential equation (PDE) In Horn-Schunck L f x L f x y L f y = 0 L u, v, u x, u y, v x, v y = I x u + I y v + I t 2 + α ux 2 + u y 2 + v x 2 + v y 2 L u = 2 I xu + I y v + I t I x L u x = 2αu x, x L = 2αu u xx, L = 2αu x u y, y y L u y = 2αu yy
41 Linear PDE The Euler-Lagrange PDE for Horn-Schunck is I x u + I y v + I t I x α u xx + u yy = 0 I x u + I y v + I t I y α v xx + v yy = 0 u xx + u yy can be obtained by a Laplacian operator: In the end, we solve a large linear system I x 2 + αl I x I y I x I y I 2 y + αl U V = I xi t I y I t
42 How to solve a large linear system? I x 2 + αl I x I y I x I y I 2 y + αl U V = I xi t I y I t With α > 0, this system is positive definite! You can use your favorite solver Gauss-Seidel, successive over-relaxation (SOR) (Pre-conditioned) conjugate gradient No need to wait for the solver to converge completely
43 Condition for convergence In the objective function u, v = arg min I x u + I y v + I t 2 + α u 2 + v 2 dxdy The displacement (u, v) has to be small for the Taylor expansion to be valid. More practically, we can estimate the optimal incremental change I x du + I y dv + I t 2 + α u + du 2 + v + dv 2 dxdy The solution becomes I x 2 + αl I x I y I x I y I 2 y + αl du dv = I xi t + αlu I y I t + αlv
44 Examples Horn-Schunck Input two frames Coarse-to-fine LK Flow visualization Coarse-to-fine LK with median filtering
45 The source of over-smoothness Horn-Schunck is a Gaussian Markov random field (GMRF) I x u + I y v + I t 2 + α u 2 + v 2 dxdy Spatial over-smoothness is caused by quadratic smoothness term Nevertheless, optical flow fields are sparse! Horn-Schunck u u x u y v v x v y Ground truth
46 Continuous Markov Random Fields Horn-Schunck started 30 years of research on continuous Markov random fields Optical flow estimation Image reconstruction, e.g. denoising, super resolution Shape from shading, inverse rendering problems Natural image priors Why continuous? Many signals are differentiable More complicated spatial relationships Fast solvers Multi-grid Preconditioned conjugate gradient FFT + annealing
47 Contents Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
48 Modification to Horn-Schunck Let x = (x, y, t), and w x = (u x, v x, 1) be the flow vector Horn-Schunck (recall) I x u + I y v + I t 2 + α u 2 + v 2 dxdy Robust estimation ψ I x + w I x 2 + αφ u 2 + v 2 dxdy Robust estimation with Lucas-Kanade g ψ I x + w I x 2 + αφ u 2 + v 2 dxdy
49 Robust functions Various forms of robust functions L1 norm: ψ z 2 = z 2 + ε 2, φ z 2 = z 2 + ε 2 Sub L1: ψ z 2 ; η = z 2 + ε 2 η, η < 0.5 Lorentzian: ψ z 2 = log(1 + z 2 ) z z 2 + ε =0.5 =0.4 =0.3 =
50 Special cases The robust objective function g ψ I x + w I x 2 + αφ u 2 + v 2 dxdy Lucas-Kanade: α = 0, ψ z 2 = z 2 Robust Lucas-Kanade: α = 0, ψ z 2 = z 2 + ε 2 Horn-Schunck: g = 1, ψ z 2 = z 2, φ z 2 = z 2 One can also learn the filters (other than gradients), and robust function ψ, φ [Roth & Black 2005]
51 Derivation strategies Euler-Lagrange Derive in continuous domain, discretize in the end Nonlinear PDE s Outer and inner fixed point iterations Cannot generalize to general filters Variational optimization Iterative reweighted least square (IRLS) Discretize first and derive in matrix form Easy to understand and derive These three approaches are equivalent!
52 Iterative reweighted least square (IRLS) Let φ z = z 2 + ε 2 η be a robust function We want to minimize the objective function n Φ Ax + b = φ a i T x + b i 2 i=1 where x R d, A = a 1 a 2 a n T R n d, b R n By setting Φ x = 0, we can derive Φ n x = φ a T 2 i x + b i = i=1 n i=1 n w ii a i T xa i + w ii b i a i = i=1 = A T WAx + A T Wb a i T w ii xa i + b i w ii a i a i T x + b i a i w ii = φ a i T x + b i 2 W = diag Φ Ax + b
53 Iterative reweighted least square (IRLS) Derivative: Φ x = AT WAx + A T Wb Iterate between reweighting and least square 1. Initialize x = x 0 2. Compute weight matrix W = diag Φ Ax + b 3. Solve the linear system A T WAx = A T Wb 4. If x converges, return; otherwise, go to 2 Convergence is guaranteed (local minima)
54 Objective function IRLS for robust optical flow g ψ I x + w I x 2 + αφ u 2 + v 2 dxdy Discretize, linearize and increment x,y g ψ( I t + I x du + I Y dv 2 ) + αφ( u + du 2 + v + dv 2 ) IRLS (initialize du = dv = 0) Weight: Least square: Ψ xx = diag g ψ I x I x, Ψ xy = diag g ψ I x I y, Ψ yy = diag g ψ I y I y, Ψ xt = diag(g ψ I x I t ), = diag(g ψ I y I t ), L = D T x Φ D x + D T y Φ D y Ψ yt Ψ xx + αl Ψ xy + αl Ψ xy Ψ yy du dv = Ψ xt + αlu I y I t + αlv
55 Examples Robust optical flow Input two frames Horn-Schunck Flow visualization Coarse-to-fine LK with median filtering
56 Contents Motion perception Motion representation Parametric motion: Lucas-Kanade Dense optical flow: Horn-Schunck Robust estimation Applications (1)
57 Video stabilization
58 Video denoising Use multiple frames for temporal coherence Non-local mean
59 Video denoising
60 Video super resolution Merge information from adjacent frames Reconstruction depends on flow accuracy
61 Summary Lucas-Kanade Parametric motion Dense flow field (with median filtering) Horn-Schunck Gaussian Markov random field Euler-Lagrange Robust flow estimation Robust function Account for outliers in data term Encourage piecewise smoothness IRLS (= nonlinear PDE = variational optimization)
62 Next time Discrete optical flow Layer motion analysis Contour motion analysis Obtaining motion ground truth SIFT flow: generalized optical flow Applications (2)
Motion Estimation (I)
Motion Estimation (I) Ce Liu celiu@microsoft.com Microsoft Research New England We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationIntroduction to motion correspondence
Introduction to motion correspondence 1 IPAM - UCLA July 24, 2013 Iasonas Kokkinos Center for Visual Computing Ecole Centrale Paris / INRIA Saclay Why estimate visual motion? 2 Tracking Segmentation Structure
More informationVideo and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)
Video and Motion Analysis 16-385 Computer Vision Carnegie Mellon University (Kris Kitani) Optical flow used for feature tracking on a drone Interpolated optical flow used for super slow-mo optical flow
More informationLucas-Kanade Optical Flow. Computer Vision Carnegie Mellon University (Kris Kitani)
Lucas-Kanade Optical Flow Computer Vision 16-385 Carnegie Mellon University (Kris Kitani) I x u + I y v + I t =0 I x = @I @x I y = @I u = dx v = dy I @y t = @I dt dt @t spatial derivative optical flow
More informationMotion estimation. Digital Visual Effects Yung-Yu Chuang. with slides by Michael Black and P. Anandan
Motion estimation Digital Visual Effects Yung-Yu Chuang with slides b Michael Black and P. Anandan Motion estimation Parametric motion image alignment Tracking Optical flow Parametric motion direct method
More informationIterative Image Registration: Lucas & Kanade Revisited. Kentaro Toyama Vision Technology Group Microsoft Research
Iterative Image Registration: Lucas & Kanade Revisited Kentaro Toyama Vision Technology Group Microsoft Research Every writer creates his own precursors. His work modifies our conception of the past, as
More informationMethods in Computer Vision: Introduction to Optical Flow
Methods in Computer Vision: Introduction to Optical Flow Oren Freifeld Computer Science, Ben-Gurion University March 22 and March 26, 2017 Mar 22, 2017 1 / 81 A Preliminary Discussion Example and Flow
More informationOptic Flow Computation with High Accuracy
Cognitive Computer Vision Colloquium Prague, January 12 13, 2004 Optic Flow Computation with High ccuracy Joachim Weickert Saarland University Saarbrücken, Germany joint work with Thomas Brox ndrés Bruhn
More informationGlobal parametric image alignment via high-order approximation
Global parametric image alignment via high-order approximation Y. Keller, A. Averbuch 2 Electrical & Computer Engineering Department, Ben-Gurion University of the Negev. 2 School of Computer Science, Tel
More informationCS4495/6495 Introduction to Computer Vision. 6B-L1 Dense flow: Brightness constraint
CS4495/6495 Introduction to Computer Vision 6B-L1 Dense flow: Brightness constraint Motion estimation techniques Feature-based methods Direct, dense methods Motion estimation techniques Direct, dense methods
More informationElaborazione delle Immagini Informazione multimediale - Immagini. Raffaella Lanzarotti
Elaborazione delle Immagini Informazione multimediale - Immagini Raffaella Lanzarotti OPTICAL FLOW Thanks to prof. Mubarak Shah,UCF 2 Video Video: sequence of frames (images) catch in the time Data: function
More informationTRACKING and DETECTION in COMPUTER VISION
Technischen Universität München Winter Semester 2013/2014 TRACKING and DETECTION in COMPUTER VISION Template tracking methods Slobodan Ilić Template based-tracking Energy-based methods The Lucas-Kanade(LK)
More informationDense Optical Flow Estimation from the Monogenic Curvature Tensor
Dense Optical Flow Estimation from the Monogenic Curvature Tensor Di Zang 1, Lennart Wietzke 1, Christian Schmaltz 2, and Gerald Sommer 1 1 Department of Computer Science, Christian-Albrechts-University
More informationEdges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise
Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity
More informationGeneralized Newton-Type Method for Energy Formulations in Image Processing
Generalized Newton-Type Method for Energy Formulations in Image Processing Leah Bar and Guillermo Sapiro Department of Electrical and Computer Engineering University of Minnesota Outline Optimization in
More informationMultigrid Acceleration of the Horn-Schunck Algorithm for the Optical Flow Problem
Multigrid Acceleration of the Horn-Schunck Algorithm for the Optical Flow Problem El Mostafa Kalmoun kalmoun@cs.fau.de Ulrich Ruede ruede@cs.fau.de Institute of Computer Science 10 Friedrich Alexander
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:
More informationComputer Vision I. Announcements
Announcements Motion II No class Wednesda (Happ Thanksgiving) HW4 will be due Frida 1/8 Comment on Non-maximal supression CSE5A Lecture 15 Shi-Tomasi Corner Detector Filter image with a Gaussian. Compute
More informationFeature extraction: Corners and blobs
Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over
More informationNonlinear Diffusion. Journal Club Presentation. Xiaowei Zhou
1 / 41 Journal Club Presentation Xiaowei Zhou Department of Electronic and Computer Engineering The Hong Kong University of Science and Technology 2009-12-11 2 / 41 Outline 1 Motivation Diffusion process
More informationTemplates, Image Pyramids, and Filter Banks
Templates, Image Pyramids, and Filter Banks 09/9/ Computer Vision James Hays, Brown Slides: Hoiem and others Review. Match the spatial domain image to the Fourier magnitude image 2 3 4 5 B A C D E Slide:
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff
More informationINTEREST POINTS AT DIFFERENT SCALES
INTEREST POINTS AT DIFFERENT SCALES Thank you for the slides. They come mostly from the following sources. Dan Huttenlocher Cornell U David Lowe U. of British Columbia Martial Hebert CMU Intuitively, junctions
More informationImage Alignment and Mosaicing
Image Alignment and Mosaicing Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic mosaicing Image Enhancement Original
More informationOptical Flow, Motion Segmentation, Feature Tracking
BIL 719 - Computer Vision May 21, 2014 Optical Flow, Motion Segmentation, Feature Tracking Aykut Erdem Dept. of Computer Engineering Hacettepe University Motion Optical Flow Motion Segmentation Feature
More informationImage Analysis. Feature extraction: corners and blobs
Image Analysis Feature extraction: corners and blobs Christophoros Nikou cnikou@cs.uoi.gr Images taken from: Computer Vision course by Svetlana Lazebnik, University of North Carolina at Chapel Hill (http://www.cs.unc.edu/~lazebnik/spring10/).
More informationAn Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections
An Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections Claudia Kondermann, Daniel Kondermann, Bernd Jähne, Christoph Garbe Interdisciplinary Center for Scientific Computing
More informationOptical flow. Subhransu Maji. CMPSCI 670: Computer Vision. October 20, 2016
Optical flow Subhransu Maji CMPSC 670: Computer Vision October 20, 2016 Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes CMPSC 670 2 Motion and perceptual organization Sometimes,
More informationImage Alignment and Mosaicing Feature Tracking and the Kalman Filter
Image Alignment and Mosaicing Feature Tracking and the Kalman Filter Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic
More informationNon-linear least squares
Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined
More informationEfficient & Robust LK for Mobile Vision
Efficient & Robust LK for Mobile Vision Instructor - Simon Lucey 16-623 - Designing Comuter Vision As Direct Method (ours) Indirect Method (ORB+RANSAC) H. Alismail, B. Browning, S. Lucey Bit-Planes: Dense
More informationNotes for CS542G (Iterative Solvers for Linear Systems)
Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,
More informationEfficient Nonlocal Regularization for Optical Flow
Efficient Nonlocal Regularization for Optical Flow Philipp Krähenbühl and Vladlen Koltun Stanford University {philkr,vladlen}@cs.stanford.edu Abstract. Dense optical flow estimation in images is a challenging
More informationErkut Erdem. Hacettepe University February 24 th, Linear Diffusion 1. 2 Appendix - The Calculus of Variations 5.
LINEAR DIFFUSION Erkut Erdem Hacettepe University February 24 th, 2012 CONTENTS 1 Linear Diffusion 1 2 Appendix - The Calculus of Variations 5 References 6 1 LINEAR DIFFUSION The linear diffusion (heat)
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationEfficient Inference in Fully Connected CRFs with Gaussian Edge Potentials
Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials by Phillip Krahenbuhl and Vladlen Koltun Presented by Adam Stambler Multi-class image segmentation Assign a class label to each
More informationNumerical Optimization Algorithms
Numerical Optimization Algorithms 1. Overview. Calculus of Variations 3. Linearized Supersonic Flow 4. Steepest Descent 5. Smoothed Steepest Descent Overview 1 Two Main Categories of Optimization Algorithms
More informationPhotometric Stereo: Three recent contributions. Dipartimento di Matematica, La Sapienza
Photometric Stereo: Three recent contributions Dipartimento di Matematica, La Sapienza Jean-Denis DUROU IRIT, Toulouse Jean-Denis DUROU (IRIT, Toulouse) 17 December 2013 1 / 32 Outline 1 Shape-from-X techniques
More informationRevisiting Horn and Schunck: Interpretation as Gauß-Newton Optimisation
ZIKIC, KAMEN, AND NAVAB: REVISITING HORN AND SCHUNCK 1 Revisiting Horn and Schunck: Interpretation as Gauß-Newton Optimisation Darko Zikic 1 zikic@in.tum.de Ali Kamen ali.kamen@siemens.com Nassir Navab
More informationMTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.
MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:
More informationCS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015
CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning First-Order Methods, L1-Regularization, Coordinate Descent Winter 2016 Some images from this lecture are taken from Google Image Search. Admin Room: We ll count final numbers
More informationRandom walks for deformable image registration Dana Cobzas
Random walks for deformable image registration Dana Cobzas with Abhishek Sen, Martin Jagersand, Neil Birkbeck, Karteek Popuri Computing Science, University of Alberta, Edmonton Deformable image registration?t
More informationChapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs
Chapter Two: Numerical Methods for Elliptic PDEs Finite Difference Methods for Elliptic PDEs.. Finite difference scheme. We consider a simple example u := subject to Dirichlet boundary conditions ( ) u
More informationAsaf Bar Zvi Adi Hayat. Semantic Segmentation
Asaf Bar Zvi Adi Hayat Semantic Segmentation Today s Topics Fully Convolutional Networks (FCN) (CVPR 2015) Conditional Random Fields as Recurrent Neural Networks (ICCV 2015) Gaussian Conditional random
More information2 Nonlinear least squares algorithms
1 Introduction Notes for 2017-05-01 We briefly discussed nonlinear least squares problems in a previous lecture, when we described the historical path leading to trust region methods starting from the
More informationSec. 1.1: Basics of Vectors
Sec. 1.1: Basics of Vectors Notation for Euclidean space R n : all points (x 1, x 2,..., x n ) in n-dimensional space. Examples: 1. R 1 : all points on the real number line. 2. R 2 : all points (x 1, x
More informationLecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.
Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference
More informationVlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems
1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Perception Concepts Vision Chapter 4 (textbook) Sections 4.3 to 4.5 What is the course
More informationEdge Detection. CS 650: Computer Vision
CS 650: Computer Vision Edges and Gradients Edge: local indication of an object transition Edge detection: local operators that find edges (usually involves convolution) Local intensity transitions are
More informationPose Tracking II! Gordon Wetzstein! Stanford University! EE 267 Virtual Reality! Lecture 12! stanford.edu/class/ee267/!
Pose Tracking II! Gordon Wetzstein! Stanford University! EE 267 Virtual Reality! Lecture 12! stanford.edu/class/ee267/!! WARNING! this class will be dense! will learn how to use nonlinear optimization
More informationNewton s Method. Javier Peña Convex Optimization /36-725
Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and
More information2.29 Numerical Fluid Mechanics Spring 2015 Lecture 9
Spring 2015 Lecture 9 REVIEW Lecture 8: Direct Methods for solving (linear) algebraic equations Gauss Elimination LU decomposition/factorization Error Analysis for Linear Systems and Condition Numbers
More informationImage restoration: numerical optimisation
Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context
More informationLevel-Set Person Segmentation and Tracking with Multi-Region Appearance Models and Top-Down Shape Information Appendix
Level-Set Person Segmentation and Tracking with Multi-Region Appearance Models and Top-Down Shape Information Appendix Esther Horbert, Konstantinos Rematas, Bastian Leibe UMIC Research Centre, RWTH Aachen
More informationScene from Lola rennt
Scene from Lola rennt (Optically guided) sensorimotor action is automatic (Lola s mind is otherwise occupied!). In this respect Lola is a zombie. This is physiology & space-time is generated by the Galilean
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationKernel Correlation for Robust Distance Minimization
Chapter 2 Kernel Correlation for Robust Distance Minimization We introduce kernel correlation between points, between a point and a set of points, and among a set of points. We show that kernel correlation
More informationKasetsart University Workshop. Multigrid methods: An introduction
Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available
More informationMath 409/509 (Spring 2011)
Math 409/509 (Spring 2011) Instructor: Emre Mengi Study Guide for Homework 2 This homework concerns the root-finding problem and line-search algorithms for unconstrained optimization. Please don t hesitate
More information17 Solution of Nonlinear Systems
17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m
More informationImage processing and Computer Vision
1 / 1 Image processing and Computer Vision Continuous Optimization and applications to image processing Martin de La Gorce martin.de-la-gorce@enpc.fr February 2015 Optimization 2 / 1 We have a function
More informationA Riemannian Framework for Denoising Diffusion Tensor Images
A Riemannian Framework for Denoising Diffusion Tensor Images Manasi Datar No Institute Given Abstract. Diffusion Tensor Imaging (DTI) is a relatively new imaging modality that has been extensively used
More informationT H E S I S. Computer Engineering May Smoothing of Matrix-Valued Data. Thomas Brox
D I P L O M A Computer Engineering May 2002 T H E S I S Smoothing of Matrix-Valued Data Thomas Brox Computer Vision, Graphics, and Pattern Recognition Group Department of Mathematics and Computer Science
More informationSuper-Resolution. Dr. Yossi Rubner. Many slides from Miki Elad - Technion
Super-Resolution Dr. Yossi Rubner yossi@rubner.co.il Many slides from Mii Elad - Technion 5/5/2007 53 images, ratio :4 Example - Video 40 images ratio :4 Example Surveillance Example Enhance Mosaics Super-Resolution
More informationNONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)
NONLINEAR PROGRAMMING (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Nonlinear Programming g Linear programming has a fundamental role in OR. In linear programming all its functions
More informationVisual SLAM Tutorial: Bundle Adjustment
Visual SLAM Tutorial: Bundle Adjustment Frank Dellaert June 27, 2014 1 Minimizing Re-projection Error in Two Views In a two-view setting, we are interested in finding the most likely camera poses T1 w
More information6.869 Advances in Computer Vision. Prof. Bill Freeman March 1, 2005
6.869 Advances in Computer Vision Prof. Bill Freeman March 1 2005 1 2 Local Features Matching points across images important for: object identification instance recognition object class recognition pose
More informationUniversity of Houston, Department of Mathematics Numerical Analysis, Fall 2005
3 Numerical Solution of Nonlinear Equations and Systems 3.1 Fixed point iteration Reamrk 3.1 Problem Given a function F : lr n lr n, compute x lr n such that ( ) F(x ) = 0. In this chapter, we consider
More informationRobustifying the Flock of Trackers
16 th Computer Vision Winter Workshop Andreas Wendel, Sabine Sternig, Martin Godec (eds.) Mitterberg, Austria, February 2-4, 2011 Robustifying the Flock of Trackers Jiří Matas and Tomáš Vojíř The Center
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationMath 265H: Calculus III Practice Midterm II: Fall 2014
Name: Section #: Math 65H: alculus III Practice Midterm II: Fall 14 Instructions: This exam has 7 problems. The number of points awarded for each question is indicated in the problem. Answer each question
More informationHIGH-ORDER ACCURATE METHODS BASED ON DIFFERENCE POTENTIALS FOR 2D PARABOLIC INTERFACE MODELS
HIGH-ORDER ACCURATE METHODS BASED ON DIFFERENCE POTENTIALS FOR 2D PARABOLIC INTERFACE MODELS JASON ALBRIGHT, YEKATERINA EPSHTEYN, AND QING XIA Abstract. Highly-accurate numerical methods that can efficiently
More informationA Third Order Fast Sweeping Method with Linear Computational Complexity for Eikonal Equations
DOI.7/s95-4-9856-7 A Third Order Fast Sweeping Method with Linear Computational Complexity for Eikonal Equations Liang Wu Yong-Tao Zhang Received: 7 September 23 / Revised: 23 February 24 / Accepted: 5
More informationProperties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context
Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Silvio Savarese Lecture 10-16-Feb-15 From the 3D to 2D & vice versa
More informationRegularization in Neural Networks
Regularization in Neural Networks Sargur Srihari 1 Topics in Neural Network Regularization What is regularization? Methods 1. Determining optimal number of hidden units 2. Use of regularizer in error function
More informationMethod 1: Geometric Error Optimization
Method 1: Geometric Error Optimization we need to encode the constraints ŷ i F ˆx i = 0, rank F = 2 idea: reconstruct 3D point via equivalent projection matrices and use reprojection error equivalent projection
More informationBasic Aspects of Discretization
Basic Aspects of Discretization Solution Methods Singularity Methods Panel method and VLM Simple, very powerful, can be used on PC Nonlinear flow effects were excluded Direct numerical Methods (Field Methods)
More informationSimultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors
Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,
More informationNumerical Analysis Comprehensive Exam Questions
Numerical Analysis Comprehensive Exam Questions 1. Let f(x) = (x α) m g(x) where m is an integer and g(x) C (R), g(α). Write down the Newton s method for finding the root α of f(x), and study the order
More informationComputer Vision Motion
Computer Vision Motion Professor Hager http://www.cs.jhu.edu/~hager 12/1/12 CS 461, Copyright G.D. Hager Outline From Stereo to Motion The motion field and optical flow (2D motion) Factorization methods
More informationAM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α
More informationChapter 6. Finite Element Method. Literature: (tiny selection from an enormous number of publications)
Chapter 6 Finite Element Method Literature: (tiny selection from an enormous number of publications) K.J. Bathe, Finite Element procedures, 2nd edition, Pearson 2014 (1043 pages, comprehensive). Available
More informationAchieving scale covariance
Achieving scale covariance Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region size that is covariant
More informationScale & Affine Invariant Interest Point Detectors
Scale & Affine Invariant Interest Point Detectors Krystian Mikolajczyk and Cordelia Schmid Presented by Hunter Brown & Gaurav Pandey, February 19, 2009 Roadmap: Motivation Scale Invariant Detector Affine
More informationUncertainty Models in Quasiconvex Optimization for Geometric Reconstruction
Uncertainty Models in Quasiconvex Optimization for Geometric Reconstruction Qifa Ke and Takeo Kanade Department of Computer Science, Carnegie Mellon University Email: ke@cmu.edu, tk@cs.cmu.edu Abstract
More informationIntroduction to Nonlinear Image Processing
Introduction to Nonlinear Image Processing 1 IPAM Summer School on Computer Vision July 22, 2013 Iasonas Kokkinos Center for Visual Computing Ecole Centrale Paris / INRIA Saclay Mean and median 2 Observations
More informationOptimization. Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison
Optimization Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison optimization () cost constraints might be too much to cover in 3 hours optimization (for big
More informationThe Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR
Appendix B The Derivative B.1 The Derivative of f In this chapter, we give a short summary of the derivative. Specifically, we want to compare/contrast how the derivative appears for functions whose domain
More informationAn Introduction to Numerical Methods for Differential Equations. Janet Peterson
An Introduction to Numerical Methods for Differential Equations Janet Peterson Fall 2015 2 Chapter 1 Introduction Differential equations arise in many disciplines such as engineering, mathematics, sciences
More informationDynamical Systems & Lyapunov Stability
Dynamical Systems & Lyapunov Stability Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University Outline Ordinary Differential Equations Existence & uniqueness Continuous dependence
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More informationMobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti
Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes
More informationIterative Reweighted Least Squares
Iterative Reweighted Least Squares Sargur. University at Buffalo, State University of ew York USA Topics in Linear Classification using Probabilistic Discriminative Models Generative vs Discriminative
More informationNeural Network Training
Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification
More informationBlobs & Scale Invariance
Blobs & Scale Invariance Prof. Didier Stricker Doz. Gabriele Bleser Computer Vision: Object and People Tracking With slides from Bebis, S. Lazebnik & S. Seitz, D. Lowe, A. Efros 1 Apertizer: some videos
More informationOverview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison of different detectors Region descriptors and
More informationHIGH-ORDER ACCURATE METHODS BASED ON DIFFERENCE POTENTIALS FOR 2D PARABOLIC INTERFACE MODELS
HIGH-ORDER ACCURATE METHODS BASED ON DIFFERENCE POTENTIALS FOR 2D PARABOLIC INTERFACE MODELS JASON ALBRIGHT, YEKATERINA EPSHTEYN, AND QING XIA Abstract. Highly-accurate numerical methods that can efficiently
More informationMath 4263 Homework Set 1
Homework Set 1 1. Solve the following PDE/BVP 2. Solve the following PDE/BVP 2u t + 3u x = 0 u (x, 0) = sin (x) u x + e x u y = 0 u (0, y) = y 2 3. (a) Find the curves γ : t (x (t), y (t)) such that that
More informationSimple Examples on Rectangular Domains
84 Chapter 5 Simple Examples on Rectangular Domains In this chapter we consider simple elliptic boundary value problems in rectangular domains in R 2 or R 3 ; our prototype example is the Poisson equation
More information