Motion estimation. Digital Visual Effects Yung-Yu Chuang. with slides by Michael Black and P. Anandan
|
|
- Natalie Parks
- 5 years ago
- Views:
Transcription
1 Motion estimation Digital Visual Effects Yung-Yu Chuang with slides b Michael Black and P. Anandan
2 Motion estimation Parametric motion image alignment Tracking Optical flow
3 Parametric motion direct method for image stitching
4 Tracking
5 Optical flow
6 Three assumptions Brightness consistenc Spatial coherence Temporal persistence
7 Brightness consistenc mage measurement e g brightness in a small region mage measurement e.g. brightness in a small region remain the same although their location ma change.
8 Spatial coherence Neighboring points in the scene tpicall belong to the same surface and hence tpicall have similar motions. Since the also project to nearb piels in the image we epect spatial coherence in image flow.
9 Temporal persistence The image motion of a surface patch changes graduall over time.
10 mage registration Goal: register a template image T and an input image where = T. warp so that it matches T mage alignment: and T are two images Tracking: T is a small patch around a point p in the image at t. is the image at time t+1. Optical flow: T and are patches th of images at t and t+1. T fied warp
11 Simple approach for translation Minimize brightness difference E u v u v T 2
12 Simple SSD algorithm For each offset u v compute Euv; Choose u v which minimizes Euv; Problems: Not efficient No sub-piel accurac
13 Lucas-Kanade algorithm
14 Newton s method Root finding for f=0 March and test signs Determine Δ small slow; large miss
15 Newton s method Root finding for f=0
16 Newton s method Root finding for f=0 T l i Talor s epansion: 2 '' 1 ' f f f f '' 2 ' f f f f ' f f f f f f f ' n n n f f ' 1 n n n f f n f
17 Newton s method Root finding for f=0 n f n f ' n n 2 1 0
18 Newton s method pick up = 0 iterate compute f f ' update b +Δ until converge Finding root is useful for optimization because Minimize g find root for f=g =0
19 Lucas-Kanade algorithm T v u v u E 2 u v v u u v T 2 v u T u E 2 0 v u T v E 2 0 v
20 Lucas-Kanade algorithm v u T E 2 0 u v u T E 2 0 v u T v 2 0 T v u T v u 2 2 T v u 2 T T v u 2 2 T v
21 Lucas-Kanade algorithm iterate shift with u v shift with uv compute gradient image compute error image T compute error image T- compute Hessian matri solve the linear sstem solve the linear sstem uv=uv+ u v til until converge T 2 T v u 2
22 Parametric model T v u v u E 2 W;p p 2 T E Our goal is to find p to minimize Ep W;p p T E d p to minimize Ep for all in T s domain T d d p d d W;p translation d d d 1 d A W;p affine T d d d 1 1 A d W;p affine T d d d d d d p
23 Parametric model minimize with respect to Δp W;p Δp T W W W;p Δp W;p Δp p W W;p Δp W;p Δp p W W;p Δp pp minimize W Δp W;p ; p T p 2 2
24 Parametric model image gradient warped image target image W 2 image gradient Δp p W p W; T Jacobian of the warp p W p W p W p W W n n p W p W p W p p p p W p p W n p p p p 2 1
25 Jacobian matri The Jacobian matri is the matri of all first-order partial derivatives of a vector-valued function partial derivatives of a vector-valued function. m n F R R : 2 1 n F n m n n f f f f f or 2 1 n F J n f f or 2 1 m f f f m m n f f n n m m 1 Δ Δ F J F F
26 Jacobian matri : R R F t r sin cos cos sin sin r v r u v u t r F t t t r cos v u u u t t r t v v v u u r u r J F v v r v cos sin sin cos sin sin sin sin cos cos cos sin r r r r 0 sin cos cos s s cos s s r
27 Parametric model image gradient warped image target image W 2 image gradient Δp p W p W; T Jacobian of the warp p W p W p W p W W n n p W p W p W p p p p W p p W n p p p p 2 1
28 Jacobian of the warp W W W W n W W W p p p W p p W 2 1 For eample for affine n p p p p 2 1 For eample for affine d d d d d d 1 1 W;p d d d d d d W;p p W p d d d d d d
29 Parametric model W W;p Δp T p arg min Δp T W W 0 W;p Δp T p p T 1 W Δp H T W;p p 2 Approimated Hessian H W W p p W T
30 Lucas-Kanade algorithm iterate 1 warp with W;p 2 compute error image T-Wp 3 compute gradient image with Wp W 4 evaluate Jacobian at ;p p W 5 compute p 6 compute Hessian W W 7 compute T T W;p p 8 solve Δp 9 update p b p+ Δp until converge T 1 W Δp H T W;p p
31 1 W Δp H T W;p p T
32 Coarse-to-fine strateg J pramid construction a in J warp J w refine + a J warp J w refine + a a pramid construction J J w warp refine a + a out
33 Application of image alignment
34 Direct vs feature-based Direct methods use all information and can be ver accurate but the depend on the fragile brightness constanc assumption. terative approaches require initialization. Not robust to illumination change and noise images. n earl das direct method is better. Feature based methods are now more robust and potentiall faster. Even better it can recognize panorama without initialization.
35 Tracking
36 Tracking t u v +u+vt+1
37 Tracking 0 1 t t v u brightness constanc 0 t t t v t u t t 0 t t v t u t 0 v u optical flow constraint equation 0 t v u optical flow constraint equation
38 Optical flow constraint equation
39 Multiple constraints
40 Area-based method Assume spatial smoothness
41 Area-based method Assume spatial smoothness E u v u v t 2
42 Area-based method must be invertible
43 Area-based method The eigenvalues tell us about the local image structure. The also tell us how well we can estimate the flow in both directions. Link to Harris corner detector.
44 Tetured area
45 Edge
46 Homogenous area
47 KLT tracking Select features b min 1 2 Monitor features b measuring dissimilariti il i
48 Aperture problem
49 Aperture problem
50 Aperture problem
51 Demo for aperture problem eathing_square.htm htm rberpole_llusion.htm
52 Aperture problem Larger window reduces ambiguit but easil violates spatial smoothness assumption
53
54
55 KLT tracking clemson edu/~stb/klt/
56 KLT tracking clemson edu/~stb/klt/
57 SFT tracking matching actuall Frame 0 Frame 10
58 SFT tracking Frame 0 Frame 100
59 SFT tracking Frame 0 Frame 200
60 KLT vs SFT tracking KLT has larger accumulating error; partl because our KLT implementation doesn t have affine transformation? SFT is surprisingl ii robust Combination of SFT and KLT eample
61 Rotoscoping Ma Fleischer
62 Tracking for rotoscoping
63 Tracking for rotoscoping
64 Waking life 2001
65 A Scanner Darkl 2006 Rotoshop a proprietar software. Each minute of animation required 500 hours of work.
66 Optical flow
67 Single-motion assumption Violated b Motion discontinuiti i Shadows Transparenc Specular reflection
68 Multiple motion
69 Multiple motion
70 Simple problem: fit a line
71 Least-square fit
72 Least-square fit
73 Robust statistics Recover the best fit for the majorit of the data Detect and reject outliers
74 Approach
75 Robust weighting Truncated quadratic
76 Robust weighting Geman & McClure
77 Robust estimation
78 Fragmented occlusion
79
80
81 Regularization and dense optical flow Neighboring points in the scene tpicall belong to the same surface and hence tpicall have similar motions. Since the also project to nearb piels in the image we epect spatial coherence in image flow.
82
83
84
85
86
87
88 nput image Horizontal motion Vertical motion
89
90
91 Application of optical flow video matching
92
93 nput for the NPR algorithm
94 Brushes
95 Edge clipping
96 Gradient
97 Smooth gradient
98 Tetured brush
99 Edge clipping
100 Temporal artifacts Frame-b-frame application of the NPR algorithm
101 Temporal coherence
102 References B.D. Lucas and T. Kanade An terative mage Registration Technique with an Application to Stereo Vision Proceedings of the 1981 DARPA mage Understanding di Workshop 1981 pp Bergen J. R. and Anandan P. and Hanna K. J. and Hingorani R. Hierarchical Model-Based Motion Estimation ECCV 1992 pp J. Shi and C. Tomasi Good Features to Track CVPR 1994 pp Michael Black and P. Anandan The Robust Estimation of Multiple Motions: Parametric and Piecewise-Smooth Flow Fields Computer Vision and mage Understanding 1996 pp S. Baker and. Matthews Lucas-Kanade 20 Years On: A Unifing Framework nternational Journal of Computer Vision pp Peter Litwinowicz Processing mages and Video for An mpressionist Effects SGGRAPH Aseem Agarwala Aaron Hertzman David Salesin and Steven Seitz Keframe-Based Tracking for Rotoscoping and Animation SGGRAPH 2004 pp
Optical flow. Subhransu Maji. CMPSCI 670: Computer Vision. October 20, 2016
Optical flow Subhransu Maji CMPSC 670: Computer Vision October 20, 2016 Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes CMPSC 670 2 Motion and perceptual organization Sometimes,
More information6.869 Advances in Computer Vision. Prof. Bill Freeman March 1, 2005
6.869 Advances in Computer Vision Prof. Bill Freeman March 1 2005 1 2 Local Features Matching points across images important for: object identification instance recognition object class recognition pose
More informationIterative Image Registration: Lucas & Kanade Revisited. Kentaro Toyama Vision Technology Group Microsoft Research
Iterative Image Registration: Lucas & Kanade Revisited Kentaro Toyama Vision Technology Group Microsoft Research Every writer creates his own precursors. His work modifies our conception of the past, as
More informationMotion Estimation (I) Ce Liu Microsoft Research New England
Motion Estimation (I) Ce Liu celiu@microsoft.com Microsoft Research New England We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationMotion Estimation (I)
Motion Estimation (I) Ce Liu celiu@microsoft.com Microsoft Research New England We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationKeypoint extraction: Corners Harris Corners Pkwy, Charlotte, NC
Kepoint etraction: Corners 9300 Harris Corners Pkw Charlotte NC Wh etract kepoints? Motivation: panorama stitching We have two images how do we combine them? Wh etract kepoints? Motivation: panorama stitching
More informationLucas-Kanade Optical Flow. Computer Vision Carnegie Mellon University (Kris Kitani)
Lucas-Kanade Optical Flow Computer Vision 16-385 Carnegie Mellon University (Kris Kitani) I x u + I y v + I t =0 I x = @I @x I y = @I u = dx v = dy I @y t = @I dt dt @t spatial derivative optical flow
More informationInstance-level l recognition. Cordelia Schmid INRIA
nstance-level l recognition Cordelia Schmid NRA nstance-level recognition Particular objects and scenes large databases Application Search photos on the web for particular places Find these landmars...in
More informationIntroduction to motion correspondence
Introduction to motion correspondence 1 IPAM - UCLA July 24, 2013 Iasonas Kokkinos Center for Visual Computing Ecole Centrale Paris / INRIA Saclay Why estimate visual motion? 2 Tracking Segmentation Structure
More informationLecture 05 Point Feature Detection and Matching
nstitute of nformatics nstitute of Neuroinformatics Lecture 05 Point Feature Detection and Matching Davide Scaramuzza 1 Lab Eercise 3 - Toda afternoon Room ETH HG E 1.1 from 13:15 to 15:00 Wor description:
More informationInstance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble
nstance-level recognition: ocal invariant features Cordelia Schmid NRA Grenoble Overview ntroduction to local features Harris interest points + SSD ZNCC SFT Scale & affine invariant interest point detectors
More informationInstance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble
nstance-level recognition: ocal invariant features Cordelia Schmid NRA Grenoble Overview ntroduction to local features Harris interest points + SSD ZNCC SFT Scale & affine invariant interest point detectors
More informationCamera calibration. Outline. Pinhole camera. Camera projection models. Nonlinear least square methods A camera calibration tool
Outline Camera calibration Camera projection models Camera calibration i Nonlinear least square methods A camera calibration tool Applications Digital Visual Effects Yung-Yu Chuang with slides b Richard
More informationVideo and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)
Video and Motion Analysis 16-385 Computer Vision Carnegie Mellon University (Kris Kitani) Optical flow used for feature tracking on a drone Interpolated optical flow used for super slow-mo optical flow
More informationCEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.
CEE598 - Visual Sensing for Civil nfrastructure Eng. & Mgmt. Session 9- mage Detectors, Part Mani Golparvar-Fard Department of Civil and Environmental Engineering 3129D, Newmark Civil Engineering Lab e-mail:
More informationPerception III: Filtering, Edges, and Point-features
Perception : Filtering, Edges, and Point-features Davide Scaramuzza Universit of Zurich Margarita Chli, Paul Furgale, Marco Hutter, Roland Siegwart 1 Toda s outline mage filtering Smoothing Edge detection
More informationElaborazione delle Immagini Informazione multimediale - Immagini. Raffaella Lanzarotti
Elaborazione delle Immagini Informazione multimediale - Immagini Raffaella Lanzarotti OPTICAL FLOW Thanks to prof. Mubarak Shah,UCF 2 Video Video: sequence of frames (images) catch in the time Data: function
More informationOptical Flow, Motion Segmentation, Feature Tracking
BIL 719 - Computer Vision May 21, 2014 Optical Flow, Motion Segmentation, Feature Tracking Aykut Erdem Dept. of Computer Engineering Hacettepe University Motion Optical Flow Motion Segmentation Feature
More informationImage Alignment and Mosaicing
Image Alignment and Mosaicing Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic mosaicing Image Enhancement Original
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff
More informationInstance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble
nstance-level recognition: ocal invariant features Cordelia Schmid NRA Grenoble Overview ntroduction to local features Harris interest t points + SSD ZNCC SFT Scale & affine invariant interest point detectors
More informationGlobal parametric image alignment via high-order approximation
Global parametric image alignment via high-order approximation Y. Keller, A. Averbuch 2 Electrical & Computer Engineering Department, Ben-Gurion University of the Negev. 2 School of Computer Science, Tel
More informationEdges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise
Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity
More informationInstance-level l recognition. Cordelia Schmid & Josef Sivic INRIA
nstance-level l recognition Cordelia Schmid & Josef Sivic NRA nstance-level recognition Particular objects and scenes large databases Application Search photos on the web for particular places Find these
More informationOptical Flow, KLT Feature Tracker.
Optical Flow, KL Feature racker E-mail: hogijung@hanang.ac.kr Motion in Computer Vision Motion Structure rom motion Detection/segmentation with direction [1] E-mail: hogijung@hanang.ac.kr Motion Field.s..
More informationComputer Vision I. Announcements
Announcements Motion II No class Wednesda (Happ Thanksgiving) HW4 will be due Frida 1/8 Comment on Non-maximal supression CSE5A Lecture 15 Shi-Tomasi Corner Detector Filter image with a Gaussian. Compute
More informationTRACKING and DETECTION in COMPUTER VISION
Technischen Universität München Winter Semester 2013/2014 TRACKING and DETECTION in COMPUTER VISION Template tracking methods Slobodan Ilić Template based-tracking Energy-based methods The Lucas-Kanade(LK)
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:
More informationInterest Operators. All lectures are from posted research papers. Harris Corner Detector: the first and most basic interest operator
Interest Operators All lectures are from posted research papers. Harris Corner Detector: the first and most basic interest operator SIFT interest point detector and region descriptor Kadir Entrop Detector
More informationImage Alignment and Mosaicing Feature Tracking and the Kalman Filter
Image Alignment and Mosaicing Feature Tracking and the Kalman Filter Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic
More informationFeature extraction: Corners and blobs
Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over
More informationImage Alignment Computer Vision (Kris Kitani) Carnegie Mellon University
Lucas Kanade Image Alignment 16-385 Comuter Vision (Kris Kitani) Carnegie Mellon University htt://www.humansensing.cs.cmu.edu/intraface/ How can I find in the image? Idea #1: Temlate Matching Slow, combinatory,
More informationINTEREST POINTS AT DIFFERENT SCALES
INTEREST POINTS AT DIFFERENT SCALES Thank you for the slides. They come mostly from the following sources. Dan Huttenlocher Cornell U David Lowe U. of British Columbia Martial Hebert CMU Intuitively, junctions
More informationCorners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros
Corners, Blobs & Descriptors With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Motivation: Build a Panorama M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 How do we build panorama?
More informationRecap: edge detection. Source: D. Lowe, L. Fei-Fei
Recap: edge detection Source: D. Lowe, L. Fei-Fei Canny edge detector 1. Filter image with x, y derivatives of Gaussian 2. Find magnitude and orientation of gradient 3. Non-maximum suppression: Thin multi-pixel
More informationLecture 13: Tracking mo3on features op3cal flow
Lecture 13: Tracking mo3on features op3cal flow Professor Fei- Fei Li Stanford Vision Lab Lecture 14-1! What we will learn today? Introduc3on Op3cal flow Feature tracking Applica3ons Reading: [Szeliski]
More informationComputer Vision Motion
Computer Vision Motion Professor Hager http://www.cs.jhu.edu/~hager 12/1/12 CS 461, Copyright G.D. Hager Outline From Stereo to Motion The motion field and optical flow (2D motion) Factorization methods
More informationFeature extraction: Corners and blobs
Featre etraction: Corners and blobs Wh etract featres? Motiation: panorama stitching We hae two images how do we combine them? Wh etract featres? Motiation: panorama stitching We hae two images how do
More informationVlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems
1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Perception Concepts Vision Chapter 4 (textbook) Sections 4.3 to 4.5 What is the course
More informationRobustifying the Flock of Trackers
16 th Computer Vision Winter Workshop Andreas Wendel, Sabine Sternig, Martin Godec (eds.) Mitterberg, Austria, February 2-4, 2011 Robustifying the Flock of Trackers Jiří Matas and Tomáš Vojíř The Center
More informationCS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015
CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature
More informationTemplates, Image Pyramids, and Filter Banks
Templates, Image Pyramids, and Filter Banks 09/9/ Computer Vision James Hays, Brown Slides: Hoiem and others Review. Match the spatial domain image to the Fourier magnitude image 2 3 4 5 B A C D E Slide:
More informationEfficient & Robust LK for Mobile Vision
Efficient & Robust LK for Mobile Vision Instructor - Simon Lucey 16-623 - Designing Comuter Vision As Direct Method (ours) Indirect Method (ORB+RANSAC) H. Alismail, B. Browning, S. Lucey Bit-Planes: Dense
More informationImage Processing 1 (IP1) Bildverarbeitung 1
MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV KOGS Image Processing 1 IP1 Bildverarbeitung 1 Lecture : Object Recognition Winter Semester 015/16 Slides: Prof. Bernd Neumann Slightly revised
More informationInvariant local features. Invariant Local Features. Classes of transformations. (Good) invariant local features. Case study: panorama stitching
Invariant local eatures Invariant Local Features Tuesday, February 6 Subset o local eature types designed to be invariant to Scale Translation Rotation Aine transormations Illumination 1) Detect distinctive
More informationEPIPOLAR GEOMETRY WITH MANY DETAILS
EPIPOLAR GEOMERY WIH MANY DEAILS hank ou for the slides. he come mostl from the following source. Marc Pollefes U. of North Carolina hree questions: (i) Correspondence geometr: Given an image point in
More informationAn Axiomatic Approach to Corner Detection
An Axiomatic Approach to Corner Detection C S Kenney M Zuliani B S Manjunath Vision Research Lab Department of Electrical and Computer Engineering University of California, Santa Barbara Abstract This
More informationMethods in Computer Vision: Introduction to Optical Flow
Methods in Computer Vision: Introduction to Optical Flow Oren Freifeld Computer Science, Ben-Gurion University March 22 and March 26, 2017 Mar 22, 2017 1 / 81 A Preliminary Discussion Example and Flow
More informationCS4495/6495 Introduction to Computer Vision. 6B-L1 Dense flow: Brightness constraint
CS4495/6495 Introduction to Computer Vision 6B-L1 Dense flow: Brightness constraint Motion estimation techniques Feature-based methods Direct, dense methods Motion estimation techniques Direct, dense methods
More informationAn Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections
An Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections Claudia Kondermann, Daniel Kondermann, Bernd Jähne, Christoph Garbe Interdisciplinary Center for Scientific Computing
More informationEfficient Computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm
Efficient Computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson, Anton van den Hengel School of Computer Science University of Adelaide,
More informationStatistical Filters for Crowd Image Analysis
Statistical Filters for Crowd Image Analysis Ákos Utasi, Ákos Kiss and Tamás Szirányi Distributed Events Analysis Research Group, Computer and Automation Research Institute H-1111 Budapest, Kende utca
More informationDifferential Motion Analysis
Differential Motion Analysis Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@ece.northwestern.edu http://www.eecs.northwestern.edu/~yingwu July 19,
More informationUncertainty Models in Quasiconvex Optimization for Geometric Reconstruction
Uncertainty Models in Quasiconvex Optimization for Geometric Reconstruction Qifa Ke and Takeo Kanade Department of Computer Science, Carnegie Mellon University Email: ke@cmu.edu, tk@cs.cmu.edu Abstract
More informationTemporal Factorization Vs. Spatial Factorization
Temporal Factorization Vs. Spatial Factorization Lihi Zelnik-Manor 1 and Michal Irani 2 1 California Institute of Technology, Pasadena CA, USA, lihi@caltech.edu, WWW home page: http://www.vision.caltech.edu/lihi
More informationRobust Motion Segmentation by Spectral Clustering
Robust Motion Segmentation by Spectral Clustering Hongbin Wang and Phil F. Culverhouse Centre for Robotics Intelligent Systems University of Plymouth Plymouth, PL4 8AA, UK {hongbin.wang, P.Culverhouse}@plymouth.ac.uk
More informationDetectors part II Descriptors
EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers Goal:
More informationImage matching. by Diva Sian. by swashford
Image matching by Diva Sian by swashford Harder case by Diva Sian by scgbt Invariant local features Find features that are invariant to transformations geometric invariance: translation, rotation, scale
More informationBeyond Newton s method Thomas P. Minka
Beyond Newton s method Thomas P. Minka 2000 (revised 7/21/2017) Abstract Newton s method for optimization is equivalent to iteratively maimizing a local quadratic approimation to the objective function.
More informationLevel-Set Person Segmentation and Tracking with Multi-Region Appearance Models and Top-Down Shape Information Appendix
Level-Set Person Segmentation and Tracking with Multi-Region Appearance Models and Top-Down Shape Information Appendix Esther Horbert, Konstantinos Rematas, Bastian Leibe UMIC Research Centre, RWTH Aachen
More informationCS9840 Learning and Computer Vision Prof. Olga Veksler. Lecture 2. Some Concepts from Computer Vision Curse of Dimensionality PCA
CS9840 Learning an Computer Vision Prof. Olga Veksler Lecture Some Concepts from Computer Vision Curse of Dimensionality PCA Some Slies are from Cornelia, Fermüller, Mubarak Shah, Gary Braski, Sebastian
More informationHarris Corner Detector
Multimedia Computing: Algorithms, Systems, and Applications: Feature Extraction By Dr. Yu Cao Department of Computer Science The University of Massachusetts Lowell Lowell, MA 01854, USA Part of the slides
More informationSuper-Resolution. Dr. Yossi Rubner. Many slides from Miki Elad - Technion
Super-Resolution Dr. Yossi Rubner yossi@rubner.co.il Many slides from Mii Elad - Technion 5/5/2007 53 images, ratio :4 Example - Video 40 images ratio :4 Example Surveillance Example Enhance Mosaics Super-Resolution
More informationLecture 12. Local Feature Detection. Matching with Invariant Features. Why extract features? Why extract features? Why extract features?
Lecture 1 Why extract eatures? Motivation: panorama stitching We have two images how do we combine them? Local Feature Detection Guest lecturer: Alex Berg Reading: Harris and Stephens David Lowe IJCV We
More informationMonte Carlo Integration I
Monte Carlo Integration I Digital Image Synthesis Yung-Yu Chuang with slides by Pat Hanrahan and Torsten Moller Introduction L o p,ωo L e p,ωo s f p,ωo,ω i L p,ω i cosθ i i d The integral equations generally
More informationAnnouncements. Tracking. Comptuer Vision I. The Motion Field. = ω. Pure Translation. Motion Field Equation. Rigid Motion: General Case
Announcements Tracking Computer Vision I CSE5A Lecture 17 HW 3 due toda HW 4 will be on web site tomorrow: Face recognition using 3 techniques Toda: Tracking Reading: Sections 17.1-17.3 The Motion Field
More informationCS5670: Computer Vision
CS5670: Computer Vision Noah Snavely Lecture 5: Feature descriptors and matching Szeliski: 4.1 Reading Announcements Project 1 Artifacts due tomorrow, Friday 2/17, at 11:59pm Project 2 will be released
More informationRobert Collins CSE486, Penn State. Lecture 25: Structure from Motion
Lecture 25: Structure from Motion Structure from Motion Given a set of flow fields or displacement vectors from a moving camera over time, determine: the sequence of camera poses the 3D structure of the
More informationAdvances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.
6.869 Advances in Computer Vision Prof. Bill Freeman March 3, 2005 Image and shape descriptors Affine invariant features Comparison of feature descriptors Shape context Readings: Mikolajczyk and Schmid;
More informationDigital Matting. Outline. Introduction to Digital Matting. Introduction to Digital Matting. Compositing equation: C = α * F + (1- α) * B
Digital Matting Outline. Introduction to Digital Matting. Bayesian Matting 3. Poisson Matting 4. A Closed Form Solution to Matting Presenting: Alon Gamliel,, Tel-Aviv University, May 006 Introduction to
More informationImage Analysis. Feature extraction: corners and blobs
Image Analysis Feature extraction: corners and blobs Christophoros Nikou cnikou@cs.uoi.gr Images taken from: Computer Vision course by Svetlana Lazebnik, University of North Carolina at Chapel Hill (http://www.cs.unc.edu/~lazebnik/spring10/).
More informationCovariance Tracking Algorithm on Bilateral Filtering under Lie Group Structure Yinghong Xie 1,2,a Chengdong Wu 1,b
Applied Mechanics and Materials Online: 014-0-06 ISSN: 166-748, Vols. 519-50, pp 684-688 doi:10.408/www.scientific.net/amm.519-50.684 014 Trans Tech Publications, Switzerland Covariance Tracking Algorithm
More informationComputer Vision & Digital Image Processing
Computer Vision & Digital Image Processing Image Segmentation Dr. D. J. Jackson Lecture 6- Image segmentation Segmentation divides an image into its constituent parts or objects Level of subdivision depends
More informationAdvanced Features. Advanced Features: Topics. Jana Kosecka. Slides from: S. Thurn, D. Lowe, Forsyth and Ponce. Advanced features and feature matching
Advanced Features Jana Kosecka Slides from: S. Thurn, D. Lowe, Forsyth and Ponce Advanced Features: Topics Advanced features and feature matching Template matching SIFT features Haar features 2 1 Features
More informationEdge Detection. Introduction to Computer Vision. Useful Mathematics Funcs. The bad news
Edge Detection Introduction to Computer Vision CS / ECE 8B Thursday, April, 004 Edge detection (HO #5) Edge detection is a local area operator that seeks to find significant, meaningful changes in image
More informationComputational Photography
Computational Photography Si Lu Spring 208 http://web.cecs.pdx.edu/~lusi/cs50/cs50_computati onal_photography.htm 04/0/208 Last Time o Digital Camera History of Camera Controlling Camera o Photography
More informationCS376 Computer Vision Lecture 6: Optical Flow
CS376 Compuer Vision Lecure 6: Opical Flow Qiing Huang Feb. 11 h 2019 Slides Credi: Krisen Grauman and Sebasian Thrun, Michael Black, Marc Pollefeys Opical Flow mage racking 3D compuaion mage sequence
More informationVision-based navigation around small bodies
Astronet-II, International Final Conference Vision-based navigation around small bodies Pawel Kicman VISION-BASED NAVIGATION IN SPACE Camera LOS (Line-of-sight) sensor Star-horizon measurements Apparent
More informationLeaving Some Stones Unturned: Dynamic Feature Prioritization for Activity Detection in Streaming Video
In Proceedings of the European Conference on Computer Vision (ECCV), 26 Leaving Some Stones Unturned: Dynamic Feature Prioritization for Activity Detection in Streaming Video Yu-Chuan Su and Kristen Grauman
More informationDeformation and Viewpoint Invariant Color Histograms
1 Deformation and Viewpoint Invariant Histograms Justin Domke and Yiannis Aloimonos Computer Vision Laboratory, Department of Computer Science University of Maryland College Park, MD 274, USA domke@cs.umd.edu,
More informationGalilean-diagonalized spatio-temporal interest operators
Galilean-diagonalized spatio-temporal interest operators Tony Lindeberg, Amir Akbarzadeh and Ivan Laptev Computational Vision and Active Perception Laboratory (CVAP) Department of Numerical Analysis and
More informationTwo conventions for coordinate systems. Left-Hand vs Right-Hand. x z. Which is which?
walters@buffalo.edu CSE 480/580 Lecture 2 Slide 3-D Transformations 3-D space Two conventions for coordinate sstems Left-Hand vs Right-Hand (Thumb is the ais, inde is the ais) Which is which? Most graphics
More informationTwo-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix
Two-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix René Vidal Stefano Soatto Shankar Sastry Department of EECS, UC Berkeley Department of Computer Sciences, UCLA 30 Cory Hall,
More informationScale-space image processing
Scale-space image processing Corresponding image features can appear at different scales Like shift-invariance, scale-invariance of image processing algorithms is often desirable. Scale-space representation
More informationSIFT: Scale Invariant Feature Transform
1 SIFT: Scale Invariant Feature Transform With slides from Sebastian Thrun Stanford CS223B Computer Vision, Winter 2006 3 Pattern Recognition Want to find in here SIFT Invariances: Scaling Rotation Illumination
More informationProbabilistic Models of the Brain: Perception and Neural Function
Probabilistic Models of the Brain: Perception and Neural Function Edited by Rajesh P. N. Rao Bruno A. Olshausen Michael S. Lewicki The MIT Press Cambridge, Massachusetts London, England c 1 Massachusetts
More informationCS4670: Computer Vision Kavita Bala. Lecture 7: Harris Corner Detec=on
CS4670: Computer Vision Kavita Bala Lecture 7: Harris Corner Detec=on Announcements HW 1 will be out soon Sign up for demo slots for PA 1 Remember that both partners have to be there We will ask you to
More informationRevisiting Horn and Schunck: Interpretation as Gauß-Newton Optimisation
ZIKIC, KAMEN, AND NAVAB: REVISITING HORN AND SCHUNCK 1 Revisiting Horn and Schunck: Interpretation as Gauß-Newton Optimisation Darko Zikic 1 zikic@in.tum.de Ali Kamen ali.kamen@siemens.com Nassir Navab
More informationSubspace Methods for Visual Learning and Recognition
This is a shortened version of the tutorial given at the ECCV 2002, Copenhagen, and ICPR 2002, Quebec City. Copyright 2002 by Aleš Leonardis, University of Ljubljana, and Horst Bischof, Graz University
More informationLecture 4.2 Finite Difference Approximation
Lecture 4. Finite Difference Approimation 1 Discretization As stated in Lecture 1.0, there are three steps in numerically solving the differential equations. They are: 1. Discretization of the domain by
More informationProperties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context
Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Silvio Savarese Lecture 10-16-Feb-15 From the 3D to 2D & vice versa
More informationEdge Detection. Image Processing - Computer Vision
Image Processing - Lesson 10 Edge Detection Image Processing - Computer Vision Low Level Edge detection masks Gradient Detectors Compass Detectors Second Derivative - Laplace detectors Edge Linking Image
More information3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:
3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by
More informationSingle Exposure Enhancement and Reconstruction. Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman
Single Exposure Enhancement and Reconstruction Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman 1 Reconstruction as an Inverse Problem Original image f Distortion & Sampling
More informationCSE 473/573 Computer Vision and Image Processing (CVIP)
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu inwogu@buffalo.edu Lecture 11 Local Features 1 Schedule Last class We started local features Today More on local features Readings for
More information* h + = Lec 05: Interesting Points Detection. Image Analysis & Retrieval. Outline. Image Filtering. Recap of Lec 04 Image Filtering Edge Features
age Analsis & Retrieval Outline CS/EE 5590 Special Topics (Class ds: 44873, 44874) Fall 06, M/W 4-5:5p@Bloch 00 Lec 05: nteresting Points Detection Recap of Lec 04 age Filtering Edge Features Hoework Harris
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationChapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence
Chapter 6 Nonlinear Equations 6. The Problem of Nonlinear Root-finding In this module we consider the problem of using numerical techniques to find the roots of nonlinear equations, f () =. Initially we
More information16.5. Maclaurin and Taylor Series. Introduction. Prerequisites. Learning Outcomes
Maclaurin and Talor Series 6.5 Introduction In this Section we eamine how functions ma be epressed in terms of power series. This is an etremel useful wa of epressing a function since (as we shall see)
More informationFeature detectors and descriptors. Fei-Fei Li
Feature detectors and descriptors Fei-Fei Li Feature Detection e.g. DoG detected points (~300) coordinates, neighbourhoods Feature Description e.g. SIFT local descriptors (invariant) vectors database of
More information