Video and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)

Size: px
Start display at page:

Download "Video and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)"

Transcription

1 Video and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)

2 Optical flow used for feature tracking on a drone

3 Interpolated optical flow used for super slow-mo

4 optical flow used for motion estimation in visual odometry

5 Roadmap (Where we have been and where we are going)

6 Image filtering Frequency domain image pyramids Image gradients Boundaries Hough Transform Image Manipulation (January)

7 Corner detection Multi-scale detection Haar-like HOG SURF SIFT Image Features (February)

8 Bag-of-words K-means Nearest Neighbor Naive Bayes SVM Object Recognition (February)

9 2D Transforms DLT RANSAC 2D Alignment (March)

10 x = PX P X camera matrix pose estimation triangulation F fundamental matrix epipolar geometry Reconstruction 2 view geometry (March)

11 Block matching Energy minimization Stereo (March)

12 What you can do now Detect lines (circles, shapes) in an image Recognize objects using a bag-of-words model Automatic image warping and basic AR Reconstruct 3D scene structure from two images

13 What you will learn next Object tracking in video

14 2 6 4 I x (p 1 ) I y (p 1 ) I x (p 2 ) I y (p 2 ).. I x (p 25 ) I y (p 25 ) apple u v = I t (p 1 ) I t (p 2 ). I t (p 25 ) min u, v X ij E d (i, j)+ E s (i, j) Constant Flow Horn Schunck Optical Flow (April)

15 Lucas Kanade (Forward additive) Baker Matthews (Inverse Compositional) Image Alignment (April)

16 KLT Mean shift Kalman Filtering SLAM Tracking in Video (April-May)

17 Brightness Constancy Computer Vision Carnegie Mellon University (Kris Kitani)

18 Optical Flow Problem Definition Given two consecutive image frames, estimate the motion of each pixel Assumptions Brightness constancy Small motion

19 Assumption 1 Brightness constancy Scene point moving through image sequence (x(2),y(2)) (x(k),y(k)) (x(1),y(1))

20 Assumption 1 Brightness constancy Scene point moving through image sequence (x(2),y(2)) (x(k),y(k)) (x(1),y(1)) I(x, y, 1) I(x, y, 2) I(x, y, k)

21 Assumption 1 Brightness constancy Scene point moving through image sequence (x(2),y(2)) (x(k),y(k)) (x(1),y(1)) I(x, y, 1) I(x, y, 2) I(x, y, k) Assumption:Brightness of the point will remain the same

22 Assumption 1 Brightness constancy Scene point moving through image sequence (x(2),y(2)) (x(k),y(k)) (x(1),y(1)) I(x, y, 1) I(x, y, 2) I(x, y, k) Assumption:Brightness of the point will remain the same I(x(t),y(t),t)=C constant

23 Assumption 2 Small motion I(x, y, t) I(x, y, t + t)

24 Assumption 2 Small motion (x + u t, y + v t) (x, y) I(x, y, t) (x, y) I(x, y, t + t)

25 Assumption 2 Small motion (x + u t, y + v t) (x, y) I(x, y, t) (x, y) I(x, y, t + t) Optical flow (velocities): (u, v) Displacement: ( x, y) =(u t, v t)

26 Assumption 2 Small motion (x + u t, y + v t) (x, y) I(x, y, t) (x, y) I(x, y, t + t) Optical flow (velocities): (u, v) Displacement: ( x, y) =(u t, v t) For a really small space-time step I(x + u t, y + v t, t + t) =I(x, y, t) the brightness between two consecutive image frames is the same

27 These assumptions yield the Brightness Constancy Equation di dx dy =0 total derivative partial derivative Where does this come from?

28 I(x + u t, y + v t, t + t) =I(x, y, t) For small space-time step, brightness of a point is the same If the time step is really small, we can linearize the intensity function

29 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b)

30 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b) I(x, t = I(x, y, t) assuming small motion

31 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b) I(x, t = I(x, y, t) assuming small motion cancel terms

32 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b) I(x, t = I(x, y, t) assuming t =0 cancel terms

33 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b) I(x, t = I(x, y, t) assuming t =0 divide by take limit t t! 0

34 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b) I(x, t = I(x, y, t) assuming t =0 divide by take limit t dx dy =0

35 I(x + u t, y + v t, t + t) =I(x, y, t) Multivariable Taylor Series Expansion (First order approximation, two variables) f(x, y) f(a, b)+f x (a, b)(x a) f y (a, b)(y b) I(x, t = I(x, y, t) assuming t =0 divide by take limit t dx dy =0 Brightness Constancy Equation

36 dx dy =0 Brightness Constancy Equation I x u + I y v + I t =0 shorthand notation (displacement) ri > v + I t =0 vector form

37 (putting the math aside for a second ) What do the term of the brightness constancy equation represent?! I x u + I y v + I t =0

38 (putting the math aside for a second ) What do the term of the brightness constancy equation represent?! I x u + I y v + I t =0 Image gradients

39 (putting the math aside for a second ) What do the term of the brightness constancy equation represent?! flow velocities I x u + I y v + I t =0 Image gradients

40 (putting the math aside for a second ) What do the term of the brightness constancy equation represent?! flow velocities I x u + I y v + I t =0 Image gradients temporal gradient How do you compute these terms?

41 I x u + I y v + I t =0 How do you compute I I spatial derivative

42 I x u + I y v + I t =0 How do you compute I I spatial derivative Forward difference Sobel filter Scharr filter

43 I x u + I y v + I t =0 How do you compute I I I spatial derivative temporal derivative Forward difference Sobel filter Scharr filter

44 I x u + I y v + I t =0 How do you compute I I I spatial derivative temporal derivative Forward difference Sobel filter Scharr filter frame differencing

45 Frame differencing t t +1 I = (example of a forward difference)

46 Example: t t +1 x x x 3 patch y y I x I x I x y y y

47 I x u + I y v + I t =0 How do you compute I I spatial derivative u = dx v = dy dt dt optical flow I temporal derivative Forward difference Sobel filter Scharr filter How do you compute this? frame differencing

48 I x u + I y v + I t =0 How do you compute I I spatial derivative u = dx v = dy dt dt optical flow I temporal derivative Forward difference Sobel filter Scharr filter We need to solve for this!! (this is the unknown in the optical flow problem) frame differencing

49 I x u + I y v + I t =0 How do you compute I I spatial derivative u = dx v = dy dt dt optical flow I temporal derivative Forward difference Sobel filter Scharr filter (u, v) Solution lies on a line Cannot be found uniquely with a single constraint frame differencing

50 I x u + I y v + I t =0 I I y u = dx v = dy t dt spatial derivative optical flow temporal derivative How can we use the brightness constancy equation to estimate the optical flow?

51 unknown I x u + I y v + I t =0 known We need at least equations to solve for 2 unknowns.

52 unknown I x u + I y v + I t =0 known Where do we get more equations (constraints)?

Lucas-Kanade Optical Flow. Computer Vision Carnegie Mellon University (Kris Kitani)

Lucas-Kanade Optical Flow. Computer Vision Carnegie Mellon University (Kris Kitani) Lucas-Kanade Optical Flow Computer Vision 16-385 Carnegie Mellon University (Kris Kitani) I x u + I y v + I t =0 I x = @I @x I y = @I u = dx v = dy I @y t = @I dt dt @t spatial derivative optical flow

More information

Motion Estimation (I) Ce Liu Microsoft Research New England

Motion Estimation (I) Ce Liu Microsoft Research New England Motion Estimation (I) Ce Liu celiu@microsoft.com Microsoft Research New England We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion

More information

Elaborazione delle Immagini Informazione multimediale - Immagini. Raffaella Lanzarotti

Elaborazione delle Immagini Informazione multimediale - Immagini. Raffaella Lanzarotti Elaborazione delle Immagini Informazione multimediale - Immagini Raffaella Lanzarotti OPTICAL FLOW Thanks to prof. Mubarak Shah,UCF 2 Video Video: sequence of frames (images) catch in the time Data: function

More information

CS4495/6495 Introduction to Computer Vision. 6B-L1 Dense flow: Brightness constraint

CS4495/6495 Introduction to Computer Vision. 6B-L1 Dense flow: Brightness constraint CS4495/6495 Introduction to Computer Vision 6B-L1 Dense flow: Brightness constraint Motion estimation techniques Feature-based methods Direct, dense methods Motion estimation techniques Direct, dense methods

More information

Image Alignment Computer Vision (Kris Kitani) Carnegie Mellon University

Image Alignment Computer Vision (Kris Kitani) Carnegie Mellon University Lucas Kanade Image Alignment 16-385 Comuter Vision (Kris Kitani) Carnegie Mellon University htt://www.humansensing.cs.cmu.edu/intraface/ How can I find in the image? Idea #1: Temlate Matching Slow, combinatory,

More information

Motion Estimation (I)

Motion Estimation (I) Motion Estimation (I) Ce Liu celiu@microsoft.com Microsoft Research New England We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion

More information

CSCI 250: Intro to Robotics. Spring Term 2017 Prof. Levy. Computer Vision: A Brief Survey

CSCI 250: Intro to Robotics. Spring Term 2017 Prof. Levy. Computer Vision: A Brief Survey CSCI 25: Intro to Robotics Spring Term 27 Prof. Levy Computer Vision: A Brief Survey What Is Computer Vision? Higher-order tasks Face recognition Object recognition (Deep Learning) What Is Computer Vision?

More information

Computer Vision I. Announcements

Computer Vision I. Announcements Announcements Motion II No class Wednesda (Happ Thanksgiving) HW4 will be due Frida 1/8 Comment on Non-maximal supression CSE5A Lecture 15 Shi-Tomasi Corner Detector Filter image with a Gaussian. Compute

More information

Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University

Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University Mean-Shift Tracker 16-385 Computer Vision (Kris Kitani) Carnegie Mellon University Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Mean Shift Algorithm A mode seeking algorithm

More information

Introduction to motion correspondence

Introduction to motion correspondence Introduction to motion correspondence 1 IPAM - UCLA July 24, 2013 Iasonas Kokkinos Center for Visual Computing Ecole Centrale Paris / INRIA Saclay Why estimate visual motion? 2 Tracking Segmentation Structure

More information

Motion estimation. Digital Visual Effects Yung-Yu Chuang. with slides by Michael Black and P. Anandan

Motion estimation. Digital Visual Effects Yung-Yu Chuang. with slides by Michael Black and P. Anandan Motion estimation Digital Visual Effects Yung-Yu Chuang with slides b Michael Black and P. Anandan Motion estimation Parametric motion image alignment Tracking Optical flow Parametric motion direct method

More information

Scale & Affine Invariant Interest Point Detectors

Scale & Affine Invariant Interest Point Detectors Scale & Affine Invariant Interest Point Detectors Krystian Mikolajczyk and Cordelia Schmid Presented by Hunter Brown & Gaurav Pandey, February 19, 2009 Roadmap: Motivation Scale Invariant Detector Affine

More information

Edges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise

Edges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity

More information

CS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015

CS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015 CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature

More information

Feature extraction: Corners and blobs

Feature extraction: Corners and blobs Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over

More information

Kalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University

Kalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University Kalman Filter 16-385 Computer Vision (Kris Kitani) Carnegie Mellon University Examples up to now have been discrete (binary) random variables Kalman filtering can be seen as a special case of a temporal

More information

Image Gradients and Gradient Filtering Computer Vision

Image Gradients and Gradient Filtering Computer Vision Image Gradients and Gradient Filtering 16-385 Computer Vision What is an image edge? Recall that an image is a 2D function f(x) edge edge How would you detect an edge? What kinds of filter would you use?

More information

Blobs & Scale Invariance

Blobs & Scale Invariance Blobs & Scale Invariance Prof. Didier Stricker Doz. Gabriele Bleser Computer Vision: Object and People Tracking With slides from Bebis, S. Lazebnik & S. Seitz, D. Lowe, A. Efros 1 Apertizer: some videos

More information

Computer Vision Motion

Computer Vision Motion Computer Vision Motion Professor Hager http://www.cs.jhu.edu/~hager 12/1/12 CS 461, Copyright G.D. Hager Outline From Stereo to Motion The motion field and optical flow (2D motion) Factorization methods

More information

Lecture 8: Interest Point Detection. Saad J Bedros

Lecture 8: Interest Point Detection. Saad J Bedros #1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:

More information

Image Alignment and Mosaicing Feature Tracking and the Kalman Filter

Image Alignment and Mosaicing Feature Tracking and the Kalman Filter Image Alignment and Mosaicing Feature Tracking and the Kalman Filter Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic

More information

Optical Flow, Motion Segmentation, Feature Tracking

Optical Flow, Motion Segmentation, Feature Tracking BIL 719 - Computer Vision May 21, 2014 Optical Flow, Motion Segmentation, Feature Tracking Aykut Erdem Dept. of Computer Engineering Hacettepe University Motion Optical Flow Motion Segmentation Feature

More information

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Corners, Blobs & Descriptors With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Motivation: Build a Panorama M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 How do we build panorama?

More information

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES V.Sridevi, P.Malarvizhi, P.Mathivannan Abstract Rain removal from a video is a challenging problem due to random spatial distribution and

More information

Multigrid Acceleration of the Horn-Schunck Algorithm for the Optical Flow Problem

Multigrid Acceleration of the Horn-Schunck Algorithm for the Optical Flow Problem Multigrid Acceleration of the Horn-Schunck Algorithm for the Optical Flow Problem El Mostafa Kalmoun kalmoun@cs.fau.de Ulrich Ruede ruede@cs.fau.de Institute of Computer Science 10 Friedrich Alexander

More information

Determining Constant Optical Flow

Determining Constant Optical Flow Determining Constant Optical Flow Berthold K.P. Horn Copyright 003 The original optical flow algorithm [1] dealt with a flow field that could vary from place to place in the image, as would typically occur

More information

Lecture 8: Interest Point Detection. Saad J Bedros

Lecture 8: Interest Point Detection. Saad J Bedros #1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff

More information

What is Motion? As Visual Input: Change in the spatial distribution of light on the sensors.

What is Motion? As Visual Input: Change in the spatial distribution of light on the sensors. What is Motion? As Visual Input: Change in the spatial distribution of light on the sensors. Minimally, di(x,y,t)/dt 0 As Perception: Inference about causes of intensity change, e.g. I(x,y,t) v OBJ (x,y,z,t)

More information

TRACKING and DETECTION in COMPUTER VISION

TRACKING and DETECTION in COMPUTER VISION Technischen Universität München Winter Semester 2013/2014 TRACKING and DETECTION in COMPUTER VISION Template tracking methods Slobodan Ilić Template based-tracking Energy-based methods The Lucas-Kanade(LK)

More information

Iterative Image Registration: Lucas & Kanade Revisited. Kentaro Toyama Vision Technology Group Microsoft Research

Iterative Image Registration: Lucas & Kanade Revisited. Kentaro Toyama Vision Technology Group Microsoft Research Iterative Image Registration: Lucas & Kanade Revisited Kentaro Toyama Vision Technology Group Microsoft Research Every writer creates his own precursors. His work modifies our conception of the past, as

More information

Optical flow. Subhransu Maji. CMPSCI 670: Computer Vision. October 20, 2016

Optical flow. Subhransu Maji. CMPSCI 670: Computer Vision. October 20, 2016 Optical flow Subhransu Maji CMPSC 670: Computer Vision October 20, 2016 Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes CMPSC 670 2 Motion and perceptual organization Sometimes,

More information

Tracking for VR and AR

Tracking for VR and AR Tracking for VR and AR Hakan Bilen November 17, 2017 Computer Graphics University of Edinburgh Slide credits: Gordon Wetzstein and Steven M. La Valle 1 Overview VR and AR Inertial Sensors Gyroscopes Accelerometers

More information

Lecture 13: Tracking mo3on features op3cal flow

Lecture 13: Tracking mo3on features op3cal flow Lecture 13: Tracking mo3on features op3cal flow Professor Fei- Fei Li Stanford Vision Lab Lecture 14-1! What we will learn today? Introduc3on Op3cal flow Feature tracking Applica3ons Reading: [Szeliski]

More information

Integration - Past Edexcel Exam Questions

Integration - Past Edexcel Exam Questions Integration - Past Edexcel Exam Questions 1. (a) Given that y = 5x 2 + 7x + 3, find i. - ii. - (b) ( 1 + 3 ) x 1 x dx. [4] 2. Question 2b - January 2005 2. The gradient of the curve C is given by The point

More information

An Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections

An Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections An Adaptive Confidence Measure for Optical Flows Based on Linear Subspace Projections Claudia Kondermann, Daniel Kondermann, Bernd Jähne, Christoph Garbe Interdisciplinary Center for Scientific Computing

More information

Image Analysis. Feature extraction: corners and blobs

Image Analysis. Feature extraction: corners and blobs Image Analysis Feature extraction: corners and blobs Christophoros Nikou cnikou@cs.uoi.gr Images taken from: Computer Vision course by Svetlana Lazebnik, University of North Carolina at Chapel Hill (http://www.cs.unc.edu/~lazebnik/spring10/).

More information

Detectors part II Descriptors

Detectors part II Descriptors EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers Goal:

More information

Image Alignment and Mosaicing

Image Alignment and Mosaicing Image Alignment and Mosaicing Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic mosaicing Image Enhancement Original

More information

Edge Detection. Introduction to Computer Vision. Useful Mathematics Funcs. The bad news

Edge Detection. Introduction to Computer Vision. Useful Mathematics Funcs. The bad news Edge Detection Introduction to Computer Vision CS / ECE 8B Thursday, April, 004 Edge detection (HO #5) Edge detection is a local area operator that seeks to find significant, meaningful changes in image

More information

Differential Motion Analysis (II)

Differential Motion Analysis (II) Differential Motion Analysis (II) Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@northwestern.edu http://www.eecs.northwestern.edu/~yingwu 1/30 Outline

More information

Visual Object Recognition

Visual Object Recognition Visual Object Recognition Lecture 2: Image Formation Per-Erik Forssén, docent Computer Vision Laboratory Department of Electrical Engineering Linköping University Lecture 2: Image Formation Pin-hole, and

More information

INTEREST POINTS AT DIFFERENT SCALES

INTEREST POINTS AT DIFFERENT SCALES INTEREST POINTS AT DIFFERENT SCALES Thank you for the slides. They come mostly from the following sources. Dan Huttenlocher Cornell U David Lowe U. of British Columbia Martial Hebert CMU Intuitively, junctions

More information

Methods in Computer Vision: Introduction to Optical Flow

Methods in Computer Vision: Introduction to Optical Flow Methods in Computer Vision: Introduction to Optical Flow Oren Freifeld Computer Science, Ben-Gurion University March 22 and March 26, 2017 Mar 22, 2017 1 / 81 A Preliminary Discussion Example and Flow

More information

SLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada

SLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM

More information

Achieving scale covariance

Achieving scale covariance Achieving scale covariance Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region size that is covariant

More information

Erkut Erdem. Hacettepe University February 24 th, Linear Diffusion 1. 2 Appendix - The Calculus of Variations 5.

Erkut Erdem. Hacettepe University February 24 th, Linear Diffusion 1. 2 Appendix - The Calculus of Variations 5. LINEAR DIFFUSION Erkut Erdem Hacettepe University February 24 th, 2012 CONTENTS 1 Linear Diffusion 1 2 Appendix - The Calculus of Variations 5 References 6 1 LINEAR DIFFUSION The linear diffusion (heat)

More information

Vision for Mobile Robot Navigation: A Survey

Vision for Mobile Robot Navigation: A Survey Vision for Mobile Robot Navigation: A Survey (February 2002) Guilherme N. DeSouza & Avinash C. Kak presentation by: Job Zondag 27 February 2009 Outline: Types of Navigation Absolute localization (Structured)

More information

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems 1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Perception Concepts Vision Chapter 4 (textbook) Sections 4.3 to 4.5 What is the course

More information

Multiple View Geometry in Computer Vision

Multiple View Geometry in Computer Vision Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Scene Planes & Homographies Lecture 19 March 24, 2005 2 In our last lecture, we examined various

More information

Subspace Methods for Visual Learning and Recognition

Subspace Methods for Visual Learning and Recognition This is a shortened version of the tutorial given at the ECCV 2002, Copenhagen, and ICPR 2002, Quebec City. Copyright 2002 by Aleš Leonardis, University of Ljubljana, and Horst Bischof, Graz University

More information

Advances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.

Advances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al. 6.869 Advances in Computer Vision Prof. Bill Freeman March 3, 2005 Image and shape descriptors Affine invariant features Comparison of feature descriptors Shape context Readings: Mikolajczyk and Schmid;

More information

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017 EKF and SLAM McGill COMP 765 Sept 18 th, 2017 Outline News and information Instructions for paper presentations Continue on Kalman filter: EKF and extension to mapping Example of a real mapping system:

More information

Deep Learning (CNNs)

Deep Learning (CNNs) 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deep Learning (CNNs) Deep Learning Readings: Murphy 28 Bishop - - HTF - - Mitchell

More information

LoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D

LoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D Achieving scale covariance Blobs (and scale selection) Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region

More information

Differential Motion Analysis

Differential Motion Analysis Differential Motion Analysis Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@ece.northwestern.edu http://www.eecs.northwestern.edu/~yingwu July 19,

More information

Advanced Features. Advanced Features: Topics. Jana Kosecka. Slides from: S. Thurn, D. Lowe, Forsyth and Ponce. Advanced features and feature matching

Advanced Features. Advanced Features: Topics. Jana Kosecka. Slides from: S. Thurn, D. Lowe, Forsyth and Ponce. Advanced features and feature matching Advanced Features Jana Kosecka Slides from: S. Thurn, D. Lowe, Forsyth and Ponce Advanced Features: Topics Advanced features and feature matching Template matching SIFT features Haar features 2 1 Features

More information

Global parametric image alignment via high-order approximation

Global parametric image alignment via high-order approximation Global parametric image alignment via high-order approximation Y. Keller, A. Averbuch 2 Electrical & Computer Engineering Department, Ben-Gurion University of the Negev. 2 School of Computer Science, Tel

More information

Optical Flow, KLT Feature Tracker.

Optical Flow, KLT Feature Tracker. Optical Flow, KL Feature racker E-mail: hogijung@hanang.ac.kr Motion in Computer Vision Motion Structure rom motion Detection/segmentation with direction [1] E-mail: hogijung@hanang.ac.kr Motion Field.s..

More information

Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context

Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Silvio Savarese Lecture 10-16-Feb-15 From the 3D to 2D & vice versa

More information

Partial Derivatives Formulas. KristaKingMath.com

Partial Derivatives Formulas. KristaKingMath.com Partial Derivatives Formulas KristaKingMath.com Domain and range of a multivariable function A function f of two variables is a rule that assigns to each ordered pair of real numbers (x, y) in a set D

More information

6.869 Advances in Computer Vision. Prof. Bill Freeman March 1, 2005

6.869 Advances in Computer Vision. Prof. Bill Freeman March 1, 2005 6.869 Advances in Computer Vision Prof. Bill Freeman March 1 2005 1 2 Local Features Matching points across images important for: object identification instance recognition object class recognition pose

More information

CSE 473/573 Computer Vision and Image Processing (CVIP)

CSE 473/573 Computer Vision and Image Processing (CVIP) CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu inwogu@buffalo.edu Lecture 11 Local Features 1 Schedule Last class We started local features Today More on local features Readings for

More information

Review for Exam 1. Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA

Review for Exam 1. Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA Review for Exam Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 0003 March 26, 204 Abstract Here are some things you need to know for the in-class

More information

Efficient & Robust LK for Mobile Vision

Efficient & Robust LK for Mobile Vision Efficient & Robust LK for Mobile Vision Instructor - Simon Lucey 16-623 - Designing Comuter Vision As Direct Method (ours) Indirect Method (ORB+RANSAC) H. Alismail, B. Browning, S. Lucey Bit-Planes: Dense

More information

Modeling Complex Temporal Composition of Actionlets for Activity Prediction

Modeling Complex Temporal Composition of Actionlets for Activity Prediction Modeling Complex Temporal Composition of Actionlets for Activity Prediction ECCV 2012 Activity Recognition Reading Group Framework of activity prediction What is an Actionlet To segment a long sequence

More information

Lecture 12 Body Pose and Dynamics

Lecture 12 Body Pose and Dynamics 15-869 Lecture 12 Body Pose and Dynamics Yaser Sheikh Human Motion Modeling and Analysis Fall 2012 Debate Advice Affirmative Team Thesis Statement: Every (good) paper has a thesis. What is the most provocative

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE

SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE Overview Motivation of Work Overview of Algorithm Scale Space and Difference of Gaussian Keypoint Localization Orientation Assignment Descriptor Building

More information

Estimating the Rotation of a Ball Using High Speed Camera

Estimating the Rotation of a Ball Using High Speed Camera 1 2 1 1 1 2 Hough ICP 3DCG 3 6% 3% 6% Estimating the Rotation of a Ball Using High Speed Camera Hiroki Kamada, 1 Yoshinori Takeuchi, 2 Tetsuya Matsumoto, 1 Hiroaki Kudo 1 and Noburu Ohnishi 1 In table

More information

CAMFPLAN. A Real-time Markerless Camera Pose Estimation System for Augmented Reality. Srinath Sridhar and Vineet R. Kamat

CAMFPLAN. A Real-time Markerless Camera Pose Estimation System for Augmented Reality. Srinath Sridhar and Vineet R. Kamat CAMFPLAN A Real-time Markerless Camera Pose Estimation System for Augmented Reality By Srinath Sridhar and Vineet R. Kamat UMCEE Report No. 2011-01 Electrical Engineering and Computer Science Civil and

More information

Learning Linear Detectors

Learning Linear Detectors Learning Linear Detectors Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Detection versus Classification Bayes Classifiers Linear Classifiers Examples of Detection 3 Learning: Detection

More information

Feature detectors and descriptors. Fei-Fei Li

Feature detectors and descriptors. Fei-Fei Li Feature detectors and descriptors Fei-Fei Li Feature Detection e.g. DoG detected points (~300) coordinates, neighbourhoods Feature Description e.g. SIFT local descriptors (invariant) vectors database of

More information

Robotics. Mobile Robotics. Marc Toussaint U Stuttgart

Robotics. Mobile Robotics. Marc Toussaint U Stuttgart Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand

More information

FINITE DIFFERENCES. Lecture 1: (a) Operators (b) Forward Differences and their calculations. (c) Backward Differences and their calculations.

FINITE DIFFERENCES. Lecture 1: (a) Operators (b) Forward Differences and their calculations. (c) Backward Differences and their calculations. FINITE DIFFERENCES Lecture 1: (a) Operators (b) Forward Differences and their calculations. (c) Backward Differences and their calculations. 1. Introduction When a function is known explicitly, it is easy

More information

Uncertainty Models in Quasiconvex Optimization for Geometric Reconstruction

Uncertainty Models in Quasiconvex Optimization for Geometric Reconstruction Uncertainty Models in Quasiconvex Optimization for Geometric Reconstruction Qifa Ke and Takeo Kanade Department of Computer Science, Carnegie Mellon University Email: ke@cmu.edu, tk@cs.cmu.edu Abstract

More information

Example: f(x) = 2x² + 1 Solution: Math 2 VM Part 5 Quadratic Functions April 25, 2017

Example: f(x) = 2x² + 1 Solution: Math 2 VM Part 5 Quadratic Functions April 25, 2017 Math 2 Variable Manipulation Part 5 Quadratic Functions MATH 1 REVIEW THE CONCEPT OF FUNCTIONS The concept of a function is both a different way of thinking about equations and a different way of notating

More information

Harris Corner Detector

Harris Corner Detector Multimedia Computing: Algorithms, Systems, and Applications: Feature Extraction By Dr. Yu Cao Department of Computer Science The University of Massachusetts Lowell Lowell, MA 01854, USA Part of the slides

More information

DESIGNING CNN GENES. Received January 23, 2003; Revised April 2, 2003

DESIGNING CNN GENES. Received January 23, 2003; Revised April 2, 2003 Tutorials and Reviews International Journal of Bifurcation and Chaos, Vol. 13, No. 10 (2003 2739 2824 c World Scientific Publishing Company DESIGNING CNN GENES MAKOTO ITOH Department of Information and

More information

Loss Functions and Optimization. Lecture 3-1

Loss Functions and Optimization. Lecture 3-1 Lecture 3: Loss Functions and Optimization Lecture 3-1 Administrative Assignment 1 is released: http://cs231n.github.io/assignments2017/assignment1/ Due Thursday April 20, 11:59pm on Canvas (Extending

More information

SIFT: Scale Invariant Feature Transform

SIFT: Scale Invariant Feature Transform 1 SIFT: Scale Invariant Feature Transform With slides from Sebastian Thrun Stanford CS223B Computer Vision, Winter 2006 3 Pattern Recognition Want to find in here SIFT Invariances: Scaling Rotation Illumination

More information

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e. Kalman Filter Localization Bayes Filter Reminder Prediction Correction Gaussians p(x) ~ N(µ,σ 2 ) : Properties of Gaussians Univariate p(x) = 1 1 2πσ e 2 (x µ) 2 σ 2 µ Univariate -σ σ Multivariate µ Multivariate

More information

Lecture 12. Local Feature Detection. Matching with Invariant Features. Why extract features? Why extract features? Why extract features?

Lecture 12. Local Feature Detection. Matching with Invariant Features. Why extract features? Why extract features? Why extract features? Lecture 1 Why extract eatures? Motivation: panorama stitching We have two images how do we combine them? Local Feature Detection Guest lecturer: Alex Berg Reading: Harris and Stephens David Lowe IJCV We

More information

Single Exposure Enhancement and Reconstruction. Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman

Single Exposure Enhancement and Reconstruction. Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman Single Exposure Enhancement and Reconstruction Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman 1 Reconstruction as an Inverse Problem Original image f Distortion & Sampling

More information

Feature detectors and descriptors. Fei-Fei Li

Feature detectors and descriptors. Fei-Fei Li Feature detectors and descriptors Fei-Fei Li Feature Detection e.g. DoG detected points (~300) coordinates, neighbourhoods Feature Description e.g. SIFT local descriptors (invariant) vectors database of

More information

CSE 598C Vision-based Tracking Seminar. Times: MW 10:10-11:00AM Willard 370 Instructor: Robert Collins Office Hours: Tues 2-4PM, Wed 9-9:50AM

CSE 598C Vision-based Tracking Seminar. Times: MW 10:10-11:00AM Willard 370 Instructor: Robert Collins Office Hours: Tues 2-4PM, Wed 9-9:50AM CSE 598C Vision-based Tracking Seminar Times: MW 10:10-11:00AM Willard 370 Instructor: Robert Collins Office Hours: Tues 2-4PM, Wed 9-9:50AM What is Tracking? typical idea: tracking a single target in

More information

Single-Image-Based Rain and Snow Removal Using Multi-guided Filter

Single-Image-Based Rain and Snow Removal Using Multi-guided Filter Single-Image-Based Rain and Snow Removal Using Multi-guided Filter Xianhui Zheng 1, Yinghao Liao 1,,WeiGuo 2, Xueyang Fu 2, and Xinghao Ding 2 1 Department of Electronic Engineering, Xiamen University,

More information

Dense Optical Flow Estimation from the Monogenic Curvature Tensor

Dense Optical Flow Estimation from the Monogenic Curvature Tensor Dense Optical Flow Estimation from the Monogenic Curvature Tensor Di Zang 1, Lennart Wietzke 1, Christian Schmaltz 2, and Gerald Sommer 1 1 Department of Computer Science, Christian-Albrechts-University

More information

Do We Really Have to Consider Covariance Matrices for Image Feature Points?

Do We Really Have to Consider Covariance Matrices for Image Feature Points? Electronics and Communications in Japan, Part 3, Vol. 86, No. 1, 2003 Translated from Denshi Joho Tsushin Gakkai Ronbunshi, Vol. J85-A, No. 2, February 2002, pp. 231 239 Do We Really Have to Consider Covariance

More information

Laplacian Filters. Sobel Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters

Laplacian Filters. Sobel Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters Sobel Filters Note that smoothing the image before applying a Sobel filter typically gives better results. Even thresholding the Sobel filtered image cannot usually create precise, i.e., -pixel wide, edges.

More information

Announcements. Tracking. Comptuer Vision I. The Motion Field. = ω. Pure Translation. Motion Field Equation. Rigid Motion: General Case

Announcements. Tracking. Comptuer Vision I. The Motion Field. = ω. Pure Translation. Motion Field Equation. Rigid Motion: General Case Announcements Tracking Computer Vision I CSE5A Lecture 17 HW 3 due toda HW 4 will be on web site tomorrow: Face recognition using 3 techniques Toda: Tracking Reading: Sections 17.1-17.3 The Motion Field

More information

Gradient Domain High Dynamic Range Compression

Gradient Domain High Dynamic Range Compression Gradient Domain High Dynamic Range Compression Raanan Fattal Dani Lischinski Michael Werman The Hebrew University of Jerusalem School of Computer Science & Engineering 1 In A Nutshell 2 Dynamic Range Quantity

More information

Bayes Classifiers. CAP5610 Machine Learning Instructor: Guo-Jun QI

Bayes Classifiers. CAP5610 Machine Learning Instructor: Guo-Jun QI Bayes Classifiers CAP5610 Machine Learning Instructor: Guo-Jun QI Recap: Joint distributions Joint distribution over Input vector X = (X 1, X 2 ) X 1 =B or B (drinking beer or not) X 2 = H or H (headache

More information

A Factorization Method for 3D Multi-body Motion Estimation and Segmentation

A Factorization Method for 3D Multi-body Motion Estimation and Segmentation 1 A Factorization Method for 3D Multi-body Motion Estimation and Segmentation René Vidal Department of EECS University of California Berkeley CA 94710 rvidal@eecs.berkeley.edu Stefano Soatto Dept. of Computer

More information

Shape of Gaussians as Feature Descriptors

Shape of Gaussians as Feature Descriptors Shape of Gaussians as Feature Descriptors Liyu Gong, Tianjiang Wang and Fang Liu Intelligent and Distributed Computing Lab, School of Computer Science and Technology Huazhong University of Science and

More information

Optic Flow Computation with High Accuracy

Optic Flow Computation with High Accuracy Cognitive Computer Vision Colloquium Prague, January 12 13, 2004 Optic Flow Computation with High ccuracy Joachim Weickert Saarland University Saarbrücken, Germany joint work with Thomas Brox ndrés Bruhn

More information

Vision-based navigation around small bodies

Vision-based navigation around small bodies Astronet-II, International Final Conference Vision-based navigation around small bodies Pawel Kicman VISION-BASED NAVIGATION IN SPACE Camera LOS (Line-of-sight) sensor Star-horizon measurements Apparent

More information

CS5670: Computer Vision

CS5670: Computer Vision CS5670: Computer Vision Noah Snavely Lecture 5: Feature descriptors and matching Szeliski: 4.1 Reading Announcements Project 1 Artifacts due tomorrow, Friday 2/17, at 11:59pm Project 2 will be released

More information

Image Filtering. Slides, adapted from. Steve Seitz and Rick Szeliski, U.Washington

Image Filtering. Slides, adapted from. Steve Seitz and Rick Szeliski, U.Washington Image Filtering Slides, adapted from Steve Seitz and Rick Szeliski, U.Washington The power of blur All is Vanity by Charles Allen Gillbert (1873-1929) Harmon LD & JuleszB (1973) The recognition of faces.

More information

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1. MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:

More information

Lecture 5. Epipolar Geometry. Professor Silvio Savarese Computational Vision and Geometry Lab. 21-Jan-15. Lecture 5 - Silvio Savarese

Lecture 5. Epipolar Geometry. Professor Silvio Savarese Computational Vision and Geometry Lab. 21-Jan-15. Lecture 5 - Silvio Savarese Lecture 5 Epipolar Geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 5-21-Jan-15 Lecture 5 Epipolar Geometry Why is stereo useful? Epipolar constraints Essential

More information