Detectors part II Descriptors

Similar documents
Extract useful building blocks: blobs. the same image like for the corners

Feature extraction: Corners and blobs

Blob Detection CSC 767

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros

Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context

SIFT keypoint detection. D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp , 2004.

Edges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise

Blobs & Scale Invariance

Feature detectors and descriptors. Fei-Fei Li

Lecture 8: Interest Point Detection. Saad J Bedros

CSE 473/573 Computer Vision and Image Processing (CVIP)

Feature detectors and descriptors. Fei-Fei Li

Overview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors

Overview. Harris interest points. Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points

Image Analysis. Feature extraction: corners and blobs

LoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D

CS5670: Computer Vision

Achieving scale covariance

Recap: edge detection. Source: D. Lowe, L. Fei-Fei

Advances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.

INTEREST POINTS AT DIFFERENT SCALES

Lecture 6: Finding Features (part 1/2)

Lecture 8: Interest Point Detection. Saad J Bedros

Scale & Affine Invariant Interest Point Detectors

Overview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors

Lecture 12. Local Feature Detection. Matching with Invariant Features. Why extract features? Why extract features? Why extract features?

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems

CS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015

Invariant local features. Invariant Local Features. Classes of transformations. (Good) invariant local features. Case study: panorama stitching

Maximally Stable Local Description for Scale Selection

Corner detection: the basic idea

Lecture 7: Finding Features (part 2/2)

Wavelet-based Salient Points with Scale Information for Classification

SIFT: Scale Invariant Feature Transform

SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE

Instance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble

Lecture 7: Finding Features (part 2/2)

SURF Features. Jacky Baltes Dept. of Computer Science University of Manitoba WWW:

Keypoint extraction: Corners Harris Corners Pkwy, Charlotte, NC

SURVEY OF APPEARANCE-BASED METHODS FOR OBJECT RECOGNITION

Harris Corner Detector

Image matching. by Diva Sian. by swashford

Given a feature in I 1, how to find the best match in I 2?

Local Features (contd.)

The Kadir Operator Saliency, Scale and Image Description. Timor Kadir and Michael Brady University of Oxford

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

Scale-space image processing

Instance-level l recognition. Cordelia Schmid & Josef Sivic INRIA

Advanced Features. Advanced Features: Topics. Jana Kosecka. Slides from: S. Thurn, D. Lowe, Forsyth and Ponce. Advanced features and feature matching

Feature extraction: Corners and blobs

Affine Adaptation of Local Image Features Using the Hessian Matrix

Scale & Affine Invariant Interest Point Detectors

Image Processing 1 (IP1) Bildverarbeitung 1

Instance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble

CS4670: Computer Vision Kavita Bala. Lecture 7: Harris Corner Detec=on

Instance-level l recognition. Cordelia Schmid INRIA

ECE 468: Digital Image Processing. Lecture 8

SIFT, GLOH, SURF descriptors. Dipartimento di Sistemi e Informatica

EE 6882 Visual Search Engine

arxiv: v1 [cs.cv] 10 Feb 2016

The state of the art and beyond

Lecture 6: Edge Detection. CAP 5415: Computer Vision Fall 2008

Instance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble

Edge Detection. CS 650: Computer Vision

Edge Detection PSY 5018H: Math Models Hum Behavior, Prof. Paul Schrater, Spring 2005

EECS150 - Digital Design Lecture 15 SIFT2 + FSM. Recap and Outline

Hilbert-Huang Transform-based Local Regions Descriptors

Orientation Map Based Palmprint Recognition

Lecture 7: Edge Detection

Face detection and recognition. Detection Recognition Sally

Feature Vector Similarity Based on Local Structure

Filtering and Edge Detection

Feature detection.

6.869 Advances in Computer Vision. Prof. Bill Freeman March 1, 2005

Computer Vision Lecture 3

Templates, Image Pyramids, and Filter Banks

Edge Detection. Computer Vision P. Schrater Spring 2003

Visual Object Recognition

Machine vision. Summary # 4. The mask for Laplacian is given

Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition

Machine vision, spring 2018 Summary 4

Taking derivative by convolution

Interest Operators. All lectures are from posted research papers. Harris Corner Detector: the first and most basic interest operator

Lesson 04. KAZE, Non-linear diffusion filtering, ORB, MSER. Ing. Marek Hrúz, Ph.D.

A New Shape Adaptation Scheme to Affine Invariant Detector

Lecture 05 Point Feature Detection and Matching

Introduction to Computer Vision

Scale Space Smoothing, Image Feature Extraction and Bessel Filters

Coding Images with Local Features

Perception III: Filtering, Edges, and Point-features

KAZE Features. 1 Introduction. Pablo Fernández Alcantarilla 1, Adrien Bartoli 1, and Andrew J. Davison 2

Rapid Object Recognition from Discriminative Regions of Interest

Introduction to Computer Vision

Human Action Recognition under Log-Euclidean Riemannian Metric

Laplacian Filters. Sobel Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters

SYMMETRY is a highly salient visual phenomenon and

Robert Collins CSE598G Mean-Shift Blob Tracking through Scale Space

Lecture 04 Image Filtering

Region Covariance: A Fast Descriptor for Detection and Classification

Transcription:

EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers

Goal: Identify interesting regions from the images (edges, corners, blobs ) Descriptors Matching / Indexing / Recognition e.g. SIFT

Repeatability The same feature can be found in several images despite geometric and photometric transformations Saliency Each feature is found at an interesting region of the image Locality A feature occupies a relatively small area of the image;

Repeatability Illumination invariance Scale invariance Pose invariance Rotation Affine

Saliency Locality

Harris Detector

Invariance Detector Illumination Rotation Scale View point Harris corner partial No No

Harris Detector: Some Properties Rotation invariance Corner response R is invariant to image rotation C=U 1 [ ] λ1 0 0 U λ2

Harris Detector: Some Properties But: non-invariant to image scale! All points will be classified as edges Corner!

Extract useful building blocks: blobs

Edge detection f d g dx d f g dx Edge Derivative of Gaussian Edge = maximum of derivative Source: S. Seitz

Edge detection as zero crossing f Edge 2 Second derivative of Gaussian (Laplacian) d g 2 dx 2 d f 2 g dx Edge = zero crossing of second derivative Source: S. Seitz

Edge detection as zero crossing edge edge

From edges to blobs Blob = superposition of nearby edges maximum Magnitude of the Laplacian response achieves a maximum at the center of the blob, provided the scale of the Laplacian is matched to the scale of the blob

From edges to blobs Blob = superposition of nearby edges maximum What if the blob is slightly thicker or slimmer?

Scale selection We want to find the characteristic scale of the blob by convolving it with Laplacians at several scales and looking for the maximum response However, Laplacian response decays as scale increases: original signal (radius=8) increasing σ This should give the max response Why does this happen?

Scale normalization The response of a derivative of Gaussian filter to a perfect step edge decreases as σ increases 1 σ 2π

Scale normalization To keep response the same (scaleinvariant), must multiply Gaussian derivative by σ Laplacian is the second Gaussian derivative, so it must be multiplied by σ2

Effect of scale normalization Original signal Unnormalized Laplacian response Scale-normalized Laplacian response Maximum

Blob detection in 2D Laplacian of Gaussian: Circularly symmetric operator for blob detection in 2D 2 2 g g g= 2 2 x y 2

Blob detection in 2D Laplacian of Gaussian: Circularly symmetric operator for blob detection in 2D 2 2 Scale-normalized: norm g=σ 2 2 g g 2 2 x y

Scale selection For a binary circle of radius r, the Laplacian achieves a maximum at σ=r / 2 image Laplacian response r r/ 2 scale (σ)

Characteristic scale We define the characteristic scale as the scale that produces peak of Laplacian response characteristic scale T. Lindeberg (1998). "Feature detection with automatic scale selection." International Journal of Computer Vision 30 (2): pp 77--116.

Scale-space blob detector 1. Convolve image with scale-normalized Laplacian at several scales 1. Find maxima of squared Laplacian response in scale-space 1. This indicate if a blob has been detected 1. And what s its intrinsic scale

Scale-space blob detector: Example

Scale-space blob detector: Example

Scale-space blob detector: Example

DOG David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04 Approximating the Laplacian with a difference of Gaussians: L = σ 2 ( Gxx ( x, y, σ ) + G yy ( x, y, σ ) ) (Laplacian) DoG = G ( x, y, kσ ) G ( x, y, σ ) (Difference of Gaussians) ***or*** Difference of gaussian blurred images at scales k σ and σ L

DOG k Output: location, scale, orientation (more later)

Example of keypoint detection

Invariance Detect or Illuminati on Rotatio n Scale View point Harris corner No No Lowe 99 (DoG) No

Harris-Laplace [Mikolajczyk & Schmid 01] Collect locations (x,y) of detected Harris features for σ = σ1 σ2 (the sigma is here comes from g g ) x, y For each detected location (x,y) and for each σ, reject detection if Laplacian(x,y, σ) is not a local maximum Output: location, scale

Invariance Detect or Illuminati on Rotatio n Scale View point Harris corner No No Lowe 99 (DoG) No Mikolajczy k& Schmid 01 No

Repeatability Illumination invariance Scale invariance Pose invariance Rotation Affine

Affine invariance K. Mikolajczyk and C. Schmid, Scale and Affine invariant interest point detectors, IJCV 60(1):63-86, 2004. Similarly to characteristic scale selection, detect the characteristic shape of the local feature

Affine adaptation M= w x,y x,y [ 2 Ix Ix I y IxIy I 2y ] [ ] 1 =R λ1 0 0 R λ2 We can visualize M as an ellipse with axis lengths determined by the eigenvalues and orientation determined by R Ellipse equation: [] u [u v ] M = const v direction of the fastest change (λmax)-1/2 direction of the slowest change (λmin)-1/2 The second moment ellipse can be viewed as the characteristic shape of a region

Affine adaptation Detect initial region with Harris Laplace Estimate affine shape with M Normalize the affine region to a circular one Re-detect the new location and scale in the normalized image Go to step 2 if the eigenvalues of the M for the new point are not equal [detector not yet adapted to the characteristic shape]

Affine adaptation Output: location, scale, affine shape, rotation (more later)

Affine adaptation example Scale-invariant regions (blobs)

Affine adaptation example Affine-adapted blobs

Invariance Detect or Illuminati on Rotatio n Scale View point Harris corner No No Lowe 99 (DoG) No Mikolajczy k& Schmid 01 No Mikolajczy k& Schmid 02

Detect or Illuminati on Rotatio n Scale View point Harris corner No No Lowe 99 (DoG) Mikolajczy k& Schmid 01, 02 Tuytelaars, 00 No ( 04 ) Kadir & Brady, 01 no Matas, 02 no

Blob detectors Invariance Descriptors

Goal: Identify interesting regions from the images (edges, corners, blobs ) Descriptors Matching / Indexing / Recognition e.g. SIFT

Matching Features (stitching images) Detect feature points in both images

Matching Features (stitching images) Detect feature points in both images Find corresponding pairs

Matching Features (stitching images) Detect feature points in both images Find corresponding pairs Use these pairs to align images

Matching Features (estimating F) Detect feature points in both images Find corresponding pairs Use these pairs to estimate F

Matching Features (recognizing objects) Detect feature points in both images Find corresponding pairs Use these pairs to match different object instances A a

Challenges Depending on the application a descriptor must incorporate information that is: Invariant w.r.t: Illumination Pose Scale Intraclass variability A a Highly distinctive (allows a single feature to find its correct match with good probability in a large database of features)

Illumination normalization Affine intensity change: I I+b ai+b Make each patch zero mean: Then make unit variance: I x (image coordinate)

Pose normalization View 1 Scale, rotation & sheer NOTE: location, scale, rotation & affine pose are given by the detector or calculated within the detected regions View 2

Pose normalization Keypoints are transformed in order to be invariant to translation, rotation, scale, and other geometrical parameters Courtesy of D. Lowe Change of scale, pose, illumination

The simplest descriptor M N 1 x NM vector of pixel intensities w= [ ] w w w n= w w Makes the descriptor invariant with respect to affine transformation of the illumination condition

Why can t we just use this? Sensitive to small variation of: location Pose Scale intra-class variability Poorly distinctive

Sensitive to pose variations Norm. corr Normalized Correlation: ' ' w w w w ' w n w n = ' ' w w w w

Detector Illumination Pose Intra-class variab. PATCH *** * *

Bank of filters image filter bank filter responses More robust but still quite sensitive to pose variations descriptor

Bank of filters - Steerable filters

Detector Illumination Pose Intra-class variab. PATCH *** * * FILTERS *** ** **

SIFT descriptor David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04 Alternative representation for image patches Location and characteristic scale s given by DoG detector s

SIFT descriptor David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04 Alternative representation for image patches Location and characteristic scale s given by DoG detector Compute gradient at each pixel N x N spatial bins Compute an histogram of M orientations for each bean θ1 θ1 θm

SIFT descriptor David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04 Alternative representation for image patches Location and characteristic scale s given by DoG detector Compute gradient at each pixel N x N spatial bins Compute an histogram of M orientations for each bean Gaussian center-weighting θ1 θ1 θm

SIFT descriptor David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04 Alternative representation for image patches Location and characteristic scale s given by DoG detector Compute gradient at each pixel N x N spatial bins Compute an histogram of M orientations for each bean Gaussian center-weighting Normalized unit norm Typically M = 8; N= 4 1 x 128 descriptor

SIFT descriptor Robust w.r.t. small variation in: Illumination (thanks to gradient & normalization) Pose (small affine variation thanks to orientation histogram ) Scale (scale is fixed by DOG) Intra-class variability (small variations thanks to histograms)

Rotational invariance Find dominant orientation by building smoothed orientation histogram Rotate all orientations by the dominant orientation 0 2 π This makes the SIFT descriptor rotational invariant

Rotational invariance

Rotational invariance

Matching using SIFT David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04

Matching using SIFT David G. Lowe. "Distinctive image features from scale-invariant keypoints. IJCV 60 (2), 04

Detector Illumination Pose Intra-class variab. PATCH *** * * FILTERS *** ** ** SIFT *** *** ***

Shape context Belongie et al. 2002 Histogram (occurrences within each bin) A 3 1 1 2 3 4 5 6 7 8 9 10 11. Bin #

Matching different instances

EECS 442 Computer vision Next lecture: Mid level representations

Other interest point detectors Scale Saliency [Kadir & Brady 01, 03]

Other interest point detectors Scale Saliency [Kadir & Brady 01, 03] Uses entropy measure of local pdf of intensities: H D s,x = p d,s,x log 2 p d,s,x. dd d D Takes local maxima in scale Weights with change of distribution with scale: W D s,x =s p d,s,x. dd s d D To get saliency measure: Y D s,x =H D s,x W D s,x

Other interest point detectors Scale Saliency [Kadir & Brady 01, 03] HD(s,x) Just using HD(s,x) WD(s,x) Using YD(s,x) = HDWD Most salient parts detected

Other interest point detectors maximum stable extremal regions [matas et al. 02] Sweep threshold of intensity from black to white Locate regions based on stability of region with respect to change of threshold

Creating features stable to viewpoint change Edelman, Intrator & Poggio (97) showed that complex cell outputs are better for 3D recognition than simple correlation

Feature stability to noise Match features after random change in image scale & orientation, with differing levels of image noise Find nearest neighbor in database of 30,000 features

Feature stability to affine change Match features after random change in image scale & orientation, with 2% image noise, and affine distortion Find nearest neighbor in database of 30,000 features

Geometric blur Berg et al. 2001

Geometric blur