The Kadir Operator Saliency, Scale and Image Description. Timor Kadir and Michael Brady University of Oxford

Similar documents
Detectors part II Descriptors

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros

Feature detectors and descriptors. Fei-Fei Li

Feature detectors and descriptors. Fei-Fei Li

CS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015

SIFT keypoint detection. D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp , 2004.

Blob Detection CSC 767

LoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D

CSE 473/573 Computer Vision and Image Processing (CVIP)

Rapid Object Recognition from Discriminative Regions of Interest

Achieving scale covariance

Maximally Stable Local Description for Scale Selection

Wavelet-based Salient Points with Scale Information for Classification

Feature extraction: Corners and blobs

Scale & Affine Invariant Interest Point Detectors

SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE

Edges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise

Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context

Overview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors

CS5670: Computer Vision

Image Compression Based on Visual Saliency at Individual Scales

Recap: edge detection. Source: D. Lowe, L. Fei-Fei

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems

Overview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors

Overview. Harris interest points. Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points

Lecture 8: Interest Point Detection. Saad J Bedros

The state of the art and beyond

Lecture 8: Interest Point Detection. Saad J Bedros

Advances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.

SURVEY OF APPEARANCE-BASED METHODS FOR OBJECT RECOGNITION

Learning features by contrasting natural images with noise

Image matching. by Diva Sian. by swashford

Object Recognition Using Local Characterisation and Zernike Moments

A New Efficient Method for Producing Global Affine Invariants

Human Action Recognition under Log-Euclidean Riemannian Metric

Lecture 12. Local Feature Detection. Matching with Invariant Features. Why extract features? Why extract features? Why extract features?

Feature Extraction and Image Processing

Loss Functions and Optimization. Lecture 3-1

Interest Operators. All lectures are from posted research papers. Harris Corner Detector: the first and most basic interest operator

Blobs & Scale Invariance

Instance-level l recognition. Cordelia Schmid INRIA

SURF Features. Jacky Baltes Dept. of Computer Science University of Manitoba WWW:

Lecture 6: Edge Detection. CAP 5415: Computer Vision Fall 2008

Deformation and Viewpoint Invariant Color Histograms

Binary Image Analysis

Probability Distributions

Image Analysis. Feature extraction: corners and blobs

arxiv: v1 [cs.cv] 10 Feb 2016

Large-scale classification of traffic signs under real-world conditions

Multimedia Databases. Previous Lecture. 4.1 Multiresolution Analysis. 4 Shape-based Features. 4.1 Multiresolution Analysis

Extract useful building blocks: blobs. the same image like for the corners

CITS 4402 Computer Vision

Shape of Gaussians as Feature Descriptors

Multimedia Databases. Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig

Instance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble

Global Scene Representations. Tilke Judd

Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition

Information Theory in Computer Vision and Pattern Recognition

6.869 Advances in Computer Vision. Prof. Bill Freeman March 1, 2005

INTEREST POINTS AT DIFFERENT SCALES

Kernel Density Topic Models: Visual Topics Without Visual Words

Clustering with k-means and Gaussian mixture distributions

Hilbert-Huang Transform-based Local Regions Descriptors

SIFT: Scale Invariant Feature Transform

INF Shape descriptors

Advanced Features. Advanced Features: Topics. Jana Kosecka. Slides from: S. Thurn, D. Lowe, Forsyth and Ponce. Advanced features and feature matching

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

Lecture 6: Finding Features (part 1/2)

Keypoint extraction: Corners Harris Corners Pkwy, Charlotte, NC

Metric-based classifiers. Nuno Vasconcelos UCSD

DESIGNING CNN GENES. Received January 23, 2003; Revised April 2, 2003

Roadmap. Introduction to image analysis (computer vision) Theory of edge detection. Applications

Comparative study of global invariant. descriptors for object recognition

Image Segmentation: Definition Importance. Digital Image Processing, 2nd ed. Chapter 10 Image Segmentation.

Filtering and Edge Detection

Lecture 7: Finding Features (part 2/2)

Robust License Plate Detection Using Covariance Descriptor in a Neural Network Framework

Scale & Affine Invariant Interest Point Detectors

Analysis on a local approach to 3D object recognition

Loss Functions and Optimization. Lecture 3-1

Multimedia Databases. 4 Shape-based Features. 4.1 Multiresolution Analysis. 4.1 Multiresolution Analysis. 4.1 Multiresolution Analysis

Tutorial on Methods for Interpreting and Understanding Deep Neural Networks. Part 3: Applications & Discussion

Corner detection: the basic idea

Instance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble

Image Processing 1 (IP1) Bildverarbeitung 1

Distinguish between different types of scenes. Matching human perception Understanding the environment

Natural Image Statistics

Luminance, disparity and range statistics in 3D natural scenes

Large Scale Environment Partitioning in Mobile Robotics Recognition Tasks

Nonlinear Classification

Invariant local features. Invariant Local Features. Classes of transformations. (Good) invariant local features. Case study: panorama stitching

Slide a window along the input arc sequence S. Least-squares estimate. σ 2. σ Estimate 1. Statistically test the difference between θ 1 and θ 2

Image Contrast Enhancement and Quantitative Measuring of Information Flow

Corner. Corners are the intersections of two edges of sufficiently different orientations.

Instance-level recognition: Local invariant features. Cordelia Schmid INRIA, Grenoble

Original Article Design Approach for Content-Based Image Retrieval Using Gabor-Zernike Features

SYMMETRY is a highly salient visual phenomenon and

Aruna Bhat Research Scholar, Department of Electrical Engineering, IIT Delhi, India

Model neurons!!poisson neurons!

Transcription:

The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford 1

The issues salient standing out from the rest, noticeable, conspicous, prominent scale find the best scale for a feature image description create a descriptor for use in object recognition 2

Early Vision Motivation pre-attentive stage: features pop out attentive stage: relationships between features and grouping 3

4

Detection of Salient Features for an Object Class 5

How do we do this? 1. fixed size windows (simple approach) 2. Harris detector, Lowe detector, etc. 3. Kadir s approach 6

Kadir s Approach Scale is intimately related to the problem of determining saliency and extracting relevant descriptions. Saliency is related to the local image complexity, ie. Shannon entropy. entropy definition H = - P i log 2 P i i in set of interest 7

Specifically x is a point on the image R x is its local neighborhood D is a descriptor and has values {d 1,... d r }. P D,Rx (d i ) is the probability of descriptor D taking the value d i in the local region R x. (The normalized histogram of the gray tones in a region estimates this probability distribution.) 8

Local Histograms of Intensity Neighborhoods with structure have flatter distributions which converts to higher entropy. 9

Problems Kadir wanted to solve 1. Scale should not be a global, preselected parameter 2. Highly textured regions can score high on entropy, but not be useful 3. The algorithm should not be sensitive to small changes in the image or noise. 10

Kadir s Methodology use a scale-space approach features will exist over multiple scales Berghoml (1986) regarded features (edges) that existed over multiple scales as best. Kadir took the opposite approach. He considers these too self-similar. Instead he looks for peaks in (weighted) entropy over the scales. 11

The Algorithm 1. For each pixel location x a. For each scale s between smin and smax i. Measure the local descriptor values within a window of scale s ii. Estimate the local PDF (use a histogram) b. Select scales (set S) for which the entropy is peaked (S may be empty) c. Weight the entropy values in S by the sum of absolute difference of the PDFs of the local descriptor around S. 12

Finding salient points the math for saliency discretized Y H W x s = ( s, x) s, x ( s, x) ( s, x) point ( s, x) W s, x ( s, x) s, x s, x s 1, x ( s, r, θ ) = ( scale, eccentricity, orientation) D = low - level feature domain p D D D = ( d) = H D d D p s2 = 2s 1 = = d D D ( d)log p 2 ( d) p p ( d) ( d) (gray tones) histogram probability of of descriptor valuesdoftaking value d in the region centered at x with scale s D in region s, x saliency entropy weight based on difference between scales s X = normalized histogram count for the bin representing gray tone d. 13

Picking salient points and their scales 14

Getting rid of texture One goal was to not select highly textured regions such as grass or bushes, which are not the type of objects the Oxford group wanted to recognize Such regions are highly salient with just entropy, because they contain a lot of gray tones in roughly equal proportions But they are similar at different scales and thus the weights make them go away 15

Salient Regions Instead of just selecting the most salient points (based on weighted entropy), select salient regions (more robust). Regions are like volumes in scale space. Kadir used clustering to group selected points into regions. We found the clustering was a critical step. 16

Kadir s clustering (VERY ad hoc) Apply a global threshold on saliency. Choose the highest salient points (50% works well). Find the K nearest neighbors (K=8 preset) Check variance at center points with these neighbors. Accept if far enough away from existant clusters and variance small enough. Represent with mean scale and spatial location of the K points Repeat with next highest salient point 17

More examples 18

Robustness Claims scale invariant (chooses its scale) rotation invariant (uses circular regions and histograms) somewhat illumination invariant (why?) not affine invariant (able to handle small changes in viewpoint) 19

More Examples 20

Temple 21

Capitol 22

Houses and Boats 23

Houses and Boats 24

Sky Scraper 25

Car 26

Trucks 27

Fish 28

Other 29

Symmetry and More 30

Benefits General feature: not tied to any specific object Can be used to detect rather complex objects that are not all one color Location invariant, rotation invariant Selects relevant scale, so scale invariant What else is good? Anything bad? 31

References Kadir, Brady; Saliency, scale and image description; IJCV 45(2), 83-105, 2001 Kadir, Brady; Scale saliency: a novel approach to salient feature and scale selection Treisman; Visual coding of features and objects: Some evidence from behavioral studies; Advances in the Modularity of Vision Selections; NAS Press, 1990 Wolfe, Treisman, Horowitz; What shall we do with the preattentive processing stage: Use it or lose it? (poster); 3 rd Annual Mtg Vis Sci Soc 32

References Dayan, Abbott; Theoretical Neuroscience; MIT Press, 2001 Lamme; Separate neural definitions of visual consciousness and visual attention; Neural Networks 17, 861-872, 2004 Di Lollo, Kawahara, Zuvic, Visser; The preattentive emperor has no clothes: A dynamic redressing; J Experimental Psych, General 130(3), 479-492 Hochstein, Ahissar; View from the top: Hierarchies and reverse hierarchies in the visual system; Neuron 36, 791-804, 2002 33