TRACKING and DETECTION in COMPUTER VISION Filtering and edge detection
|
|
- Jonas Long
- 5 years ago
- Views:
Transcription
1 Technischen Universität München Winter Semester 0/0 TRACKING and DETECTION in COMPUTER VISION Filtering and edge detection Slobodan Ilić
2 Overview Image formation Convolution Non-liner filtering: Median and Bilateral filters Gaussian smoothing Image derivatives: Gradients and Laplacians Edge detection
3 Image formation Basics of image formation
4 Image formation Image formation occurs when a sensor registers radiation that has interacted with physical objects. It consists of several components: Basics of image formation
5 Image formation Image formation occurs when a sensor registers radiation that has interacted with physical objects. It consists of several components: An imaging function - a fundamental abstraction of an image Basics of image formation
6 Image formation Image formation occurs when a sensor registers radiation that has interacted with physical objects. It consists of several components: An imaging function - a fundamental abstraction of an image A geometric model - projection of the D world into D representation Basics of image formation
7 Image formation Image formation occurs when a sensor registers radiation that has interacted with physical objects. It consists of several components: An imaging function - a fundamental abstraction of an image A geometric model - projection of the D world into D representation A radiometric model - describes how light radiated from the object is measured by the imaging sensor. Basics of image formation
8 Image formation Image formation occurs when a sensor registers radiation that has interacted with physical objects. It consists of several components: An imaging function - a fundamental abstraction of an image A geometric model - projection of the D world into D representation A radiometric model - describes how light radiated from the object is measured by the imaging sensor. A color model - describes how different spectral measurements are related to image colors Basics of image formation
9 Imaging function I is an imaging function defined on the area positive real numbers: Because of the discretization we have: Or in general: I : Ω R R + ;(x, y) I(x, y). and takes values in the Ω = [, 640] [, 480] Z, R + = [0, 55] Z + I = I(x, y) ={I R (x, y),i G (x, y),i B (x, y)} Ω Basics of image formation 4
10 Geometric model The pinhole camera model (a) Projection geometry (b) Image plane in front. (c) Notations for pinhole camera model Basics of image formation 5
11 Perspective projection Basics of image formation 6
12 Perspective projection Wold coordinate system is identical to the camera coordinate system with the origin at the center of projection. Under perspective projection, the object point with world coordinates projects to the image point with ideal image coordinates: Ideal image coordinates are expressed in terms of the image coordinate system with the origin at the optical center. Basics of image formation 6
13 Perspective projection Wold coordinate system is identical to the camera coordinate system with the origin at the center of projection. Under perspective projection, the object point with world coordinates P =(X, Y, Z) projects to the image point with ideal image coordinates: x = f X Z y = f Y Z Ideal image coordinates are expressed in terms of the image coordinate system with the origin at the optical center. Basics of image formation 6
14 Lenses and Discrepancies from the Pinhole Camera d f D thin lens equation D + d = f n = f a aperture measure is f-number: D inf f = d 7
15 Aperture Aperture measure is f-number: n = f a Large(wide) aperture, small f-number a = f 4. Small aperture, large f-number a = f 9 Why use wide aperture if images can be made sharp with small aperture? Basics of image formation 8
16 Improving images Noise removal Contrast increase 9
17 Detecting edges Steps in edge detection 0
18 Image processing Image processing is in general not part of Computer Vision, but it is sometimes a very necessary preprocessing step. Image processing provides a number of methods to convert an image into something suitable for analysis. There are two main approaches: Processing in the spacial domain: Point processing (brightness, contrast, histogram equalization etc.) Filtering Processing in the frequency domain: Fourier transform
19 Image processing Image processing is in general not part of Computer Vision, but it is sometimes a very necessary preprocessing step. Image processing provides a number of methods to convert an image into something suitable for analysis. There are two main approaches: Processing in the spacial domain: Point processing (brightness, contrast, histogram equalization etc.) Filtering Processing in the frequency domain: Fourier transform
20 Filtering Original image Blurred image From the draft of Szeliski s book: Filtering is based on neighborhood operations.
21 Neighborhood filtering Correlation and Convolution From the draft of Szeliski s book: H(i, j), i, j convolution kernel J(x, y) I(x, y), x, y x dim,y dim J(x, y) =H I = J(x, y) =H I = i= H(i, j)i(x + i, y + j) j= i= j= H(i, j)i(x i, y j) - correlation - convolution
22 Convolution Example Rotate the convolution kernel H Rotate I - Apply correlation - - J(x, y) =H I = H(i, j)i(x + i, y + j) i= j= 4
23 Convolution Example Step H I H y - - I x J(x, y) =H I 5
24 Convolution Example Step H I H y - - I x J(x, y) =H I 6
25 Convolution Example Step H I H y - - I J(x, y) =H I x 7
26 Convolution Example Step 4 H I H y - - I J(x, y) =H I x 8
27 Convolution Example Step 5 H I H y I x J(x, y) =H I 9
28 Convolution Example Step 6 H I H y I x J(x, y) =H I 0
29 Example * =
30 Example * =
31 Expressing convolution Mathematically convolution can be expresses as: + + J(x, y) =H I = H(i, j)i(x i, y j) i= j= if we change the variables i x i and j y j we get: + + J(x, y) =H I = H(x i, y j)i(i, j) i= j= with the image and the kernel playing interchangeable roles.
32 Convolution by shifting, copying and multiplying the image For each element H(i,j) in the kernel matrix, image I is copied into a zero-padded image, P starting at (i,j). Each P is m u l t i p l i e d b y t h e corresponding weight H(i,j). All the P images are summed pixel-wise, then divided by the sum of the elements of h. The result is copped out of the center of the accumulated P s. J(x, y) =H I = + + i= j= H(x i, y j)i(i, j) 4
33 Shift invariance δ a,b (i, j) = {, for i = a and j = b 0, elsewhere H 5 δ a,b H(x, y) =H(x a, c b) The kernel(point spread function) H is shift invariant or space-invariant because the entries in H do not depend on the position (x, y) in the output image.
34 Convolution properties Commutativity Associativity Distributivity H I = I H H (G I) =(H G) I H (I + J) =(H I)+(H J) Associativity with scalar multiplication a(h I) =(ah) I = H (ai) 6
35 Padding (border effects) zero clamp mirror 7
36 Non-linear filtering Linear filters combine input pixels in a way that depends on where a pixel is in the image and not on its value Non-linear filters take into account input pixel values before deciding how to use them in the output. 8
37 Linear vs. non liner filters original image salt&pepper noise added average filter median filter 9
38 Median Filtering Sort: ( ) Example Original signal: Noisy signal: Filter by [ ]/: Filter by x median filter:
39 Median Filter Median filters are nonlinear Median filtering reduces noise without blurring edges and other sharp details Median filtering is particularly effective when the noise pattern consists of strong, spike-like components. (Salt-and-pepper noise.)
40 Bilateral filter It is nonlinear filtering technique based on: Domain(smoothing) filter - It is based on closeness function which accounts for spacial distance between the central pixel x and its neighbors ξ c(ξ, x) =e d(ξ,x) σ d, d(ξ, x) = ξ x Range filter - It is based on the similarity function between image intensities between the central pixel and its neighbors I(ξ) I(x) σ d s(ξ, x) =e ( δ(i(ξ),i(x)) σr ), δ(i(ξ),i(x)) = I(ξ) I(x) - desired amount of spacial smoothing σ r - desired amount of combining of pixel values
41 Bilateral filter Combining similarity and closeness functions we obtain: h(x) = c ξ Ω I(ξ)c(ξ, x)s((i(ξ),i(x)) c = ξ Ω c(ξ, x)s((i(ξ),i(x)) This filter is know to reduce noise and to preserve edges
42 Examples C.Tomasi, R. Manduchi, "Bilateral Filtering for gray and color images", Sixth International Conference on Computer Vision, pp 89-46, New Delhi, India,
43 Gaussian smoothing G(u, v) =e u +v σ Gaussian kernel Gaussian function 5
44 Gaussian smoothing example Original image Smoothed σ = Smoothed Smoothed σ = σ =4 6
45 Separability of Gaussians G(u, v) =g(u)g(v) G(u, v) =e u +v σ = G(u, v) =e ( u σ ) e ( v σ ) Gaussian separation speeds up computation, because with the separable kernel G the convolution can be separated in two onedimensional convolutions like: n n J(x, y) = g(u) g(v)i(x u, y v) J(x, y) = n u= n u= n v= n g(u)φ(x u, y) φ(x, y) = this requires m multiplications and (m-) additions compared to m^ multiplications and m^- additions with D kernel, where m=n+. n v= n g(v)i(x, y v) 7
46 Gaussian normalization G(x) = πσ e x σ G(x, y) = πσ e x +y σ Since we discretize in practice we normalize not by the integral, but by the sum G(x, y) = c e x +y σ c = n i= n n j= n G(i, j) 8
47 Image derivatives Remember that the image can be represented as a function of its pixel intensities I(x) I(x 0 ) x 0 consequently we should be able to compute derivatives of this function... x 9
48 Linear approximation I(x) I(x 0 ) I(x 0 ) x x x 0 x 0 + x x I(x 0 + x) I(x 0 )+ I(x 0) x x + h.o.t I(x 0 ) x = lim x 0 I(x 0 + x) I(x 0 ) x I(x 0 + x) I(x 0 ) 40
49 Image D profile example D image line profile D image line profile smoothed with Gaussian 4
50 Example images Discontinuity 4 Smoothed with Gaussian
51 Example images derivatives 0 0 Edges correspond to the fast changes in the intensities Magnitude of the derivatives is large -0 4
52 Derivatives as linear filters Image derivative: I(x 0 ) x = lim x 0 I(x 0 + x) I(x 0 ) x I(x 0 + x) I(x 0 ) can be implemented as linear filters: H I =[ ] [I(x 0 ) I(x 0 + x)] T or in its symmetric version: H I = [ 0 ] [I(x 0 x) I(x 0 ) I(x 0 + x)] T 44
53 Partial image derivatives If we consider an image I as a function of vector variables I(x,y) we write image gradient using partial derivatives as: I(x, y) = I x i x + I y i y This can be written using gradient filters: I(x, y) = (D x I) i x +(D y I) i y And we can define gradient intensity and direction: I(x, y) = (D x I) +(D y I) ψ( I) = arctan ( Dy I D x I ) 45
54 Gradient filters Basic derivative filters: D x = 0 0 D y = Prewitt gradient filter Sobel gradient filter D x = 0 0 D x = D y = D y =
55 Gradients smoothed with Gaussians Images smoothed with Gaussians and then filtered with derivatives filters: I x = D x (G I) I y = D y (G I) Because of the associativity: I x =(D x G) I = G x I I y =(D y G) I = G y I images can be directly convolved by the derivatives of the Gaussian filters. 47
56 Derivatives of Gaussians in D G(x) = πσ e x σ G x = G x = x σ πσ e x σ 48
57 Derivatives of Gaussians in D Images are from Carlo Tomasi Computer Vision lecture notes G(x, y) = πσ e x +y σ G x = G x = x πσ 4 e x +y σ G y = G y = y πσ 4 e x +y σ 49
58 Practical Aspects and normalization Derivatives without normalization and due to separability: G x (x, y) = x σ e G y (x, y) = y σ e in discreet case: I x (x, y) = = I y (x, y) = = n i= n j= n n i= n n n i= n n G x (x i) n i= n j= n G(x i) n j= n n j= n 50 x +y σ x +y σ I(i, j)g x (x i, y j) = I(i, j)g(y j) = I(i, j)g y (x i, y j) = I(i, j)g y (y j) = = G x (x)g(y) = G y (y)g(x) n j= n n j= n G(y j) G y (y j) n i= n n i= n I(i, j)g x (x i) I(i, j)g(x i)
59 With normalization I x (x, y) = I y (x, y) = n i= n n i= n G x (x i) G(x i) n j= n n j= n I(i, j)g(y j) I(i, j)g y (y j) Normalize by the sum of the filter samples: G u (u) =k g G u (u), where G u (u) = ue ( u σ ) and k d = n ug u (u) u= n G(u) =k G(u), where G(u) =e ( u σ ) and k = n G u (u) u= n 5
60 Image directional gradients Gradient in x direction Gx Gradient in y direction Gy 5
61 Gradient magnitude I(x, y) = (G x I) +(G y I) σ =0.5 σ =.0 σ =.5 σ =.0 5
62 Gradient orientation T h e u p p e r l e f t corner of the image of 50x50 pixels large. Gradient orientation computed as: ψ( I) = arctan ( G x I G y I ) 54
63 Second order derivatives Mathematically we have: I(x 0 ) x = lim x 0 I (x 0 + x) I (x 0 ) x I (x 0 + x) I (x 0 ) Further simplified: I(x 0 ) x = I = I(x 0 + x) I(x 0 + x)+i(x 0 ) In D: I(x,y) x = I(x, y) = I x + I y 55
64 Example images derivatives 0 Original image 0 first order derivative -0 5 Smoothed with Gaussian 0 second order derivative 56-5
65 Second order derivative filters Second order derivative or Laplacian: can be implemented as simple linear filter: H = I(x, y) = I x + I y I(x 0 ) x = I(x 0 + x) I(x 0 + x)+i(x 0 ) I = H [I(x 0 + x) I(x 0 + x) I(x 0 )] T or by separately in x and y direction: H xx = H T yy = [ ] I = H xx I + H T yy I 57
66 Edge detection To compute image edges we need: Image smoothing Image derivatives Non-maximal suppression Thresholding 58
67 D edge detection 59
68 D edge detection 60
69 Convolving with the first order derivative of Gaussian Smoothing removes noise and amplifies signal at the edge discontinuities. 6
70 Zero crossings and convolution with Laplacian Looking at minima or maxima of fist order image derivatives is the same as looking for zero-crossings of second order image derivatives. 6
71 Multi-scale edge detection The amount of smoothing controls the scale at which we analyze the image. Small smoothing brings edges at a fine scale Signal noise is not suppressed 6 Increased smoothing suppresses noise
72 D edge detection Original image Gradient image smoothed with Gaussian Gradient image smoothed with Gaussian 64
73 D edge detection Original image Gradient image smoothed with Gaussian Gradient image smoothed with Gaussian σ =.0 64
74 D edge detection Original image Gradient image smoothed with Gaussian σ =0.5 Gradient image smoothed with Gaussian σ =.0 64
75 Non-maximal suppression Non-maxima suppression does thinning of the edges. The gradient orientations are quantized into four bins original quantization of gradient orientations Pick neighboring pixels of the edge pixel in the direction of the gradient and take the one with the maximum gradient magnitude, while the others are set to zero. gradient magnitude gradients after non-maximum suppression 65
76 Canny edge detection Canny edge detection composes of all already mentioned steps which are: Image smoothing Image derivatives Non-maximal suppression Hysteresis thresholding 66
77 Hysteresis thresholding Apply two thresholds Thigh and Tlow to follow edges Algorithm:. Search for the pixel with the gradient magnitude value in the non-maximum suppressed image is higher then Thigh. Recursively search its neighbors and assign them to the edge if their gradient magnitude value in the non-maximum suppressed image is higher then the Tlow.. otherwise, stop if the gradient magnitude is bellow Tlow or the pixel is already visited and assigned to be on an edge and go to. 67
78 Canny examples Gradient image Th=50, Tl=0 Th=00, Tl=0 Th=50, Tl=40 68
Lecture 04 Image Filtering
Institute of Informatics Institute of Neuroinformatics Lecture 04 Image Filtering Davide Scaramuzza 1 Lab Exercise 2 - Today afternoon Room ETH HG E 1.1 from 13:15 to 15:00 Work description: your first
More informationLecture 7: Edge Detection
#1 Lecture 7: Edge Detection Saad J Bedros sbedros@umn.edu Review From Last Lecture Definition of an Edge First Order Derivative Approximation as Edge Detector #2 This Lecture Examples of Edge Detection
More informationLaplacian Filters. Sobel Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters
Sobel Filters Note that smoothing the image before applying a Sobel filter typically gives better results. Even thresholding the Sobel filtered image cannot usually create precise, i.e., -pixel wide, edges.
More informationFiltering and Edge Detection
Filtering and Edge Detection Local Neighborhoods Hard to tell anything from a single pixel Example: you see a reddish pixel. Is this the object s color? Illumination? Noise? The next step in order of complexity
More informationEdge Detection. CS 650: Computer Vision
CS 650: Computer Vision Edges and Gradients Edge: local indication of an object transition Edge detection: local operators that find edges (usually involves convolution) Local intensity transitions are
More informationMachine vision. Summary # 4. The mask for Laplacian is given
1 Machine vision Summary # 4 The mask for Laplacian is given L = 0 1 0 1 4 1 (6) 0 1 0 Another Laplacian mask that gives more importance to the center element is L = 1 1 1 1 8 1 (7) 1 1 1 Note that the
More informationMachine vision, spring 2018 Summary 4
Machine vision Summary # 4 The mask for Laplacian is given L = 4 (6) Another Laplacian mask that gives more importance to the center element is given by L = 8 (7) Note that the sum of the elements in the
More informationITK Filters. Thresholding Edge Detection Gradients Second Order Derivatives Neighborhood Filters Smoothing Filters Distance Map Image Transforms
ITK Filters Thresholding Edge Detection Gradients Second Order Derivatives Neighborhood Filters Smoothing Filters Distance Map Image Transforms ITCS 6010:Biomedical Imaging and Visualization 1 ITK Filters:
More informationReading. 3. Image processing. Pixel movement. Image processing Y R I G Q
Reading Jain, Kasturi, Schunck, Machine Vision. McGraw-Hill, 1995. Sections 4.-4.4, 4.5(intro), 4.5.5, 4.5.6, 5.1-5.4. 3. Image processing 1 Image processing An image processing operation typically defines
More informationLinear Diffusion. E9 242 STIP- R. Venkatesh Babu IISc
Linear Diffusion Derivation of Heat equation Consider a 2D hot plate with Initial temperature profile I 0 (x, y) Uniform (isotropic) conduction coefficient c Unit thickness (along z) Problem: What is temperature
More informationLecture 6: Edge Detection. CAP 5415: Computer Vision Fall 2008
Lecture 6: Edge Detection CAP 5415: Computer Vision Fall 2008 Announcements PS 2 is available Please read it by Thursday During Thursday lecture, I will be going over it in some detail Monday - Computer
More informationAdvanced Edge Detection 1
Advanced Edge Detection 1 Lecture 4 See Sections 2.4 and 1.2.5 in Reinhard Klette: Concise Computer Vision Springer-Verlag, London, 2014 1 See last slide for copyright information. 1 / 27 Agenda 1 LoG
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff
More informationEdge Detection. Introduction to Computer Vision. Useful Mathematics Funcs. The bad news
Edge Detection Introduction to Computer Vision CS / ECE 8B Thursday, April, 004 Edge detection (HO #5) Edge detection is a local area operator that seeks to find significant, meaningful changes in image
More informationToday s lecture. Local neighbourhood processing. The convolution. Removing uncorrelated noise from an image The Fourier transform
Cris Luengo TD396 fall 4 cris@cbuuse Today s lecture Local neighbourhood processing smoothing an image sharpening an image The convolution What is it? What is it useful for? How can I compute it? Removing
More informationReview Smoothing Spatial Filters Sharpening Spatial Filters. Spatial Filtering. Dr. Praveen Sankaran. Department of ECE NIT Calicut.
Spatial Filtering Dr. Praveen Sankaran Department of ECE NIT Calicut January 7, 203 Outline 2 Linear Nonlinear 3 Spatial Domain Refers to the image plane itself. Direct manipulation of image pixels. Figure:
More informationEdge Detection PSY 5018H: Math Models Hum Behavior, Prof. Paul Schrater, Spring 2005
Edge Detection PSY 5018H: Math Models Hum Behavior, Prof. Paul Schrater, Spring 2005 Gradients and edges Points of sharp change in an image are interesting: change in reflectance change in object change
More informationEdge Detection in Computer Vision Systems
1 CS332 Visual Processing in Computer and Biological Vision Systems Edge Detection in Computer Vision Systems This handout summarizes much of the material on the detection and description of intensity
More informationSpatial Enhancement Region operations: k'(x,y) = F( k(x-m, y-n), k(x,y), k(x+m,y+n) ]
CEE 615: Digital Image Processing Spatial Enhancements 1 Spatial Enhancement Region operations: k'(x,y) = F( k(x-m, y-n), k(x,y), k(x+m,y+n) ] Template (Windowing) Operations Template (window, box, kernel)
More informationImage as a signal. Luc Brun. January 25, 2018
Image as a signal Luc Brun January 25, 2018 Introduction Smoothing Edge detection Fourier Transform 2 / 36 Different way to see an image A stochastic process, A random vector (I [0, 0], I [0, 1],..., I
More informationComputer Vision & Digital Image Processing
Computer Vision & Digital Image Processing Image Restoration and Reconstruction I Dr. D. J. Jackson Lecture 11-1 Image restoration Restoration is an objective process that attempts to recover an image
More informationSURF Features. Jacky Baltes Dept. of Computer Science University of Manitoba WWW:
SURF Features Jacky Baltes Dept. of Computer Science University of Manitoba Email: jacky@cs.umanitoba.ca WWW: http://www.cs.umanitoba.ca/~jacky Salient Spatial Features Trying to find interest points Points
More informationEdge Detection. Computer Vision P. Schrater Spring 2003
Edge Detection Computer Vision P. Schrater Spring 2003 Simplest Model: (Canny) Edge(x) = a U(x) + n(x) U(x)? x=0 Convolve image with U and find points with high magnitude. Choose value by comparing with
More informationDigital Image Processing. Lecture 6 (Enhancement) Bu-Ali Sina University Computer Engineering Dep. Fall 2009
Digital Image Processing Lecture 6 (Enhancement) Bu-Ali Sina University Computer Engineering Dep. Fall 009 Outline Image Enhancement in Spatial Domain Spatial Filtering Smoothing Filters Median Filter
More informationVlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems
1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Perception Concepts Vision Chapter 4 (textbook) Sections 4.3 to 4.5 What is the course
More informationCSE 473/573 Computer Vision and Image Processing (CVIP)
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu inwogu@buffalo.edu Lecture 11 Local Features 1 Schedule Last class We started local features Today More on local features Readings for
More informationAnnouncements. Filtering. Image Filtering. Linear Filters. Example: Smoothing by Averaging. Homework 2 is due Apr 26, 11:59 PM Reading:
Announcements Filtering Homework 2 is due Apr 26, :59 PM eading: Chapter 4: Linear Filters CSE 52 Lecture 6 mage Filtering nput Output Filter (From Bill Freeman) Example: Smoothing by Averaging Linear
More informationComputational Photography
Computational Photography Si Lu Spring 208 http://web.cecs.pdx.edu/~lusi/cs50/cs50_computati onal_photography.htm 04/0/208 Last Time o Digital Camera History of Camera Controlling Camera o Photography
More informationEdge Detection. Image Processing - Computer Vision
Image Processing - Lesson 10 Edge Detection Image Processing - Computer Vision Low Level Edge detection masks Gradient Detectors Compass Detectors Second Derivative - Laplace detectors Edge Linking Image
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:
More informationIntroduction to Computer Vision. 2D Linear Systems
Introduction to Computer Vision D Linear Systems Review: Linear Systems We define a system as a unit that converts an input function into an output function Independent variable System operator or Transfer
More informationECE Digital Image Processing and Introduction to Computer Vision. Outline
2/9/7 ECE592-064 Digital Image Processing and Introduction to Computer Vision Depart. of ECE, NC State University Instructor: Tianfu (Matt) Wu Spring 207. Recap Outline 2. Sharpening Filtering Illustration
More informationRoadmap. Introduction to image analysis (computer vision) Theory of edge detection. Applications
Edge Detection Roadmap Introduction to image analysis (computer vision) Its connection with psychology and neuroscience Why is image analysis difficult? Theory of edge detection Gradient operator Advanced
More informationEdges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise
Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity
More informationGaussian derivatives
Gaussian derivatives UCU Winter School 2017 James Pritts Czech Tecnical University January 16, 2017 1 Images taken from Noah Snavely s and Robert Collins s course notes Definition An image (grayscale)
More informationLocal Enhancement. Local enhancement
Local Enhancement Local Enhancement Median filtering (see notes/slides, 3.5.2) HW4 due next Wednesday Required Reading: Sections 3.3, 3.4, 3.5, 3.6, 3.7 Local Enhancement 1 Local enhancement Sometimes
More informationMedical Image Analysis
Medical Image Analysis CS 593 / 791 Computer Science and Electrical Engineering Dept. West Virginia University 20th January 2006 Outline 1 Discretizing the heat equation 2 Outline 1 Discretizing the heat
More informationLinear Diffusion and Image Processing. Outline
Outline Linear Diffusion and Image Processing Fourier Transform Convolution Image Restoration: Linear Filtering Diffusion Processes for Noise Filtering linear scale space theory Gauss-Laplace pyramid for
More informationBlobs & Scale Invariance
Blobs & Scale Invariance Prof. Didier Stricker Doz. Gabriele Bleser Computer Vision: Object and People Tracking With slides from Bebis, S. Lazebnik & S. Seitz, D. Lowe, A. Efros 1 Apertizer: some videos
More informationCAP 5415 Computer Vision
CAP 545 Computer Vision Dr. Mubarak Sa Univ. o Central Florida Filtering Lecture-2 Contents Filtering/Smooting/Removing Noise Convolution/Correlation Image Derivatives Histogram Some Matlab Functions General
More informationLocal enhancement. Local Enhancement. Local histogram equalized. Histogram equalized. Local Contrast Enhancement. Fig 3.23: Another example
Local enhancement Local Enhancement Median filtering Local Enhancement Sometimes Local Enhancement is Preferred. Malab: BlkProc operation for block processing. Left: original tire image. 0/07/00 Local
More informationCS4670: Computer Vision Kavita Bala. Lecture 7: Harris Corner Detec=on
CS4670: Computer Vision Kavita Bala Lecture 7: Harris Corner Detec=on Announcements HW 1 will be out soon Sign up for demo slots for PA 1 Remember that both partners have to be there We will ask you to
More informationComputer Vision Lecture 3
Computer Vision Lecture 3 Linear Filters 03.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Demo Haribo Classification Code available on the class website...
More informationEmpirical Mean and Variance!
Global Image Properties! Global image properties refer to an image as a whole rather than components. Computation of global image properties is often required for image enhancement, preceding image analysis.!
More informationMedical Image Analysis
Medical Image Analysis CS 593 / 791 Computer Science and Electrical Engineering Dept. West Virginia University 23rd January 2006 Outline 1 Recap 2 Edge Enhancement 3 Experimental Results 4 The rest of
More informationTaking derivative by convolution
Taking derivative by convolution Partial derivatives with convolution For 2D function f(x,y), the partial derivative is: For discrete data, we can approximate using finite differences: To implement above
More informationTemplates, Image Pyramids, and Filter Banks
Templates, Image Pyramids, and Filter Banks 09/9/ Computer Vision James Hays, Brown Slides: Hoiem and others Review. Match the spatial domain image to the Fourier magnitude image 2 3 4 5 B A C D E Slide:
More informationConjugate gradient acceleration of non-linear smoothing filters Iterated edge-preserving smoothing
Cambridge, Massachusetts Conjugate gradient acceleration of non-linear smoothing filters Iterated edge-preserving smoothing Andrew Knyazev (knyazev@merl.com) (speaker) Alexander Malyshev (malyshev@merl.com)
More informationImage Gradients and Gradient Filtering Computer Vision
Image Gradients and Gradient Filtering 16-385 Computer Vision What is an image edge? Recall that an image is a 2D function f(x) edge edge How would you detect an edge? What kinds of filter would you use?
More informationMultimedia Databases. Previous Lecture. 4.1 Multiresolution Analysis. 4 Shape-based Features. 4.1 Multiresolution Analysis
Previous Lecture Multimedia Databases Texture-Based Image Retrieval Low Level Features Tamura Measure, Random Field Model High-Level Features Fourier-Transform, Wavelets Wolf-Tilo Balke Silviu Homoceanu
More informationMultimedia Databases. Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig
Multimedia Databases Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 4 Previous Lecture Texture-Based Image Retrieval Low
More informationColorado School of Mines Image and Multidimensional Signal Processing
Image and Multidimensional Signal Processing Professor William Hoff Department of Electrical Engineering and Computer Science Spatial Filtering Main idea Spatial filtering Define a neighborhood of a pixel
More informationMultimedia Databases. 4 Shape-based Features. 4.1 Multiresolution Analysis. 4.1 Multiresolution Analysis. 4.1 Multiresolution Analysis
4 Shape-based Features Multimedia Databases Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 4 Multiresolution Analysis
More informationCITS 4402 Computer Vision
CITS 4402 Computer Vision Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh, CEO at Mapizy (www.mapizy.com) and InFarm (www.infarm.io) Lecture 04 Greyscale Image Analysis Lecture 03 Summary Images as 2-D signals
More informationFeature detection.
Feature detection Kim Steenstrup Pedersen kimstp@itu.dk The IT University of Copenhagen Feature detection, The IT University of Copenhagen p.1/20 What is a feature? Features can be thought of as symbolic
More informationImage Segmentation: Definition Importance. Digital Image Processing, 2nd ed. Chapter 10 Image Segmentation.
: Definition Importance Detection of Discontinuities: 9 R = wi z i= 1 i Point Detection: 1. A Mask 2. Thresholding R T Line Detection: A Suitable Mask in desired direction Thresholding Line i : R R, j
More informationCAP 5415 Computer Vision Fall 2011
CAP 545 Computer Vision Fall 2 Dr. Mubarak Sa Univ. o Central Florida www.cs.uc.edu/~vision/courses/cap545/all22 Oice 247-F HEC Filtering Lecture-2 General Binary Gray Scale Color Binary Images Y Row X
More informationLinear Operators and Fourier Transform
Linear Operators and Fourier Transform DD2423 Image Analysis and Computer Vision Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 13, 2013
More informationI Chen Lin, Assistant Professor Dept. of CS, National Chiao Tung University. Computer Vision: 4. Filtering
I Chen Lin, Assistant Professor Dept. of CS, National Chiao Tung University Computer Vision: 4. Filtering Outline Impulse response and convolution. Linear filter and image pyramid. Textbook: David A. Forsyth
More informationAdvanced Features. Advanced Features: Topics. Jana Kosecka. Slides from: S. Thurn, D. Lowe, Forsyth and Ponce. Advanced features and feature matching
Advanced Features Jana Kosecka Slides from: S. Thurn, D. Lowe, Forsyth and Ponce Advanced Features: Topics Advanced features and feature matching Template matching SIFT features Haar features 2 1 Features
More informationFeature extraction: Corners and blobs
Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over
More informationCOMP344 Digital Image Processing Fall 2007 Final Examination
COMP344 Digital Image Processing Fall 2007 Final Examination Time allowed: 2 hours Name Student ID Email Question 1 Question 2 Question 3 Question 4 Question 5 Question 6 Total With model answer HK University
More informationLecture Outline. Basics of Spatial Filtering Smoothing Spatial Filters. Sharpening Spatial Filters
1 Lecture Outline Basics o Spatial Filtering Smoothing Spatial Filters Averaging ilters Order-Statistics ilters Sharpening Spatial Filters Laplacian ilters High-boost ilters Gradient Masks Combining Spatial
More informationImage Enhancement: Methods. Digital Image Processing. No Explicit definition. Spatial Domain: Frequency Domain:
Image Enhancement: No Explicit definition Methods Spatial Domain: Linear Nonlinear Frequency Domain: Linear Nonlinear 1 Spatial Domain Process,, g x y T f x y 2 For 1 1 neighborhood: Contrast Enhancement/Stretching/Point
More informationOrientation Map Based Palmprint Recognition
Orientation Map Based Palmprint Recognition (BM) 45 Orientation Map Based Palmprint Recognition B. H. Shekar, N. Harivinod bhshekar@gmail.com, harivinodn@gmail.com India, Mangalore University, Department
More informationAchieving scale covariance
Achieving scale covariance Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region size that is covariant
More informationScale Space Smoothing, Image Feature Extraction and Bessel Filters
Scale Space Smoothing, Image Feature Extraction and Bessel Filters Sasan Mahmoodi and Steve Gunn School of Electronics and Computer Science, Building 1, Southampton University, Southampton, SO17 1BJ, UK
More informationIntroduction to Nonlinear Image Processing
Introduction to Nonlinear Image Processing 1 IPAM Summer School on Computer Vision July 22, 2013 Iasonas Kokkinos Center for Visual Computing Ecole Centrale Paris / INRIA Saclay Mean and median 2 Observations
More information3.8 Combining Spatial Enhancement Methods 137
3.8 Combining Spatial Enhancement Methods 137 a b FIGURE 3.45 Optical image of contact lens (note defects on the boundary at 4 and 5 o clock). (b) Sobel gradient. (Original image courtesy of Mr. Pete Sites,
More informationErrata for Robot Vision
Errata for Robot Vision This is a list of known nontrivial bugs in Robot Vision 1986) by B.K.P. Horn, MIT Press, Cambridge, MA ISBN 0-6-08159-8 and McGraw-Hill, New York, NY ISBN 0-07-030349-5. Thanks
More informationLoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D
Achieving scale covariance Blobs (and scale selection) Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region
More informationHistogram Processing
Histogram Processing The histogram of a digital image with gray levels in the range [0,L-] is a discrete function h ( r k ) = n k where r k n k = k th gray level = number of pixels in the image having
More informationConvolution and Linear Systems
CS 450: Introduction to Digital Signal and Image Processing Bryan Morse BYU Computer Science Introduction Analyzing Systems Goal: analyze a device that turns one signal into another. Notation: f (t) g(t)
More informationSIFT: Scale Invariant Feature Transform
1 SIFT: Scale Invariant Feature Transform With slides from Sebastian Thrun Stanford CS223B Computer Vision, Winter 2006 3 Pattern Recognition Want to find in here SIFT Invariances: Scaling Rotation Illumination
More informationImage Alignment and Mosaicing
Image Alignment and Mosaicing Image Alignment Applications Local alignment: Tracking Stereo Global alignment: Camera jitter elimination Image enhancement Panoramic mosaicing Image Enhancement Original
More informationIntroduction to Linear Systems
cfl David J Fleet, 998 Introduction to Linear Systems David Fleet For operator T, input I, and response R = T [I], T satisfies: ffl homogeniety: iff T [ai] = at[i] 8a 2 C ffl additivity: iff T [I + I 2
More informationImage Noise: Detection, Measurement and Removal Techniques. Zhifei Zhang
Image Noise: Detection, Measurement and Removal Techniques Zhifei Zhang Outline Noise measurement Filter-based Block-based Wavelet-based Noise removal Spatial domain Transform domain Non-local methods
More informationDetectors part II Descriptors
EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers Goal:
More informationSingle Exposure Enhancement and Reconstruction. Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman
Single Exposure Enhancement and Reconstruction Some slides are from: J. Kosecka, Y. Chuang, A. Efros, C. B. Owen, W. Freeman 1 Reconstruction as an Inverse Problem Original image f Distortion & Sampling
More informationGeneralized Laplacian as Focus Measure
Generalized Laplacian as Focus Measure Muhammad Riaz 1, Seungjin Park, Muhammad Bilal Ahmad 1, Waqas Rasheed 1, and Jongan Park 1 1 School of Information & Communications Engineering, Chosun University,
More informationImage Degradation Model (Linear/Additive)
Image Degradation Model (Linear/Additive),,,,,,,, g x y h x y f x y x y G uv H uv F uv N uv 1 Source of noise Image acquisition (digitization) Image transmission Spatial properties of noise Statistical
More informationImage Characteristics
1 Image Characteristics Image Mean I I av = i i j I( i, j 1 j) I I NEW (x,y)=i(x,y)-b x x Changing the image mean Image Contrast The contrast definition of the entire image is ambiguous In general it is
More informationImage Analysis. Feature extraction: corners and blobs
Image Analysis Feature extraction: corners and blobs Christophoros Nikou cnikou@cs.uoi.gr Images taken from: Computer Vision course by Svetlana Lazebnik, University of North Carolina at Chapel Hill (http://www.cs.unc.edu/~lazebnik/spring10/).
More informationIntensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER? Intensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER?
: WHICH ONE LOOKS BETTER? 3.1 : WHICH ONE LOOKS BETTER? 3.2 1 Goal: Image enhancement seeks to improve the visual appearance of an image, or convert it to a form suited for analysis by a human or a machine.
More informationErrata for Robot Vision
Errata for Robot Vision This is a list of known nontrivial bugs in Robot Vision 1986) by B.K.P. Horn, MIT Press, Cambridge, MA ISBN 0-6-08159-8 and McGraw-Hill, New York, NY ISBN 0-07-030349-5. If you
More informationIntroduction to Computer Vision
Introduction to Computer Vision Michael J. Black Sept 2009 Lecture 8: Pyramids and image derivatives Goals Images as functions Derivatives of images Edges and gradients Laplacian pyramids Code for lecture
More informationComputer Vision. Filtering in the Frequency Domain
Computer Vision Filtering in the Frequency Domain Filippo Bergamasco (filippo.bergamasco@unive.it) http://www.dais.unive.it/~bergamasco DAIS, Ca Foscari University of Venice Academic year 2016/2017 Introduction
More informationNONLINEAR DIFFUSION PDES
NONLINEAR DIFFUSION PDES Erkut Erdem Hacettepe University March 5 th, 0 CONTENTS Perona-Malik Type Nonlinear Diffusion Edge Enhancing Diffusion 5 References 7 PERONA-MALIK TYPE NONLINEAR DIFFUSION The
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationImage Processing 1 (IP1) Bildverarbeitung 1
MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV (KOGS) Image Processing 1 (IP1) Bildverarbeitung 1 Lecture 7 Spectral Image Processing and Convolution Winter Semester 2014/15 Slides: Prof. Bernd
More informationChapter 16. Local Operations
Chapter 16 Local Operations g[x, y] =O{f[x ± x, y ± y]} In many common image processing operations, the output pixel is a weighted combination of the gray values of pixels in the neighborhood of the input
More informationImage preprocessing in spatial domain
Image preprocessing in spatial domain Sharpening, image derivatives, Laplacian, edges Revision: 1.2, dated: May 25, 2007 Tomáš Svoboda Czech Technical University, Faculty of Electrical Engineering Center
More informationSampling in 1D ( ) Continuous time signal f(t) Discrete time signal. f(t) comb
Sampling in 2D 1 Sampling in 1D Continuous time signal f(t) Discrete time signal t ( ) f [ k] = f( kt ) = f( t) δ t kt s k s f(t) comb k 2 Nyquist theorem (1D) At least 2 sample/period are needed to represent
More informationBasics on 2-D 2 D Random Signal
Basics on -D D Random Signal Spring 06 Instructor: K. J. Ray Liu ECE Department, Univ. of Maryland, College Park Overview Last Time: Fourier Analysis for -D signals Image enhancement via spatial filtering
More informationCS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015
CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature
More informationFeature Extraction and Image Processing
Feature Extraction and Image Processing Second edition Mark S. Nixon Alberto S. Aguado :*авш JBK IIP AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO SINGAPORE SYDNEY TOKYO
More informationLecture 7: Finding Features (part 2/2)
Lecture 7: Finding Features (part 2/2) Dr. Juan Carlos Niebles Stanford AI Lab Professor Fei- Fei Li Stanford Vision Lab 1 What we will learn today? Local invariant features MoPvaPon Requirements, invariances
More informationMotion Estimation (I)
Motion Estimation (I) Ce Liu celiu@microsoft.com Microsoft Research New England We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationSIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE
SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE Overview Motivation of Work Overview of Algorithm Scale Space and Difference of Gaussian Keypoint Localization Orientation Assignment Descriptor Building
More information