A Method for Blur and Similarity Transform Invariant Object Recognition
|
|
- Julie Thornton
- 5 years ago
- Views:
Transcription
1 A Method for Blur and Similarity Transform Invariant Object Recognition Ville Ojansivu and Janne Heikkilä Machine Vision Group, Department of Electrical and Information Engineering, University of Oulu, PO Box 4500, 90014, Finland Abstract In this paper, we propose novel blur and similarity transform (i.e. rotation, scaling and translation) invariant features for the recognition of objects in images. The features are based on blur invariant forms of the log-polar sampled phase-only bispectrum and are invariant to centrally symmetric blur, including linear motion and out of focus blur. An additional advantage of using the phase-only bispectrum is the invariance to uniform illumination changes. According to our knowledge, the invariants of this paper are the first blur and similarity transform invariants in the Fourier domain. We have compared our features to the blur invariants based on complex image moments using simulated and real data. The moment invariants have not been evaluated earlier in the case of similarity transform. The results show that our invariants can recognize objects better in the presence of noise. 1. Introduction Recognition of objects and patterns in images is a fundamental part of computer vision, with numerous applications. The task is difficult as the objects rarely look exactly the same in different conditions. In real applications, images contain various artifacts such as geometrical and convolutional degradations. Image analysis systems should be able to operate also in these nonideal conditions. There has been a vast amount of research in this field of invariant pattern and object recognition [14]. However, the invariant recognition of objects degraded by blur is a much less studied topic. In various applications, images may contain blur, which can result, for example, from atmospheric turbulence, being out of focus, or relative motion between the camera and the scene. This degradation process can be modeled as a linear shift-invariant system in which the relation between an ideal image f(x) and an observed image g(x) is given by g(x) =f(x) h(x)+n(x), (1) where x is a 2-D spatial coordinate vector, h(x) the point spread function (PSF) of the system, n(x) additive noise, and denotes 2-D convolution. The point spread function h(x) represents blur while other degradations are captured by the noise term n(x). The analysis of blurred images is often carried out by first deblurring the images and then applying standard methods for further analysis. Unfortunately, image deblurring is a difficult problem. Conventional solutions include the estimation of the PSF of the blur and deconvolution of the image using that PSF. When the PSF is known, the latter illposed problem can be solved using approaches which use regularization [1]. In practice, the PSF is often unknown and very hard to estimate accurately. In this case, blind image restoration algorithms are used [8]. The analysis of the blurred images can also be performed without deblurring using features which are invariant to blur. Flusser and Suk proposed the first blur invariant features in [5]. These invariants are based on geometric moments or the spectrum of the image. Of these, the invariants based on central moments are also invariant to translation. The moment invariants have been applied to template matching [5], recognition of defocused objects [6] and registration of X-ray images [2]. In [9], new Fourier domain invariants based on the phase of the image spectrum or bispectrum were proposed. The bispectrum invariants are also invariant to a shift of the image. In [7], blur, rotation, scale and shift invariants based on complex moments were proposed and in [15, 12], the theory was extended to blur and affine moment invariants. A shortcoming of the moment invariants is their sensitivity to noise and background clutter. On the other hand, the invariance to geometrical transformations of the Fourier domain features is limited to image shift. In this paper, we propose novel Fourier domain blur in-
2 variants, which are invariant to rotation, scaling and translation i.e. similarity transform. We call them blur and similarity transform invariants (BSI). Similarly to above mentioned former invariants, also the BSIs are invariant to centrally symmetric convolution. Many practical blur forms, such as linear motion or defocus blur fall into this category. The invariants are based on the blur invariant form of the log-polar sampled phase-only bispectrum. 2. Blur invariants in the Fourier domain In this section, we first show the conditions of blur invariance in the Fourier domain in Section 2.1 and then present the blur-similarity transform invariants in Section 2.2. The implementation of the invariants is discussed in Section Blur invariance If noise n(x) is neglected, (1) can be expressed in the Fourier domain using the convolution theorem by G(u) =F (u) H(u), where G(u), F (u),andh(u) are the Fourier transforms of the observed image, ideal image and the PSF. In the phasor form, the equation becomes G(u) = G(u) e iφg(u), where u is a vector in the 2-D frequency space. If the Fourier transform G(u) is normalized by its magnitude, only the complex exponential containing the phase remains, namely G(u) G(u) = e iφg(u) = e i[φ f (u)+φ h (u)], (2) where φ f (u) is the phase of the original image f(x) and φ h (u) the phase of the blur PSF h(x). Since h(x) is assumed to be centrally symmetric, its Fourier transform H(u) is real and its phase φ h (u) has only two possible values φ h (u) =0 φ h (u) =π. It follows from this, and from the periodicity of the complex argument, that the equality [e iφg(u) ] 2n = e i2nφg(u) holds for any integer n. = e i2nφ f (u) e i2nφ h(u) = [e iφ f (u) ] 2n (3) Thus, any even power of the normalized Fourier transform, i.e. e i2nφ(u), of the observed image is invariant to the convolution of the original image with any centrally symmetric PSF Blur and similarity transform invariants The phase spectrum of an image, that is used in Section 2.1, is not invariant to translation. The amplitude spectrum is invariant to translation, but blur invariants cannot be constructed from it, as they are based on phase. For this reason, we have to turn to the higher order spectra defined by Ψ m (u 1, u 2,, u n )=F (s) m F (u i ), (4) i=1 where u i with i = 1,,m are vectors in the 2-D frequency space, and s = u 1 + u u m. It can be easily shown that Ψ m is shift invariant [3]. When m =1in (4) we get the power spectrum, and further with the value m =2, we get the bispectrum, namely Ψ 2 (u 1, u 2 )=F(u 1 )F (u 2 )F (u 1 + u 2 ). (5) The bispectrum contains both amplitude and phase information and is shift invariant. It is possible to construct a blur invariant phase-only bispectrum where the exponentials containing the phase are raised to the second power, similar to (3), namely [ ] 2 [ ] 2 [ F (u1 ) F (u2 ) F ] 2 (u 1 + u 2 ) P (u 1, u 2 )= F (u 1 ) F (u 2 ) F (u 1 + u 2 ) = e i2φ(u1) e i2φ(u2) e i2φ(u1+u2) = e i2[φ(u1)+φ(u2) φ(u1+u2)]. (6) The bispectrum (5) as well as its blur invariant form (6) are functions of two vector arguments, containing a total of four scalar variables. Assuming that F (u) is an N-by-N discrete Fourier transform (DFT) of an image f(x), the bispectrum becomes a four-dimensional N-by-N-by-N-by- N matrix. Fortunately, it is not necessary to evaluate the whole bispectrum. It is possible to take only 2-D slices of the original bispectrum, which contain basically the same information. There are various ways of defining the slices [4, 10]. We define the slices as S k (u) =Ψ 2 (u,ku) k R. (7) Similarly to (6) but using only one slice (7) we form a blurtranslation invariant phase-only bispectrum slice P k (u) = [ ] 2 Sk (u) = e i2[φ(u)+φ(ku) φ((k+1)u)]. (8) S k (u)
3 The blur-translation invariants we proposed in [9] were based on (8). A pure bispectrum has also been used for rotation, scale and translation invariant object recognition [13]. An advantage of using the phase-only bispectrum is its robustness to illumination changes. The bispectrum is also fairly insensitivity to noise. The bispectrum slices (7) have the following scaling and rotation properties: f(ax, by) B 1 ab 3 S k( u a, v b ) (9) f(r, θ + α) B S k (ω, φ + α), (10) where B denotes the transformation from spatial image f(x, y) to its bispectrum slice S k (u, v), f(r, θ) and S k (ω, φ) are the corresponding representations in the polar coordinates, and a, b,andα are some constants. (9) and (10) correspond to the scaling and rotation properties of the Fourier transform, except for the term 1 which is 1 ab 3 ab for the Fourier transform. This means that the bispectrum slices can be used to form rotation, scaling and translation invariant image representation similar to the amplitude spectrum in the case of Fourier-Mellin invariant descriptors [11]. Using log-polar representation, the property (9) transforms the scale difference of the images into displacement of their logarithmically sampled bispectrum slice. The property (10) converts the rotation of the image into displacement along the φ-axis in the polar coordinates. As a result, both scale and orientation changes of the image turn into shifts in the log-polar sampled bispectrum slice. We denote this log-polar resampled version of the k th blur invariant bispectrum slice (8) as R k (r, θ). The shift of R k (r, θ) can be eliminated by computing its bispectrum slice which is shift invariant. We denote the l th bispectrum slice (7) of R k (r, θ) as ζ kl (v), wherek and l identify the first and second successive slices. ζ kl (v) is now invariant to blur and similarity transform (rotation, scaling and translation) of the original image. The final BSI features that we used for object classification task are acquired by normalizing the slice, namely I kl (v) = ζ kl(v) ζ kl (v). (11) The computation of the BSIs can be summarized into following steps, which are discussed in detail in Section 2.3: 1) compute the Fourier transform of image:f(x) F (u), 2) compute the k th bispectrum slice: F (u) S k (u), 3) normalize and raise to a 2 nd power: S k (u) P k (u), 4) log-polar transform: P k (u) R k (r, θ), 5) compute the Fourier transform: R k (r, θ) F{R k (r, θ)}, 6) compute the l th bispectrum slice: F{R k (r, θ)} ζ lk (v), 7) normalize to get the invariants: ζ lk (v) I lk (v). The resulting BSI features are invariant to similarity transform and centrally symmetric blur, similar to the moment invariants of [7]. This includes linear motion blur and out of focus blur which are common. The features possess also invariance to uniform illumination changes and, because of the nature of the bispectrum, are fairly insensitive to noise Implementation To compute an arbitrary bispectrum slice S k (u) in step 2, we need frequency samples in points u but also at ku and (k +1)u. The samples in the two latter cases can be extracted from the DFT corresponding to F (u) by utilizing its conjugate symmetry and periodicity. The same is true in step 6 when another bispectrum slice is computed. After computation of the blur invariant phase-only slice P k (u), follows the log-polar sampling using bilinear interpolation. The sampling points are selected in the same way as in [11], namely u i = t ri/t cos θ i v i = t ri/t sin θ i, where t = N/2 1, and(r i,θ i ) are the points in the new uniform sampling lattice. The size of the DFT is N N. The resulting log-polar sampled bispectrum slice is denoted by R k (r i,θ i ). The accuracy of the final invariants depends on the resolution of the DFT and interpolation. In the experiments, the DFT dimensions were eight times the corresponding image dimensions, with values r i {0, 2,...,t} and θ i {0, 0.25,...,360}. The computation of the l th bispectrum slice ζ kl (v) of R k (r, θ) is achieved using a complex DFT and (7). It is possible to obtain slightly better classification results by using invariants from multiple slices k and l in the steps 2 and 6, respectively. In the experiments, we used only one slice in each step with values k =1and l =1.If the DFT size is N N, this results in approximately N 2 /2 non-redundant invariants I 11 (v). Itseemsthattheinvariants corresponding to the lower frequencies have a higher signal-to-noise ratio. In the experiments, we used the invariants of (11) within a radius ρ from the zero frequency, but without using the conjugate symmetric or zero frequency components.
4 100 Figure 1. Top row: images of the first four letters in the alphabet. Bottom row: motion blurred, rotated, scaled, translated and noisy versions of the same images. PSNRs are from left to right: 50, 40, 30 and 20 db. Noise is removed from the background using thresholding and connectivity analysis. Classification accuracy [%] BSI Mom inv PSNR [db] Figure 2. Classification accuracy of the nearest neighbor classification of distorted images of the letters of alphabet. The distance between the images f(x) and g(x) to be classified is computed as D = e(v 1,v 2 ), (12) v 2 where v 1 e(v) = I (f) 11 (v) I(g) 11 (v) 2, and where I (f) 11 (v) and I(g) 11 (v) are the BSIs of (11) for images f and g, respectively, v =[v 1,v 2 ],and v1 2 + v2 2 ρ. 3. Experiments In the experiments, we classified blurred, rotated, scaled and shifted images using the nearest neighbor classifier. We evaluated the classification accuracy of the BSIs against complex moment invariants which exhibit similar invariance properties. The moment invariants, although possessing invariance to the similarity transform, have not been evaluated previously so that scaling is present. Moment invariants were implemented as proposed in [7] using nine invariants of the orders 3, 5 and 7. To ensure roughly the same range of values, the moment invariants were normalized by their standard deviations for the whole test data in each experiment. The BSIs were implemented as explained in Section 2.3. We used the values k =1, l =1and ρ =30. In the first experiment, the target classes were formed by the images of the letters of the alphabet. The test images were distorted versions of the same images. Some of the original and distorted images are shown in Figure 1. The distortion included motion blur of length 30 pixels in a random direction, rotation of {0, 15, 30, 45, 60, 75, 90} degrees, scaling by a factor of {0.8, 0.9, 1.0, 1.1, 1.2},random translation in the range [-5, 5] and noise. Because the moment invariants are known to be sensitive to noise, we removed the noise from the background by using thresholding and connectivity analysis. At the same time, the image information at the border of the object is distorted. Each letter underwent all combinations of the above mentioned disturbances and was each time classified using the BSIs and moment invariants. The resulting percentage of correct classification is shown in Figure 2 for five different peak signal-to-noise ratios (PSNR). As was expected, both methods can classify images correctly at a low noise level. When the noise level is increased, the BSIs become more robust compared to the moment invariants. In the second experiment, the test material included images of fish. These original images formed the target classes into which the distorted versions of the images were to be classified. Some original and distorted fish images are shown in Figure 3. The images underwent similar distortions as those in the first experiment, but the blur was now circular and created by a disk shaped PSF with a radius of 15 pixels. The blur was also smaller relative to the image size compared to the first experiment. We performed similar thresholding and connectivity analysis to that in the first experiment to remove the background noise. The classification results are shown in the diagram of Figure 4. Again, both methods classify all cases correctly when the noise level is low, but when noise level increases the performance of the BSIs remains almost unchanged while the
5 Figure 3. Top row: four examples of the 90 fish images. Bottom row: circularly blurred, rotated, scaled, translated and noisy versions of the same images. PSNRs are from left to right: 50, 40, 30 and 20 db. Noise is removed from the background using thresholding and connectivity analysis. ( images are cropped from the images that were used.) Classification accuracy [%] BSI Mom. inv PSNR [db] Figure 4. Classification accuracy of the nearest neighbor classification of distorted images of 90 fish. moments break down after 30 db. In the final experiment, we used real images of four playing cards, the number seven from each suit (Seven of Spades, Hearts, Diamonds, and Clubs), which look very similar. Sharp images of these cards formed the four target classes of the classification. The cards were photographed on a black background using a Sony DFM-X710 video camera with a zoom lens. The image size was pixels. The background clutter was removed from the images using thresholding and connectivity analysis similarly to the simulated experiments. At the same time, this operation causes inevitable but realistic distortion at the image borders. We photographed four motion blurred and four out of focus blurred versions of each card rotating, zooming and translating the camera between the shots. Motion blur was generated by linearly panning the camera and out of focus blur by defocusing the lens. Altogether, this resulted in 32 degraded images to be classified into one of the four classes. Each row in Figure 5 shows a sharp and two degraded images of one class. As can be seen, the blur is so severe that it is impossible to recognize the cards visually. However, the BSI features classified all the distorted images into the correct classes, performing perfectly. The moment invariants were able to classify only 47 % of the images correctly. This result is in line with the results of the simulated experiments. 4. Conclusions In this paper, we have proposed novel Fourier domain blur and similarity transform invariants called BSIs. The features are invariant to blur or any other convolution of the image having centrally symmetric PSF. Special cases of this are linear motion blur and out of focus blur. As demonstrated, the invariants can be used to recognize blurred images or objects. Because features are based on a phase only bispectrum, they are also robust to illumination changes and as bispectrum, they are fairly insensitive to noise. We compared BSIs to the complex moment invariants proposed in [7] for classification of blurred and noisy objects under similarity transform. The results show that our BSIs perform better when the objects are noisy. 5. Acknowledgments The authors would like to thank the Academy of Finland (project no ), and Prof. Petrou and Dr. Kadyrov for providing us with the fish image database.
6 Figure 5. The rows contain one sharp and two blurred and transformed versions of the cards Seven of Hearts and Seven of Diamonds, respectively. Cards are photographed on a black background. Background clutter is removed using thresholding and connectivity analysis. References [1] M. R. Banham and A. K. Katsaggelos. Digital image restoration. IEEE Signal Processing Magazine, 14(2):24 41, Mar [2] Y. Bentoutou, N. Taleb, M. C. E. Mezouar, M. Taleb, and L. Jetto. An invariant approach for image registration in digital subtraction angiography. Pattern Recognition, 35(12): , [3] V. Chandran, B. Carswell, B. Boashash, and S. Elgar. Pattern recognition using invariants defined from higher order spectra: 2-d image inputs. IEEE Trans. Image Processing, 6(5): , [4] S. A. Dianat and R. M. Rao. Fast algorithms for phase and magnitude reconstruction from bispectra. Optical Engineering, 29(5): , [5] J. Flusser and T. Suk. Degraded image analysis: An invariant approach. IEEE Trans. Pattern Anal. Machine Intell., 20(6): , [6] J. Flusser, T. Suk, and S. Saic. Recognition of blurred images by the method of moments. IEEE Trans. Image Processing, 5(3): , [7] J. Flusser and B. Zitová. Combined invariants to linear filtering and rotation. Int. J. Pattern Recognition and Artificial Intelligence, 13(8): , [8] D. Kundur and D. Hatzinakos. Blind image deconvolution. IEEE Signal Processing Magazine, 13(3):43 64, May [9] V. Ojansivu and J. Heikkilä. Object recognition using frequency domain blur invariant features. In Proc. Scandinavian Conference on Image Analysis (SCIA 07), Aalborg, Denmark, June [10] A. P. Petropulu and H. Pozidis. Phase reconstruction from bispectrum slices. IEEE Trans. Signal Processing, 46(2): , [11] Q. sheng Chen, M. Defrise, and F. Deconinck. Symmetric phase-only matched filtering of fourier-mellin transforms for image registration and recognition. IEEE Trans. Pattern Anal. Machine Intell., 16(12): , [12] T. Suk and J. Flusser. Combined blur and affine moment invariants and their use in pattern recognition. Pattern Recognition, 26(12): , Dec [13] M. K. Tsatsanis and G. B. Giannakis. Object and texture classification using higher order statistics. IEEE Trans. Pattern Anal. Machine Intell., 14(7): , [14] J. Wood. Invariant pattern recognition: A review. Pattern Recognition, 29(1):1 17, [15] Y. Zhang, C. Wen, Y. Zhang, and Y. C. Soh. Determination of blur and affine combined invariants by normalization. Pattern recognition, 35(1): , Jan
Blur Insensitive Texture Classification Using Local Phase Quantization
Blur Insensitive Texture Classification Using Local Phase Quantization Ville Ojansivu and Janne Heikkilä Machine Vision Group, Department of Electrical and Information Engineering, University of Oulu,
More informationScience Insights: An International Journal
Available online at http://www.urpjournals.com Science Insights: An International Journal Universal Research Publications. All rights reserved ISSN 2277 3835 Original Article Object Recognition using Zernike
More informationA New Efficient Method for Producing Global Affine Invariants
A New Efficient Method for Producing Global Affine Invariants Esa Rahtu, Mikko Salo 2, and Janne Heikkilä Machine Vision Group, Department of Electrical and Information Engineering, P.O. Box 45, 94 University
More informationA new class of shift-invariant operators
1 A new class of shift-invariant operators Janne Heiilä Machine Vision Group Department of Electrical and Information Engineering P.O. Box 4500, 90014 University of Oulu, Finland Tel.: +358 8 553 2786,
More informationINSERM : LABORATOIRE INTERNATIONAL ASSOCI É, Universit é de Rennes I, SouthEast University, Rennes,FR
Combined invariants to similarity transformation and to blur using orthogonal Zernike moments Chen Beijing 1 2, Huazhong Shu 1 2 *, Hui Zhang 1 2, Gouenou Coatrieux 3, Limin Luo 1 2 *, Jean-Louis Coatrieux
More informationAffine Normalization of Symmetric Objects
Affine Normalization of Symmetric Objects Tomáš Suk and Jan Flusser Institute of Information Theory and Automation, Academy of Sciences of the Czech Republic, Pod vodárenskou věží 4, 182 08 Prague 8, Czech
More informationMultiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition
Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Esa Rahtu Mikko Salo Janne Heikkilä Department of Electrical and Information Engineering P.O. Box 4500, 90014 University of
More informationANALYSIS and interpretation of an image which was
Moment Invariants in Image Analysis Jan Flusser Abstract This paper aims to present a survey of object recognition/classification methods based on image moments. We review various types of moments (geometric
More informationImage Recognition Using Modified Zernike Moments
Sensors & Transducers 204 by IFSA Publishing S. L. http://www.sensorsportal.com Image Recognition Using Modified ernike Moments Min HUANG Yaqiong MA and Qiuping GONG School of Computer and Communication
More informationA New Affine Invariant Image Transform Based on Ridgelets
A New Affine Invariant Image Transform Based on idgelets Esa ahtu Janne Heikkilä Department of Electrical and Information Engineering P.O. Box 45, 914 University of Oulu, Finland {erahtu,jth}@ee.oulu.fi
More informationIMAGE INVARIANTS. Andrej Košir, Jurij F. Tasič
IMAGE INVARIANTS Andrej Košir, Jurij F. Tasič University of Ljubljana Faculty of Electrical Engineering Tržaška 25, Ljubljana, Slovenia andrej.kosir@fe.uni-lj.si ABSTRACT A pre-processing technique for
More informationResearch Article Cutting Affine Moment Invariants
Mathematical Problems in Engineering Volume 2012, Article ID 928161, 12 pages doi:10.1155/2012/928161 Research Article Cutting Affine Moment Invariants Jianwei Yang, 1 Ming Li, 2 Zirun Chen, 1 and Yunjie
More informationROTATION, TRANSLATION AND SCALING INVARIANT WATERMARKING USING A GENERALIZED RADON TRANSFORMATION
ROTATION, TRANSLATION AND SCALING INVARIANT WATERMARKING USING A GENERALIZED RADON TRANSFORMATION D. SIMITOPOULOS, A. OIKONOMOPOULOS, AND M. G. STRINTZIS Aristotle University of Thessaloniki Electrical
More informationComputer Vision & Digital Image Processing
Computer Vision & Digital Image Processing Image Restoration and Reconstruction I Dr. D. J. Jackson Lecture 11-1 Image restoration Restoration is an objective process that attempts to recover an image
More informationAccurate Orthogonal Circular Moment Invariants of Gray-Level Images
Journal of Computer Science 7 (5): 75-722, 20 ISSN 549-3636 20 Science Publications Accurate Orthogonal Circular Moment Invariants of Gray-Level Images Khalid Mohamed Hosny Department of Computer Science,
More informationSpatially adaptive alpha-rooting in BM3D sharpening
Spatially adaptive alpha-rooting in BM3D sharpening Markku Mäkitalo and Alessandro Foi Department of Signal Processing, Tampere University of Technology, P.O. Box FIN-553, 33101, Tampere, Finland e-mail:
More informationInvariant Pattern Recognition using Dual-tree Complex Wavelets and Fourier Features
Invariant Pattern Recognition using Dual-tree Complex Wavelets and Fourier Features G. Y. Chen and B. Kégl Department of Computer Science and Operations Research, University of Montreal, CP 6128 succ.
More informationEmpirical Analysis of Invariance of Transform Coefficients under Rotation
International Journal of Engineering Research and Development e-issn: 2278-67X, p-issn: 2278-8X, www.ijerd.com Volume, Issue 5 (May 25), PP.43-5 Empirical Analysis of Invariance of Transform Coefficients
More informationAffine Invariant Pattern Recognition Using Multi-Scale Autoconvolution
Affine Invariant Pattern Recognition Using Multi-Scale Autoconvolution Esa Rahtu, Mikko Salo and Janne Heikkilä Abstract This article presents a new affine invariant image transform called Multi-Scale
More informationSYSTEM RECONSTRUCTION FROM SELECTED HOS REGIONS. Haralambos Pozidis and Athina P. Petropulu. Drexel University, Philadelphia, PA 19104
SYSTEM RECOSTRUCTIO FROM SELECTED HOS REGIOS Haralambos Pozidis and Athina P. Petropulu Electrical and Computer Engineering Department Drexel University, Philadelphia, PA 94 Tel. (25) 895-2358 Fax. (25)
More informationDeformation and Viewpoint Invariant Color Histograms
1 Deformation and Viewpoint Invariant Histograms Justin Domke and Yiannis Aloimonos Computer Vision Laboratory, Department of Computer Science University of Maryland College Park, MD 274, USA domke@cs.umd.edu,
More informationPhase-Correlation Motion Estimation Yi Liang
EE 392J Final Project Abstract Phase-Correlation Motion Estimation Yi Liang yiliang@stanford.edu Phase-correlation motion estimation is studied and implemented in this work, with its performance, efficiency
More informationOn the FPA infrared camera transfer function calculation
On the FPA infrared camera transfer function calculation (1) CERTES, Université Paris XII Val de Marne, Créteil, France (2) LTM, Université de Bourgogne, Le Creusot, France by S. Datcu 1, L. Ibos 1,Y.
More informationA RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES
A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES V.Sridevi, P.Malarvizhi, P.Mathivannan Abstract Rain removal from a video is a challenging problem due to random spatial distribution and
More informationVarious Shape Descriptors in Image Processing A Review
Various Shape Descriptors in Image Processing A Review Preetika D Silva 1, P. Bhuvaneswari 2 1 PG Student, Department of Electronics and Communication, Rajarajeswari College of Engineering, Bangalore,
More informationObject Recognition Using Local Characterisation and Zernike Moments
Object Recognition Using Local Characterisation and Zernike Moments A. Choksuriwong, H. Laurent, C. Rosenberger, and C. Maaoui Laboratoire Vision et Robotique - UPRES EA 2078, ENSI de Bourges - Université
More informationChapter 1 INTRODUCTION TO MOMENTS. 1.1 Motivation
Chapter 1 INTRODUCTION TO MOMENTS 1.1 Motivation In our everyday life, each of us nearly constantly receives, processes, and analyzes a huge amount of information of various kind, significance and quality
More informationA Modified Moment-Based Image Watermarking Method Robust to Cropping Attack
A Modified Moment-Based Watermarking Method Robust to Cropping Attack Tianrui Zong, Yong Xiang, and Suzan Elbadry School of Information Technology Deakin University, Burwood Campus Melbourne, Australia
More informationEdges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise
Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:
More informationAnalysis of Hu's Moment Invariants on Image Scaling and Rotation
Edith Cowan University Research Online ECU Publications Pre. 11 1 Analysis of Hu's Moent Invariants on Iage Scaling and Rotation Zhihu Huang Edith Cowan University Jinsong Leng Edith Cowan University 1.119/ICCET.1.548554
More informationQuaternion Wiener Deconvolution for Noise Robust Color Image Registration
1278 IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 9, SEPTEMBER 2015 Quaternion Wiener Deconvolution for Noise Robust Color Image Registration Matteo Pedone, Eduardo Bayro-Corrochano, Jan Flusser, and Janne
More informationA Robust Modular Wavelet Network Based Symbol Classifier
A Robust Modular Wavelet Network Based Symbol Classifier A.K. Mishra, P.W. Fieguth, D.A. Clausi University of Waterloo, Vision and Image Processing Group akmishra, pfieguth, dclausi@uwaterloo.ca Abstract.
More informationGeneralized Laplacian as Focus Measure
Generalized Laplacian as Focus Measure Muhammad Riaz 1, Seungjin Park, Muhammad Bilal Ahmad 1, Waqas Rasheed 1, and Jongan Park 1 1 School of Information & Communications Engineering, Chosun University,
More informationA Recognition System for 3D Embossed Digits on Non-Smooth Metallic Surface
2011 International Conference on elecommunication echnology and Applications Proc.of CSI vol.5 (2011) (2011) IACSI Press, Singapore A Recognition System for 3D Embossed Digits on Non-Smooth Metallic Surface
More informationClassifying Galaxy Morphology using Machine Learning
Julian Kates-Harbeck, Introduction: Classifying Galaxy Morphology using Machine Learning The goal of this project is to classify galaxy morphologies. Generally, galaxy morphologies fall into one of two
More informationAtmospheric Turbulence Effects Removal on Infrared Sequences Degraded by Local Isoplanatism
Atmospheric Turbulence Effects Removal on Infrared Sequences Degraded by Local Isoplanatism Magali Lemaitre 1, Olivier Laligant 1, Jacques Blanc-Talon 2, and Fabrice Mériaudeau 1 1 Le2i Laboratory, University
More informationCorners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros
Corners, Blobs & Descriptors With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Motivation: Build a Panorama M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 How do we build panorama?
More informationEmpirical Mean and Variance!
Global Image Properties! Global image properties refer to an image as a whole rather than components. Computation of global image properties is often required for image enhancement, preceding image analysis.!
More informationwhere =0,, 1, () is the sample at time index and is the imaginary number 1. Then, () is a vector of values at frequency index corresponding to the mag
Efficient Discrete Tchebichef on Spectrum Analysis of Speech Recognition Ferda Ernawan and Nur Azman Abu Abstract Speech recognition is still a growing field of importance. The growth in computing power
More informationAIR FORCE RESEARCH LABORATORY Directed Energy Directorate 3550 Aberdeen Ave SE AIR FORCE MATERIEL COMMAND KIRTLAND AIR FORCE BASE, NM
AFRL-DE-PS-JA-2007-1004 AFRL-DE-PS-JA-2007-1004 Noise Reduction in support-constrained multi-frame blind-deconvolution restorations as a function of the number of data frames and the support constraint
More informationComputer Vision. Filtering in the Frequency Domain
Computer Vision Filtering in the Frequency Domain Filippo Bergamasco (filippo.bergamasco@unive.it) http://www.dais.unive.it/~bergamasco DAIS, Ca Foscari University of Venice Academic year 2016/2017 Introduction
More informationCITS 4402 Computer Vision
CITS 4402 Computer Vision Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh, CEO at Mapizy (www.mapizy.com) and InFarm (www.infarm.io) Lecture 04 Greyscale Image Analysis Lecture 03 Summary Images as 2-D signals
More informationDetectors part II Descriptors
EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers Goal:
More informationFast and Automatic Watermark Resynchronization based on Zernike. Moments
Fast and Automatic Watermark Resynchronization based on Zernike Moments Xiangui Kang a, Chunhui Liu a, Wenjun Zeng b, Jiwu Huang a, Congbai Liu a a Dept. of ECE., Sun Yat-Sen Univ. (Zhongshan Univ.), Guangzhou
More informationAdvances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.
6.869 Advances in Computer Vision Prof. Bill Freeman March 3, 2005 Image and shape descriptors Affine invariant features Comparison of feature descriptors Shape context Readings: Mikolajczyk and Schmid;
More informationEnhanced Fourier Shape Descriptor Using Zero-Padding
Enhanced ourier Shape Descriptor Using Zero-Padding Iivari Kunttu, Leena Lepistö, and Ari Visa Tampere University of Technology, Institute of Signal Processing, P.O. Box 553, I-330 Tampere inland {Iivari.Kunttu,
More informationIntroduction to Computer Vision. 2D Linear Systems
Introduction to Computer Vision D Linear Systems Review: Linear Systems We define a system as a unit that converts an input function into an output function Independent variable System operator or Transfer
More informationImproved system blind identification based on second-order cyclostationary statistics: A group delay approach
SaÅdhanaÅ, Vol. 25, Part 2, April 2000, pp. 85±96. # Printed in India Improved system blind identification based on second-order cyclostationary statistics: A group delay approach P V S GIRIDHAR 1 and
More informationSuper-Resolution. Dr. Yossi Rubner. Many slides from Miki Elad - Technion
Super-Resolution Dr. Yossi Rubner yossi@rubner.co.il Many slides from Mii Elad - Technion 5/5/2007 53 images, ratio :4 Example - Video 40 images ratio :4 Example Surveillance Example Enhance Mosaics Super-Resolution
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff
More informationINTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY
[Gaurav, 2(1): Jan., 2013] ISSN: 2277-9655 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY Face Identification & Detection Using Eigenfaces Sachin.S.Gurav *1, K.R.Desai 2 *1
More informationImage enhancement. Why image enhancement? Why image enhancement? Why image enhancement? Example of artifacts caused by image encoding
13 Why image enhancement? Image enhancement Example of artifacts caused by image encoding Computer Vision, Lecture 14 Michael Felsberg Computer Vision Laboratory Department of Electrical Engineering 12
More informationFeature extraction: Corners and blobs
Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over
More informationECG782: Multidimensional Digital Signal Processing
Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu ECG782: Multidimensional Digital Signal Processing Filtering in the Frequency Domain http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Background
More informationVlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems
1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Perception Concepts Vision Chapter 4 (textbook) Sections 4.3 to 4.5 What is the course
More informationNo. of dimensions 1. No. of centers
Contents 8.6 Course of dimensionality............................ 15 8.7 Computational aspects of linear estimators.................. 15 8.7.1 Diagonalization of circulant andblock-circulant matrices......
More informationA Comparison of Multi-Frame Blind Deconvolution and Speckle Imaging Energy Spectrum Signal-to-Noise Ratios-- Journal Article (Preprint)
AFRL-RD-PS- TP-2008-1005 AFRL-RD-PS- TP-2008-1005 A Comparison of Multi-Frame Blind Deconvolution and Speckle Imaging Energy Spectrum Signal-to-Noise Ratios-- Journal Article (Preprint) Charles L. Matson
More informationRole of Assembling Invariant Moments and SVM in Fingerprint Recognition
56 Role of Assembling Invariant Moments SVM in Fingerprint Recognition 1 Supriya Wable, 2 Chaitali Laulkar 1, 2 Department of Computer Engineering, University of Pune Sinhgad College of Engineering, Pune-411
More information1. INTRODUCTION. Yen-Chung Chiu. Wen-Hsiang Tsai.
Copyright Protection by Watermarking for Color Images against Rotation and Scaling Attacks Using Coding and Synchronization of Peak Locations in DFT Domain Yen-Chung Chiu Department of Computer and Information
More information! # %& () +,,. + / 0 & ( , 3, %0203) , 3, &45 & ( /, 0203 & ( 4 ( 9
! # %& () +,,. + / 0 & ( 0111 0 2 0+, 3, %0203) 0111 0 2 0+, 3, &45 & ( 6 7 2. 2 0111 48 5488 /, 0203 & ( 4 ( 9 : BLIND IMAGE DECONVOLUTION USING THE SYLVESTER RESULTANT MATRIX Nora Alkhaldi and Joab Winkler
More informationEvaluation of OCR Algorithms for Images with Different Spatial Resolutions and Noises
Evaluation of OCR Algorithms for Images with Different Spatial Resolutions and Noises by Qing Chen A thesis submitted to the School of Graduate Studies and Research in partial fulfillment of the requirements
More informationSampling in 1D ( ) Continuous time signal f(t) Discrete time signal. f(t) comb
Sampling in 2D 1 Sampling in 1D Continuous time signal f(t) Discrete time signal t ( ) f [ k] = f( kt ) = f( t) δ t kt s k s f(t) comb k 2 Nyquist theorem (1D) At least 2 sample/period are needed to represent
More informationWhat is Image Deblurring?
What is Image Deblurring? When we use a camera, we want the recorded image to be a faithful representation of the scene that we see but every image is more or less blurry, depending on the circumstances.
More informationUsing Entropy and 2-D Correlation Coefficient as Measuring Indices for Impulsive Noise Reduction Techniques
Using Entropy and 2-D Correlation Coefficient as Measuring Indices for Impulsive Noise Reduction Techniques Zayed M. Ramadan Department of Electronics and Communications Engineering, Faculty of Engineering,
More informationShape Classification Using Zernike Moments
Shape Classification Using Zernike Moments Michael Vorobyov icamp at University of California Irvine August 5, 20 Abstract Zernike moments have mathematical properties, make them ideal image features to
More informationRESTORATION OF VIDEO BY REMOVING RAIN
RESTORATION OF VIDEO BY REMOVING RAIN Sajitha Krishnan 1 and D.Venkataraman 1 1 Computer Vision and Image Processing, Department of Computer Science, Amrita Vishwa Vidyapeetham University, Coimbatore,
More informationSimultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors
Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,
More informationFeature Vector Similarity Based on Local Structure
Feature Vector Similarity Based on Local Structure Evgeniya Balmachnova, Luc Florack, and Bart ter Haar Romeny Eindhoven University of Technology, P.O. Box 53, 5600 MB Eindhoven, The Netherlands {E.Balmachnova,L.M.J.Florack,B.M.terHaarRomeny}@tue.nl
More informationTensor Method for Constructing 3D Moment Invariants
Tensor Method for Constructing 3D Moment Invariants Tomáš Suk and Jan Flusser Institute of Information Theory and Automation of the ASCR {suk,flusser}@utia.cas.cz Abstract. A generalization from 2D to
More informationIMAGE ENHANCEMENT II (CONVOLUTION)
MOTIVATION Recorded images often exhibit problems such as: blurry noisy Image enhancement aims to improve visual quality Cosmetic processing Usually empirical techniques, with ad hoc parameters ( whatever
More informationAN INVESTIGATION ON EFFICIENT FEATURE EXTRACTION APPROACHES FOR ARABIC LETTER RECOGNITION
AN INVESTIGATION ON EFFICIENT FEATURE EXTRACTION APPROACHES FOR ARABIC LETTER RECOGNITION H. Aboaisha 1, Zhijie Xu 1, I. El-Feghi 2 z.xu@hud.ac.uk Idrisel@ee.edu.ly 1 School of Computing and Engineering,
More informationFARSI CHARACTER RECOGNITION USING NEW HYBRID FEATURE EXTRACTION METHODS
FARSI CHARACTER RECOGNITION USING NEW HYBRID FEATURE EXTRACTION METHODS Fataneh Alavipour 1 and Ali Broumandnia 2 1 Department of Electrical, Computer and Biomedical Engineering, Qazvin branch, Islamic
More informationSpatial Transformation
Spatial Transformation Presented by Liqun Chen June 30, 2017 1 Overview 2 Spatial Transformer Networks 3 STN experiments 4 Recurrent Models of Visual Attention (RAM) 5 Recurrent Models of Visual Attention
More informationINTEREST POINTS AT DIFFERENT SCALES
INTEREST POINTS AT DIFFERENT SCALES Thank you for the slides. They come mostly from the following sources. Dan Huttenlocher Cornell U David Lowe U. of British Columbia Martial Hebert CMU Intuitively, junctions
More informationMichal Kuneš
A (Zernike) moment-based nonlocal-means algorithm for image denoising Michal Kuneš xkunes@utia.cas.cz ZOI UTIA, ASCR, Friday seminar 13.03.015 Introduction - Uses nonlocal (NL) means filter - Introduce
More information3-D Projective Moment Invariants
Journal of Information & Computational Science 4: * (2007) 1 Available at http://www.joics.com 3-D Projective Moment Invariants Dong Xu a,b,c,d,, Hua Li a,b,c a Key Laboratory of Intelligent Information
More informationIEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART C: APPLICATIONS AND REVIEWS, VOL. 30, NO. 1, FEBRUARY I = N
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART C: APPLICATIONS AND REVIEWS, VOL. 30, NO. 1, FEBRUARY 2000 125 Translation, Rotation, and Scale-Invariant Object Recognition L. A. Torres-Méndez,
More informationMultiscale Image Transforms
Multiscale Image Transforms Goal: Develop filter-based representations to decompose images into component parts, to extract features/structures of interest, and to attenuate noise. Motivation: extract
More informationScale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract
Scale-Invariance of Support Vector Machines based on the Triangular Kernel François Fleuret Hichem Sahbi IMEDIA Research Group INRIA Domaine de Voluceau 78150 Le Chesnay, France Abstract This paper focuses
More informationFeature Extraction Using Zernike Moments
Feature Extraction Using Zernike Moments P. Bhaskara Rao Department of Electronics and Communication Engineering ST.Peter's Engineeing college,hyderabad,andhra Pradesh,India D.Vara Prasad Department of
More informationHigh-Resolution. Transmission. Electron Microscopy
Part 4 High-Resolution Transmission Electron Microscopy 186 Significance high-resolution transmission electron microscopy (HRTEM): resolve object details smaller than 1nm (10 9 m) image the interior of
More informationarxiv: v1 [cs.cv] 10 Feb 2016
GABOR WAVELETS IN IMAGE PROCESSING David Bařina Doctoral Degree Programme (2), FIT BUT E-mail: xbarin2@stud.fit.vutbr.cz Supervised by: Pavel Zemčík E-mail: zemcik@fit.vutbr.cz arxiv:162.338v1 [cs.cv]
More informationDESIGN OF MULTI-DIMENSIONAL DERIVATIVE FILTERS. Eero P. Simoncelli
Published in: First IEEE Int l Conf on Image Processing, Austin Texas, vol I, pages 790--793, November 1994. DESIGN OF MULTI-DIMENSIONAL DERIVATIVE FILTERS Eero P. Simoncelli GRASP Laboratory, Room 335C
More informationBananagrams Tile Extraction and Letter Recognition for Rapid Word Suggestion
Bananagrams Tile Extraction and Letter Recognition for Rapid Word Suggestion Steven Leung, Steffi Perkins, and Colleen Rhoades Department of Bioengineering Stanford University Stanford, CA Abstract Bananagrams
More informationToday. MIT 2.71/2.710 Optics 11/10/04 wk10-b-1
Today Review of spatial filtering with coherent illumination Derivation of the lens law using wave optics Point-spread function of a system with incoherent illumination The Modulation Transfer Function
More informationNotes on Regularization and Robust Estimation Psych 267/CS 348D/EE 365 Prof. David J. Heeger September 15, 1998
Notes on Regularization and Robust Estimation Psych 67/CS 348D/EE 365 Prof. David J. Heeger September 5, 998 Regularization. Regularization is a class of techniques that have been widely used to solve
More informationConvolution Spatial Aliasing Frequency domain filtering fundamentals Applications Image smoothing Image sharpening
Frequency Domain Filtering Correspondence between Spatial and Frequency Filtering Fourier Transform Brief Introduction Sampling Theory 2 D Discrete Fourier Transform Convolution Spatial Aliasing Frequency
More information1. Abstract. 2. Introduction/Problem Statement
Advances in polarimetric deconvolution Capt. Kurtis G. Engelson Air Force Institute of Technology, Student Dr. Stephen C. Cain Air Force Institute of Technology, Professor 1. Abstract One of the realities
More informationOn New Radon-Based Translation, Rotation, and Scaling Invariant Transform for Face Recognition
On New Radon-Based Translation, Rotation, and Scaling Invariant Transform for Face Recognition Tomasz Arodź 1,2 1 Institute of Computer Science, AGH, al. Mickiewicza 30, 30-059 Kraków, Poland 2 Academic
More informationSupplementary Figure 1: Example non-overlapping, binary probe functions P1 (~q) and P2 (~q), that add to form a top hat function A(~q).
Supplementary Figures P(q) A(q) + Function Value P(q) qmax = Supplementary Figure : Example non-overlapping, binary probe functions P (~q) and P (~q), that add to form a top hat function A(~q). qprobe
More informationImage Processing 1 (IP1) Bildverarbeitung 1
MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV (KOGS) Image Processing 1 (IP1) Bildverarbeitung 1 Lecture 7 Spectral Image Processing and Convolution Winter Semester 2014/15 Slides: Prof. Bernd
More informationHalf-Pel Accurate Motion-Compensated Orthogonal Video Transforms
Flierl and Girod: Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms, IEEE DCC, Mar. 007. Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms Markus Flierl and Bernd Girod Max
More informationImage Degradation Model (Linear/Additive)
Image Degradation Model (Linear/Additive),,,,,,,, g x y h x y f x y x y G uv H uv F uv N uv 1 Source of noise Image acquisition (digitization) Image transmission Spatial properties of noise Statistical
More informationOriginal Article Design Approach for Content-Based Image Retrieval Using Gabor-Zernike Features
International Archive of Applied Sciences and Technology Volume 3 [2] June 2012: 42-46 ISSN: 0976-4828 Society of Education, India Website: www.soeagra.com/iaast/iaast.htm Original Article Design Approach
More informationFourier Series. Combination of Waves: Any PERIODIC function f(t) can be written: How to calculate the coefficients?
Fourier Series Combination of Waves: Any PERIODIC function f(t) can be written: How to calculate the coefficients? What is the Fourier Transform It is the representation of an arbitrary NON-periodic signal
More informationCITS 4402 Computer Vision
CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh Lecture 06 Object Recognition Objectives To understand the concept of image based object recognition To learn how to match images
More informationImage preprocessing in spatial domain
Image preprocessing in spatial domain Sampling theorem, aliasing, interpolation, geometrical transformations Revision: 1.4, dated: May 25, 2006 Tomáš Svoboda Czech Technical University, Faculty of Electrical
More informationCS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015
CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature
More informationNonparametric Rotational Motion Compensation Technique for High-Resolution ISAR Imaging via Golden Section Search
Progress In Electromagnetics Research M, Vol. 36, 67 76, 14 Nonparametric Rotational Motion Compensation Technique for High-Resolution ISAR Imaging via Golden Section Search Yang Liu *, Jiangwei Zou, Shiyou
More information