Motion estimations based on invariant moments for frames interpolation in stereovision

Similar documents
Science Insights: An International Journal

Feature Extraction Using Zernike Moments

Shape of Gaussians as Feature Descriptors

Various Shape Descriptors in Image Processing A Review

Accurate Orthogonal Circular Moment Invariants of Gray-Level Images

Image Recognition Using Modified Zernike Moments

Single-Image-Based Rain and Snow Removal Using Multi-guided Filter

Object Recognition Using Local Characterisation and Zernike Moments

A Method for Blur and Similarity Transform Invariant Object Recognition

Empirical Analysis of Invariance of Transform Coefficients under Rotation

Affine Normalization of Symmetric Objects

Enhanced Fourier Shape Descriptor Using Zero-Padding

RESTORATION OF VIDEO BY REMOVING RAIN

Using Entropy and 2-D Correlation Coefficient as Measuring Indices for Impulsive Noise Reduction Techniques

A Modified Moment-Based Image Watermarking Method Robust to Cropping Attack

Feature Vector Similarity Based on Local Structure

Phase-Correlation Motion Estimation Yi Liang

A Novel Activity Detection Method

AN INVESTIGATION ON EFFICIENT FEATURE EXTRACTION APPROACHES FOR ARABIC LETTER RECOGNITION

Representing regions in 2 ways:

A Recognition System for 3D Embossed Digits on Non-Smooth Metallic Surface

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES

Rotational Invariants for Wide-baseline Stereo

Deformation and Viewpoint Invariant Color Histograms

Classifying Galaxy Morphology using Machine Learning

Robust Speed Controller Design for Permanent Magnet Synchronous Motor Drives Based on Sliding Mode Control

Role of Assembling Invariant Moments and SVM in Fingerprint Recognition

Research Article Cutting Affine Moment Invariants

Basic Concepts of. Feature Selection

A New Efficient Method for Producing Global Affine Invariants

Lecture 8: Interest Point Detection. Saad J Bedros

MOMENT functions are used in several computer vision

Affine Structure From Motion

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors

CHAPTER 1. Over 50 Years of Image Moments and Moment Invariants

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

Michal Kuneš

Covariance Tracking Algorithm on Bilateral Filtering under Lie Group Structure Yinghong Xie 1,2,a Chengdong Wu 1,b

Lecture 8: Interest Point Detection. Saad J Bedros

OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS

Inverse Problems in Image Processing

The Efficient Discrete Tchebichef Transform for Spectrum Analysis of Speech Recognition

International Journal of Computer Engineering and Applications, Volume XII, Special Issue, August 18, ISSN

A HYBRID MOMENT METHOD FOR IDENTIFICATION OF 3D-OBJECTS FROM 2D-MONOCROMATIC IMAGES

The New Graphic Description of the Haar Wavelet Transform

Super-Resolution. Dr. Yossi Rubner. Many slides from Miki Elad - Technion

Invariant Pattern Recognition using Dual-tree Complex Wavelets and Fourier Features

OSCILLATION OF THE MEASUREMENT ACCURACY OF THE CENTER LOCATION OF AN ELLIPSE ON A DIGITAL IMAGE

Vectors and Matrices

Analysis of Hu's Moment Invariants on Image Scaling and Rotation

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 8, AUGUST

Evaluation of OCR Algorithms for Images with Different Spatial Resolutions and Noises

Statistical Filters for Crowd Image Analysis

Original Article Design Approach for Content-Based Image Retrieval Using Gabor-Zernike Features

Comparative study of global invariant. descriptors for object recognition

Object Recognition Using a Neural Network and Invariant Zernike Features

Fast and Automatic Watermark Resynchronization based on Zernike. Moments

Invariant Feature Extraction from Fingerprint Biometric Using Pseudo Zernike Moments

Orientation Map Based Palmprint Recognition

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Tensor Method for Constructing 3D Moment Invariants

An Efficient Algorithm for Fast Computation of Pseudo-Zernike Moments

Available online at ScienceDirect. Procedia Engineering 136 (2016 )

Open Access Measuring Method for Diameter of Bearings Based on the Edge Detection Using Zernike Moment

Degeneracies, Dependencies and their Implications in Multi-body and Multi-Sequence Factorizations

Temporal Factorization Vs. Spatial Factorization

OBJECT BASED IMAGE ANALYSIS FOR URBAN MAPPING AND CITY PLANNING IN BELGIUM. P. Lemenkova

Capability Assessment of Finite Element Software in Predicting the Last Ply Failure of Composite Laminates

Dominant Feature Vectors Based Audio Similarity Measure

Singlets. Multi-resolution Motion Singularities for Soccer Video Abstraction. Katy Blanc, Diane Lingrand, Frederic Precioso

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Available online at ScienceDirect. Physics Procedia 67 (2015 )

Edge Detection in Computer Vision Systems

Asaf Bar Zvi Adi Hayat. Semantic Segmentation

SIMPLE GABOR FEATURE SPACE FOR INVARIANT OBJECT RECOGNITION

A METHOD OF FINDING IMAGE SIMILAR PATCHES BASED ON GRADIENT-COVARIANCE SIMILARITY

FARSI CHARACTER RECOGNITION USING NEW HYBRID FEATURE EXTRACTION METHODS

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier

Human Pose Tracking I: Basics. David Fleet University of Toronto

A Robust Modular Wavelet Network Based Symbol Classifier

Theoretical Computer Science

Generalized Laplacian as Focus Measure

Blur Image Edge to Enhance Zernike Moments for Object Recognition

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

K. Zainuddin et al. / Procedia Engineering 20 (2011)

Convolutional Associative Memory: FIR Filter Model of Synapse

INTEREST POINTS AT DIFFERENT SCALES

Available online at ScienceDirect. Procedia Computer Science 20 (2013 ) 90 95

No. of dimensions 1. No. of centers

Sound Recognition in Mixtures

Unsupervised Learning with Permuted Data

Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms

Procedia - Social and Behavioral Sciences 174 ( 2015 ) INTE 2014

Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network

Multiple Similarities Based Kernel Subspace Learning for Image Classification

INSERM : LABORATOIRE INTERNATIONAL ASSOCI É, Universit é de Rennes I, SouthEast University, Rennes,FR

ROTATION, TRANSLATION AND SCALING INVARIANT WATERMARKING USING A GENERALIZED RADON TRANSFORMATION

Available online at ScienceDirect. Procedia Computer Science 70 (2015 ) Bengal , India

Hu and Zernike Moments for Sign Language Recognition

Transcription:

Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science (013 ) 110 1111 17 th International Conference in Knowledge Based and Intelligent Information and Engineering Systems - Abstract KES013 Motion estimations based on invariant moments for frames interpolation in stereovision Margarita Favorskaya*, Dmitriy Pyankov, Aleksei Popov Siberian State Aerospace University, 31 Krasnoyarsky Rabochy av., Krasnoyarsk, 660014 Russian Federation At present, stereo video sequences are actively used in the movie industry, in geographical information systems, and in navigation systems, among others. A novel method improves the frames interpolation by forming an invariant set of local motion vectors. First, the motion in a scene is estimated by block-matching algorithm. Second, accurate estimations by using Hu moments (for a noisy video sequence Zernike moments) are calculated. Such approach provides smooth motion that significantly improves the resulting stereo video sequence. Experimental results show the efficiency of frames interpolation based on such approach. The detection of local motion vectors achieves 86 % accuracy. 013 The Authors. Published by Elsevier B.V. Selection and peer-review under responsibility of KES International. Open access under CC BY-NC-ND license. Keyword: Motion estimation; Hu moments; Zernike moments; frame interpolation; stereovision 1. Introduction The basis of stereo shooting is a preparation of two similar, but not identical video sequences which are received from two survey points relatively an object. In ideal case, the axes of video cameras and objective lens systems ought to be calibrated. A stereo video sequence is formed by overlapping of two video sequences which are received from left ant right video cameras. Then the resulting stereo video sequence may be transformed for required device: an anaglyph, a linear polarization system, a locked liquid crystal display, etc. At present, a stereo shooting received by the home video cameras is possible in an automated mode through understanding that video sequences will be manually synchronized because of a non-synchronous start of shooting and different access time to data storages and writing devices. The development of spatio-temporal correction of stereo video sequences permits to solve these problems. A video sequence correction based on a * Corresponding author. Tel.: +7-391-91-940; fax: +7-391-91-9147. E-mail address: favorskaya@sibsau.ru. 1877-0509 013 The Authors. Published by Elsevier B.V. Open access under CC BY-NC-ND license. Selection and peer-review under responsibility of KES International doi:10.1016/j.procs.013.09.196

Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 1103 frequency frames transformation is the popular approach for interpolation of missing frames. Such approach requires the accurate motion estimations in a scene that may be realized by the usage of the moment theory. The first application of geometric moments (or regular moments) was introduced by Hu [1] as a mean of invariant visual pattern recognition. Moment functions transform D function describing an image into a feature space with representing the certain shape characteristics by the transformed coefficients. Moment functions are categorized into two main groups depending of their definition into a polynomial subspace: Cartesian space and polar coordinate space. Geometric, Legendre, and some others moments are concerned to the first group and have a difficulty with a calculation of rotational invariants. Zernike, pseudo-zernike, and affine moments are involved in the second group. All kinds of moments are not well suitable for image stretching or compression if parameters of such transformation do not know. The different methods of motion estimation were investigated. The experimental results show that the most appropriative approach is based on a two-stage method. First, the preliminary regions with any motion are found by fast and approximate Block-Matching Algorithm (BMA). Second, an accurate motion in detected moving regions is determined by usage the invariant moments. The calculation of any moments is a high computer cost procedure. However the processes of correction and interpolation do not concern to the real-tine applications. Also the usage of pseudo-zernike moments instead of Zernike moments reduces a calculation cost. This paper is organized as follows. A brief description of various applications of moments in a digital image analysis is situated in Section. The problem statement is discussed in Section 3. Section 4 describes a background model building. In Section 5, the proposed two-stage method of motion estimations based on BMA and invariants moments is presented. The task of frame interpolation is provided by Section 6. Experimental results by using test video sequences are situated in Section 7. Conclusions are located in Section 8.. Related Work The motion estimation requires the high computational resources. Therefore various algorithms reducing computation complexity were developed. Usually existing methods of frames interpolation use the modified BMA for various types of motion. For slow motion estimation into rarely changeable regions, a frame recalculation with quarter-pixel accuracy can be applied. Sometimes methods of motion estimation are based on the calculation of wavelet coefficients in each frequency sub-range. Revaud et al. [] proposed the method of a global motion compensation for frames interpolation. This method processed a fast motion of determined type (rotation, zoom-in, and zoom-out relative to video camera). However a sensitivity of local objects motion may be lost that causes an interrupted motion into a scene. Various types of BMA were in detail represented by Favorskaya [3]. BMA is divided in two main groups: with fixed sizes of blocks and with variable sizes of blocks. In the first case, a frame is partitioned on nonoverlapping blocks with fixed sizes. The best algorithm is the algorithm of full search which requires the high computational cost. In recent standards of video coding (H.64/AVC), methods of motion estimation with variable sizes of blocks (the second case) are used for more qualitative and efficient coding. However such methods are more complex in the algorithmic realization. The hierarchical structure of block sizes is used in the standard H.64/AVC. Each macro-block with sizes 16 16 pixels is divided into blocks 16 8, 8 16, or 8 8 pixels. On following hierarchical level, a macro-block with sizes 8 8 pixels are subdivided on blocks 8 4, 4 8, or 4 4 pixels. Blocks with large sizes are applied for homogeneous regions with a slow motion, and blocks with small sizes are oriented on regions with a fast motion. In methods with variable sizes of blocks, the motion search is executed for each from 41 sub-blocks; after that the optimal motion vector and the optimal block size are determined. The standard H.64/AVC provides several methods of motion estimations with variable sizes of blocks, including Full Search (FS), Fast Full Search (FFS), and a non-symmetrical crossed search by hexagon

1104 Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 (UMHexagonS). Motion estimations in FS are executed for each of 41 sub-blocks independently. FFS reduces computational cost relatively to FS by calculation the sums of absolute differences of pixels for blocks 4 4 pixels and for large sizes blocks by summation of sub-blocks with small sizes. The non-symmetrical crossed search by hexagon requires a prediction of initial point of search, early search termination, and several search patterns for fast motion search. However the accurate motion estimations into a scene need in more exact methods. Methods based on various types of moments are well suitable for frames interpolation. The typical applications of moments in image analysis theory connect with various tasks. Chen et al. [4] proposed to use the quaternion Zernike moments and their invariants for color image analysis and object recognition. The recognition of blurred image by Legendre moment invariants was investigated by Zhang et al. [5]. The application of Zernike phase-based descriptor for image representation and matching was proposed by Chen and Sun [6]. Yang and Pei [7] designed a fast algorithm of sub-pixel edge detection based on Zernike moments. Hu moment invariants for image scaling and rotation were analyzed by Huang and Leng [8]. Sanjeev et al. [9] applied Zernike moments for affine global motion estimations in a task of digital image stabilization. Shutler and Nixon [10] received the moving features for sequence-based description by usage Zernike velocity moments. The geometric transform of invariant texture analysis based on the modified Zernike moments was developed by Vyas and Rege [11]. Frejlichowski [1] analyzed objects shapes based on Zernike moments. Application of pseudo-zernike moments was proposed by Yang and Guo [13]. Zhang et al. [14] designed the approach based on Legendre moment invariants for image watermarking. The complex Zernike moment features for face recognition were proposed by Chandan et al. [15]. A methodology of content based image retrieval using the exact Legendre moments was developed by Rao et al. [16]. Some investigations are connected with a fast computation of geometric and Zernike moments. Methods of fast computation of geometric moments using a symmetric kernel and discrete orthogonal image moments were developed by Wee et al. [17] and Papakostas and Koulouriotis [18]. Boveiri [19] proposed a pattern classification using the statistical pseudo Zernike moments. Current researches presenting in this paper connect with the accurate motion estimations from frame to frame by usage Hu and Zernike moments into the predetermined motion regions. 3. Problem Statement For effective frames interpolation in stereovision, it is required to receive the exact motion estimations into sequential frames. Classification of motion types and tracking of dynamic objects trajectories based on such exact estimations permit to realize a qualitative frames interpolation. The proposed method for motion estimation is based on two-stage procedure. On the first preliminary stage, BMA is applied. The improved estimations based on Hu moments are received during the second final stage when local vectors in moving regions are determined. For noisy video sequences, the application of Zernike moments shows the better results. Noises, camera jitters, object rotations are essential problems for motion estimation. The development of motion estimation method which is invariant to specified issues is an actual task. Let us suppose that two video sequences were received from left V L ={I L 1, I L,, I L t,, I L k} and right V R ={I R 1, I R,, I R t,, I R k} video cameras, where I L t and I R t are frames from left and right video sequences respectively in time moment t. Let the right video sequence is a permanent, and d frames from I L t to I L t+d of initial left video sequence are analyzed (in the following discussion, indexes denoting left or right video sequences will not be pointed). The intermediate (interpolated) frames {I in } between adjacent frames {I t, I t+1 }, {I t+1, I t+ },, {I t+d 1, I t+d }of initial left video sequence ought to be found. Then such procedure is applied to right video sequence. Such frames interpolation needs in the building of background scene model, the application of motion estimation method, and the generation of required number of interpolated frames (usually from up to 5 interpolated frames).

Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 1105 4. Background Model Building The approaches as a background subtraction model, the probability methods, and the methods of time differences may be successfully used for a background model building. They have different complexity, computer cost, and functionality. A hybrid method based on probability and time differences show the good experimental results in fast changeable scenes. For scene with shadows and slow changeable brightness, a background subtraction model jointly with color information was used. Let us suppose that a camera noise into three color channels has a normal distribution with noise dispersions Rn, Gn, and Bn. For each pixel with coordinates (x, y), averages and dispersions of intensity function I(x, y) are calculated by usage the series of initial frames. Then a set of features [ R, G, B, R, G, B ] is formed. Update of these parameters is carried out by Eqs. 1 and, where t and t are average and dispersion in time moment t, t+1 and t+1 are average and dispersion in time moment (t+1), is an empirical constant. x,y x,y I x,y (1) t1 t 1 t1 I x,y x, x,y x,y x,y x,y () t1 t t1 t 1 t1 t1 y Let R x,y, G x,y, and B x,y are the values of RGB components for a pixel with coordinates (x, y). It is consider that a current pixel owns to a foreground if any of three inequalities is satisfied: R G B x,y x,y x,y R G B x,y 3max R, Rn x,y 3max G, Gn x,y 3max, B Bn (3) If anyone inequality from Eq. 3 is not fulfilled then a pixel owns to background. Also Eqs. (1) (3) permit to estimate a noisiness of video sequence. 5. Two-stage Method of Motion Estimation When a background model is built, the motions in a scene can be estimated. For this purpose, the approach of approximate detection of moving regions by usage BMA (Section 5.1) and accurate motion estimation based on Hu and Zernike moments (Section 5.) were used. 5.1. Block-matching motion estimation The preliminary motion estimation permits to detect not only the moving regions but also to determine a type of motion into sequential frames of initial video sequence. A calculation of local motion vectors by BMA with fixed size of block is implemented by the following schema: A current frame I t (x, y) and a following frame I t+1 (x, y) are divided into equaled blocks B c (x, y) and R f (x, y) with sizes s_block s_block. For a current frame, a sum S of intensity values of each pixel in the selected block B c (x, y) is computed.

1106 Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 In a following frame I t+1 (x, y), the adjacent blocks (diagonal, vertical, and horizontal) with a displacement on value shift relatively a current block B c (x, y) are pointed. Then sums S n, n {0, 1,, 8} of intensity values of each pixel in adjacent blocks are calculated. Sums S n, n {0, 1,, 8} are compared alternately with sum S of block B c (x, y) by function based on SSD (Sum of Squared Differences) metric by Eq. 4. 8 move min S S n (4) n0 The block with a minimal error value indicates a direction of motion vector for block B c (x, y) in a following frame. Such approach is not robust to displacement, rotation, and scaling of object. However it permits to detect the moving regions which will be analyzed by more exact methods. From all affine transformations, displacements and rotations are actual for interpolation task; a sharp scaling is practically absent in two adjacent frames. Thereby a motion classification includes only two types of motion a translation and/or a rotation. By usage BMA, the local motion vectors are built from frame to frame, then directions of these vectors are analyzed, and a motion type is determined by a voting method. If a motion type is known then the objects trajectories are tracked on a stage of resulting motion estimation. 5.. Accurate motion estimation based on Hu and Zernike moments The accurate motion estimation is a complex stage based on the usage of Hu moments or Zernike moments. If a video sequence is a noisy sequence then Zernike moments are used. The tracking of objects trajectories, which were determined on the preliminary stage, reduces a computational cost of interpolation task. If an image describing by D discrete intensity function I(x, y) with non-zero values in the finite part of XOY plane then the geometrical moments of all orders (p + q) are presented by Eq. 5 [1], where M and N are the image dimensions. M 1 N 1 x0 y0 p q m x y I x, y (5) pq Eq. 6 provides a calculation of geometrical central moments of order (p + q), where x and y are the gravity center coordinates of object image which are invariant to translations and determined by Eq. 7. pq M 1 N 1 p q x y y I x, y x0 y0 x (6) x m (7) 10 m00 y m01 m00 The point x, y is a centroid of image. A scale invariance is achieved by normalizations of moments γ 1 p q according to Eq. 8, where. η μ μ (8) pq pq pq It is known seven non-linear functions 1 7 of normalized invariant Hu moments [1] calculating by Eq. 9.

Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 1107 η 1 3 4 5 6 7 0 η η0 η0 4η11 η30 3η1 3η 1 η0 η30 η1 η1 η03 η30 3η1 η30 η1 η30 η1 3η 1 η03 3η 1 η03η1 η033η 30 η1 η1 η03 η0 η0 η30 η1 η1 η03 4η11 η30 3η 1 η03η30 η1 η30 η1 3η1 η03 η 3η η η 3η η η η 30 0 1 η1 η1 η03 1 03 30 1 1 03 (9) During experiments, a motion was estimated by usage of all 7 invariant Hu moments (Eq. 9) for a block B c (x, y) in a current frame I t, and for 9 blocks R f (x, y) in a following frame I t+1 which had a displacement value shift in diagonals, vertical, and horizontal directions (8 adjacent blocks and 1 central block). A similarity metric is based on the Euclid distance providing by Eq. 10, where B c denotes a block B c (x, y) in a current frame I t, R f denotes a block R f (x, y) in a following frame I t+1. R f R f sign n ln n 7 Bc Bc Bc,R f sign n ln n M (10) n1 If M(B c, R f ) value is less then the blocks in two adjacent frames are more similar. Eq. 11 determines the direction of motion vector as minimum from 9 correspondences calculating by Eq. 10. n min M nbc,rf 0,1,,3,..., 8 movehu (11) Zernike moments are coefficients mapping of initial function on a basis of radial polynomials. Property of a polynomial orthogonality of Zernike moments promotes essentially a restoration process of the initial image in comparison with a restoration process based on Hu moments. A set of Zernike polynomials {V nm (x, y)} which are determined in inner area of a unit circumference x + y = 1 may be written by Eq. 1, where n is an order of polynomial (n 0), m is a polynomial frequency which ought to satisfy following conditions: n m is an even number and m n, ρ is a distance from a center of a unit circumference to pixel with coordinates (x, y), θ is an angle between a motion vector and OX axis in anticlockwise direction, R nm (ρ) are radial polynomials. V R nm ρ nm x,yv, R ρexpimθ nm x y n m ρ 1 s0 s nm n s! ns ρ n m n m s! s! s! (1) The normalized Zernike moments {Z nm } with n order and m frequency for discrete image I(x, y) are calculated by Eq. 13.

1108 Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 n m * Znm Ix,yVnmx,y, x y 1 (13) π x y The 9 Zernike moments with various order and frequency of radial polynomial for each compared block R 00 (ρ), R 11 (ρ), R 0 (ρ), R (ρ), R 31 (ρ), R 33 (ρ), R 40 (ρ), R 4 (ρ), and R 44 (ρ) were analyzed during experiments. The block matching, the detection of the most similar block, and the motion vector direction were accomplished by using Eqs. 10 11. The moment s amplitudes are invariant to rotation. The robustness to transitions and scaling is achieved by normalization. Zernike moments are more robust for a noise as it was shown by Sanjeev et al. in research [9]. Thereby it is better to use Zernike moments for the noisy video sequences analysis. Let us notice that a motion type is determined on a preliminary stage. Therefore it is not needed to calculate all 9 matching of blocks (Eq. 10) when a motion direction is known. 6. Frames Interpolation Let d is a number of interpolated frames which will be added between frames I t and I t+1 (usually the parameter d is chosen from up to 5). The procedure includes the analysis of each pixel С B from block B c (x, y) in a current frame I t and each pixel С R from block R f (x, y) in a following frame I t+1. Previously, the color RGBspace values are recalculated to the YUV-space values for the explicit definition of Y intensity component. In an interpolated frame, each pixel С in changes smoothly its value of each component (Y, U, V) according to a motion vector direction. Therefore a smooth frames interpolation becomes a linear interpolation of each color components for each pixel determined by Eq. 14, where C (Y,U,V) B are values of components (Y, U, V) of pixel with coordinates (x, y) from a block B c (x, y) in a current frame I t, C (Y,U,V) R are values of components (Y, U, V) (Y,U,V) of pixel with coordinates (x, y) from a block R f (x, y) in a following frame I t+1, C in are values of components (Y, U, V) of pixel with coordinates (x, y) in an interpolated frame I in, k d is a parameter of linear displacement of pixels in each interpolated frame. Y, U, V Y, U, V Y, U, V Y, U, V C C C C k, k 0... 1 (14) in B R B d d Parameter k d is calculated by Eq. 15 and changed into interval k d [0 1] that provides a smooth displacement of color components for each interpolated frame. k k f k f 1,,,..., d (15) d 1 d 3 Such procedure repeats until all blocks will not be interpolated. Thereby d interpolated frames between I t and I t+1 frames of original video sequence will be received. 7. Experimental Researches For a motion estimation of moving objects, two test video sequences were used: a video sequence Coastguard [0] and a video sequence Horse from a test set of stereo video sequences [1]. Several frames from these test video sequences are situated in Figs. 1 and. Video sequence Coastguard includes the object with translation motion along OX and OY axes with different speed and a rotation motion is absent. Video sequence Horse represents the object with translation and rotation motions along OX and OY axes with different speed and angle rotation (from 5 to 90).

Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 1109 a b c Fig. 1. (a) frame 53; (b) frame 73; (c) frame 93 a b c Fig.. (a) frame 19; (b) frame 65; (c) frame 117 The tests were executed on PC with computer speed 15 Gflops. The software tool was development on language C++. A moving object was inscribed into a rectangle with sizes 44 44 pixels. By using the designed software tool, the sample local motion vectors were calculated. During investigations, the following parameters were considered: block sizes 4 4, 8 8, 16 16 pixels; values of displacement as all of block pixels, half of block pixels, and 1 pixel; and various speeds of objects. The 8 directions of error motion vectors relatively the sample vectors were analyzed. If these differences had been increased then amplitudes of error motion vectors were multiplied by the increasing coefficient [0 ] with step 0.5 (0 means the accurate direction and means the error direction). Then total errors are summed for each frame. Maximums of error motion vectors are situated in Table 1. Table 1. Maximums of error motion vectors Block sizes, pixels Displacement between blocks, pixels Video sequence Сoastguard Error, maximum value Video sequence Horse Error, maximum value BMA Hu moments Zernike moments BMA Hu moments Zernike moments 44 4 max max max max 1 max max 88 8 max 4 max max max 1 max max max 1616 16 max max 8 max max 1 max max

1110 Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 It is show from Table 1 that Hu moments give a maximum number of error motion vectors, Zernike moments provide a minimum number error motion vectors. The motion vectors estimations for test video sequences Сoastguard and Horse with various speed values and type of motion are situated in Tables and 3 respectively. Table. Accuracy of motion vectors for a video sequence Сoastguard Block sizes, pixels Displacement between blocks, pixels Accuracy, % BMA Hu moments Zernike moments 44 4 70.88 70.59 71.15 7.58 7.37 7.70 1 73.33 7.48 73.38 88 8 70.67 7.93 73.60 4 7.50 75.03 76.95 1 74.88 75.61 75.0 1616 16 77.54 76.50 79.71 8 75.04 74.46 75.71 1 73.79 7.96 80.7 Table 3. Accuracy of motion vectors for a video sequence Horse Block sizes, pixels Displacement between blocks, pixels Accuracy, % BMA Hu moments Zernike moments 44 4 7.94 73.51 79.03 7.96 74.6 81.5 1 74.46 79.47 8.60 88 8 75.70 77.83 83.78 4 70.08 78.6 84.1 1 7.56 79.09 85.55 1616 16 79.95 73.3 83.8 8 77.8 76.93 84.04 1 76.0 78.33 86.94 Experimental results representing in Tables and 3 show that motion estimation based on Hu moments is a non-accurate procudure for a rotation motion but it is faster then motion estimation based on Zernike moments. If a rotation is absent then BMA is a more appropriative method. The motion vectors become more accurate vectors if a slow motion and large block sizes occur in a video sequence. 8. Conclusion The correction of left and right video sequences is based on a transformation of frequency frames. The investigations demonstrate the necessity of developed methods for accurate motion estimations in complex scenes. A two-stage method of motion estimations based on BMA and Hu or Zernike invariant moments was

Margarita Favorskaya et al. / Procedia Computer Science ( 013 ) 110 1111 1111 proposed. As a result, the invariance to transitions and rotations of motion vectors set is achieved. The local motion vectors predict the smooth displacements of color components for each pixel into interpolated frames. The analysis of test video sequences show that Zernike moments are the most accurate method for object rotation (86 % accuracy). A method based on Hu moments is a robust approach for object rotation in comparison to BMA. Hu moments provide a less accuracy of detected motion vectors as compared with Zernike moments. Computation costs of all test video sequences are equaled because the algorithm analyzes the blocks with the same sizes. In future, an adaptive algorithm for blocks with variable sizes will be designed. Also the usage of pseudo Zernike moments will decrease a computational cost of the developed algorithm. References [1] Hu MK. Visual Pattern Recognition by Moment Invariant. IRE Trans. Info. Theory 196;IT 8:179 187. [] Revaud J, Lavoué G, Baskurt A Improving Zernike moments comparison for optimal similarity and rotation angle retrieval. IEEE Transactions on Pattern Analysis and Machine Intelligence 009;31(4):67 636. [3] Favorskaya M. Motion Estimation for Object Analysis and Detection in Videos. In: Kountchev R, Nakamatsu K, editors. Advances in reasoning-based image processing, analysis and intelligent systems: Conventional and intelligent paradigms, Springer-Verlag, Berlin Heidelberg; 01, p. 11 53. [4] Chen BJ, Shu HZ, Zhang H, Chen G, Toumoulin C, Dillenseger JL, Luo LM. Quaternion Zernike moments and their invariants for color image analysis and object recognition. Signal Processing 01;9():308 318. [5] Zhang H, Shu HZ, Han GN, Coatrieux G, Luo LM, Coatrieux JL. Blurred image recognition by Legendre moment invariants. IEEE Trans. Image Process 010;19(3):596 611. [6] Chen Z, Sun SK. A Zernike moment phase-based descriptor for local image representation and matching. IEEE Trans. Image Process 010;19:05-19. [7] Yang H.,, Pei, L., 011. Fast algorithm of subpixel edge detection based on Zernike moments. 4 th International Congress on Image and Signal Processing (CISP), p. 136 140. [8] Huang, Z., Leng, J., 010. Analysis of Hu s Moment Invariants on Image Scaling and Rotation. nd International Conference on Computer Engineering and Technology (ICCET), p. 476 480. [9] Sanjeev K, Haleh A, Mainak B, Truong N. Real-Time Affine Global Motion Estimation Using Phase Correlation and its Application for Digital Image Stabilization. IEEE Transactions on Image Processing 011;0(1):3406 3418. [10] Shutler JD, Nixon MS. Zernike velocity moments for sequence-based description of moving features. Image and Vision Computing 006;4:343 356. [11] Vyas VS, Rege PP. Geometric transform Invariant Texture Analysis based on Modified Zernike Moments. J. Fundamenta Informaticae 008;88(1 ):177 19. [1] Frejlichowski D. Application of Zernike Moments to the problem of General Shape Analysis. Control & Cybernetics 011;40():5 15. [13] Yang, Z.-L., Guo, B.-L., 008. Image Registration Using Feature Points Extraction and Pseudo-Zernike Moments. International Conference on Intelligent Information Hiding and Multimedia Signal Processing, p. 75 755. [14] Zhang H, Shu H, Coatrieux G, Zhu J, Wu QMJ, Zhang Y, Zhu H, Luo L. Affine Legendre Moment Invariants for Image Watermarking Robust to Geometric Distortions. IEEE Transactions on Image Processing 011;0(8):189 199. [15] Chandan S, Neerja M, Ekta W. Face recognition using Zernike and complex Zernike moment features. Pattern Recognition and Image Analysis 011;1(1):71 81. [16] Rao ChS, Kumar SS, Mohan BC. Content Based Image Retrieval using Exact Legendre Moments and Support Vector Machine. Int. J. of Multimedia & Its Applications 010;():69 79. [17] Wee CY, Paramesran R, Mukundan R. Fast computation of geometric moments using a symmetric kernel. Pattern Recognition 008;41:369 380. [18] Papakostas GA, Koulouriotis DE. A unified methodology for the efficient computation of discrete orthogonal image moments. Inform. Sci. 009;176:3619 3633. [19] Boveiri HR. On Pattern Classification Using Statistical Moments. Int. J. of Signal Processing, Image Processing and Pattern Recognition 010;3(4):15 4. [0] Video Test Media of Xiph.Org Foundation, Available online, at http://media.xiph.org/video/derf/ [1] Stereo-video content. Mobile 3DTV content delivery optimization over DVB-H system, Available online, at http://sp.cs.tut.fi/mobile3dtv/stereo-video/