The structure tensor in projective spaces

Similar documents
On the Equivariance of the Orientation and the Tensor Field Representation Klas Nordberg Hans Knutsson Gosta Granlund Computer Vision Laboratory, Depa

The GET Operator. LiTH-ISY-R-2633

Channel Representation of Colour Images

Using Local 3D Structure for Segmentation of Bone from Computer Tomography Images

A Probabilistic Definition of Intrinsic Dimensionality for Images

Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix

Algorithms for Computing a Planar Homography from Conics in Correspondence

Two-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix

Department of Mathematics, University of California, Berkeley. GRADUATE PRELIMINARY EXAMINATION, Part A Fall Semester 2014

Trinocular Geometry Revisited

Spectral Algorithms I. Slides based on Spectral Mesh Processing Siggraph 2010 course

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

EE Camera & Image Formation

Statistical Geometry Processing Winter Semester 2011/2012

Backprojection of Some Image Symmetries Based on a Local Orientation Description

Camera Models and Affine Multiple Views Geometry

Image enhancement. Why image enhancement? Why image enhancement? Why image enhancement? Example of artifacts caused by image encoding

Tangent spaces, normals and extrema

Linear Algebra Massoud Malek

The Multibody Trifocal Tensor: Motion Segmentation from 3 Perspective Views

Reconstruction from projections using Grassmann tensors

series. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques..

Review of Some Concepts from Linear Algebra: Part 2

Multiple View Geometry in Computer Vision

Algebra Workshops 10 and 11

Sequential Filter Trees for Ecient D 3D and 4D Orientation Estimation Mats ndersson Johan Wiklund Hans Knutsson omputer Vision Laboratory Linkoping Un

Spectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms

Seminar on Linear Algebra

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Dynamic P n to P n Alignment

Estimators for Orientation and Anisotropy in Digitized Images


A Factorization Method for 3D Multi-body Motion Estimation and Segmentation

GENERALIZED CAYLEY-CHOW COORDINATES AND COMPUTER VISION

Linear Algebra. Session 12

Example Linear Algebra Competency Test

Lecture 2: Linear Algebra Review

Parameterizing the Trifocal Tensor

Outline. Linear Algebra for Computer Vision

Representation and Learning of. Klas Nordberg Gosta Granlund Hans Knutsson

where m is the maximal ideal of O X,p. Note that m/m 2 is a vector space. Suppose that we are given a morphism

which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i

Chapter Two Elements of Linear Algebra

here, this space is in fact infinite-dimensional, so t σ ess. Exercise Let T B(H) be a self-adjoint operator on an infinitedimensional

Linear vector spaces and subspaces.

Multiple View Geometry in Computer Vision

Laplace-Beltrami Eigenfunctions for Deformation Invariant Shape Representation

Linear Algebra and Robot Modeling

Coding the Matrix Index - Version 0

Lecture 9: Low Rank Approximation

Review and problem list for Applied Math I

Feature Extraction and Image Processing

Two-View Multibody Structure from Motion

A geometric interpretation of the homogeneous coordinates is given in the following Figure.

Linear Algebra Review

Linear Algebra Practice Problems

7. Symmetric Matrices and Quadratic Forms

April 13, We now extend the structure of the horseshoe to more general kinds of invariant. x (v) λ n v.

Math 1553, Introduction to Linear Algebra

Linear Algebra- Final Exam Review

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Homogeneous Coordinates

A UNIFIED THEORY FOR STEERABLE AND QUADRATURE FILTERS

Some global properties of neural networks. L. Accardi and A. Aiello. Laboratorio di Cibernetica del C.N.R., Arco Felice (Na), Italy

2.3. VECTOR SPACES 25

2. Every linear system with the same number of equations as unknowns has a unique solution.

Final Exam Due on Sunday 05/06

L26: Advanced dimensionality reduction

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

arxiv: v1 [math.ag] 25 Dec 2015

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Functional Analysis I

From Wikipedia, the free encyclopedia

Inverse differential kinematics Statics and force transformations

II. DIFFERENTIABLE MANIFOLDS. Washington Mio CENTER FOR APPLIED VISION AND IMAGING SCIENCES

ALG - Algebra

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

MAT Linear Algebra Collection of sample exams

Reproducing Kernel Hilbert Spaces

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Robust Motion Segmentation by Spectral Clustering

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

12.2 Dimensionality Reduction

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Title Project Summary

Online Kernel PCA with Entropic Matrix Updates

SURF Features. Jacky Baltes Dept. of Computer Science University of Manitoba WWW:

CALCULUS ON MANIFOLDS. 1. Riemannian manifolds Recall that for any smooth manifold M, dim M = n, the union T M =

On John type ellipsoids

EE Technion, Spring then. can be isometrically embedded into can be realized as a Gram matrix of rank, Properties:

(v, w) = arccos( < v, w >

Confidence and curvature estimation of curvilinear structures in 3-D

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem

Tensors. Notes by Mateusz Michalek and Bernd Sturmfels for the lecture on June 5, 2018, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra

Analysis Preliminary Exam Workshop: Hilbert Spaces

Transcription:

The structure tensor in projective spaces Klas Nordberg Computer Vision Laboratory Department of Electrical Engineering Linköping University Sweden Abstract The structure tensor has been used mainly for representation of local orientation in spaces of arbitrary dimensions, where the eigenvectors represent the orientation and the corresponding eigenvalues indicate the type of structure which is represented. Apart from being local, the structure tensor may be referred to as object centered since it describes the corresponding structure relative to a local reference system. This paper proposes that the basic properties of the structure tensor can be extended to a tensor defined in a projective space rather than in a local Euclidean space. The result, the projective tensor, is symmetric in the same way as the structure tensor, and also uses the eigensystem to carry the relevant information. However, instead of orientation, the projective tensor represents geometrical primitives such as points, lines, and planes (depending on dimensionality of the underlying space). Furthermore, this representation has the useful property of mapping the operation of forming the affine hull of points and lines to tensor summation, e.g., the sum of two projective tensors which represent two points amounts to a projective tensor that represent the line which passes through the two points, etc. The projective tensor may be referred to as view centered since each tensor, which still may be defined on a local scale, represents a geometric primitive relative to a global image based reference system. This implies that two such tensors may be combined, e.g., using summation, in a meaningful way over large regions. 1. Introduction The structure tensor was introduced at the end of the 80 s for the representation of local orientation. One of main advantages of the structure tensor is that it provides a consistent representation of both local orientation and type of structure in spaces of arbitrary dimensionality. In practice, the structure tensor can be represented by a symmetric n n matrix for n-dimensional signal. For simples signals, s(x) =g(x T ^n); (1) where x; ^n 2 R n, the corresponding structure tensor is T = A ^n ^n T ; (2) where A>0. This indicates that orientation of the signal, represented by the normal vector ±^n, is given by any normalized eigenvector of T that has a non-zero eigenvalue. In the Fourier domain, the simple signal is concentrated to a one-dimensional subspace parallel to ^n, and consequently the eigenvectors of T with non-zero eigenvalues span this subspace. This can be generalized to arbitrary signals; if the Fourier transform of the local signal is confined to a subspace, then the eigenvectors of T with non-zero eigenvalues span this subspace. In practice, non-zero must be interpreted as relatively large compared to other eigenvalues that vanishes in a first order approximation. It follows directly from the previous discussion that it is the eigensystem of the structure tensor which carries the orientation representation. The eigenvalues reflect the type of structure, e.g., a simple signal corresponds to a single non-zero eigenvalue, and the eigenvectors represent the orientation of that structure. Hence, a rotation of the structure should always be reflected in the same rotation of the tensor s eigenvectors. Over the years several methods for approximating the structure tensor have been presented. A method by Bigün [1] is based on estimating the inertia matrix of the signal s energy in the Fourier domain. Knutsson [5] [4] uses a set of carefully designed quadrature filters in obtain better phase invariance. Farnebäck [2] makes a polynomial approximation of the local signal and use separable polynomial filters in combination with normalized convolution to get lower computational complexity. The structure tensor suffers from one weakness, it is object centered. It represents the orientation of a structure relative to a reference system attached to the structure itself, which means that two identical structures at different global positions are represented by identical tensors. This 1

is not a problem as long a the tensor is being used for local operations, e.g., averaging. But even at the local scale, averaging may be problematic, e.g., at a sharp corner where two distinct orientations should be represented. In this case, the local tensor average will, at best, indicate that there is more than one dominant orientation by having two non-zero eigenvalues. In general however, the corresponding pair of eigenvectors will not represent the two distinct orientation. The summation of structure tensors over some region will either indicate a consistent orientation by having a single non-zero eigenvalue or the presence of two or more orientations by having two or more non-zero eigenvalues, and only in the first case will the eigenvectors carry relevant orientation information. 2. Projective spaces A projective space can be thought of as a vector space where the primary elements are equivalence classes of vectors rather than vectors, and where two vectors are considered equivalent if they differ by a non-zero multiplicative constant (the zero vector is typically not considered as an element of a projective space). For example, if we start with the vector space R n and then consider all one-dimensional linear subspaces, corresponding to the above defined equivalence classes, the subspaces define a projective space of dimension n 1, denoted P n 1.Thisimpliesthatanyvector space R n can be embedded in a projective space P n+1. A canonical (but not unique) way to do this is to map any vector x 2 R n to a vector x H 2 R n+1 as follows x H = x 1 : (3) However, x H represents also an element in P n, the onedimensional subspace spanned by x H. This element is unique in the sense that each x 2 R n maps to a distinct element of P n. However, not all elements of P n represent an element of R n. Any element y H 2 R n+1 of the form y H = y 0 cannot be proportional to an x H defined in Equation (3), and therefore does have a corresponding x 2 R n. This can be overcome by having these y H 2 P n represent points at infinity in the direction of y, a construction which proves useful for certain applications. Projective spaces have applications is many fields, and one if their main advantage is that they allow affine and bilinear transformations on R n to be expressed as linear transformations on P n, which in most cases simplifies the mathematics considerably and makes the mathematical expressions more compact. A typical example is the equations related to projecting points in a 3D scene onto the camera (4) image in the standard pin-hole camera model. If standard coordinates are being used, the mapping from scene coordinates in R 3 to image coordinates in R 2 is expresses by means of a bilinear transformation. By mapping both the 3D coordinates and the 2D coordinates, to P 3 and P 2,respectively, by means of Equation (3) (often referred to as using homogeneous coordinates), the bilinear transformation can be written as a linear mapping from P 3 to P 2. 3. Tensors in projective spaces The idea of using tensors on projective spaces is not new. In fact, one of the more prominent examples is related to the projective transformation from a 3D point in a scene to a 2D image point. If two cameras at different locations are being used, the mapping from P 3 to P 2 will be different for the two camera images, and result in two points x 1 ; x 2 2 P 2 which, in general, are different. However, as was shown in [3], x 1 and x 2 must always satisfy the following equation x T 1 Fx 2 =0; (5) where F is the so-called fundamental matrix or tensor (3 3) that is dependent only on the geometrical relation between the two cameras and their internal parameters. The tensor nature of F is reflected by the fact that for any affine transformation of the 3D space, there is a corresponding linear transformation that transforms the tensor. Extensions has also been made to define a similar tensor for the case of corresponding points in three images, the so-called trilinear tensor [6]. It should be mentioned that the above mentioned tensors on projective spaces are rather different from the structure tensor. The fundamental tensor and its relatives are not symmetric which means that an analysis of its eigensystem has not been considered relevant (though singular value decomposition is). Furthermore, the estimation procedures of the structure tensor mentioned above are all stable from a numerical point of view. The fundamental tensor, on the other hand, can only be robustly estimated under certain conditions which may not apply in general. What we will do next, is to see how the properties of the structure tensor can be combined with the simple mathematics of projective spaces. 4. Symmetric tensors on projective spaces A symmetric tensor, in practice represented by a symmetric matrix, has a well-defined eigensystem in terms of a set of eigenvalues, given by the roots of the corresponding characteristic polynomial, and for each distinct eigenvalue there is a linear subspace of corresponding eigenvectors. The dimensionality of the subspace is the same as the multiplicity 2

of the eigenvalue as root of the characteristic polynomial. This means that any eigenvector e may in fact be identified with, at least, all vectors which are parallel to it since they will also be eigenvectors and have the same eigenvalue as e. This observation may be taken as a first indication that the eigensystems of symmetric tensors have a relation to projective spaces. In the case of symmetric tensors we know that two eigenvectors that have different eigenvalues are always orthogonal, a direct consequence of the spectral theorem. In particular this means that if we, somehow, can divide the eigenvalues into non-zero (or large) and zero (or small) eigenvalues, the eigenvectors corresponding to the non-zero eigenvalues are always orthogonal to the eigenvectors with zero eigenvalues. If we restrict ourself to only positive semi-definite tensors, i.e., to tensors which have eigenvalues equal to or larger than zero, the operation of summation (using non-negative weights if desired) is closed on this set. Furthermore, if T s is the summation of a set of positive semi-definite tensors ft k g,theneis an eigenvector of zero eigenvalue relative T s if and only if this is true also for all T k. Furthermore, since the elements of the projective space P n are linear subspaces of R n+1, the concept of orthogonality of vectors in R n+1 is transfered directly to orthogonality between elements of P n. 5. The projective tensor The idea that will be explored in the following is based embedding a vector space R n in a projective space P n,using Equation (3). Geometrical relations in R n are then introduced which can be written as orthogonality relations in P n. These relations are then encoded in a symmetric tensor defined on P n, such that the two orthogonal vectors fall into different classes of eigenvectors, either with non-zero eigenvalue or zero eigenvalue. Only the case n = 3 will be discussed, since this is the simples that has most applications, but the results can be generalized to spaces of arbitrary dimension. Also, the degenerate cases related to the interpretation of the projective tensor when non-zero eigenvectors of the type described in Equation (4) appear will not be discussed here. Consider first a point x 2 R 3. Using homogeneous coordinates, Equation (3), this point can be mapped to x H 2 R 4. A suitable tensor representation of this point is the rank one tensor T point = A x H x T H ; (6) where A > 0. Note that we may consider x H as an element of P 3, and take any vector in R 4 parallel to x H instead of x H in Equation (6). Alternatively, any eigenvector of T point with non-zero eigenvalue represent the same element of P 3, which in turn is a representation of x. Let us now consider a plane in R 3 that includes x. Given a normal vector ^n of the plane, and its distance d to the origin in the direction of ^n, the fact that x lies in the plane can be written as (d ^n x) T ^n =0: (7) This relation can also be written as an orthogonality relation between two vectors in R 4, where x H is given by Equation (3), and x T H p =0; (8) p = ^n d : (9) According to the previous discussion the orthogonality relation applies also to x H and p as elements of P 3, and consequently p 2 P 3 can be used as a representation of the plane. Furthermore, it follows also that p must be an eigenvector of T point with eigenvalue zero. In fact, since T point is represented by a 4 4 matrix, and has a single non-zero eigenvalue, there is a three-dimensional subspace of R 4 that contains eigenvectors of T point with eigenvalue zero. This corresponds to a two-dimensional projective subspace that contains representations of all planes that include the point x. This observation corresponds to the well-known duality principle which implies, e.g., that a point can be represented by the set of all planes which includes the point, or that a plane can be represented by the set of points which it includes. Let us now consider the summation of two point tensors T 1 and T 2, as defined in Equation (6), which represent two distinct points x 1 and x 2. For example, let us define, T line = T 1 + T 2 : (10) This tensor is of rank two, and from the discussion above it follows that the eigenvectors that have eigenvalue zero relative to T line is exactly the set of vectors which are eigenvectors of eigenvalue zero relative to both T 1 and T 2.This set of eigenvectors must then be the representations of all planes which includes both x 1 and x 2. The duality principle allows us to identify this set with the unique line that includes both x 1 and x 2. However, any summation of any set of point tensors that represent points which lie on the same line will result in a tensor T line of rank two which has the same set of eigenvectors with eigenvalue zero. Consequently, any rank two tensor can be identified with a line in R 3 either by having the two eigenvectors of nonzero eigenvalue represent two distinct points in R 3 that uniquely defines the line, or, which is equivalent, by having 3

the set of eigenvectors which have eigenvalue zero represent a set of planes which all intersect at the line. Finally, consider the summation of three point tensors and apply the same arguments again. For example, for T 1, T 2 and T 3 that represent the points x 1 ; x 2 ; x 3 2 R 3, not lying on a line, define T plane = T 1 + T 2 + T 3 : (11) This tensor is of rank three and any eigenvector of eigenvalue zero relative to T plane must also be an eigenvector of eigenvalue zero relative to T 1, T 2 and T 3. These eigenvectors must be the representations of all planes which include all three points, which in fact must be a unique plane. However, any summation of any set of point tensors that represent points which lie on the same plane will result in a tensor T plane of rank three which has the same set of eigenvectors with eigenvalue zero. This includes summation of point and line tensors as long as the lines lie in the plane too. Consequently, a rank three tensor can be identified with a plane in R 3 either by having the three eigenvectors of nonzero eigenvalue represent three distinct points in R 3 that uniquely defines the plane, or, which is equivalent, by considering the set of eigenvectors which have eigenvalue zero as a single element of P 3 that represent the plane. The summation of point tensors which represent four or more points that do not lie in a plane always results in a tensor of full rank. This tensor has no eigenvector of eigenvalue zero which is a natural consequence of the fact that there is no plane which includes all these points. Given a set of points P = fx k 2 R 3 g, the corresponding affine hull AfP g is defined as AfP g = f P k ff k x k ; x k 2 P; P k ff k =1g: (12) AfP g is the smallest affine space that contains all points in P. It follows directly that, in this case, AfP g may also be defined as the intersection of all planes that contain all points in P. To summaries this section, we may conclude that given a set of points P, where each point x k is represented by a point tensor T k, defined in Equation (6), then the operation of taking a summation of the tensors T k results in a tensor T s that is of rank one, two, three, or four. However, in all four cases, the set eigenvectors of T s that have eigenvalue zero (which may be an empty set) represents the intersection of all planes which includes all points in P, and this is the same as the affine hull of P. The tensors described in this section, of different ranks for the representation of different geometric primitives, are all referred to as projective tensors. 6. Applications of the projective tensor An immediate application of the projective tensor is the representation of geometrical primitives in a 3D scene or in 3D data that somehow has been registered by an image device, e.g., a camera or a CT scanner. By first obtaining local estimates of the projective tensor, either in the form of point tensors (rank one), or of arbitrary ranks, these can then be summed in a following step to form robust descriptors of linear or planar structures together with their absolute positions. This can be done over local neighborhoods as well as larger regions since there are no restrictions on the region imposed for the summation. Furthermore, the projective tensor may be used to control region formation (or segmentation) by providing evidence that each region contains a consistent structure (e.g. line or plane). In general, the projective tensor allows us to robustly represent, e.g., a line or plane structure in 3D data or a 3D scene. The structure may not have to be represented by a dense set of points, and can even be partially occluded. As long as the summation is made over a region where all points belong to the same structure, the result will be useful. 7. Estimation of the projective tensor We next turn our focus on how the projective tensor can be estimated. A direct consequence of the results derived in Section 5 is that one way is to start by estimating projective tensors of rank one which represent various points of interest. Then the summation of these over some region will then result in projective tensors that represent lines or planes, or possibly the case that the corresponding points do not lie in a plane at all. However, this is not the only approach that is possible, and we may for example want to have the initial estimation of projective tensors to result in tensors of higher rank than one. It should be noted that each distinct geometric primitive is represented by equivalence classes of projective tensors of a particular rank that share a common null space. This makes the design of the estimation process rather flexible since there is no unique projective tensor for each geometric primitive. Two different approaches for estimating projective tensors are presented below. 7.1. Starting with structure tensors The 3D structure tensor, described in Section 1, carries information about the orientation of geometric primitive, together with the type of primitive. The only additional information needed to obtain a corresponding projective tensor is the absolute position of the corresponding local structure, which is implicitly given by where in a global, camera or sensor centered, coordinate system the tensor is located. For 4

each tensor, we can do an analysis of the tensor s eigenvalues in order to determine which of the cases plane, line, and isotropic we have. Based on this analysis, an appropriate projective tensor can be assigned to the image position based on the eigenvectors of the structure tensor and the position. 7.2. Base-line or motion stereo Using standard methods for estimating the depth of image points, typically using base-line or motion stereo, we can obtain projective tensors of point type at the image points for which the depth can be estimated with sufficient certainty. The certainty can, for example, be represented by the non-zero eigenvalue A in Equation (6), which implies that sum of these tensors result in a meaningful projective tensor. 8. Summary A projective tensor defined on a projective space has been presented which can represent, in the 3D case, points, lines, andplanesin termsof a 4 4 symmetric matrix. Compared to the usual structure tensor, the projective tensor carries information not only about orientation and type of structure, but also about the structure s absolute position. Projective tensors of rank one represent points, rank two tensors represent lines, and rank three tensors represent planes in R 3. Furthermore, it has been shown that the representation of these geometrical primitives is consistent with the operation of adding the projective tensors (possibly using non-negative weights) in the sense that summation of point tensors that all lie on a line or a plane amount to the corresponding line or plane tensor. The projective tensor can be used for the representation of structures, e.g., in three-dimensional image data or scenes, when integration of evidence for a particular structure has to be made over a larger region. It can also be used for image segmentation by indicating constistency of local tensor data relative to a particular structure in a growing region. Estmation of the 3D projective tensor can be done either by starting with the usual 3D structure tensor and build a projective tensor from that, or by estimating 3D image or scene points, e.g., using base-line or motion stereo, and then integrate the corresponding point tensors over suitable regions. IEEE First International Conference on Computer Vision, pages 433 438, London, Great Britain, June 1987. [2] G. Farnebäck. Spatial Domain Methods for Orientation and Velocity Estimation. Lic. Thesis LiU-Tek- Lic-1999:13, Dept. EE, Linköping University, SE-581 83 Linköping, Sweden, March 1999. Thesis No. 755, ISBN 91-7219-441-3. [3] O. Faugeras. What can be seen in three dimensions from an uncalibrated stereo rig? In Proceedings of the 2nd European Conference on Computer Vision, Santa Margherita Ligure, Italy, 1992, 1992. [4] G. H. Granlund and H. Knutsson. Signal Processing for Computer Vision. Kluwer Academic Publishers, 1995. ISBN 0-7923-9530-1. [5] H. Knutsson. Representing local structure using tensors. In Proceedings of the 6th Scandinavian Conference on Image Analysis, pages 244 251, Oulu, Finland, June 1989. Report LiTH-ISY-I-1019, Computer Vision Laboratory, Linköping University, Sweden, 1989. [6] A. Shashua and M. Werman. On the trilinear tensor of three perspective views and its underlying geomtry. In International Conference on Computer Vision, 1995, 1995. References [1] J. Bigün and G. H. Granlund. Optimal Orientation Detection of Linear Symmetry. In Proceedings of the 5