Robust Detection, Classification and Positioning of Traffic Signs from Street-Level Panoramic Images for Inventory Purposes

Size: px
Start display at page:

Download "Robust Detection, Classification and Positioning of Traffic Signs from Street-Level Panoramic Images for Inventory Purposes"

Transcription

1 Robust Detection, Classification and Positioning of Traffic Signs from Street-Level Panoramic Images for Inventory Purposes Lykele Hazelhoff and Ivo Creusen CycloMedia Technology B.V. Achterweg 38, 4181 AE Waardenburg, The Netherlands Peter H.N. de With Eindhoven University of Technology Den Dolech 2, 5600 MB Eindhoven, The Netherlands Abstract Accurate inventories of traffic signs are required for road maintenance and increase of the road safety. These inventories can be performed efficiently based on street-level panoramic images. However, this is a challenging problem, as these images are captured under a wide range of weather conditions. Besides this, occlusions and sign deformations occur and many sign look-a-like objects exist. Our approach is based on detecting present signs in panoramic images, both to derive a classification code and to combine multiple detections into an accurate position of the signs. It starts with detecting the present signs in each panoramic image. Then, all detections are classified to obtain the specific sign type, where also false detections are identified. Afterwards, detections from multiple images are combined to calculate the sign positions. The performance of this approach is extensively evaluated in a large, geographical region, where over 85% of the 3, 341 signs are automatically localized, with only 3.2% false detections. As nearly all missed signs are detected in at least a single image, only very limited manual interactions have to be supplied to safeguard the performance for highly accurate inventories. 1. Introduction Nowadays, several companies record street-level panoramic images, which provide a recent and accurate overview of the road infrastructure. Within The Netherlands, these images are captured by private companies (e.g. CycloMedia Technology and Google), where each public road is recaptured annually. The resulting image databases enable efficient inventories of street furniture to support maintenance and cost control. Computer vision techniques facilitate the automatic creation of such inventories and thereby reduce human interaction compared to manual inventories, where all objects are searched and annotated by hand. These inventories are of interest to governmental organizations tasked with road maintenance. Especially traffic signs are of interest, as their presence directly influences road safety. They require accurate and up-to-date inventories, as the sign visibility may be degraded due to e.g. vandalism, vegetation coverage, aging and accidents. This paper describes a framework for road-sign inventories based on computer vision techniques, aiming at retrieving the sign code and position of all traffic signs in a region. Although traffic signs are designed to attract visual attention, automatic detection and classification of road signs is a complicated problem for several reasons. The first is related to capturing from a driving vehicle. As the signs are captured from a wide range of distances, large viewpoint deviations exist, and signs may be occluded by e.g. other road users. Furthermore, capturing outside implies varying weather conditions, including e.g. fog. The second complication comes from the sign features. Signs vary in size, and there are many similar traffic signs, which are sometimes custom versions of official signs, and some signs are designed to contain custom text or symbols. Moreover, the visibility of traffic signs may be lowered due to aforementioned reasons, while especially these signs are of importance for sign maintenance. Thirdly, many sign look-a-like objects exist, including directives for customer parking or restrictions for dog access, which are not traffic signs. Examples of these complicating factors are displayed in Fig Related work In literature, detection and recognition of traffic signs is studied for many years. For example, [15] describes a cascade detection of speed signs within a large scale project, achieving a detection rate of 98.7%. However, this paper only addresses detection of a single sign type. This is also the case for [1], where the image is prefiltered with a color-version of the Viola-Jones algorithm, followed by analysis of Histogram of Oriented Gradient fea- 313

2 (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) (k) (l) (m) (n) (o) Figure 1. Examples of factors complicating detection and classification of traffic signs. (a)-(e): occlusions; (f)-(i): lowered sign visibility. (j)-(m): sign-like objects. (n)-(o): official sign with custom version. tures with a neural network. The authors test their system for triangular warning signs, and report that 260 of the 265 present signs are correctly identified, where 241 signs are correctly classified. Examples of systems focusing on multiple sign types are e.g. [12], where images captured from a car-mounted camera are exploited for sign detection. After threshold-based color-segmentation and shape analysis, the sign type is recognized based upon a grayscale version of the located blobs. The authors report that all the 104 signs are detected at least twice. Color segmentation is also exploited for extraction of sign regions in [13]. Afterwards, the shape of the sign in each blob is extracted, which is subject to classification, based on Support Vector Machines, exploiting the known sign shape. It is reported that 98 of the 102 signs are detected at least once. These proposals detect and classify traffic signs based on single images. A tracking system is proposed in [10], reducing the false alarms by tracking the signs over the frames. The authors report that all 6 signs are detected correctly. Another approach is described by Timofte et al. [16], where a van with 8 cameras is employed for capturing. Their method employs both single-image and multi-view analysis. A fraction of 95.6% of the 269 signs are positioned correctly, where 97.7% of the detected signs are also successfully recognized Our approach This paper presents an approach for performing largescale inventories of traffic signs based on computer vision techniques, where signs are detected, classified and positioned. This is a very challenging problem, and we have experienced that the performance of state-of-the-art algorithms for detection and classification are insufficient for a fully automated inventory in real-world conditions and at a large scale, which is similar as Frome et al. [6] discuss for face detection. Therefore, we aim at a system focusing on both automatic and semi-automatic inventories. This paper describes the automatic version, where we concentrate on seven different sign classes, covering 92 different sign types. The involved classes are displayed in Fig. 2. Instead of constructing a custom capturing device, as some proposals in literature, we exploit the already existing street-level panoramic images. These images are captured on all public roads with a calibrated recording system with a capturing interval of 5 meters. The capturing cars are typically employed efficiently, resulting in images captured in a very wide range of weather conditions, including even fog, which makes the problem even more challenging. Sign appearances vary greatly across the images due to the large variation in weather conditions and due to differences between different camera systems. Therefore, instead of focusing at color, which is common in literature, we investigate color gradients, since they are more robust against these situations. Next to this, we aim at a generic, learningbased system, as this allows adaption to other sign appearances, e.g. in other countries. Our system consists of three stages: sign detection, sign classification and sign positioning. At first, the signs are grouped into classes, e.g. red triangular, and their generic properties are exploited for detection. Due to the genericity of this stage, customized versions of standard signs can also be found, which is beneficial in the semi-automatic approach. For detection, a custom variant of the popular Histogram of Oriented Gradients [5] is applied, which operates with color gradient information to exploit the characteristic sign colors. Then, the minor differences between the signs are analyzed, and all detections are classified using a variant of the popular Bag of Words (BoW) technique [4]. The standard BoW approach is modified to both filter out falsely detected signs and to deal with the large intra-class similarities of the signs. Afterwards, in the sign positioning stage, the sign positions are calculated by combining the detections across multiple images. The performance of this inventory system is evaluated by a large-scale experiment, where an inventory is applied to a large geographical region, containing over 3, 340 traffic signs. We should note that this validation size is rather uncommon in related literature, as is the fact that we also take signs not directly located along the road into account. Besides this, the performance of the individual detection and classification stages are also assessed. The remainder of the paper is organized as follows. Section 2 contains the system overview. Section 3 describes the sign detection stage, Section 4 describes the classification stage, followed by the positioning procedure in Section 5. The performance evaluation can be found in Section 6, followed by the conclusions in Section

3 sign detection sign classification sign detection sign classification red triangular signs red triangular signs give-way signs red circular signs red circular signs Overlap analysis and correction redblue circular signs redblue circular signs no-entry signs blue circular signs blue circular signs yellow diamond signs yellow diamond signs sign detection sign positioning sign classification Figure 2. System overview of our inventory system. 2. System overview is combined with all nearby images. Next, hypotheses of sign positions are obtained by pair wise combinations of detections with both an identical sign code, where combinations of sign codes and wildcards give supporting evidence. Then, the final sign positions are obtained by clustering of these hypotheses. The system overview of our inventory process is depicted in Fig. 2. The system consists of three primary modules, which are briefly described below. 1. Sign detection: At first, each panoramic image is analyzed and present signs are detected by multiple, independent detectors, each focusing at a specific class of signs. These detectors are kept very generic to allow detection of distorted signs and sign-like objects. As some detectors focus at quite similar sign classes, their output may overlap. These overlapping samples are analyzed and a detection fusion step is applied to assign the correct class label. 3. Sign detection The first stage of our inventory system consists of localizing traffic signs within the individual panoramic images. As many traffic signs have similar color and shape (such as e.g. all blue circular direction signs), similar signs are grouped into sign classes (such as blue circular) and detection is performed for each class independently. The class division is displayed in Fig. 2. Since traffic signs are intended to attract attention based on their colors and shape, many traffic sign detection systems in literature start by color filtering of the image, and extracting regions with colors corresponding to the signs. However, we have found that the color and contrast of the signs varies significantly with the capturing conditions, and therefore instead we exploit color differences and shape information, which we have found to be more consistent over the varying circumstances. We apply detectors based on the popular Histogram of Oriented Gradients (HOG) algorithm, originally proposed by Dalal and Triggs [5]. As the standard algorithm extracts the maximum gradient over the color channels for each pixel, it neglects the correlation of gradients over the color channels, and thereby neglects the discriminative color of the traffic signs. We have extended the standard HOG approach with the use of color information, as described in [3]. 2. Sign classification: During the sign detection stage, all detected signs are assigned a sign class label, e.g. red triangular. Next, each detection is assigned a sign code, such as warning sign for a dangerous crossing. As some detections have an insufficient resolution for classification, these small samples are not classified, but assigned a class-specific wildcard. Furthermore, at this stage, the false findings given by the detectors are identified by inclusion of an additional class to the classification procedure, representing the false detections of the respective sign class. These detections are also assigned the wildcard code. Classes with only a single contained sign are not subject to classification. 3. Sign positioning: In the sign positioning stage, detections from multiple images are combined to calculate the position of the traffic signs based on the geometric properties of our source data. For this, each image 315

4 The modified HOG algorithm works as follows. First, the image is divided into cells of 8 8 pixels, where for each of these cells a histogram of the gradient orientation is calculated. These histograms are normalized w.r.t. adjacent histograms. Next, a sliding window is moved over the cells, covering 5 5 cells, and all included histograms are concatenated. As we perform detection on color images, the histograms of all the color channels are appended. The resulting 1, 200-dimensional feature vector is used for classification by means of a linear Support Vector Machine (SVM). Since multiple classes of signs are detected independently from each other, the same features are exploited for all classes, where each class is found by an individual SVM. As the feature extraction stage is the most timeconsuming task, adding additional classes does not affect processing time significantly. This procedure is repeated at multiple scales to obtain scale invariance, resulting in detections with a size ranging from pixels up to pixels, corresponding to a typical sign-to-car distance ranging from about 19 to 1 meters, respectively. Although the different detectors operate independently, cross-overlap may exist between the detectors for the different classes, especially when the visual difference is low. This causes signs to be detected by multiple detectors, which is especially the case for the red circular and redblue circular signs, which are both circular with a red border. Therefore, we employ a specific step to distinguish between these two classes. Each sample detected by both detectors is analyzed and assigned a single class label. Since we strive for a generic system, we apply a learning-based approach, exploiting the differences in color distribution between the sign classes. This method first transforms the input samples into the HSV color space, then extracts a color histogram from the signs, followed by classification based on a linear SVM. Afterwards, each sample is assigned the appropriate class label, which is exploited during the classification. 4. Sign classification Each detector focuses at detecting a specific sign class, which typically consists of more than one sign type. Therefore, the detector output is analyzed to obtain the sign code, e.g. danger, crossing children, where detectors that directly locate a sign type are not subject to classification. Whereas the detection stage exploits the generic characteristics of a sign class, such as the sign borders, the classification stage should discriminate between signs based upon the very minor differences within the inner template, as visualized in Fig. 4. This complicates the classification task, especially since the resolution of the discriminative part of the signs is quite low. Therefore, we ignore samples with insufficient resolution and assign them a wildcard code (its use is explained later), and only classify samples larger than pixels. The employed classification approach is based on (a) (b) (c) (d) Figure 4. Examples of signs with only very minor differences. Figure 5. Visualization of the modular codebook, containing the concatenation of individual visual dictionaries. Bag of Words [4] (BoW), and is described in [7]. For completion, we will briefly redescribe the key features of the system below. The original BoW approach [4] represents each image by a histogram containing the occurrence frequency of the elements of the visual dictionary. These elements, called visual words, are obtained by clustering the features extracted from training samples, in order to obtain features that occur frequently in all training samples. However, these words may not be the most discriminative [8], which will especially be the case when words occur in all sign types, e.g. representing the sign borders. Therefore, we altered the construction of the visual dictionary, where we generate a separate dictionary for each sign type to ensure that characteristics common for the sign types are extracted. The individual dictionaries are concatenated to a modular codebook. This approach also enables handling of unbalanced training data and allows for easy addition of extra classes without recomputing the complete codebook. As our sign detector also outputs false detections, we have extended the modular codebook with a large dictionary representing them. This prevents false detections from being assigned to a random class, and instead, have those mapped to the false detection class. The resulting modular codebook is visualized in Fig. 5 and is used, as would it be the regular codebook in the standard BoW approach. Based on this modular codebook, new samples are classified as follows. First, the input sample is normalized to a standard size. Since the sign class is known, irrelevant parts are removed based on a standard shape, similar to [13]. As the pattern is more discriminative than the color, each sample is converted to a single color image, where for almost all classes grayscale is used. However, the redblue circular signs are converted to a custom color space, given by the difference between the red and blue color channels, as the red rim and blue inner part have similar grayscale values. Then, SIFT features [11] are extracted using dense sampling, which is shown to outperform the use of interest point operators [14]. We have applied SIFT, as these features are invariant against the commonly occurring ro- 316

5 fn S Above car, fh = ½p Horizon, fh = 0 p Below car, fh = -½p -½p E ½p N W fh S -p Figure 3. Example panoramic image including a visualization of the geometric properties of the panorama, where ϕn denotes the angle w.r.t. the northern direction and ϕh the angle w.r.t. the horizon. tation and illumination changes. Each extracted feature is matched against the modular codebook, and a word histogram is constructed, which is used for classification based on linear SVMs in a One-vs-All fashion. When none or multiple classifiers recognize the sample, we consider it as unreliable, and assign it a wildcard code. this, we exploit the fact that our source data is geometrically correct, i.e. the angular orientations are linear with the pixel coordinates. This is achieved by extensive calibration of the capturing system. Furthermore, each panoramic image is divided into two equal parts by both the northern direction and the horizon, as visualized in Fig. 3. As a result, when two points corresponding to the same object are known in two images, the position of that object can be calculated by straightforward geometric calculations. The sign positions are retrieved per sign class, where detections from individual classes are combined with other detections, either having an identical sign code or a wildcard code. Although the detections themselves can be correlated, multiple identical signs can result in wrong correspondences. Furthermore, the capturing interval of 5 meters causes large differences in perspective, scale and background, thereby complicating pixel correspondences. Therefore, we only exploit the geometric information, where all the combined pairs give a position estimation, which are passed through when this estimate is closer than 45 meter w.r.t. both images. Each estimate gives a hypothesis of the sign location, which are clustered around the real sign position, as visualized in Fig. 6. These clusters are recovered using the MeanShift algorithm [2]. Afterwards, the clusters are processed from large to small cardinality, where only clusters containing at least 3 or more detections are taken into account. For each cluster, the position estimates given by the combination of all contained detections should be close to the cluster center, and when not valid, the detection with the largest mean position deviation w.r.t. this center is removed. The same rule applies when multiple detections from the same image are present. After a cluster is accepted, all contained detections are removed, and the procedure is repeated. 5. Sign positioning After the classification stage, all detections are assigned a sign code, either corresponding to a specific sign type or to a class-specific wildcard. Next, the positions of the signs are calculated by combining the detections corresponding to the same physical sign appearing in multiple images. For XXX X X XX X XX X X X Figure 6. Visualization of the sign positioning process, where two identical give-way signs are present (indicated by the give-way symbols). The positions of the capturings are displayed by blue circles, the position estimates given by the pair-wise combination of all detections are drawn as red crosses and the actual position of the traffic signs are indicated by the two sign symbols. 317

6 Precision redblue circle red circle red triangular 0.2 yellow diamond give way blue circle Recall Figure 7. Recall-precision curves for the individual traffic sign detectors. The classes indicated correspond to the classes in Fig. 2. Precision redblue circle with detection fusion 0.2 redblue circle without detection fusion red circle with detection fusion red circle without detection fusion Recall Figure 8. Recall-precision curves showing the effect of the detection fusion stage for the red and redblue circular signs. 6. Experiments and results The described system is employed to perform inventories of multiple geographical regions, containing both rural and city environments. The output is manually verified, and the resulting ground-truth data is employed to assess the performance of the detection and classification stages of the system. Furthermore, the performance of the complete inventory system is evaluated for a representative geographical area, containing multiple towns within a rural surrounding Detection performance analysis The performance of the sign-detection module is evaluated on about 33, 000 panoramic images, each containing at least a single traffic sign. This set covers both rural and city environments, and all present signs are manually annotated. However, it should be noted that the ground truth was constructed subjectively, so that there is some mismatch for the smaller signs, which may be detected but not annotated. The set is processed within 8.5 hours on a cluster of 4 computers and the resulting recall-precision curves for the different sign classes are shown in Fig. 7. As follows, the detectors are able to localize over 90% of the traffic signs in the individual panoramic images. This figure includes the detection fusion step for the red circular and redblue circular signs. The effect of this step is displayed in Fig. 8, which clearly shows that the additional fusion stage significantly reduces the number of false detections, especially for the redblue circular signs. This indicates that there is indeed a large overlap between the classes. As the percentage of detected signs is not significantly affected, our approach successfully assigns the correct class label in most cases Classification performance analysis The performance of our classification module is evaluated using the detector output on a large set of panoramic images, which is constructed such that a maximum number of different sign types are included with sufficient cardinality. The minimum size of the considered detections is pixels, where occurring false detections are included. The classification performance is analyzed by using 10-fold cross-validation, as this approaches the system performance when all samples have been subject to training [9]. Since our test set contains multiple samples of the same physical sign, captured from different viewpoints, these occurrences are forced into the same partition to prevent testing with the training set. Table 1 summarizes the key performance numbers for the classes of interest. Classification of an unseen sample takes about 1 second. It can be concluded that our classification approach successfully discriminates between the different sign types for all sign classes, where about 97% of the returned sign types is correct (neglecting the samples classified as unreliable), regardless whether 2 or 25 different sign types are involved. Next to this, only about 1 1.5% of the samples are incorrectly classified and around 1 2% are classified as unreliable. Furthermore, false detections are identified with high accuracy, while only a few real samples are labeled as background. Also, we have noted that skewed signs, which do occur quite often in practice, do not influence the classification accuracy, since rotation-invariant features are used. We are aware of the fact that this disables discrimination between rotated instances of identical signs, but this approach could be followed without complications, as the officially approved signs do not contain this kind of instances Inventory system performance analysis The performance of the complete inventory system is evaluated for a geographical region containing several towns within a rural surrounding intersected by a highway. Capturings are taken every 5 meters, covering all public roads within this region, resulting in about 147, 000 panoramic images. The described inventory system is applied and the results are manually verified, where all traffic signs are considered for which the front side is visible in at least a single panoramic image. Table 2 lists the number of correct, missing, falsely detected and signs obtained with incorrect sign code for the different classes. As can be no- 318

7 Sign class: #Total detections (signs+background) 17, , 319 7, , 296 4, 251 #Signs detected (true positives) 15, , 935 6, 752 7, 565 3, 627 #Signs correctly classified 14, , 309 6, 661 7, 049 3, 508 #Signs falsely classified #Signs classified as unrealiable #Signs classified as background #Background detected (false positives) 1, 406 3, , #Background correctly classified 1, 224 3, , #Background classified as sign #Background classified as unrealiable Table 1. Key performance numbers of our classification module. ticed, over 85% of the present signs are localized correctly. We have observed that the performance varies slightly over the sign classes, but is significantly lower for the red circular signs. Therefore, we will separately analyze the performance of this class and the other classes below. Most signs obtained with wrong sign codes correspond to signs located at some distance to the road, and to signs with degraded visibility, e.g. damaged or besmeared, thereby complicating discrimination between similar sign types. Moreover, as shown in Fig. 9(a)-9(e), some red circular signs contain a custom template with an arbitrary number, contributing to a lowered classification score. The falsely detected signs are mainly caused by two reasons. The first comes from GPS inaccuracies, where signs located along roads that are captured in both directions are identified twice, e.g. at 0.25 meter apart. Fusion of both signs is not a solution, since two identical signs may be present, as shown in Fig. 9(f). Second, objects very similar to traffic signs exist, which is especially the case for the red circular signs. There are not only custom prohibition signs, but also the red letter O is often recognized as a traffic sign. Examples of these are displayed in Fig. 9(g)-9(j). For the other sign classes, the number of falsely detected signs is quite low, indicating that our classification module successfully filters out almost all present false detections. (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) Figure 9. Examples. (a)-(e): identical signs with metadata; (f): example of two nearby, identical signs; (g)-(j): sign-like objects. When analyzing the missed signs, we have noted that almost all missed signs (except for two very degraded signs) are detected in at least a single image. However, our approach requires 3 detections, which is disabled by e.g. occlusions by other traffic, a sign orientation parallel to the road and a sign position far away from the road. The latter is especially the case for the blue circular signs when they indicate the start or end of a bicycle path, as displayed in Fig. 10(a). For the red circular signs, another phenomenon causes the rather large amount of missed signs. We have noted that framed speed signs, for which an example is displayed in Fig. 10(b), have a very small red circle, which complicates detection from a distance. This causes more than half of the missed signs for this class. Detection of the frame and sign combination would form a possible solution. As from almost all signs a single detection is obtained, and over 85% of the signs is automatically located with a correct sign code, highly accurate inventories can be realized by addition of a limited amount of manual interaction. This consists of checking all detections classified as a sign, but not part of a sign, and of checking all located signs for correctness. Both checks can be performed efficiently, and allow for addition of other added-value attributes, including subsign texts and sign states (skewed, stickered, etc). 7. Conclusions and future work This paper has described an inventory system for traffic signs from street-level panoramic images, which is a challenging problem as capturing conditions vary, signs may be deformed and many sign look-a-like objects exist. The system starts with localizing the signs in the individual images, using independent detectors for the different sign classes. Then, each detection is classified to obtain the sign code, where also falsely detected signs are identified. Afterwards, detections from multiple images are combined to calculate the sign positions. The performance of the proposed system is evaluated by performing an inventory of a large geographical region, 319

8 Sign class: Total signs Correct signs Wrong sign code Falsely det. signs Missed signs Total 3, 341 2, Table 2. Performance overview of the complete inventory system for the seven different sign classes. (a) [4] G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray. Visual categorization with bags of keypoints. In Proc. European Conference on Computer Vision (ECCV), May [5] N. Dalal and B. Triggs. Histogram of oriented gradients for human detection. In Proc. IEEE Computer Vision and Pattern Recognition (CVPR), volume 1, pages , June [6] A. Frome, G. Cheung, A. Abdulkader, M. Zennaro, B. Wu, A. Bissacco, H. Adam, H. Neven, and L. Vincent. Largescale privacy protection in google street view. In Proc. IEEE International Conference on Computer Vision (ICCV), pages , October [7] L. Hazelhoff, I. M. Creusen, D. W. J. M. van de Wouw, and P. H. N. de With. Large-scale classification of trafic signs under real-world conditions. In Proc. SPIE 8304B-34, [8] F. Jurie and B. Triggs. Creating efficient codebooks for visual recognition. In ICCV, pages , [9] R. Kohavi. A study of cross-validation and bootstrap for accuracy estimation and model selection. In IJCAI, [10] S. Lafuente-Arroyo, S. Maldonado-Bascon, P. Gil-Jimenez, J. Acevedo-Rodriguez, and R. Lopez-Sastre. A tracking system for automated inventory of road signs. In Intelligent Vehicles Symposium, 2007 IEEE, pages , june [11] D. G. Lowe. Distinctive image features from scale-invariant keypoints. Int. Journal of Computer Vision (IJCV), 60(2), January [12] S. Maldonado-Bascon, S. Lafuente-Arroyo, P. Gil-Jimenez, H. Gomez-Moreno, and F. Lopez-Ferreras. Road-sign detection and recognition based on support vector machines. Intelligent Transportation Systems, IEEE Transactions on, 8(2): , june [13] S. Maldonado-Bascon, S. Lafuente-Arroyo, P. Siegmann, H. Gomez-Moreno, and F. Acevedo-Rodriguez. Traffic sign recognition system for inventory purposes. In Intelligent Vehicles Symposium, 2008 IEEE, pages , june [14] E. Nowak, F. Jurie, and B. Triggs. Sampling strategies for bag-of-features image classification. In Proc. European Conference on Computer Vision (ECCV), pages Springer, [15] G. Overett and L. Petersson. Large scale sign detection using hog feature variants. In Intelligent Vehicles Symposium (IV), 2011 IEEE, pages , june [16] R. Timofte, K. Zimmermann, and L. V. Gool. Multi-view traffic sign detection, recognition, and 3d localisation. In Applications of Computer Vision (WACV), 2009 Workshop on, pages 1 8, dec (b) Figure 10. Examples of specific situations. (a): missed sign, located far away from capturing locations (there are no capturings present at the bicycle path); (b): framed speed sign. where over 85% of the 3, 341 signs are correctly localized. Despite the high number of sign look-a-like objects, only a limited number of objects are falsely detected as sign. Furthermore, nearly all missed signs are detected in at least a single image, where position retrieval is mainly limited by the capturing interval of 5 meters. As this performance is achieved at the large scale for a complete geographic region, where also signs not located directly along the road are taken into account, we consider this as an accurate result, especially since signs may be damaged, besmeared or partly occluded. By allowing a limited amount of manual interaction, a highly accurate inventory can be realized, with additional added-values such as indications about the sign state and possible subsign texts. In the future, we will extend the system with additional sign types, including framed (speed) signs and subsigns. Furthermore, we will perform additional validation experiments, possibly including an evaluation with a lower capturing interval of e.g. 2.5 m. References [1] I. Bonaci, I. Kusalic, I. Kovacek, Z. Kalafatic, and S. Segvic. Addressing false alarms and localization inaccuracy in traffic sign detection and recognition. In 16th computer vision winter workshop, pages 1 8, [2] D. Comanicu and P. Meer. Mean shift: A robust approach toward feature space analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24: , [3] I. M. Creusen, R. G. J. Wijnhoven, E. Herbschleb, and P. H. N. de With. Color exploitation in hog-based traffic sign detection. In Proc. IEEE International Conference on Image Processing (ICIP), pages , September

Large-scale classification of traffic signs under real-world conditions

Large-scale classification of traffic signs under real-world conditions Large-scale classification of traffic signs under real-world conditions Lykele Hazelhoff a,b, Ivo Creusen a,b, Dennis van de Wouw a,b and Peter H.N. de With a,b a CycloMedia Technology B.V., Achterweg

More information

Wavelet-based Salient Points with Scale Information for Classification

Wavelet-based Salient Points with Scale Information for Classification Wavelet-based Salient Points with Scale Information for Classification Alexandra Teynor and Hans Burkhardt Department of Computer Science, Albert-Ludwigs-Universität Freiburg, Germany {teynor, Hans.Burkhardt}@informatik.uni-freiburg.de

More information

Discriminative part-based models. Many slides based on P. Felzenszwalb

Discriminative part-based models. Many slides based on P. Felzenszwalb More sliding window detection: ti Discriminative part-based models Many slides based on P. Felzenszwalb Challenge: Generic object detection Pedestrian detection Features: Histograms of oriented gradients

More information

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Corners, Blobs & Descriptors With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Motivation: Build a Panorama M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 How do we build panorama?

More information

Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context

Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Silvio Savarese Lecture 10-16-Feb-15 From the 3D to 2D & vice versa

More information

JOINT INTERPRETATION OF ON-BOARD VISION AND STATIC GPS CARTOGRAPHY FOR DETERMINATION OF CORRECT SPEED LIMIT

JOINT INTERPRETATION OF ON-BOARD VISION AND STATIC GPS CARTOGRAPHY FOR DETERMINATION OF CORRECT SPEED LIMIT JOINT INTERPRETATION OF ON-BOARD VISION AND STATIC GPS CARTOGRAPHY FOR DETERMINATION OF CORRECT SPEED LIMIT Alexandre Bargeton, Fabien Moutarde, Fawzi Nashashibi and Anne-Sophie Puthon Robotics Lab (CAOR),

More information

Urban land use information retrieval based on scene classification of Google Street View images

Urban land use information retrieval based on scene classification of Google Street View images Urban land use information retrieval based on scene classification of Google Street View images Xiaojiang Li 1, Chuanrong Zhang 1 1 Department of Geography, University of Connecticut, Storrs Email: {xiaojiang.li;chuanrong.zhang}@uconn.edu

More information

Advances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.

Advances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al. 6.869 Advances in Computer Vision Prof. Bill Freeman March 3, 2005 Image and shape descriptors Affine invariant features Comparison of feature descriptors Shape context Readings: Mikolajczyk and Schmid;

More information

Feature extraction: Corners and blobs

Feature extraction: Corners and blobs Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over

More information

Lecture 13 Visual recognition

Lecture 13 Visual recognition Lecture 13 Visual recognition Announcements Silvio Savarese Lecture 13-20-Feb-14 Lecture 13 Visual recognition Object classification bag of words models Discriminative methods Generative methods Object

More information

CS5670: Computer Vision

CS5670: Computer Vision CS5670: Computer Vision Noah Snavely Lecture 5: Feature descriptors and matching Szeliski: 4.1 Reading Announcements Project 1 Artifacts due tomorrow, Friday 2/17, at 11:59pm Project 2 will be released

More information

Global Scene Representations. Tilke Judd

Global Scene Representations. Tilke Judd Global Scene Representations Tilke Judd Papers Oliva and Torralba [2001] Fei Fei and Perona [2005] Labzebnik, Schmid and Ponce [2006] Commonalities Goal: Recognize natural scene categories Extract features

More information

A Discriminatively Trained, Multiscale, Deformable Part Model

A Discriminatively Trained, Multiscale, Deformable Part Model A Discriminatively Trained, Multiscale, Deformable Part Model P. Felzenszwalb, D. McAllester, and D. Ramanan Edward Hsiao 16-721 Learning Based Methods in Vision February 16, 2009 Images taken from P.

More information

SUBJECTIVE EVALUATION OF IMAGE UNDERSTANDING RESULTS

SUBJECTIVE EVALUATION OF IMAGE UNDERSTANDING RESULTS 18th European Signal Processing Conference (EUSIPCO-2010) Aalborg, Denmark, August 23-27, 2010 SUBJECTIVE EVALUATION OF IMAGE UNDERSTANDING RESULTS Baptiste Hemery 1, Hélène Laurent 2, and Christophe Rosenberger

More information

Shape of Gaussians as Feature Descriptors

Shape of Gaussians as Feature Descriptors Shape of Gaussians as Feature Descriptors Liyu Gong, Tianjiang Wang and Fang Liu Intelligent and Distributed Computing Lab, School of Computer Science and Technology Huazhong University of Science and

More information

Edges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise

Edges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity

More information

Analysis on a local approach to 3D object recognition

Analysis on a local approach to 3D object recognition Analysis on a local approach to 3D object recognition Elisabetta Delponte, Elise Arnaud, Francesca Odone, and Alessandro Verri DISI - Università degli Studi di Genova - Italy Abstract. We present a method

More information

Automatic localization of tombs in aerial imagery: application to the digital archiving of cemetery heritage

Automatic localization of tombs in aerial imagery: application to the digital archiving of cemetery heritage Automatic localization of tombs in aerial imagery: application to the digital archiving of cemetery heritage M. Chaumont 1,2, L. Tribouillard 2, G. Subsol 2, F. Courtade 2, J. Pasquet 2, M. Derras 3 (1)

More information

38 1 Vol. 38, No ACTA AUTOMATICA SINICA January, Bag-of-phrases.. Image Representation Using Bag-of-phrases

38 1 Vol. 38, No ACTA AUTOMATICA SINICA January, Bag-of-phrases.. Image Representation Using Bag-of-phrases 38 1 Vol. 38, No. 1 2012 1 ACTA AUTOMATICA SINICA January, 2012 Bag-of-phrases 1, 2 1 1 1, Bag-of-words,,, Bag-of-words, Bag-of-phrases, Bag-of-words DOI,, Bag-of-words, Bag-of-phrases, SIFT 10.3724/SP.J.1004.2012.00046

More information

Traffic accidents and the road network in SAS/GIS

Traffic accidents and the road network in SAS/GIS Traffic accidents and the road network in SAS/GIS Frank Poppe SWOV Institute for Road Safety Research, the Netherlands Introduction The first figure shows a screen snapshot of SAS/GIS with part of the

More information

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES V.Sridevi, P.Malarvizhi, P.Mathivannan Abstract Rain removal from a video is a challenging problem due to random spatial distribution and

More information

Visual Object Detection

Visual Object Detection Visual Object Detection Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@northwestern.edu http://www.eecs.northwestern.edu/~yingwu 1 / 47 Visual Object

More information

Lecture 8: Interest Point Detection. Saad J Bedros

Lecture 8: Interest Point Detection. Saad J Bedros #1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:

More information

Tracking Human Heads Based on Interaction between Hypotheses with Certainty

Tracking Human Heads Based on Interaction between Hypotheses with Certainty Proc. of The 13th Scandinavian Conference on Image Analysis (SCIA2003), (J. Bigun and T. Gustavsson eds.: Image Analysis, LNCS Vol. 2749, Springer), pp. 617 624, 2003. Tracking Human Heads Based on Interaction

More information

Maximally Stable Local Description for Scale Selection

Maximally Stable Local Description for Scale Selection Maximally Stable Local Description for Scale Selection Gyuri Dorkó and Cordelia Schmid INRIA Rhône-Alpes, 655 Avenue de l Europe, 38334 Montbonnot, France {gyuri.dorko,cordelia.schmid}@inrialpes.fr Abstract.

More information

Feature Vector Similarity Based on Local Structure

Feature Vector Similarity Based on Local Structure Feature Vector Similarity Based on Local Structure Evgeniya Balmachnova, Luc Florack, and Bart ter Haar Romeny Eindhoven University of Technology, P.O. Box 53, 5600 MB Eindhoven, The Netherlands {E.Balmachnova,L.M.J.Florack,B.M.terHaarRomeny}@tue.nl

More information

Image Processing 1 (IP1) Bildverarbeitung 1

Image Processing 1 (IP1) Bildverarbeitung 1 MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV KOGS Image Processing 1 IP1 Bildverarbeitung 1 Lecture : Object Recognition Winter Semester 015/16 Slides: Prof. Bernd Neumann Slightly revised

More information

Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll

Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll 08.02.2017 By 2030 road traffic deaths will be the fifth leading cause of death

More information

LoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D

LoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D Achieving scale covariance Blobs (and scale selection) Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region

More information

Achieving scale covariance

Achieving scale covariance Achieving scale covariance Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region size that is covariant

More information

Robust License Plate Detection Using Covariance Descriptor in a Neural Network Framework

Robust License Plate Detection Using Covariance Descriptor in a Neural Network Framework MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Robust License Plate Detection Using Covariance Descriptor in a Neural Network Framework Fatih Porikli, Tekin Kocak TR2006-100 January 2007

More information

Orientation Map Based Palmprint Recognition

Orientation Map Based Palmprint Recognition Orientation Map Based Palmprint Recognition (BM) 45 Orientation Map Based Palmprint Recognition B. H. Shekar, N. Harivinod bhshekar@gmail.com, harivinodn@gmail.com India, Mangalore University, Department

More information

CS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015

CS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015 CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature

More information

Sound Recognition in Mixtures

Sound Recognition in Mixtures Sound Recognition in Mixtures Juhan Nam, Gautham J. Mysore 2, and Paris Smaragdis 2,3 Center for Computer Research in Music and Acoustics, Stanford University, 2 Advanced Technology Labs, Adobe Systems

More information

The state of the art and beyond

The state of the art and beyond Feature Detectors and Descriptors The state of the art and beyond Local covariant detectors and descriptors have been successful in many applications Registration Stereo vision Motion estimation Matching

More information

Quality and Coverage of Data Sources

Quality and Coverage of Data Sources Quality and Coverage of Data Sources Objectives Selecting an appropriate source for each item of information to be stored in the GIS database is very important for GIS Data Capture. Selection of quality

More information

Boosting: Algorithms and Applications

Boosting: Algorithms and Applications Boosting: Algorithms and Applications Lecture 11, ENGN 4522/6520, Statistical Pattern Recognition and Its Applications in Computer Vision ANU 2 nd Semester, 2008 Chunhua Shen, NICTA/RSISE Boosting Definition

More information

Face detection and recognition. Detection Recognition Sally

Face detection and recognition. Detection Recognition Sally Face detection and recognition Detection Recognition Sally Face detection & recognition Viola & Jones detector Available in open CV Face recognition Eigenfaces for face recognition Metric learning identification

More information

Detectors part II Descriptors

Detectors part II Descriptors EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers Goal:

More information

Loss Functions and Optimization. Lecture 3-1

Loss Functions and Optimization. Lecture 3-1 Lecture 3: Loss Functions and Optimization Lecture 3-1 Administrative Assignment 1 is released: http://cs231n.github.io/assignments2017/assignment1/ Due Thursday April 20, 11:59pm on Canvas (Extending

More information

EE 6882 Visual Search Engine

EE 6882 Visual Search Engine EE 6882 Visual Search Engine Prof. Shih Fu Chang, Feb. 13 th 2012 Lecture #4 Local Feature Matching Bag of Word image representation: coding and pooling (Many slides from A. Efors, W. Freeman, C. Kambhamettu,

More information

OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS

OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS Y. Li 1,*, M. Sakamoto 1, T. Shinohara 1, T. Satoh 1 1 PASCO CORPORATION, 2-8-10 Higashiyama, Meguro-ku, Tokyo 153-0043,

More information

Visibility Estimation of Traffic Signals under Rainy Weather Conditions for Smart Driving Support

Visibility Estimation of Traffic Signals under Rainy Weather Conditions for Smart Driving Support 2012 15th International IEEE Conference on Intelligent Transportation Systems Anchorage, Alaska, USA, September 16-19, 2012 Visibility Estimation of Traffic Signals under Rainy Weather Conditions for Smart

More information

Multimodal context analysis and prediction

Multimodal context analysis and prediction Multimodal context analysis and prediction Valeria Tomaselli (valeria.tomaselli@st.com) Sebastiano Battiato Giovanni Maria Farinella Tiziana Rotondo (PhD student) Outline 2 Context analysis vs prediction

More information

Blob Detection CSC 767

Blob Detection CSC 767 Blob Detection CSC 767 Blob detection Slides: S. Lazebnik Feature detection with scale selection We want to extract features with characteristic scale that is covariant with the image transformation Blob

More information

A Contrario Detection of False Matches in Iris Recognition

A Contrario Detection of False Matches in Iris Recognition A Contrario Detection of False Matches in Iris Recognition Marcelo Mottalli, Mariano Tepper, and Marta Mejail Departamento de Computación, Universidad de Buenos Aires, Argentina Abstract. The pattern of

More information

FPGA Implementation of a HOG-based Pedestrian Recognition System

FPGA Implementation of a HOG-based Pedestrian Recognition System MPC Workshop Karlsruhe 10/7/2009 FPGA Implementation of a HOG-based Pedestrian Recognition System Sebastian Bauer sebastian.bauer@fh-aschaffenburg.de Laboratory for Pattern Recognition and Computational

More information

Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images

Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images Yunxiang Mao and Zhaozheng Yin (B) Computer Science, Missouri University

More information

Blobs & Scale Invariance

Blobs & Scale Invariance Blobs & Scale Invariance Prof. Didier Stricker Doz. Gabriele Bleser Computer Vision: Object and People Tracking With slides from Bebis, S. Lazebnik & S. Seitz, D. Lowe, A. Efros 1 Apertizer: some videos

More information

CSE 473/573 Computer Vision and Image Processing (CVIP)

CSE 473/573 Computer Vision and Image Processing (CVIP) CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu inwogu@buffalo.edu Lecture 11 Local Features 1 Schedule Last class We started local features Today More on local features Readings for

More information

DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION

DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA Shizhi Chen and YingLi Tian Department of Electrical Engineering The City College of

More information

CS 231A Section 1: Linear Algebra & Probability Review

CS 231A Section 1: Linear Algebra & Probability Review CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability

More information

DM-Group Meeting. Subhodip Biswas 10/16/2014

DM-Group Meeting. Subhodip Biswas 10/16/2014 DM-Group Meeting Subhodip Biswas 10/16/2014 Papers to be discussed 1. Crowdsourcing Land Use Maps via Twitter Vanessa Frias-Martinez and Enrique Frias-Martinez in KDD 2014 2. Tracking Climate Change Opinions

More information

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang CS 231A Section 1: Linear Algebra & Probability Review Kevin Tang Kevin Tang Section 1-1 9/30/2011 Topics Support Vector Machines Boosting Viola Jones face detector Linear Algebra Review Notation Operations

More information

Asaf Bar Zvi Adi Hayat. Semantic Segmentation

Asaf Bar Zvi Adi Hayat. Semantic Segmentation Asaf Bar Zvi Adi Hayat Semantic Segmentation Today s Topics Fully Convolutional Networks (FCN) (CVPR 2015) Conditional Random Fields as Recurrent Neural Networks (ICCV 2015) Gaussian Conditional random

More information

Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network

Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network Marko Tscherepanow and Franz Kummert Applied Computer Science, Faculty of Technology, Bielefeld

More information

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi

More information

Lecture 8: Interest Point Detection. Saad J Bedros

Lecture 8: Interest Point Detection. Saad J Bedros #1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff

More information

Learning theory. Ensemble methods. Boosting. Boosting: history

Learning theory. Ensemble methods. Boosting. Boosting: history Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over

More information

Active Detection via Adaptive Submodularity

Active Detection via Adaptive Submodularity Active Detection via Adaptive Submodularity Yuxin Chen, Hiroaki Shioi, Cesar Antonio Fuentes Montesinos! Lian Pin Koh, Serge Wich and Andreas Krause! ICML Beijing June 23, 2014! Motivating Example: Biodiversity

More information

EXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL IMAGE IN URBAN AREAS. Received September 2015; revised January 2016

EXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL IMAGE IN URBAN AREAS. Received September 2015; revised January 2016 International Journal of Innovative Computing, Information and Control ICIC International c 2016 ISSN 1349-4198 Volume 12, Number 2, April 2016 pp. 371 383 EXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL

More information

Sky Segmentation in the Wild: An Empirical Study

Sky Segmentation in the Wild: An Empirical Study Sky Segmentation in the Wild: An Empirical Study Radu P. Mihail 1 Scott Workman 2 Zach Bessinger 2 Nathan Jacobs 2 rpmihail@valdosta.edu scott@cs.uky.edu zach@cs.uky.edu jacobs@cs.uky.edu 1 Valdosta State

More information

Towards Fully-automated Driving

Towards Fully-automated Driving Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, postdoc, project manager, software engineer,

More information

Road Surface Condition Analysis from Web Camera Images and Weather data. Torgeir Vaa (SVV), Terje Moen (SINTEF), Junyong You (CMR), Jeremy Cook (CMR)

Road Surface Condition Analysis from Web Camera Images and Weather data. Torgeir Vaa (SVV), Terje Moen (SINTEF), Junyong You (CMR), Jeremy Cook (CMR) Road Surface Condition Analysis from Web Camera Images and Weather data Torgeir Vaa (SVV), Terje Moen (SINTEF), Junyong You (CMR), Jeremy Cook (CMR) Motivation Cameras installed along roads for surface

More information

Action-Decision Networks for Visual Tracking with Deep Reinforcement Learning

Action-Decision Networks for Visual Tracking with Deep Reinforcement Learning Action-Decision Networks for Visual Tracking with Deep Reinforcement Learning Sangdoo Yun 1 Jongwon Choi 1 Youngjoon Yoo 2 Kimin Yun 3 and Jin Young Choi 1 1 ASRI, Dept. of Electrical and Computer Eng.,

More information

Region Moments: Fast invariant descriptors for detecting small image structures

Region Moments: Fast invariant descriptors for detecting small image structures Region Moments: Fast invariant descriptors for detecting small image structures Gianfranco Doretto Yi Yao Visualization and Computer Vision Lab, GE Global Research, Niskayuna, NY 39 doretto@research.ge.com

More information

Riemannian Metric Learning for Symmetric Positive Definite Matrices

Riemannian Metric Learning for Symmetric Positive Definite Matrices CMSC 88J: Linear Subspaces and Manifolds for Computer Vision and Machine Learning Riemannian Metric Learning for Symmetric Positive Definite Matrices Raviteja Vemulapalli Guide: Professor David W. Jacobs

More information

Pedestrian Density Estimation by a Weighted Bag of Visual Words Model

Pedestrian Density Estimation by a Weighted Bag of Visual Words Model Pedestrian Density Estimation by a Weighted Bag of Visual Words Model Shilin Zhang and Xunyuan Zhang image representation termed bag of visual words integrating weighting scheme and spatial pyramid co-occurrence,

More information

Fisher Vector image representation

Fisher Vector image representation Fisher Vector image representation Machine Learning and Category Representation 2014-2015 Jakob Verbeek, January 9, 2015 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15 A brief recap on kernel

More information

Clustering with k-means and Gaussian mixture distributions

Clustering with k-means and Gaussian mixture distributions Clustering with k-means and Gaussian mixture distributions Machine Learning and Category Representation 2012-2013 Jakob Verbeek, ovember 23, 2012 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.12.13

More information

CITS 4402 Computer Vision

CITS 4402 Computer Vision CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh Lecture 06 Object Recognition Objectives To understand the concept of image based object recognition To learn how to match images

More information

Recurrent Autoregressive Networks for Online Multi-Object Tracking. Presented By: Ishan Gupta

Recurrent Autoregressive Networks for Online Multi-Object Tracking. Presented By: Ishan Gupta Recurrent Autoregressive Networks for Online Multi-Object Tracking Presented By: Ishan Gupta Outline Multi Object Tracking Recurrent Autoregressive Networks (RANs) RANs for Online Tracking Other State

More information

Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition

Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Esa Rahtu Mikko Salo Janne Heikkilä Department of Electrical and Information Engineering P.O. Box 4500, 90014 University of

More information

KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE. Israel -

KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE. Israel - KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE Ammatzia Peled a,*, Michael Gilichinsky b a University of Haifa, Department of Geography and Environmental Studies,

More information

RESTORATION OF VIDEO BY REMOVING RAIN

RESTORATION OF VIDEO BY REMOVING RAIN RESTORATION OF VIDEO BY REMOVING RAIN Sajitha Krishnan 1 and D.Venkataraman 1 1 Computer Vision and Image Processing, Department of Computer Science, Amrita Vishwa Vidyapeetham University, Coimbatore,

More information

Vision for Mobile Robot Navigation: A Survey

Vision for Mobile Robot Navigation: A Survey Vision for Mobile Robot Navigation: A Survey (February 2002) Guilherme N. DeSouza & Avinash C. Kak presentation by: Job Zondag 27 February 2009 Outline: Types of Navigation Absolute localization (Structured)

More information

A Hierarchical Convolutional Neural Network for Mitosis Detection in Phase-Contrast Microscopy Images

A Hierarchical Convolutional Neural Network for Mitosis Detection in Phase-Contrast Microscopy Images A Hierarchical Convolutional Neural Network for Mitosis Detection in Phase-Contrast Microscopy Images Yunxiang Mao and Zhaozheng Yin (B) Department of Computer Science, Missouri University of Science and

More information

Human Action Recognition under Log-Euclidean Riemannian Metric

Human Action Recognition under Log-Euclidean Riemannian Metric Human Action Recognition under Log-Euclidean Riemannian Metric Chunfeng Yuan, Weiming Hu, Xi Li, Stephen Maybank 2, Guan Luo, National Laboratory of Pattern Recognition, Institute of Automation, CAS, Beijing,

More information

Object Recognition Using Local Characterisation and Zernike Moments

Object Recognition Using Local Characterisation and Zernike Moments Object Recognition Using Local Characterisation and Zernike Moments A. Choksuriwong, H. Laurent, C. Rosenberger, and C. Maaoui Laboratoire Vision et Robotique - UPRES EA 2078, ENSI de Bourges - Université

More information

Loss Functions and Optimization. Lecture 3-1

Loss Functions and Optimization. Lecture 3-1 Lecture 3: Loss Functions and Optimization Lecture 3-1 Administrative: Live Questions We ll use Zoom to take questions from remote students live-streaming the lecture Check Piazza for instructions and

More information

Automatic estimation of crowd size and target detection using Image processing

Automatic estimation of crowd size and target detection using Image processing Automatic estimation of crowd size and target detection using Image processing Asst Prof. Avinash Rai Dept. of Electronics and communication (UIT-RGPV) Bhopal avinashrai@rgtu.net Rahul Meshram Dept. of

More information

SIFT keypoint detection. D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp , 2004.

SIFT keypoint detection. D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp , 2004. SIFT keypoint detection D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (), pp. 91-110, 004. Keypoint detection with scale selection We want to extract keypoints with characteristic

More information

MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No.

MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No. MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No. 2211) Submitted to: Ginger McGovern, P.E. Planning and Research Division Engineer

More information

Multiple Similarities Based Kernel Subspace Learning for Image Classification

Multiple Similarities Based Kernel Subspace Learning for Image Classification Multiple Similarities Based Kernel Subspace Learning for Image Classification Wang Yan, Qingshan Liu, Hanqing Lu, and Songde Ma National Laboratory of Pattern Recognition, Institute of Automation, Chinese

More information

Fantope Regularization in Metric Learning

Fantope Regularization in Metric Learning Fantope Regularization in Metric Learning CVPR 2014 Marc T. Law (LIP6, UPMC), Nicolas Thome (LIP6 - UPMC Sorbonne Universités), Matthieu Cord (LIP6 - UPMC Sorbonne Universités), Paris, France Introduction

More information

SYMMETRY is a highly salient visual phenomenon and

SYMMETRY is a highly salient visual phenomenon and JOURNAL OF L A T E X CLASS FILES, VOL. 6, NO. 1, JANUARY 2011 1 Symmetry-Growing for Skewed Rotational Symmetry Detection Hyo Jin Kim, Student Member, IEEE, Minsu Cho, Student Member, IEEE, and Kyoung

More information

Parking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition-

Parking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition- Parking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition- Sanngoen Wanayuth, Akihisa Ohya and Takashi Tsubouchi Abstract The automated

More information

Real-time image-based parking occupancy detection using deep learning. Debaditya Acharya, Weilin Yan & Kourosh Khoshelham The University of Melbourne

Real-time image-based parking occupancy detection using deep learning. Debaditya Acharya, Weilin Yan & Kourosh Khoshelham The University of Melbourne Real-time image-based parking occupancy detection using deep learning Debaditya Acharya, Weilin Yan & Kourosh Khoshelham The University of Melbourne Slide 1/20 Prologue People spend on Does average that

More information

INTEREST POINTS AT DIFFERENT SCALES

INTEREST POINTS AT DIFFERENT SCALES INTEREST POINTS AT DIFFERENT SCALES Thank you for the slides. They come mostly from the following sources. Dan Huttenlocher Cornell U David Lowe U. of British Columbia Martial Hebert CMU Intuitively, junctions

More information

Image Analysis. Feature extraction: corners and blobs

Image Analysis. Feature extraction: corners and blobs Image Analysis Feature extraction: corners and blobs Christophoros Nikou cnikou@cs.uoi.gr Images taken from: Computer Vision course by Svetlana Lazebnik, University of North Carolina at Chapel Hill (http://www.cs.unc.edu/~lazebnik/spring10/).

More information

SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS

SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS Hans-Jürgen Winkler ABSTRACT In this paper an efficient on-line recognition system for handwritten mathematical formulas is proposed. After formula

More information

Rapid Object Recognition from Discriminative Regions of Interest

Rapid Object Recognition from Discriminative Regions of Interest Rapid Object Recognition from Discriminative Regions of Interest Gerald Fritz, Christin Seifert, Lucas Paletta JOANNEUM RESEARCH Institute of Digital Image Processing Wastiangasse 6, A-81 Graz, Austria

More information

Major Crime Map Help Documentation

Major Crime Map Help Documentation Major Crime Map Help Documentation This web application is designed to make it easier to visualize and understand crime trends in Overland Park. The data for this application are generally limited to relatively

More information

Overview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors

Overview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison of different detectors Region descriptors and

More information

Real Time Face Detection and Recognition using Haar - Based Cascade Classifier and Principal Component Analysis

Real Time Face Detection and Recognition using Haar - Based Cascade Classifier and Principal Component Analysis Real Time Face Detection and Recognition using Haar - Based Cascade Classifier and Principal Component Analysis Sarala A. Dabhade PG student M. Tech (Computer Egg) BVDU s COE Pune Prof. Mrunal S. Bewoor

More information

Distinguish between different types of scenes. Matching human perception Understanding the environment

Distinguish between different types of scenes. Matching human perception Understanding the environment Scene Recognition Adriana Kovashka UTCS, PhD student Problem Statement Distinguish between different types of scenes Applications Matching human perception Understanding the environment Indexing of images

More information

Adaptive Binary Integration CFAR Processing for Secondary Surveillance Radar *

Adaptive Binary Integration CFAR Processing for Secondary Surveillance Radar * BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 9, No Sofia 2009 Adaptive Binary Integration CFAR Processing for Secondary Surveillance Radar Ivan Garvanov, Christo Kabakchiev

More information

Introduction to GIS I

Introduction to GIS I Introduction to GIS Introduction How to answer geographical questions such as follows: What is the population of a particular city? What are the characteristics of the soils in a particular land parcel?

More information

2D Image Processing Face Detection and Recognition

2D Image Processing Face Detection and Recognition 2D Image Processing Face Detection and Recognition Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information