Robust Detection, Classification and Positioning of Traffic Signs from Street-Level Panoramic Images for Inventory Purposes
|
|
- Alice Phillips
- 5 years ago
- Views:
Transcription
1 Robust Detection, Classification and Positioning of Traffic Signs from Street-Level Panoramic Images for Inventory Purposes Lykele Hazelhoff and Ivo Creusen CycloMedia Technology B.V. Achterweg 38, 4181 AE Waardenburg, The Netherlands Peter H.N. de With Eindhoven University of Technology Den Dolech 2, 5600 MB Eindhoven, The Netherlands Abstract Accurate inventories of traffic signs are required for road maintenance and increase of the road safety. These inventories can be performed efficiently based on street-level panoramic images. However, this is a challenging problem, as these images are captured under a wide range of weather conditions. Besides this, occlusions and sign deformations occur and many sign look-a-like objects exist. Our approach is based on detecting present signs in panoramic images, both to derive a classification code and to combine multiple detections into an accurate position of the signs. It starts with detecting the present signs in each panoramic image. Then, all detections are classified to obtain the specific sign type, where also false detections are identified. Afterwards, detections from multiple images are combined to calculate the sign positions. The performance of this approach is extensively evaluated in a large, geographical region, where over 85% of the 3, 341 signs are automatically localized, with only 3.2% false detections. As nearly all missed signs are detected in at least a single image, only very limited manual interactions have to be supplied to safeguard the performance for highly accurate inventories. 1. Introduction Nowadays, several companies record street-level panoramic images, which provide a recent and accurate overview of the road infrastructure. Within The Netherlands, these images are captured by private companies (e.g. CycloMedia Technology and Google), where each public road is recaptured annually. The resulting image databases enable efficient inventories of street furniture to support maintenance and cost control. Computer vision techniques facilitate the automatic creation of such inventories and thereby reduce human interaction compared to manual inventories, where all objects are searched and annotated by hand. These inventories are of interest to governmental organizations tasked with road maintenance. Especially traffic signs are of interest, as their presence directly influences road safety. They require accurate and up-to-date inventories, as the sign visibility may be degraded due to e.g. vandalism, vegetation coverage, aging and accidents. This paper describes a framework for road-sign inventories based on computer vision techniques, aiming at retrieving the sign code and position of all traffic signs in a region. Although traffic signs are designed to attract visual attention, automatic detection and classification of road signs is a complicated problem for several reasons. The first is related to capturing from a driving vehicle. As the signs are captured from a wide range of distances, large viewpoint deviations exist, and signs may be occluded by e.g. other road users. Furthermore, capturing outside implies varying weather conditions, including e.g. fog. The second complication comes from the sign features. Signs vary in size, and there are many similar traffic signs, which are sometimes custom versions of official signs, and some signs are designed to contain custom text or symbols. Moreover, the visibility of traffic signs may be lowered due to aforementioned reasons, while especially these signs are of importance for sign maintenance. Thirdly, many sign look-a-like objects exist, including directives for customer parking or restrictions for dog access, which are not traffic signs. Examples of these complicating factors are displayed in Fig Related work In literature, detection and recognition of traffic signs is studied for many years. For example, [15] describes a cascade detection of speed signs within a large scale project, achieving a detection rate of 98.7%. However, this paper only addresses detection of a single sign type. This is also the case for [1], where the image is prefiltered with a color-version of the Viola-Jones algorithm, followed by analysis of Histogram of Oriented Gradient fea- 313
2 (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) (k) (l) (m) (n) (o) Figure 1. Examples of factors complicating detection and classification of traffic signs. (a)-(e): occlusions; (f)-(i): lowered sign visibility. (j)-(m): sign-like objects. (n)-(o): official sign with custom version. tures with a neural network. The authors test their system for triangular warning signs, and report that 260 of the 265 present signs are correctly identified, where 241 signs are correctly classified. Examples of systems focusing on multiple sign types are e.g. [12], where images captured from a car-mounted camera are exploited for sign detection. After threshold-based color-segmentation and shape analysis, the sign type is recognized based upon a grayscale version of the located blobs. The authors report that all the 104 signs are detected at least twice. Color segmentation is also exploited for extraction of sign regions in [13]. Afterwards, the shape of the sign in each blob is extracted, which is subject to classification, based on Support Vector Machines, exploiting the known sign shape. It is reported that 98 of the 102 signs are detected at least once. These proposals detect and classify traffic signs based on single images. A tracking system is proposed in [10], reducing the false alarms by tracking the signs over the frames. The authors report that all 6 signs are detected correctly. Another approach is described by Timofte et al. [16], where a van with 8 cameras is employed for capturing. Their method employs both single-image and multi-view analysis. A fraction of 95.6% of the 269 signs are positioned correctly, where 97.7% of the detected signs are also successfully recognized Our approach This paper presents an approach for performing largescale inventories of traffic signs based on computer vision techniques, where signs are detected, classified and positioned. This is a very challenging problem, and we have experienced that the performance of state-of-the-art algorithms for detection and classification are insufficient for a fully automated inventory in real-world conditions and at a large scale, which is similar as Frome et al. [6] discuss for face detection. Therefore, we aim at a system focusing on both automatic and semi-automatic inventories. This paper describes the automatic version, where we concentrate on seven different sign classes, covering 92 different sign types. The involved classes are displayed in Fig. 2. Instead of constructing a custom capturing device, as some proposals in literature, we exploit the already existing street-level panoramic images. These images are captured on all public roads with a calibrated recording system with a capturing interval of 5 meters. The capturing cars are typically employed efficiently, resulting in images captured in a very wide range of weather conditions, including even fog, which makes the problem even more challenging. Sign appearances vary greatly across the images due to the large variation in weather conditions and due to differences between different camera systems. Therefore, instead of focusing at color, which is common in literature, we investigate color gradients, since they are more robust against these situations. Next to this, we aim at a generic, learningbased system, as this allows adaption to other sign appearances, e.g. in other countries. Our system consists of three stages: sign detection, sign classification and sign positioning. At first, the signs are grouped into classes, e.g. red triangular, and their generic properties are exploited for detection. Due to the genericity of this stage, customized versions of standard signs can also be found, which is beneficial in the semi-automatic approach. For detection, a custom variant of the popular Histogram of Oriented Gradients [5] is applied, which operates with color gradient information to exploit the characteristic sign colors. Then, the minor differences between the signs are analyzed, and all detections are classified using a variant of the popular Bag of Words (BoW) technique [4]. The standard BoW approach is modified to both filter out falsely detected signs and to deal with the large intra-class similarities of the signs. Afterwards, in the sign positioning stage, the sign positions are calculated by combining the detections across multiple images. The performance of this inventory system is evaluated by a large-scale experiment, where an inventory is applied to a large geographical region, containing over 3, 340 traffic signs. We should note that this validation size is rather uncommon in related literature, as is the fact that we also take signs not directly located along the road into account. Besides this, the performance of the individual detection and classification stages are also assessed. The remainder of the paper is organized as follows. Section 2 contains the system overview. Section 3 describes the sign detection stage, Section 4 describes the classification stage, followed by the positioning procedure in Section 5. The performance evaluation can be found in Section 6, followed by the conclusions in Section
3 sign detection sign classification sign detection sign classification red triangular signs red triangular signs give-way signs red circular signs red circular signs Overlap analysis and correction redblue circular signs redblue circular signs no-entry signs blue circular signs blue circular signs yellow diamond signs yellow diamond signs sign detection sign positioning sign classification Figure 2. System overview of our inventory system. 2. System overview is combined with all nearby images. Next, hypotheses of sign positions are obtained by pair wise combinations of detections with both an identical sign code, where combinations of sign codes and wildcards give supporting evidence. Then, the final sign positions are obtained by clustering of these hypotheses. The system overview of our inventory process is depicted in Fig. 2. The system consists of three primary modules, which are briefly described below. 1. Sign detection: At first, each panoramic image is analyzed and present signs are detected by multiple, independent detectors, each focusing at a specific class of signs. These detectors are kept very generic to allow detection of distorted signs and sign-like objects. As some detectors focus at quite similar sign classes, their output may overlap. These overlapping samples are analyzed and a detection fusion step is applied to assign the correct class label. 3. Sign detection The first stage of our inventory system consists of localizing traffic signs within the individual panoramic images. As many traffic signs have similar color and shape (such as e.g. all blue circular direction signs), similar signs are grouped into sign classes (such as blue circular) and detection is performed for each class independently. The class division is displayed in Fig. 2. Since traffic signs are intended to attract attention based on their colors and shape, many traffic sign detection systems in literature start by color filtering of the image, and extracting regions with colors corresponding to the signs. However, we have found that the color and contrast of the signs varies significantly with the capturing conditions, and therefore instead we exploit color differences and shape information, which we have found to be more consistent over the varying circumstances. We apply detectors based on the popular Histogram of Oriented Gradients (HOG) algorithm, originally proposed by Dalal and Triggs [5]. As the standard algorithm extracts the maximum gradient over the color channels for each pixel, it neglects the correlation of gradients over the color channels, and thereby neglects the discriminative color of the traffic signs. We have extended the standard HOG approach with the use of color information, as described in [3]. 2. Sign classification: During the sign detection stage, all detected signs are assigned a sign class label, e.g. red triangular. Next, each detection is assigned a sign code, such as warning sign for a dangerous crossing. As some detections have an insufficient resolution for classification, these small samples are not classified, but assigned a class-specific wildcard. Furthermore, at this stage, the false findings given by the detectors are identified by inclusion of an additional class to the classification procedure, representing the false detections of the respective sign class. These detections are also assigned the wildcard code. Classes with only a single contained sign are not subject to classification. 3. Sign positioning: In the sign positioning stage, detections from multiple images are combined to calculate the position of the traffic signs based on the geometric properties of our source data. For this, each image 315
4 The modified HOG algorithm works as follows. First, the image is divided into cells of 8 8 pixels, where for each of these cells a histogram of the gradient orientation is calculated. These histograms are normalized w.r.t. adjacent histograms. Next, a sliding window is moved over the cells, covering 5 5 cells, and all included histograms are concatenated. As we perform detection on color images, the histograms of all the color channels are appended. The resulting 1, 200-dimensional feature vector is used for classification by means of a linear Support Vector Machine (SVM). Since multiple classes of signs are detected independently from each other, the same features are exploited for all classes, where each class is found by an individual SVM. As the feature extraction stage is the most timeconsuming task, adding additional classes does not affect processing time significantly. This procedure is repeated at multiple scales to obtain scale invariance, resulting in detections with a size ranging from pixels up to pixels, corresponding to a typical sign-to-car distance ranging from about 19 to 1 meters, respectively. Although the different detectors operate independently, cross-overlap may exist between the detectors for the different classes, especially when the visual difference is low. This causes signs to be detected by multiple detectors, which is especially the case for the red circular and redblue circular signs, which are both circular with a red border. Therefore, we employ a specific step to distinguish between these two classes. Each sample detected by both detectors is analyzed and assigned a single class label. Since we strive for a generic system, we apply a learning-based approach, exploiting the differences in color distribution between the sign classes. This method first transforms the input samples into the HSV color space, then extracts a color histogram from the signs, followed by classification based on a linear SVM. Afterwards, each sample is assigned the appropriate class label, which is exploited during the classification. 4. Sign classification Each detector focuses at detecting a specific sign class, which typically consists of more than one sign type. Therefore, the detector output is analyzed to obtain the sign code, e.g. danger, crossing children, where detectors that directly locate a sign type are not subject to classification. Whereas the detection stage exploits the generic characteristics of a sign class, such as the sign borders, the classification stage should discriminate between signs based upon the very minor differences within the inner template, as visualized in Fig. 4. This complicates the classification task, especially since the resolution of the discriminative part of the signs is quite low. Therefore, we ignore samples with insufficient resolution and assign them a wildcard code (its use is explained later), and only classify samples larger than pixels. The employed classification approach is based on (a) (b) (c) (d) Figure 4. Examples of signs with only very minor differences. Figure 5. Visualization of the modular codebook, containing the concatenation of individual visual dictionaries. Bag of Words [4] (BoW), and is described in [7]. For completion, we will briefly redescribe the key features of the system below. The original BoW approach [4] represents each image by a histogram containing the occurrence frequency of the elements of the visual dictionary. These elements, called visual words, are obtained by clustering the features extracted from training samples, in order to obtain features that occur frequently in all training samples. However, these words may not be the most discriminative [8], which will especially be the case when words occur in all sign types, e.g. representing the sign borders. Therefore, we altered the construction of the visual dictionary, where we generate a separate dictionary for each sign type to ensure that characteristics common for the sign types are extracted. The individual dictionaries are concatenated to a modular codebook. This approach also enables handling of unbalanced training data and allows for easy addition of extra classes without recomputing the complete codebook. As our sign detector also outputs false detections, we have extended the modular codebook with a large dictionary representing them. This prevents false detections from being assigned to a random class, and instead, have those mapped to the false detection class. The resulting modular codebook is visualized in Fig. 5 and is used, as would it be the regular codebook in the standard BoW approach. Based on this modular codebook, new samples are classified as follows. First, the input sample is normalized to a standard size. Since the sign class is known, irrelevant parts are removed based on a standard shape, similar to [13]. As the pattern is more discriminative than the color, each sample is converted to a single color image, where for almost all classes grayscale is used. However, the redblue circular signs are converted to a custom color space, given by the difference between the red and blue color channels, as the red rim and blue inner part have similar grayscale values. Then, SIFT features [11] are extracted using dense sampling, which is shown to outperform the use of interest point operators [14]. We have applied SIFT, as these features are invariant against the commonly occurring ro- 316
5 fn S Above car, fh = ½p Horizon, fh = 0 p Below car, fh = -½p -½p E ½p N W fh S -p Figure 3. Example panoramic image including a visualization of the geometric properties of the panorama, where ϕn denotes the angle w.r.t. the northern direction and ϕh the angle w.r.t. the horizon. tation and illumination changes. Each extracted feature is matched against the modular codebook, and a word histogram is constructed, which is used for classification based on linear SVMs in a One-vs-All fashion. When none or multiple classifiers recognize the sample, we consider it as unreliable, and assign it a wildcard code. this, we exploit the fact that our source data is geometrically correct, i.e. the angular orientations are linear with the pixel coordinates. This is achieved by extensive calibration of the capturing system. Furthermore, each panoramic image is divided into two equal parts by both the northern direction and the horizon, as visualized in Fig. 3. As a result, when two points corresponding to the same object are known in two images, the position of that object can be calculated by straightforward geometric calculations. The sign positions are retrieved per sign class, where detections from individual classes are combined with other detections, either having an identical sign code or a wildcard code. Although the detections themselves can be correlated, multiple identical signs can result in wrong correspondences. Furthermore, the capturing interval of 5 meters causes large differences in perspective, scale and background, thereby complicating pixel correspondences. Therefore, we only exploit the geometric information, where all the combined pairs give a position estimation, which are passed through when this estimate is closer than 45 meter w.r.t. both images. Each estimate gives a hypothesis of the sign location, which are clustered around the real sign position, as visualized in Fig. 6. These clusters are recovered using the MeanShift algorithm [2]. Afterwards, the clusters are processed from large to small cardinality, where only clusters containing at least 3 or more detections are taken into account. For each cluster, the position estimates given by the combination of all contained detections should be close to the cluster center, and when not valid, the detection with the largest mean position deviation w.r.t. this center is removed. The same rule applies when multiple detections from the same image are present. After a cluster is accepted, all contained detections are removed, and the procedure is repeated. 5. Sign positioning After the classification stage, all detections are assigned a sign code, either corresponding to a specific sign type or to a class-specific wildcard. Next, the positions of the signs are calculated by combining the detections corresponding to the same physical sign appearing in multiple images. For XXX X X XX X XX X X X Figure 6. Visualization of the sign positioning process, where two identical give-way signs are present (indicated by the give-way symbols). The positions of the capturings are displayed by blue circles, the position estimates given by the pair-wise combination of all detections are drawn as red crosses and the actual position of the traffic signs are indicated by the two sign symbols. 317
6 Precision redblue circle red circle red triangular 0.2 yellow diamond give way blue circle Recall Figure 7. Recall-precision curves for the individual traffic sign detectors. The classes indicated correspond to the classes in Fig. 2. Precision redblue circle with detection fusion 0.2 redblue circle without detection fusion red circle with detection fusion red circle without detection fusion Recall Figure 8. Recall-precision curves showing the effect of the detection fusion stage for the red and redblue circular signs. 6. Experiments and results The described system is employed to perform inventories of multiple geographical regions, containing both rural and city environments. The output is manually verified, and the resulting ground-truth data is employed to assess the performance of the detection and classification stages of the system. Furthermore, the performance of the complete inventory system is evaluated for a representative geographical area, containing multiple towns within a rural surrounding Detection performance analysis The performance of the sign-detection module is evaluated on about 33, 000 panoramic images, each containing at least a single traffic sign. This set covers both rural and city environments, and all present signs are manually annotated. However, it should be noted that the ground truth was constructed subjectively, so that there is some mismatch for the smaller signs, which may be detected but not annotated. The set is processed within 8.5 hours on a cluster of 4 computers and the resulting recall-precision curves for the different sign classes are shown in Fig. 7. As follows, the detectors are able to localize over 90% of the traffic signs in the individual panoramic images. This figure includes the detection fusion step for the red circular and redblue circular signs. The effect of this step is displayed in Fig. 8, which clearly shows that the additional fusion stage significantly reduces the number of false detections, especially for the redblue circular signs. This indicates that there is indeed a large overlap between the classes. As the percentage of detected signs is not significantly affected, our approach successfully assigns the correct class label in most cases Classification performance analysis The performance of our classification module is evaluated using the detector output on a large set of panoramic images, which is constructed such that a maximum number of different sign types are included with sufficient cardinality. The minimum size of the considered detections is pixels, where occurring false detections are included. The classification performance is analyzed by using 10-fold cross-validation, as this approaches the system performance when all samples have been subject to training [9]. Since our test set contains multiple samples of the same physical sign, captured from different viewpoints, these occurrences are forced into the same partition to prevent testing with the training set. Table 1 summarizes the key performance numbers for the classes of interest. Classification of an unseen sample takes about 1 second. It can be concluded that our classification approach successfully discriminates between the different sign types for all sign classes, where about 97% of the returned sign types is correct (neglecting the samples classified as unreliable), regardless whether 2 or 25 different sign types are involved. Next to this, only about 1 1.5% of the samples are incorrectly classified and around 1 2% are classified as unreliable. Furthermore, false detections are identified with high accuracy, while only a few real samples are labeled as background. Also, we have noted that skewed signs, which do occur quite often in practice, do not influence the classification accuracy, since rotation-invariant features are used. We are aware of the fact that this disables discrimination between rotated instances of identical signs, but this approach could be followed without complications, as the officially approved signs do not contain this kind of instances Inventory system performance analysis The performance of the complete inventory system is evaluated for a geographical region containing several towns within a rural surrounding intersected by a highway. Capturings are taken every 5 meters, covering all public roads within this region, resulting in about 147, 000 panoramic images. The described inventory system is applied and the results are manually verified, where all traffic signs are considered for which the front side is visible in at least a single panoramic image. Table 2 lists the number of correct, missing, falsely detected and signs obtained with incorrect sign code for the different classes. As can be no- 318
7 Sign class: #Total detections (signs+background) 17, , 319 7, , 296 4, 251 #Signs detected (true positives) 15, , 935 6, 752 7, 565 3, 627 #Signs correctly classified 14, , 309 6, 661 7, 049 3, 508 #Signs falsely classified #Signs classified as unrealiable #Signs classified as background #Background detected (false positives) 1, 406 3, , #Background correctly classified 1, 224 3, , #Background classified as sign #Background classified as unrealiable Table 1. Key performance numbers of our classification module. ticed, over 85% of the present signs are localized correctly. We have observed that the performance varies slightly over the sign classes, but is significantly lower for the red circular signs. Therefore, we will separately analyze the performance of this class and the other classes below. Most signs obtained with wrong sign codes correspond to signs located at some distance to the road, and to signs with degraded visibility, e.g. damaged or besmeared, thereby complicating discrimination between similar sign types. Moreover, as shown in Fig. 9(a)-9(e), some red circular signs contain a custom template with an arbitrary number, contributing to a lowered classification score. The falsely detected signs are mainly caused by two reasons. The first comes from GPS inaccuracies, where signs located along roads that are captured in both directions are identified twice, e.g. at 0.25 meter apart. Fusion of both signs is not a solution, since two identical signs may be present, as shown in Fig. 9(f). Second, objects very similar to traffic signs exist, which is especially the case for the red circular signs. There are not only custom prohibition signs, but also the red letter O is often recognized as a traffic sign. Examples of these are displayed in Fig. 9(g)-9(j). For the other sign classes, the number of falsely detected signs is quite low, indicating that our classification module successfully filters out almost all present false detections. (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) Figure 9. Examples. (a)-(e): identical signs with metadata; (f): example of two nearby, identical signs; (g)-(j): sign-like objects. When analyzing the missed signs, we have noted that almost all missed signs (except for two very degraded signs) are detected in at least a single image. However, our approach requires 3 detections, which is disabled by e.g. occlusions by other traffic, a sign orientation parallel to the road and a sign position far away from the road. The latter is especially the case for the blue circular signs when they indicate the start or end of a bicycle path, as displayed in Fig. 10(a). For the red circular signs, another phenomenon causes the rather large amount of missed signs. We have noted that framed speed signs, for which an example is displayed in Fig. 10(b), have a very small red circle, which complicates detection from a distance. This causes more than half of the missed signs for this class. Detection of the frame and sign combination would form a possible solution. As from almost all signs a single detection is obtained, and over 85% of the signs is automatically located with a correct sign code, highly accurate inventories can be realized by addition of a limited amount of manual interaction. This consists of checking all detections classified as a sign, but not part of a sign, and of checking all located signs for correctness. Both checks can be performed efficiently, and allow for addition of other added-value attributes, including subsign texts and sign states (skewed, stickered, etc). 7. Conclusions and future work This paper has described an inventory system for traffic signs from street-level panoramic images, which is a challenging problem as capturing conditions vary, signs may be deformed and many sign look-a-like objects exist. The system starts with localizing the signs in the individual images, using independent detectors for the different sign classes. Then, each detection is classified to obtain the sign code, where also falsely detected signs are identified. Afterwards, detections from multiple images are combined to calculate the sign positions. The performance of the proposed system is evaluated by performing an inventory of a large geographical region, 319
8 Sign class: Total signs Correct signs Wrong sign code Falsely det. signs Missed signs Total 3, 341 2, Table 2. Performance overview of the complete inventory system for the seven different sign classes. (a) [4] G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray. Visual categorization with bags of keypoints. In Proc. European Conference on Computer Vision (ECCV), May [5] N. Dalal and B. Triggs. Histogram of oriented gradients for human detection. In Proc. IEEE Computer Vision and Pattern Recognition (CVPR), volume 1, pages , June [6] A. Frome, G. Cheung, A. Abdulkader, M. Zennaro, B. Wu, A. Bissacco, H. Adam, H. Neven, and L. Vincent. Largescale privacy protection in google street view. In Proc. IEEE International Conference on Computer Vision (ICCV), pages , October [7] L. Hazelhoff, I. M. Creusen, D. W. J. M. van de Wouw, and P. H. N. de With. Large-scale classification of trafic signs under real-world conditions. In Proc. SPIE 8304B-34, [8] F. Jurie and B. Triggs. Creating efficient codebooks for visual recognition. In ICCV, pages , [9] R. Kohavi. A study of cross-validation and bootstrap for accuracy estimation and model selection. In IJCAI, [10] S. Lafuente-Arroyo, S. Maldonado-Bascon, P. Gil-Jimenez, J. Acevedo-Rodriguez, and R. Lopez-Sastre. A tracking system for automated inventory of road signs. In Intelligent Vehicles Symposium, 2007 IEEE, pages , june [11] D. G. Lowe. Distinctive image features from scale-invariant keypoints. Int. Journal of Computer Vision (IJCV), 60(2), January [12] S. Maldonado-Bascon, S. Lafuente-Arroyo, P. Gil-Jimenez, H. Gomez-Moreno, and F. Lopez-Ferreras. Road-sign detection and recognition based on support vector machines. Intelligent Transportation Systems, IEEE Transactions on, 8(2): , june [13] S. Maldonado-Bascon, S. Lafuente-Arroyo, P. Siegmann, H. Gomez-Moreno, and F. Acevedo-Rodriguez. Traffic sign recognition system for inventory purposes. In Intelligent Vehicles Symposium, 2008 IEEE, pages , june [14] E. Nowak, F. Jurie, and B. Triggs. Sampling strategies for bag-of-features image classification. In Proc. European Conference on Computer Vision (ECCV), pages Springer, [15] G. Overett and L. Petersson. Large scale sign detection using hog feature variants. In Intelligent Vehicles Symposium (IV), 2011 IEEE, pages , june [16] R. Timofte, K. Zimmermann, and L. V. Gool. Multi-view traffic sign detection, recognition, and 3d localisation. In Applications of Computer Vision (WACV), 2009 Workshop on, pages 1 8, dec (b) Figure 10. Examples of specific situations. (a): missed sign, located far away from capturing locations (there are no capturings present at the bicycle path); (b): framed speed sign. where over 85% of the 3, 341 signs are correctly localized. Despite the high number of sign look-a-like objects, only a limited number of objects are falsely detected as sign. Furthermore, nearly all missed signs are detected in at least a single image, where position retrieval is mainly limited by the capturing interval of 5 meters. As this performance is achieved at the large scale for a complete geographic region, where also signs not located directly along the road are taken into account, we consider this as an accurate result, especially since signs may be damaged, besmeared or partly occluded. By allowing a limited amount of manual interaction, a highly accurate inventory can be realized, with additional added-values such as indications about the sign state and possible subsign texts. In the future, we will extend the system with additional sign types, including framed (speed) signs and subsigns. Furthermore, we will perform additional validation experiments, possibly including an evaluation with a lower capturing interval of e.g. 2.5 m. References [1] I. Bonaci, I. Kusalic, I. Kovacek, Z. Kalafatic, and S. Segvic. Addressing false alarms and localization inaccuracy in traffic sign detection and recognition. In 16th computer vision winter workshop, pages 1 8, [2] D. Comanicu and P. Meer. Mean shift: A robust approach toward feature space analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24: , [3] I. M. Creusen, R. G. J. Wijnhoven, E. Herbschleb, and P. H. N. de With. Color exploitation in hog-based traffic sign detection. In Proc. IEEE International Conference on Image Processing (ICIP), pages , September
Large-scale classification of traffic signs under real-world conditions
Large-scale classification of traffic signs under real-world conditions Lykele Hazelhoff a,b, Ivo Creusen a,b, Dennis van de Wouw a,b and Peter H.N. de With a,b a CycloMedia Technology B.V., Achterweg
More informationWavelet-based Salient Points with Scale Information for Classification
Wavelet-based Salient Points with Scale Information for Classification Alexandra Teynor and Hans Burkhardt Department of Computer Science, Albert-Ludwigs-Universität Freiburg, Germany {teynor, Hans.Burkhardt}@informatik.uni-freiburg.de
More informationDiscriminative part-based models. Many slides based on P. Felzenszwalb
More sliding window detection: ti Discriminative part-based models Many slides based on P. Felzenszwalb Challenge: Generic object detection Pedestrian detection Features: Histograms of oriented gradients
More informationCorners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros
Corners, Blobs & Descriptors With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Motivation: Build a Panorama M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 How do we build panorama?
More informationProperties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context
Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Silvio Savarese Lecture 10-16-Feb-15 From the 3D to 2D & vice versa
More informationJOINT INTERPRETATION OF ON-BOARD VISION AND STATIC GPS CARTOGRAPHY FOR DETERMINATION OF CORRECT SPEED LIMIT
JOINT INTERPRETATION OF ON-BOARD VISION AND STATIC GPS CARTOGRAPHY FOR DETERMINATION OF CORRECT SPEED LIMIT Alexandre Bargeton, Fabien Moutarde, Fawzi Nashashibi and Anne-Sophie Puthon Robotics Lab (CAOR),
More informationUrban land use information retrieval based on scene classification of Google Street View images
Urban land use information retrieval based on scene classification of Google Street View images Xiaojiang Li 1, Chuanrong Zhang 1 1 Department of Geography, University of Connecticut, Storrs Email: {xiaojiang.li;chuanrong.zhang}@uconn.edu
More informationAdvances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.
6.869 Advances in Computer Vision Prof. Bill Freeman March 3, 2005 Image and shape descriptors Affine invariant features Comparison of feature descriptors Shape context Readings: Mikolajczyk and Schmid;
More informationFeature extraction: Corners and blobs
Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over
More informationLecture 13 Visual recognition
Lecture 13 Visual recognition Announcements Silvio Savarese Lecture 13-20-Feb-14 Lecture 13 Visual recognition Object classification bag of words models Discriminative methods Generative methods Object
More informationCS5670: Computer Vision
CS5670: Computer Vision Noah Snavely Lecture 5: Feature descriptors and matching Szeliski: 4.1 Reading Announcements Project 1 Artifacts due tomorrow, Friday 2/17, at 11:59pm Project 2 will be released
More informationGlobal Scene Representations. Tilke Judd
Global Scene Representations Tilke Judd Papers Oliva and Torralba [2001] Fei Fei and Perona [2005] Labzebnik, Schmid and Ponce [2006] Commonalities Goal: Recognize natural scene categories Extract features
More informationA Discriminatively Trained, Multiscale, Deformable Part Model
A Discriminatively Trained, Multiscale, Deformable Part Model P. Felzenszwalb, D. McAllester, and D. Ramanan Edward Hsiao 16-721 Learning Based Methods in Vision February 16, 2009 Images taken from P.
More informationSUBJECTIVE EVALUATION OF IMAGE UNDERSTANDING RESULTS
18th European Signal Processing Conference (EUSIPCO-2010) Aalborg, Denmark, August 23-27, 2010 SUBJECTIVE EVALUATION OF IMAGE UNDERSTANDING RESULTS Baptiste Hemery 1, Hélène Laurent 2, and Christophe Rosenberger
More informationShape of Gaussians as Feature Descriptors
Shape of Gaussians as Feature Descriptors Liyu Gong, Tianjiang Wang and Fang Liu Intelligent and Distributed Computing Lab, School of Computer Science and Technology Huazhong University of Science and
More informationEdges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise
Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity
More informationAnalysis on a local approach to 3D object recognition
Analysis on a local approach to 3D object recognition Elisabetta Delponte, Elise Arnaud, Francesca Odone, and Alessandro Verri DISI - Università degli Studi di Genova - Italy Abstract. We present a method
More informationAutomatic localization of tombs in aerial imagery: application to the digital archiving of cemetery heritage
Automatic localization of tombs in aerial imagery: application to the digital archiving of cemetery heritage M. Chaumont 1,2, L. Tribouillard 2, G. Subsol 2, F. Courtade 2, J. Pasquet 2, M. Derras 3 (1)
More information38 1 Vol. 38, No ACTA AUTOMATICA SINICA January, Bag-of-phrases.. Image Representation Using Bag-of-phrases
38 1 Vol. 38, No. 1 2012 1 ACTA AUTOMATICA SINICA January, 2012 Bag-of-phrases 1, 2 1 1 1, Bag-of-words,,, Bag-of-words, Bag-of-phrases, Bag-of-words DOI,, Bag-of-words, Bag-of-phrases, SIFT 10.3724/SP.J.1004.2012.00046
More informationTraffic accidents and the road network in SAS/GIS
Traffic accidents and the road network in SAS/GIS Frank Poppe SWOV Institute for Road Safety Research, the Netherlands Introduction The first figure shows a screen snapshot of SAS/GIS with part of the
More informationA RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES
A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES V.Sridevi, P.Malarvizhi, P.Mathivannan Abstract Rain removal from a video is a challenging problem due to random spatial distribution and
More informationVisual Object Detection
Visual Object Detection Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@northwestern.edu http://www.eecs.northwestern.edu/~yingwu 1 / 47 Visual Object
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:
More informationTracking Human Heads Based on Interaction between Hypotheses with Certainty
Proc. of The 13th Scandinavian Conference on Image Analysis (SCIA2003), (J. Bigun and T. Gustavsson eds.: Image Analysis, LNCS Vol. 2749, Springer), pp. 617 624, 2003. Tracking Human Heads Based on Interaction
More informationMaximally Stable Local Description for Scale Selection
Maximally Stable Local Description for Scale Selection Gyuri Dorkó and Cordelia Schmid INRIA Rhône-Alpes, 655 Avenue de l Europe, 38334 Montbonnot, France {gyuri.dorko,cordelia.schmid}@inrialpes.fr Abstract.
More informationFeature Vector Similarity Based on Local Structure
Feature Vector Similarity Based on Local Structure Evgeniya Balmachnova, Luc Florack, and Bart ter Haar Romeny Eindhoven University of Technology, P.O. Box 53, 5600 MB Eindhoven, The Netherlands {E.Balmachnova,L.M.J.Florack,B.M.terHaarRomeny}@tue.nl
More informationImage Processing 1 (IP1) Bildverarbeitung 1
MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV KOGS Image Processing 1 IP1 Bildverarbeitung 1 Lecture : Object Recognition Winter Semester 015/16 Slides: Prof. Bernd Neumann Slightly revised
More informationMaarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll
Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll 08.02.2017 By 2030 road traffic deaths will be the fifth leading cause of death
More informationLoG Blob Finding and Scale. Scale Selection. Blobs (and scale selection) Achieving scale covariance. Blob detection in 2D. Blob detection in 2D
Achieving scale covariance Blobs (and scale selection) Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region
More informationAchieving scale covariance
Achieving scale covariance Goal: independently detect corresponding regions in scaled versions of the same image Need scale selection mechanism for finding characteristic region size that is covariant
More informationRobust License Plate Detection Using Covariance Descriptor in a Neural Network Framework
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Robust License Plate Detection Using Covariance Descriptor in a Neural Network Framework Fatih Porikli, Tekin Kocak TR2006-100 January 2007
More informationOrientation Map Based Palmprint Recognition
Orientation Map Based Palmprint Recognition (BM) 45 Orientation Map Based Palmprint Recognition B. H. Shekar, N. Harivinod bhshekar@gmail.com, harivinodn@gmail.com India, Mangalore University, Department
More informationCS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015
CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature
More informationSound Recognition in Mixtures
Sound Recognition in Mixtures Juhan Nam, Gautham J. Mysore 2, and Paris Smaragdis 2,3 Center for Computer Research in Music and Acoustics, Stanford University, 2 Advanced Technology Labs, Adobe Systems
More informationThe state of the art and beyond
Feature Detectors and Descriptors The state of the art and beyond Local covariant detectors and descriptors have been successful in many applications Registration Stereo vision Motion estimation Matching
More informationQuality and Coverage of Data Sources
Quality and Coverage of Data Sources Objectives Selecting an appropriate source for each item of information to be stored in the GIS database is very important for GIS Data Capture. Selection of quality
More informationBoosting: Algorithms and Applications
Boosting: Algorithms and Applications Lecture 11, ENGN 4522/6520, Statistical Pattern Recognition and Its Applications in Computer Vision ANU 2 nd Semester, 2008 Chunhua Shen, NICTA/RSISE Boosting Definition
More informationFace detection and recognition. Detection Recognition Sally
Face detection and recognition Detection Recognition Sally Face detection & recognition Viola & Jones detector Available in open CV Face recognition Eigenfaces for face recognition Metric learning identification
More informationDetectors part II Descriptors
EECS 442 Computer vision Detectors part II Descriptors Blob detectors Invariance Descriptors Some slides of this lectures are courtesy of prof F. Li, prof S. Lazebnik, and various other lecturers Goal:
More informationLoss Functions and Optimization. Lecture 3-1
Lecture 3: Loss Functions and Optimization Lecture 3-1 Administrative Assignment 1 is released: http://cs231n.github.io/assignments2017/assignment1/ Due Thursday April 20, 11:59pm on Canvas (Extending
More informationEE 6882 Visual Search Engine
EE 6882 Visual Search Engine Prof. Shih Fu Chang, Feb. 13 th 2012 Lecture #4 Local Feature Matching Bag of Word image representation: coding and pooling (Many slides from A. Efors, W. Freeman, C. Kambhamettu,
More informationOBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS
OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS Y. Li 1,*, M. Sakamoto 1, T. Shinohara 1, T. Satoh 1 1 PASCO CORPORATION, 2-8-10 Higashiyama, Meguro-ku, Tokyo 153-0043,
More informationVisibility Estimation of Traffic Signals under Rainy Weather Conditions for Smart Driving Support
2012 15th International IEEE Conference on Intelligent Transportation Systems Anchorage, Alaska, USA, September 16-19, 2012 Visibility Estimation of Traffic Signals under Rainy Weather Conditions for Smart
More informationMultimodal context analysis and prediction
Multimodal context analysis and prediction Valeria Tomaselli (valeria.tomaselli@st.com) Sebastiano Battiato Giovanni Maria Farinella Tiziana Rotondo (PhD student) Outline 2 Context analysis vs prediction
More informationBlob Detection CSC 767
Blob Detection CSC 767 Blob detection Slides: S. Lazebnik Feature detection with scale selection We want to extract features with characteristic scale that is covariant with the image transformation Blob
More informationA Contrario Detection of False Matches in Iris Recognition
A Contrario Detection of False Matches in Iris Recognition Marcelo Mottalli, Mariano Tepper, and Marta Mejail Departamento de Computación, Universidad de Buenos Aires, Argentina Abstract. The pattern of
More informationFPGA Implementation of a HOG-based Pedestrian Recognition System
MPC Workshop Karlsruhe 10/7/2009 FPGA Implementation of a HOG-based Pedestrian Recognition System Sebastian Bauer sebastian.bauer@fh-aschaffenburg.de Laboratory for Pattern Recognition and Computational
More informationTwo-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images
Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images Yunxiang Mao and Zhaozheng Yin (B) Computer Science, Missouri University
More informationBlobs & Scale Invariance
Blobs & Scale Invariance Prof. Didier Stricker Doz. Gabriele Bleser Computer Vision: Object and People Tracking With slides from Bebis, S. Lazebnik & S. Seitz, D. Lowe, A. Efros 1 Apertizer: some videos
More informationCSE 473/573 Computer Vision and Image Processing (CVIP)
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu inwogu@buffalo.edu Lecture 11 Local Features 1 Schedule Last class We started local features Today More on local features Readings for
More informationDETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION
DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA Shizhi Chen and YingLi Tian Department of Electrical Engineering The City College of
More informationCS 231A Section 1: Linear Algebra & Probability Review
CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability
More informationDM-Group Meeting. Subhodip Biswas 10/16/2014
DM-Group Meeting Subhodip Biswas 10/16/2014 Papers to be discussed 1. Crowdsourcing Land Use Maps via Twitter Vanessa Frias-Martinez and Enrique Frias-Martinez in KDD 2014 2. Tracking Climate Change Opinions
More informationCS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang
CS 231A Section 1: Linear Algebra & Probability Review Kevin Tang Kevin Tang Section 1-1 9/30/2011 Topics Support Vector Machines Boosting Viola Jones face detector Linear Algebra Review Notation Operations
More informationAsaf Bar Zvi Adi Hayat. Semantic Segmentation
Asaf Bar Zvi Adi Hayat Semantic Segmentation Today s Topics Fully Convolutional Networks (FCN) (CVPR 2015) Conditional Random Fields as Recurrent Neural Networks (ICCV 2015) Gaussian Conditional random
More informationSubcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network
Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network Marko Tscherepanow and Franz Kummert Applied Computer Science, Faculty of Technology, Bielefeld
More informationHYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH
HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff
More informationLearning theory. Ensemble methods. Boosting. Boosting: history
Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over
More informationActive Detection via Adaptive Submodularity
Active Detection via Adaptive Submodularity Yuxin Chen, Hiroaki Shioi, Cesar Antonio Fuentes Montesinos! Lian Pin Koh, Serge Wich and Andreas Krause! ICML Beijing June 23, 2014! Motivating Example: Biodiversity
More informationEXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL IMAGE IN URBAN AREAS. Received September 2015; revised January 2016
International Journal of Innovative Computing, Information and Control ICIC International c 2016 ISSN 1349-4198 Volume 12, Number 2, April 2016 pp. 371 383 EXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL
More informationSky Segmentation in the Wild: An Empirical Study
Sky Segmentation in the Wild: An Empirical Study Radu P. Mihail 1 Scott Workman 2 Zach Bessinger 2 Nathan Jacobs 2 rpmihail@valdosta.edu scott@cs.uky.edu zach@cs.uky.edu jacobs@cs.uky.edu 1 Valdosta State
More informationTowards Fully-automated Driving
Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, postdoc, project manager, software engineer,
More informationRoad Surface Condition Analysis from Web Camera Images and Weather data. Torgeir Vaa (SVV), Terje Moen (SINTEF), Junyong You (CMR), Jeremy Cook (CMR)
Road Surface Condition Analysis from Web Camera Images and Weather data Torgeir Vaa (SVV), Terje Moen (SINTEF), Junyong You (CMR), Jeremy Cook (CMR) Motivation Cameras installed along roads for surface
More informationAction-Decision Networks for Visual Tracking with Deep Reinforcement Learning
Action-Decision Networks for Visual Tracking with Deep Reinforcement Learning Sangdoo Yun 1 Jongwon Choi 1 Youngjoon Yoo 2 Kimin Yun 3 and Jin Young Choi 1 1 ASRI, Dept. of Electrical and Computer Eng.,
More informationRegion Moments: Fast invariant descriptors for detecting small image structures
Region Moments: Fast invariant descriptors for detecting small image structures Gianfranco Doretto Yi Yao Visualization and Computer Vision Lab, GE Global Research, Niskayuna, NY 39 doretto@research.ge.com
More informationRiemannian Metric Learning for Symmetric Positive Definite Matrices
CMSC 88J: Linear Subspaces and Manifolds for Computer Vision and Machine Learning Riemannian Metric Learning for Symmetric Positive Definite Matrices Raviteja Vemulapalli Guide: Professor David W. Jacobs
More informationPedestrian Density Estimation by a Weighted Bag of Visual Words Model
Pedestrian Density Estimation by a Weighted Bag of Visual Words Model Shilin Zhang and Xunyuan Zhang image representation termed bag of visual words integrating weighting scheme and spatial pyramid co-occurrence,
More informationFisher Vector image representation
Fisher Vector image representation Machine Learning and Category Representation 2014-2015 Jakob Verbeek, January 9, 2015 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15 A brief recap on kernel
More informationClustering with k-means and Gaussian mixture distributions
Clustering with k-means and Gaussian mixture distributions Machine Learning and Category Representation 2012-2013 Jakob Verbeek, ovember 23, 2012 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.12.13
More informationCITS 4402 Computer Vision
CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh Lecture 06 Object Recognition Objectives To understand the concept of image based object recognition To learn how to match images
More informationRecurrent Autoregressive Networks for Online Multi-Object Tracking. Presented By: Ishan Gupta
Recurrent Autoregressive Networks for Online Multi-Object Tracking Presented By: Ishan Gupta Outline Multi Object Tracking Recurrent Autoregressive Networks (RANs) RANs for Online Tracking Other State
More informationMultiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition
Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Esa Rahtu Mikko Salo Janne Heikkilä Department of Electrical and Information Engineering P.O. Box 4500, 90014 University of
More informationKNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE. Israel -
KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE Ammatzia Peled a,*, Michael Gilichinsky b a University of Haifa, Department of Geography and Environmental Studies,
More informationRESTORATION OF VIDEO BY REMOVING RAIN
RESTORATION OF VIDEO BY REMOVING RAIN Sajitha Krishnan 1 and D.Venkataraman 1 1 Computer Vision and Image Processing, Department of Computer Science, Amrita Vishwa Vidyapeetham University, Coimbatore,
More informationVision for Mobile Robot Navigation: A Survey
Vision for Mobile Robot Navigation: A Survey (February 2002) Guilherme N. DeSouza & Avinash C. Kak presentation by: Job Zondag 27 February 2009 Outline: Types of Navigation Absolute localization (Structured)
More informationA Hierarchical Convolutional Neural Network for Mitosis Detection in Phase-Contrast Microscopy Images
A Hierarchical Convolutional Neural Network for Mitosis Detection in Phase-Contrast Microscopy Images Yunxiang Mao and Zhaozheng Yin (B) Department of Computer Science, Missouri University of Science and
More informationHuman Action Recognition under Log-Euclidean Riemannian Metric
Human Action Recognition under Log-Euclidean Riemannian Metric Chunfeng Yuan, Weiming Hu, Xi Li, Stephen Maybank 2, Guan Luo, National Laboratory of Pattern Recognition, Institute of Automation, CAS, Beijing,
More informationObject Recognition Using Local Characterisation and Zernike Moments
Object Recognition Using Local Characterisation and Zernike Moments A. Choksuriwong, H. Laurent, C. Rosenberger, and C. Maaoui Laboratoire Vision et Robotique - UPRES EA 2078, ENSI de Bourges - Université
More informationLoss Functions and Optimization. Lecture 3-1
Lecture 3: Loss Functions and Optimization Lecture 3-1 Administrative: Live Questions We ll use Zoom to take questions from remote students live-streaming the lecture Check Piazza for instructions and
More informationAutomatic estimation of crowd size and target detection using Image processing
Automatic estimation of crowd size and target detection using Image processing Asst Prof. Avinash Rai Dept. of Electronics and communication (UIT-RGPV) Bhopal avinashrai@rgtu.net Rahul Meshram Dept. of
More informationSIFT keypoint detection. D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp , 2004.
SIFT keypoint detection D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (), pp. 91-110, 004. Keypoint detection with scale selection We want to extract keypoints with characteristic
More informationMODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No.
MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No. 2211) Submitted to: Ginger McGovern, P.E. Planning and Research Division Engineer
More informationMultiple Similarities Based Kernel Subspace Learning for Image Classification
Multiple Similarities Based Kernel Subspace Learning for Image Classification Wang Yan, Qingshan Liu, Hanqing Lu, and Songde Ma National Laboratory of Pattern Recognition, Institute of Automation, Chinese
More informationFantope Regularization in Metric Learning
Fantope Regularization in Metric Learning CVPR 2014 Marc T. Law (LIP6, UPMC), Nicolas Thome (LIP6 - UPMC Sorbonne Universités), Matthieu Cord (LIP6 - UPMC Sorbonne Universités), Paris, France Introduction
More informationSYMMETRY is a highly salient visual phenomenon and
JOURNAL OF L A T E X CLASS FILES, VOL. 6, NO. 1, JANUARY 2011 1 Symmetry-Growing for Skewed Rotational Symmetry Detection Hyo Jin Kim, Student Member, IEEE, Minsu Cho, Student Member, IEEE, and Kyoung
More informationParking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition-
Parking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition- Sanngoen Wanayuth, Akihisa Ohya and Takashi Tsubouchi Abstract The automated
More informationReal-time image-based parking occupancy detection using deep learning. Debaditya Acharya, Weilin Yan & Kourosh Khoshelham The University of Melbourne
Real-time image-based parking occupancy detection using deep learning Debaditya Acharya, Weilin Yan & Kourosh Khoshelham The University of Melbourne Slide 1/20 Prologue People spend on Does average that
More informationINTEREST POINTS AT DIFFERENT SCALES
INTEREST POINTS AT DIFFERENT SCALES Thank you for the slides. They come mostly from the following sources. Dan Huttenlocher Cornell U David Lowe U. of British Columbia Martial Hebert CMU Intuitively, junctions
More informationImage Analysis. Feature extraction: corners and blobs
Image Analysis Feature extraction: corners and blobs Christophoros Nikou cnikou@cs.uoi.gr Images taken from: Computer Vision course by Svetlana Lazebnik, University of North Carolina at Chapel Hill (http://www.cs.unc.edu/~lazebnik/spring10/).
More informationSYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS
SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS Hans-Jürgen Winkler ABSTRACT In this paper an efficient on-line recognition system for handwritten mathematical formulas is proposed. After formula
More informationRapid Object Recognition from Discriminative Regions of Interest
Rapid Object Recognition from Discriminative Regions of Interest Gerald Fritz, Christin Seifert, Lucas Paletta JOANNEUM RESEARCH Institute of Digital Image Processing Wastiangasse 6, A-81 Graz, Austria
More informationMajor Crime Map Help Documentation
Major Crime Map Help Documentation This web application is designed to make it easier to visualize and understand crime trends in Overland Park. The data for this application are generally limited to relatively
More informationOverview. Introduction to local features. Harris interest points + SSD, ZNCC, SIFT. Evaluation and comparison of different detectors
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison of different detectors Region descriptors and
More informationReal Time Face Detection and Recognition using Haar - Based Cascade Classifier and Principal Component Analysis
Real Time Face Detection and Recognition using Haar - Based Cascade Classifier and Principal Component Analysis Sarala A. Dabhade PG student M. Tech (Computer Egg) BVDU s COE Pune Prof. Mrunal S. Bewoor
More informationDistinguish between different types of scenes. Matching human perception Understanding the environment
Scene Recognition Adriana Kovashka UTCS, PhD student Problem Statement Distinguish between different types of scenes Applications Matching human perception Understanding the environment Indexing of images
More informationAdaptive Binary Integration CFAR Processing for Secondary Surveillance Radar *
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 9, No Sofia 2009 Adaptive Binary Integration CFAR Processing for Secondary Surveillance Radar Ivan Garvanov, Christo Kabakchiev
More informationIntroduction to GIS I
Introduction to GIS Introduction How to answer geographical questions such as follows: What is the population of a particular city? What are the characteristics of the soils in a particular land parcel?
More information2D Image Processing Face Detection and Recognition
2D Image Processing Face Detection and Recognition Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More information