Mapping of hyperspectral AVIRIS data using machine-learning algorithms

Size: px
Start display at page:

Download "Mapping of hyperspectral AVIRIS data using machine-learning algorithms"

Transcription

1 Can. J. Remote Sensing, Vol. 35, Suppl. 1, pp. S106 S116, 2009 Mapping of hyperspectral AVIRIS data using machine-learning algorithms Björn Waske, Jon Atli Benediktsson, Kolbeinn Árnason, and Johannes R. Sveinsson Abstract. Hyperspectral imaging provides detailed spectral and spatial information from the land cover that enables a precise differentiation between various surface materials. On the other hand, the performance of traditional and widely used statistical classification methods is often limited in this context, and thus alternative methods are required. In the study presented here, the performance of two machine-learning techniques, namely support vector machines (SVMs) and random forests (RFs), is investigated and the classification results are compared with those from well-known methods (i.e., maximum likelihood classifier and spectral angle mapper). The classifiers are applied to an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) dataset that was acquired near the Hekla volcano in Iceland. The results clearly show the advantages of the two proposed classifier algorithms in terms of accuracy. They significantly outperform the other methods and achieve overall accuracies of approximately 90%. Although SVM and RF show some diversity in the classification results, the global performance of the two classifiers is very similar. Thus, both methods can be considered attractive for the classification of hyperspectral data. Résumé. Les images hyperspectrales fournissent une information spectrale et spatiale détaillée sur le couvert permettant ainsi une différentiation précise des différents matériaux de surface. D autre part, la performance des méthodes de classification statistique traditionnelles et les plus couramment utilisées est souvent limitée à cet égard. Ainsi, des méthodes alternatives sont nécessaires. Dans cette étude, les performances de deux techniques d apprentissage automatique séparateurs à vaste marge (SVM) et forêts aléatoires (RF) sont analysées et les résultats de classification sont comparés à ceux de méthodes bien connues (c.-à-d. classificateur par maximum de vraisemblance et méthode SAM («spectral angle mapper»). Les classificateurs sont appliqués à un ensemble de données d AVIRIS («Airborne Visible/Infrared Imaging Spectrometer») acquis près du volcan Hekla, en Islande. Les résultats démontrent clairement les avantages des deux algorithmes de classification proposés en termes de précision. Ceux-ci affichent une performance nettement supérieure à celle des autres méthodes et permettent d atteindre une précision globale d environ 90 %. Bien que les méthodes SVM et RF affichent une certaine diversité au plan des résultats de classification, la performance globale des deux classificateurs est très semblable. Ainsi, les deux méthodes peuvent être considérées comme intéressantes pour la classification des données hyperspectrales. [Traduit par la Rédaction] Introduction Waske et al 116 Hyperspectral imaging is characterized by high spectral resolution with up to hundreds of narrow bands. Hyperspectral sensors provide spectrally continuous spatial information of the land surface, ranging from the visible to the short-wave infrared regions of the electromagnetic spectrum, which enables detailed discrimination between similar surface materials (Clark, 1999; van der Meer and de Jong, 2001). Imaging spectroscopy has been used for more than two decades in the field of geology, and the reflectance properties of various surface materials have been addressed in several studies (Goetz et al., 1982; Kruse, 1988). Hyperspectral data have been successfully used for classification problems that require very precise description in spectral feature space, e.g., for mapping of lithology (Harris et al., 2005; Chen et al., 2007; Sahoo et al., 2007). The operational availability of hyperspectral data will increase through future spaceborne missions, such as the German EnMAP, thus providing more importance on hyperspectral research with a focus on classification of hyperspectral data. However, the performance of widely used traditional classification methods like the maximum likelihood classifier (MLC) can be limited when applied to hyperspectral datasets because of the high-dimensional feature space and a finite number of training samples. An increasing number of bands can often decrease classification accuracies, i.e., the well-known Hughes phenomena (Hughes, 1968), which becomes even more critical as soon as ground-truth and Received 15 January Accepted 21 July Published on the Web at on 5 March B. Waske, 1 J.A. Benediktsson, and J.R. Sveinsson. Faculty of Electrical and Computer Engineering, University of Iceland, 107 Reykjavík, Iceland. K. Árnason. National Land Survey of Iceland, Stillholti 16-18, 300 Akranes, Iceland. 1 Corresponding author. Present address: Faculty of Agriculture Institute of Geodesy and Geoinformation, University of Bonn, Bonn, Germany ( bwaske@uni-bonn.de). S106

2 Canadian Journal of Remote Sensing / Journal canadien de télédétection reference data are limited, as in remote study regions that are difficult to access. The MLC is derived from the Bayes rule when classes have equal prior probabilities and each pixel is classified to the class that has the highest likelihood. The approach is based on the assumption that the probability density function for each class is multivariate, and often a Gaussian distribution is assumed. Many processes can be described by this model. Moreover, this type of model simplifies the approach because the model is only described by the mean vector and the covariance matrix. However, the MLC requires a theoretical limit of (number of bands + 1) training pixels per class to avoid the covariance matrix becoming singular. Thus, alternative classifier algorithms may be better suited to these cases. The spectral angle mapper (SAM) is an example of a common technique used for classifying hyperspectral imagery (Kruse et al., 1993; Dennison et al., 2004; Chen et al., 2007). This method, which is based on spectral similarity, compares unknown image spectra with known endmember spectra by calculating the angle between the corresponding vectors in feature space (Kruse et al., 1993). The similarity measure does not consider the absolute reflectance values, but refers only to the shape of the different spectra. Although SAM can achieve good results and can outperform statistical classifiers in terms of accuracy (Chen et al., 2007), recent developments from the field of machine learning have been generating more accurate results. SAM is not particularly well suited when applied to complex feature spaces (Harken and Sugumaran, 2005; van der Linden et al., 2007a). A description of MLC and SAM is given in Richards and Jia (2006). Support vector machines (SVMs) are well known in the field of machine learning and pattern recognition and were introduced to the field of remote sensing during the last few years (Huang et al., 2002; Melgani and Bruzzone, 2004). The SVM classifier is based on generating a hyperplane between the training samples which separates two classes in a multidimensional feature space. The optimization procedure aims at maximizing the margin between the closest training samples and the hyperplane. The approach performs well with small training sets, even when high-dimensional datasets are classified, because it only considers training data close to the class boundary (Melgani and Bruzzone, 2004; Pal and Mather, 2006; Fauvel et al., 2006). In Fauvel et al. (2006), for example, the number of available training samples was successively reduced when SVMs were applied to an urban hyperspectral dataset. In this study the overall accuracy that is achieved by a training set with 100 samples per class is only slightly decreased (i.e., 1.5%) compared with the results achieved by a small training set with 10 samples per class. Although SVMs perform well with a small number of training samples, a large number of samples may guarantee that adequate samples are included in the training set (Foody and Mathur, 2004). However, appropriate sampling strategies can decrease the training set size considerably without any significant loss in classification accuracy (Foody et al., 2006). SVMs have been used successfully in challenging hyperspectral applications such as the classification of heterogeneous urban areas (van der Linden et al., 2007b) and the mapping of similar lithologic classes (Sahoo et al., 2007). Another important development in the field of machine learning and image classification is the concept of multiple classifier systems (Polikar, 2006). Multiple classifier systems (MCSs) and classifier ensembles do not refer to a specific classification algorithm but describe a more general classification strategy. In contrast to traditional classification approaches, ensembles use several classifiers and combine the different outputs to generate a final result. This idea was introduced recently in the context of remote sensing (Benediktsson and Swain, 1992; Benediktsson et al., 2007) and is particularly useful in the context of hyperspectral and multisource applications (Briem et al., 2002; Ham et al., 2006; Waske and Benediktsson 2007; Cheung-Wai Chan and Paelinckx, 2008). Whereas some MCSs are based on the combination of various algorithms, others employ variants of the same algorithm. The individual classifier is trained on modified input data (e.g., subsets of training samples or input features), and the different outputs are combined by a voting strategy. Self-learning decision tree (DT) classifiers are particularly interesting in this context because of their simple handling and fast training times. A DT employs a hierarchical classification concept, whereby the training data are successively separated into smaller, increasingly homogeneous subclasses using a set of tests. These test rules, which are defined at each node of the tree, are generally based on the training data and an impurity measurement (Safavian and Landgrebe, 1991; Friedl and Brodley, 1997). Breiman s random forest (RF) classifier (Breiman, 2001), which has been used successfully for the classification of hyperspectral and multisource imagery (Gislason et al., 2006; Ham et al., 2006; Cheung-Wai Chan and Paelinckx, 2008; Waske and van der Linden, 2008), is one example of a decision tree based classifier system. The performance of SVMs and RFs for classification of hyperspectral imagery is addressed in the study presented here. The study site is the area surrounding Hekla volcano in southcentral Iceland. The area is characterized by recent volcanic activity comprising various historic lava flows, which form the most prominent features. The very similar spectral patterns of these mostly nonvegetated volcanic surface classes result in a challenging pattern recognition and classification problem. The results achieved by SVMs and RFs are compared and discussed. The present study addresses the following research questions: (1) Can advanced classification techniques improve the accuracy and reliability of the classification achieved by well-known methods such as MLC and SAM? (2) How do two different machine-learning approaches perform in comparison to each other? The study site and dataset are described, and then the classifier algorithms are introduced. The actual methods for the application of the classifiers to the data set are explained, and S107

3 Vol. 35, Suppl. 1, 2009 the results are presented. The final section of the paper is a discussion of the results and presents the conclusions. Study site and data The study site is the region surrounding Hekla volcano, one of the most active volcanoes in Iceland. Hekla has erupted quite regularly every 10 years since 1970, in 1970, , 1991, and The volcano is located on the southwestern margin of the Eastern volcanic zone in south Iceland. Hekla is a ridge built by repeated eruptions on a narrow volcanic fissure zone and comprises mainly andesitic and basaltic lavas and tephra. The elevation of the volcano is 1488 m, and it rises some m above the surrounding area (Soosalu et al., 2003). Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) data collected on a cloud-free day (17 June 1991) were used for the classification. The sensor operates in the visible, nearinfrared, and mid-infrared portions of the electromagnetic spectrum, with its sensitivity range spanning wavelengths from 0.4 to 2.4 µm. The sensor system has 224 data channels, utilizing four spectrometers, and each spectral band is approximately 10 nm in width (Green et al., 1998). During the image acquisition, spectrometer 4 was not working properly. This particular spectrometer operates in the wavelength range from 1.84 to 2.40 µm (64 bands). These 64 bands were deleted from the imagery along with the first channels for all the other spectrometers, leaving 157 data channels for analysis. The image strip is 2048 pixels 614 pixels (see Figure 1), with a spatial resolution of 20 m (Benediktsson and Kanellopoulos, 1999). The dataset contains calibrated image radiance returned by the AVIRIS instrument during data acquisition. The units are microwatts per square centimetre per nanometre per steradian multiplied by 200 to maintain precision in the 16 bit data. The mapping procedure focuses on 22 lithologies and land cover classes, mainly lava flows from different eruptions and older hyaloclastites (erupted subglacially during the last periods of the Ice Age). Most of the area is completely void of vegetation, and only some of the lava in the southern part of the area (bottom part of Figure 1) is wholly or partly covered by vegetation. The reference data were generated based on detailed geological and vegetation maps, expert knowledge, and image interpretation. The reference information for each class was equally divided into independent training and test data. Afterward, different training sets were generated to investigate the impact of the number of training samples on the classification accuracy. Using an equalized random sampling approach, which guarantees that all classes are equally included in the sample set, 100, 200, and up to 400 samples per class were selected out of the whole training dataset (from now on referred to as tr100, tr200, and tr400; see Table 1). S108 Classification algorithms Support vector machines The basic concept of SVMs for a linearly nonseparable case aims at the definition of a separating hyperplane in a multidimensional feature space mapped by a kernel function, in which all the computations are completed in the original feature space. The redistributed data enable the fitting of a linear hyperplane between the training samples of two classes. The optimization problem attempts to maximize the margins between the hyperplane and the closest training samples (i.e., the support vectors) and minimize the error of the training samples that cannot be differentiated (Vapnik, 1998). The influence of the nonseparable samples is controlled by a regularization parameter C. A detailed introduction to the general concept of SVMs is given by Burges (1998) and Schölkopf and Smola (2002). A brief discussion is given as follows. Let x i d (i =1,2,, L) for a binary classification problem in a d-dimensional feature space d a training set of L samples with the corresponding class labels y i {1, 1}. The separating hyperplane f(x) is described by a normal vector w d and the bias b, where b / w is the distance between the hyperplane and the origin, with w as the Euclidean norm from w: f(x) =w x + b (1) The support vectors are located on two hyperplanes w x + b =±1 that are parallel to the separating hyperplane. The approach aims at maximizing the margins, between the closest samples and the hyperplane, which leads to the following optimization problem: min w 2 L C i 2 + ξ (2) i= 1 where ξ i are slack variables and C is a regularization parameter, which are used to handle misclassified samples in nonseparable cases. The parameter C is used as a penalty for samples that lie on the wrong side of the hyperplane. In fact, it controls the shape of the solution and consequently affects the generalization capability of the classifier. Thus, a large value of C might cause an overfitting to the training data. The linear SVM approach is extended for nonlinear separable cases by socalled kernel methods. During nonlinear mapping the data are transferred by a kernel function into a higher dimensional feature space. The final hyperplane decision function can be defined as L f( x) α iyik( xi, x j) + b i= 1 where α i are Lagrange multipliers and k denotes the kernel function. (3)

4 Canadian Journal of Remote Sensing / Journal canadien de télédétection Figure 1. Overview of the location of the study site, covering approximately 41 km 12 km. AVIRIS dataset (the red, green, and blue bands (RGB) are represented by principal components PC1, PC2, and PC3, respectively) and locations of ground-truth information for the 22 classes. A widely used kernel function is the Gaussian radial basis function (RBF) kernel (Vapnik, 1998; Schölkopf and Smola, 2002), which is given by k(x i, x j ) = exp[ γ x i x j 2 ] (4) The SVM training requires the estimation of the kernel parameter γ and the regularization parameter C. The two parameters are usually determined by a grid search, testing possible combinations of C and γ in a user-defined range. Lastly, the best combinations for γ and C were selected based on cross-validation. In contrast to DT classifiers that directly provide a final class membership and the well-known maximum likelihood classifier, which is based on likelihood of classes, the output images of SVM provided by Equation (3) contain distances between each pixel and the separating hyperplane. These distance images can then be used to determine the final class membership, using a simple majority vote and the absolute distances. SVM is a binary method in that a single SVM just separates two classes. However, remote sensing applications normally deal with multiclass problems, and there are two main strategies for multiclass extensions of SVMs. A classification problem with c classes is divided into several binary subproblems. The one-against-one (OAO) method generates a set of c(c 1)/2 individual SVM classifiers, one for each possible pair of classes. A majority vote is applied to the c(c 1)/2 rule images to compute the final class membership. In contrast, the one-against-all (OAA) strategy is based on c binary classifiers, separating each class from the remaining ones. The absolute maximum of Equation (3) defines the final class label. Alternatively, the SVM classifier can be directly described as a multiclass problem (Hsu and Lin, 2002) that requires fewer support vectors and seems adequate to derive accurate S109

5 Vol. 35, Suppl. 1, 2009 Table 1. Land cover classes and number of samples within training set tr400 and the test set. Class classification accuracies with a small sample set (Mathur and Foody, 2004). Random forests The strategy of a classifier ensemble, such as RF, is based on the assumption that independent classifiers produce individual errors, which are not produced by the majority of other classifiers. In the RF, each individual DT within the ensemble is trained on a bootstrapped sample (i.e., subsets of training data) of the original training samples. This approach is based on random selection of training samples from the whole training sample set with replacement. The method randomly selects n samples from a training sample set of same size n. Because the selection is done with replacement, some training samples can be selected several times for a specific training set, whereas other samples are not considered in this particular sample set. In contrast, the random selection without replacement selects m samples out of a set with size n, where m < n. Thereby, each training sample can be selected only once in each sample subset. In addition, the split rule at each node of a DT is derived using only a randomly selected feature subset of the input data. Lastly, a majority vote is used to create the final classification result. The computational complexity is simplified by the reduction of the number of input features at each internal node. This enables RF to deal with high-dimensional datasets, and the algorithm is generally computationally lighter than conventional ensemble strategies (Gislason et al., 2006). S110 No. of samples Training set tr400 Test set Andesite lava Andesite lava 1980 I Andesite lava 1980 II Andesite lava 1991 I Andesite lava 1991 II Andesite lava with birch bushes Andesite lava with sparse moss cover Andesite lava with thick moss cover Old unvegetated andesite lava I Old unvegetated andesite lava II Old unvegetated andesite lava III Hyaloclastite formation I Hyaloclastite formation II Hyaloclastite formation III Hyaloclastite formation IV Lava covered with tephra and scoria Lichen covered basalt lava Rhyolite Scoria Volcanic tephra Firn and glacier ice Snow Different methods exist for the generation of the DT classifiers which usually are based on the measurement of impurity of the data within potential nodes. The RF algorithm is based on the Gini index of Breiman (2001). This measurement describes the impurity for a given node. The method aims to find the largest homogeneous subclass within the training samples and discriminate it from the remaining training dataset (Zambon et al., 2006). The Gini index is summed and the split rule that results in a maximum reduction of impurity is used as a decision rule (Duda et al., 2001). The Gini index is given by c ωi i= 1 Gini( t) = p ( 1 p ) ω i where c is the number of classes, and pω i is the probability or the relative frequency of class ω i at node t and is defined as p ω i (5) lωi = (6) L in which lω is the number of samples belonging to class ω i i, and L is the total number of training samples. Methods To simplify the parameter selection for the SVM, the imagery was scaled between 0 and 1 before training. The OAO approach was used to solve the multiclass problem using a set of binary classifiers with optimized parameters for γ and C. The training of the SVM with a Gaussian kernel and the classification were performed using imagesvm (available from (Janz et al., 2007). ImageSVM is a freely available IDL/ENVI implementation that uses the LIBSVM approach (Chang and Lin, 2001) for the training of the SVM. The kernel parameters C and γ are determined by a grid search, using a three-fold cross-validation. Possible combinations of C and γ are tested in a user-defined range, and the best parameter combinations were used for the final classification. Freely available Fortran code was used for the RF classification (available from ~breiman/randomforests/). To investigate the influence of the number of iterations (i.e., number of DTs within the ensemble), 25, 50, 75, and 100 iterations were performed. The size of the feature subset, which is user defined, is set to the square root of the number of input features (i.e., 13), which has shown good results in previous studies (Gislason et al., 2006; Cheung-Wai Chan and Paelinckx, 2008). Both classifiers were trained with the same training sample set, containing 100, 200, and up to 400 randomly selected samples per class. For comparison, SAM and an MLC were applied to the dataset using both training datasets. For SAM, the spectra of the training samples of each lithological unit class were averaged to construct a final class spectrum.

6 Canadian Journal of Remote Sensing / Journal canadien de télédétection The classification results were evaluated by a detailed accuracy assessment and visual interpretation. Accuracy assessment was performed using overall accuracies and confusion matrices to derivate the producer s and user s accuracies. In addition, the kappa coefficient of agreement was derived. For a statistical-based comparison of the produced maps, the significance of differences between two different maps was assessed using a McNemar test with a 95% confidence interval. This nonparametric test, which is based on the standardized normal test statistics, is defined as follows: z = f f f + f where f 12 is the number of samples correctly classified by classifier 1 but misclassified by classifier 2, and f 21 is the number of samples misclassified by classifier 1 but correctly classified by classifier 2 (Foody, 2004). Besides the McNemar test, the maps produced by RF and SVM were compared by Spearman s rank correlation coefficient. Classification results Accuracy assessment Comparing the test areas, the accuracy assessment demonstrates that both machine-learning algorithms perform very accurately and achieve overall accuracies of 94.7% (SVM) and 88.7% (RF) using a small training dataset from 100 samples per class (see Table 2). In contrast, the SAM classifier results in significantly lower accuracy, and the MLC cannot be applied in this case. 2 As expected, the more accurate results are generated by a larger training set, resulting in overall accuracies of 96.9% (SVM) and 90.7% (RF). Comparison of these results with the accuracies achieved by the MLC (67.7%) and SAM (69.7%) indicates that the two machine-learning algorithms SVM and RF significantly outperform traditional methods in terms of overall accuracy. The inclusion of additional samples in the training process (i.e., tr400) further increases the classification accuracy. However, most obvious is that the accuracy achieved by the MLC is increased by up to 13% by increasing the size of the training set. On the other hand, the accuracies of the SVM and RF increased by only 1.0% and 2.5%, respectively. Compared with the overall accuracies, the kappa coefficients show similar characteristics (see Table 2). In addition, a maximum likelihood classification was performed using a dimensionality-reduced dataset. A conventional principal component analysis was applied to the original image. One hundred bands (principal components, PCs) were selected, which explain approximately 99% of the total variance of the dataset. Whereas the overall accuracy achieved with the largest training sample set remains similar to the accuracy achieved on the whole dataset, the accuracy significantly increased up to 88% using the training (7) set with 200 samples per class. Similar results were achieved when the training set with 100 samples per class was applied to selected principal components. These results support the (wellknown) dependency of the MLC approach on the number of training samples (with regard to the dimension of the dataset). The classification results achieved by SVM and RF were compared using the McNemar test with a 95% confidence interval. The test demonstrates that the two methods provide significantly different results. The values of Spearman s rank correlation coefficient ρ (between 0.57 and 0.65) underline this observation. However, these comparisons need to be investigated in more detail. To investigate the impact of the number of iterations within the RF (i.e., number of individual DTs in the ensemble) on the classification accuracy, classifications with 25, 50, 75, and 100 iterations were performed. Twenty-five iterations within the RF ensemble resulted in 89.5% accuracy when compared with that for the test areas (Table 3), and additional iterations within the ensemble have smaller and nonsignificant effects on the overall accuracy. The producer s and user s accuracies demonstrate the balanced performance between the two machine-learning techniques in terms of accuracy (Table 4). The two classes Firn and glacier ice and Snow are classified most accurately, irrespective of the algorithms. One reason for this is the strong difference between these two classes in comparison with other classes within the spectral feature space. The SVM and the RF perform weaker on the four land cover types Old unvegetated andesite lava III, Hyaloclastite formation I, Hyaloclastite formation III, and Scoria in terms of classification accuracy, resulting in producer s accuracies between 86.6% and 91.6%. These values are much lower than those achieved for the other classes (96.5% on average). The producer s accuracies for these difficult classes achieved by the RF are even lower than those with the SVM (e.g., 61.7% for Hyaloclastite formation III ). In addition, the differences between the producer s and user s accuracies were calculated to evaluate the balance between the two measurements (Figure 2). Whereas many classes show balanced accuracies, some classes are strongly biased, particularly towards the producer s accuracies. The balance between Old unvegetated andesite lava III and Hyaloclastite formation I is highest for the RF, whereas SVM cause the highest differences for Andesite lava 1990 II and Hyaloclastite formation III. Moreover, the results produced by the RF seem to be more imbalanced than the accuracies achieved by the SVM. One reason for the lower and imbalanced accuracies might be the similarity between the spectral responses of different classes, which results in confusion between these lithological units. The spectra of hyaloclastite units, as Hyaloclastite formation I and Old unvegetated andesite lava III, appear similar; consequently, differentiation between these two classes is more difficult. 2 The MLC requires a theoretical limit of (number of bands + 1) training pixels per class to avoid a singular covariance matrix. S111

7 Vol. 35, Suppl. 1, 2009 Table 2. Overall accuracy and kappa coefficient using different classifier algorithms and training datasets. tr100 tr200 tr400 Classifier Accuracy (%) Kappa Accuracy (%) Kappa Accuracy (%) Kappa SVM RF MLC SAM Visual interpretation The visual interpretation supports the results of the statistical accuracy assessment; the maps appear similar in many regions (Figure 3) but also show some differences. Particularly large homogeneous regions do not show significant differences. However, the main discrepancy appears to be that some linear structures of the class Volcanic tephra are more distinct on the RF map than on the SVM map. However, it is assumed that the linear features are related to artifacts. Parallel linear features of Hyaloclastite formation III and Old unvegetated andesite lava II are visible in the southeast of the map (i.e., lower right-hand corner). The two maps are significantly different in this region, and in this case the structures appear more distinct in the SVM result than in the RF classification. In addition, some differences are visible in smaller regions. Discussion and conclusion The performance of two algorithms (support vector machines (SVMs) and random forests (RFs) from the field of machine learning) on the classification of hyperspectral imagery was assessed in this paper. Generally, SVMs and RFs have the potential to be more accurate in the classification of hyperspectral data than the maximum likelihood classifier because of the limited number of training samples that are usually available and the high dimensionality of the dataset. The low accuracy achieved by the spectral angle mapper (SAM) for the current study area is not necessarily a general limitation of the method. SAM has performed well in several studies, particularly when applied to spectroscopy data. Moreover, the approach can be improved, e.g., by using spectral libraries. Nevertheless, for a typical supervised classification approach, the two proposed algorithms may be superior because they provide better results independent of the size of the training sample sets. The high producer s and user s accuracies underline the good performance, and on average the individual lithological units are classified with average accuracies of approximately 90%. As mentioned previously, the methods perform weaker on some land cover types (e.g., S112 Table 3. Overall accuracy using the RF with a different number of iterations and training set tr200. No. of iterations Accuracy (%) Old unvegetated andesite lava III, Hyaloclastite formation I, Hyaloclastite formation III, and Scoria ) in terms of classification accuracy, resulting in values that are much lower than the average accuracies achieved for the other classes. One reason for this observation might be the similarity between the spectral responses of various classes, which results in confusion between different classes. In terms of the overall and class accuracies, SVMs outperform RFs. Moreover, SVMs achieve more balanced producer s and user s accuracies. Nevertheless, the RFs achieve accurate results with overall accuracies of 90% and higher. The differences in the produced maps were also underlined by the McNemar test and Spearman s rank correlation coefficients. However, it is interesting to note that the two classifiers do not generate identical results, and the diversity between the two classifiers is interesting. Diversity is the basic assumption of an adequate classifier combination, whereas combining similar classification results would not further improve the overall accuracy. Thus it could be interesting to combine both algorithms to further increase the classification accuracy (Waske and van der Linden, 2008). The main reason for the discrepancies between the maps may be the two different classifier concepts. SVM classifiers are based on a few relatively complex decision boundaries. In linear nonseparable cases, the data are transferred into a higher dimensional feature space that enables the generation of a decision boundary that appears nonlinear in the original feature space. In contrast, the RF approach is based on several DT classifiers, which construct many rather simple decision boundaries that are parallel to the feature axis. On the whole, the two classification approaches are useful for classification of hyperspectral data. They perform well, even with small training sample sets. This observation is particularly important for classification of remote and isolated environments with difficult access and lack of infrastructure (i.e., northern environments). In such regions, fieldwork is often limited (and very costly), and mapping strategies that are based on limited ground-truth data seem favorable. Adequate sampling strategies can further decrease the number of required training samples (Foody et al., 2006). Although the SVM results in higher overall accuracies, the RF is also attractive as an accurate and simple approach. The RF is easy to handle because it mainly depends on only two user-defined values (i.e., the number of selected features and the number of iterations). On the other hand, the parameters for the SVM cannot be as simply defined as those for the RF, but an automatic grid search in the SVM case enables

8 Canadian Journal of Remote Sensing / Journal canadien de télédétection Table 4. Producer s and user s accuracies using SVM and RF (100 iterations) with training set tr200. Class SVM RF Producer s accuracy (%) User s accuracy (%) Producer s accuracy (%) Andesite lava Andesite lava 1980 I Andesite lava 1980 II Andesite lava 1991 I Andesite lava 1991 II Andesite lava with birch bushes Andesite lava with sparse moss cover Andesite lava with thick moss cover Old unvegetated andesite lava I Old unvegetated andesite lava II Old unvegetated andesite lava III Hyaloclastite formation I Hyaloclastite formation II Hyaloclastite formation III Hyaloclastite formation IV Lava with tephra and scoria Lichen covered basalt lava Rhyolite Scoria Volcanic tephra Firn and glacier ice Snow User s accuracy (%) Figure 2. Differences between producer s and user s accuracy generated by the SVM and RF classifications. S113

9 Vol. 35, Suppl. 1, 2009 Figure 3. Classification results achieved by (a) SVM and (b) RF. parameter selection. In conclusion, both classification algorithms are useful for discriminating different volcanic units in Iceland based on hyperspectral datasets. It is anticipated that these classification techniques will also be appropriate for most northern environments where hyperspectral data are used for mapping lithology. Acknowledgements This work was partially funded by the Research Fund of the University of Iceland and the Icelandic Research Fund. We appreciate the comments of the anonymous reviewers who helped us significantly improve the paper. S114 References Benediktsson, J.A., and Kanellopoulos, I Classification of multisource and hyperspectral data based on decision fusion. IEEE Transactions on Geoscience and Remote Sensing, Vol. 37, No. 3, pp Benediktsson, J.A., and Swain, P.H Consensus theoretic classification methods. IEEE Transactions on Systems, Man and Cybernetics, Vol. 22, No. 4, pp Benediktsson, J.A., Chanussot, J., and Fauvel, M Multiple classifier systems in remote sensing: from basics to recent developments. In Multiple Classifier Systems, Proceedings of the 7th International Workshop, MCS 2007, Prague. Lecture Notes in Computer Science Edited by M. Haindl, J. Kittler, and F. Roli. Springer, Berlin.

10 Canadian Journal of Remote Sensing / Journal canadien de télédétection Breiman, L Random forests. Machine Learning, Vol. 45, No. 1, pp Briem, G.J., Benediktsson, J.A., and Sveinsson, J.R Multiple classifiers applied to multisource remote sensing data. IEEE Transactions on Geoscience and Remote Sensing, Vol. 40, No. 10, pp Burges, C.J.C A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, Vol. 2, pp Chang, C.-C., and Lin, C.-J LIBSVM: a library for support vector machines. Available at Chen, X., Warner, T.A., and Campagna, D.J Integrating visible, nearinfrared and short-wave infrared hyperspectral and multispectral thermal imagery for geological mapping at Cuprite, Nevada. Remote Sensing of Environment, Vol. 110, No. 3, pp Cheung-Wai Chan, J., and Paelinckx, D Evaluation of Random Forest and Adaboost tree-based ensemble classification and spectral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sensing of Environment, Vol. 112, No. 6, pp Clark, R.N Spectroscopy of rocks and minerals, and principles of spectroscopy, remote sensing for the Earth sciences. In Manual of remote sensing. Edited by A.N. Rencz. 3rd ed. John Wiley and Sons, New York. Vol. 3, pp Dennison, P.E., Halligan, K.Q., and Roberts, D.A A comparison of error metrics and constraints for multiple endmember spectral mixture analysis and spectral angle mapper. Remote Sensing of Environment, Vol. 93, No. 3, pp Duda, R.O., Hart, P.E., and Stork, D.G. (Editors) Pattern classification. 2nd ed. John Wiley & Sons Inc., Chichester, N.Y. Fauvel, M., Chanussot, J., and Benediktsson, J.A Evaluation of kernels for multiclass classification of hyperspectral remote sensing data. In ICASSP 2006, Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, May 2006, Toulouse, France. IEEE, New York. Vol. 2, pp Foody, G.M Thematic map comparison: evaluating the statistical significance of differences in classification accuracy. Photogrammetric Engineering & Remote Sensing, Vol. 70, pp Foody, G.M., and Mathur, A A relative evaluation of multiclass image classification of support vector machines. IEEE Transactions on Geoscience and Remote Sensing, Vol. 42, pp Foody, G.M., Mathur, A., Sanchez-Hernandez, C., and Boyd, D.S Training set size requirements for the classification of a specific class. Remote Sensing of Environment, Vol. 104, No. 1, pp Friedl, M.A., and Brodley, C.E Decision tree classification of land cover from remotely sensed data. Remote Sensing of Environment, Vol. 61, No. 3, pp Green, R.O., Eastwood, M.L., Sarture, C.M., Chrien, T.G., Aronsson, M., Chippendale, B.J., Faust, J.A., Pavri, B.E., Chovit, C.J., Solis, M., Olah, M.R., and Williams, O Imaging spectroscopy and the airborne visible/infrared imaging spectrometer (AVIRIS). Remote Sensing of Environment, Vol. 65, No. 3, pp Gislason, P.O., Benediktsson, J.A., and Sveinsson, J.R Random Forests for land cover classification. Pattern Recognition Letters, Vol. 27, pp Goetz, A.F.H., Rowan, L.C., and Kingston, M.J Mineral identification from orbit: initial results from the shuttle multispectral infrared radiometer. Science (Washington, D.C.), Vol. 218, pp Ham, J., Chen, Y., and Crawford, M.M Investigation of the Random Forest framework for classification of hyperspectral data. IEEE Transactions on Geoscience and Remote Sensing, Vol. 43, No. 3, pp Harken, J., and Sugumaran, R Classification of Iowa wetlands using an airborne hyperspectral image: a comparison of the spectral angle mapper classifier and an object-oriented approach. Canadian Journal of Remote Sensing, Vol. 31, No. 2, pp Harris, J.R., Rogge, D., Hitchcock, R., Ijewliw, O., and Wright, D Mapping lithology in Canada s Arctic: application of hyperspectral data using the minimum noise fraction transformation and matched filtering. Canadian Journal of Earth Sciences, Vol. 42, No. 12, pp Hsu, C.-W., and Lin, C.-J A comparison of methods for multi-class support vector machines. IEEE Transactions on Neural Networks, Vol. 13, No. 2, pp Huang, C., Davis, L.S., and Townshend, J.R An assessment of support vector machines for land cover classification. International Journal of Remote Sensing, Vol. 23, No. 4, pp Hughes, G.F On the mean accuracy of statistical pattern recognition. IEEE Transactions on Information Theory, Vol. 14, pp Janz, A., Schiefer, S., Waske, B., and Hostert, P imagesvm A useroriented tool for advanced classification of hyperspectral data using support vector machines. In Imaging Spectroscopy: Innovation in Environmental Research: Proceedings of the 5th Workshop of the EARSeL Special Interest Group on Imaging Spectroscopy, April 2007, Bruges, Belgium. EARSeL Secretariat, Hannover, Germany. Kruse, F.A Use of airborne imaging spectrometer data to map minerals associated with hydrothermally altered rocks in the northern Grapevine Mountains, Nevada and California: Remote Sensing of Environment, Vol. 24, No. 1, pp Kruse, F.A., Lefkoff, A.B., Boardman, J.W., Heidebrecht, K.B., Shapiro, A.T., Barloon, P.J., and Goetz, A.F.H The Spectral Image Processing System (SIPS). Interactive visualization and analysis of imaging spectrometer data. Remote Sensing of Environment, Vol. 44, No. 2 3, pp Mathur, A., and Foody, G.M Multiclass and binary SVM classification: implications for training and classification users. IEEE Geoscience and Remote Sensing Letters, Vol. 5, No. 2, pp Melgani, F., and Bruzzone, L Classification of hyperspectral remote sensing images with support vector machines. IEEE Transactions on Geoscience and Remote Sensing, Vol. 42, No. 8, pp Pal, M., and Mather, P.M Some issues in the classification of DAIS hyperspectral data. International Journal of Remote Sensing, Vol. 27, No. 14, pp Polikar, R Ensemble based systems in decision making. IEEE Circuits and Systems Magazine, Vol. 6, pp Richards, J.A., and Jia, X Remote sensing digital image analysis: an introduction. Springer, New York. Safavian, S.R., and Landgrebe, D.A A survey of decision tree classifier methodology. IEEE Transactions on Systems, Man, and Cybernetics, Vol. 21, pp Sahoo, B.C., Oommen, T., Misra, D., and Newby, G Using the onedimensional S-transform as a discrimination tool in classification of hyperspectral images. Canadian Journal of Remote Sensing, Vol. 33, No. 6, pp S115

11 Vol. 35, Suppl. 1, 2009 Schölkopf, B., and Smola, A Learning with kernels. MIT Press, Cambridge, Mass. Soosalu, H., Einarsson, P., and Jakobsdóttir S Volcanic tremor related to the 1991 eruption of the Hekla volcano, Iceland. Bulletin of Volcanology, Vol. 65, No. 8, pp Vapnik, V.V Statistical learning theory. Wiley, New York. van der Linden, S., Waske, B., and Hostert, P. 2007a. Towards an optimized use of the spectral angle space. In Imaging Spectroscopy: Innovation in Environmental Research: Proceedings of the 5th Workshop of the EARSeL Special Interest Group on Imaging Spectroscopy, April 2007, Bruges, Belgium. EARSeL Secretariat, Hannover, Germany. van der Linden, S., Janz, A., Waske, B., Eiden, M., and Hostert, P. 2007b. Classifying segmented hyperspectral data from a heterogeneous urban environment using support vector machines. Journal of Applied Remote Sensing [online], Vol. 1, No van der Meer, F., and de Jong, S.M Imaging spectrometry: basic principles and prospective applications. Kluwer Academic Publishers, London, UK. Waske, B., and Benediktsson, J.A Fusion of support vector machines for classification of multisensor data. IEEE Transactions on Geoscience and Remote Sensing, Vol. 45, No. 12, pp Waske, B., and van der Linden, S Classifying multilevel imagery from SAR and optical sensors by decision fusion. IEEE Transactions on Geoscience and Remote Sensing, Vol. 46, No. 5, pp Zambon, M., Lawrence, R., Bunn, A., and Powell, S Effect of alternative splitting rules on image processing using classification tree analysis. Photogrammetric Engineering & Remote Sensing, Vol. 72, pp S116

Hyperspectral image classification using Support Vector Machine

Hyperspectral image classification using Support Vector Machine Journal of Physics: Conference Series OPEN ACCESS Hyperspectral image classification using Support Vector Machine To cite this article: T A Moughal 2013 J. Phys.: Conf. Ser. 439 012042 View the article

More information

SECOND ORDER STATISTICS FOR HYPERSPECTRAL DATA CLASSIFICATION. Saoussen Bahria and Mohamed Limam

SECOND ORDER STATISTICS FOR HYPERSPECTRAL DATA CLASSIFICATION. Saoussen Bahria and Mohamed Limam Manuscrit auteur, publié dans "42èmes Journées de Statistique (2010)" SECOND ORDER STATISTICS FOR HYPERSPECTRAL DATA CLASSIFICATION Saoussen Bahria and Mohamed Limam LARODEC laboratory- High Institute

More information

Spectral and Spatial Methods for the Classification of Urban Remote Sensing Data

Spectral and Spatial Methods for the Classification of Urban Remote Sensing Data Spectral and Spatial Methods for the Classification of Urban Remote Sensing Data Mathieu Fauvel gipsa-lab/dis, Grenoble Institute of Technology - INPG - FRANCE Department of Electrical and Computer Engineering,

More information

DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION

DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA Shizhi Chen and YingLi Tian Department of Electrical Engineering The City College of

More information

Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion

Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion Journal of Advances in Information Technology Vol. 8, No. 1, February 2017 Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion Guizhou Wang Institute of Remote Sensing

More information

Hyperspectral Data as a Tool for Mineral Exploration

Hyperspectral Data as a Tool for Mineral Exploration 1 Hyperspectral Data as a Tool for Mineral Exploration Nahid Kavoosi, PhD candidate of remote sensing kavoosyn@yahoo.com Nahid Kavoosi Abstract In Geology a lot of minerals and rocks have characteristic

More information

1 University of Portland Portland OR

1 University of Portland Portland OR Effect Of -To-Reflectance Transformation And Atmosphere Removal On Maximum ikelihood Classification Accuracy Of High-Dimensional Remote Sensing Data Joseph P. Hoffbeck and David A. andgrebe 2 University

More information

LAND COVER classification is one of the widest used

LAND COVER classification is one of the widest used 3858 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 45, NO. 12, DECEMBER 2007 Fusion of Support Vector Machines for Classification of Multisensor Data Björn Waske, Student Member, IEEE, and Jón

More information

RVM-based multi-class classification of remotely sensed data

RVM-based multi-class classification of remotely sensed data RVM-based multi-class classification of remotely sensed data Foody, G. M. International Journal of Remote Sensing, 29, 1817-1823 (2008). The manuscript of the above article revised after peer review and

More information

HYPERSPECTRAL IMAGING

HYPERSPECTRAL IMAGING 1 HYPERSPECTRAL IMAGING Lecture 9 Multispectral Vs. Hyperspectral 2 The term hyperspectral usually refers to an instrument whose spectral bands are constrained to the region of solar illumination, i.e.,

More information

Chemometrics: Classification of spectra

Chemometrics: Classification of spectra Chemometrics: Classification of spectra Vladimir Bochko Jarmo Alander University of Vaasa November 1, 2010 Vladimir Bochko Chemometrics: Classification 1/36 Contents Terminology Introduction Big picture

More information

Classification of Pavia University Hyperspectral Image using Gabor and SVM Classifier

Classification of Pavia University Hyperspectral Image using Gabor and SVM Classifier Kavitha K, Arivazhagan S and Suriya B 9 Classification of Pavia University Hyperspectral Image using Gabor and SVM Classifier Kavitha K. 1, Arivazhagan S. 2, Suriya B. 3 Abstract: Hyperspectral remote

More information

Support Vector Machine & Its Applications

Support Vector Machine & Its Applications Support Vector Machine & Its Applications A portion (1/3) of the slides are taken from Prof. Andrew Moore s SVM tutorial at http://www.cs.cmu.edu/~awm/tutorials Mingyue Tan The University of British Columbia

More information

Vegetation Change Detection of Central part of Nepal using Landsat TM

Vegetation Change Detection of Central part of Nepal using Landsat TM Vegetation Change Detection of Central part of Nepal using Landsat TM Kalpana G. Bastakoti Department of Geography, University of Calgary, kalpanagb@gmail.com Abstract This paper presents a study of detecting

More information

Support Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Support Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI Support Vector Machines II CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Outline Linear SVM hard margin Linear SVM soft margin Non-linear SVM Application Linear Support Vector Machine An optimization

More information

A Tutorial on Support Vector Machine

A Tutorial on Support Vector Machine A Tutorial on School of Computing National University of Singapore Contents Theory on Using with Other s Contents Transforming Theory on Using with Other s What is a classifier? A function that maps instances

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014 Exam policy: This exam allows two one-page, two-sided cheat sheets (i.e. 4 sides); No other materials. Time: 2 hours. Be sure to write

More information

KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE. Israel -

KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE. Israel - KNOWLEDGE-BASED CLASSIFICATION OF LAND COVER FOR THE QUALITY ASSESSEMENT OF GIS DATABASE Ammatzia Peled a,*, Michael Gilichinsky b a University of Haifa, Department of Geography and Environmental Studies,

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Tobias Pohlen Selected Topics in Human Language Technology and Pattern Recognition February 10, 2014 Human Language Technology and Pattern Recognition Lehrstuhl für Informatik 6

More information

Jeff Howbert Introduction to Machine Learning Winter

Jeff Howbert Introduction to Machine Learning Winter Classification / Regression Support Vector Machines Jeff Howbert Introduction to Machine Learning Winter 2012 1 Topics SVM classifiers for linearly separable classes SVM classifiers for non-linearly separable

More information

Deriving Uncertainty of Area Estimates from Satellite Imagery using Fuzzy Land-cover Classification

Deriving Uncertainty of Area Estimates from Satellite Imagery using Fuzzy Land-cover Classification International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 10 (2013), pp. 1059-1066 International Research Publications House http://www. irphouse.com /ijict.htm Deriving

More information

Linear Classification and SVM. Dr. Xin Zhang

Linear Classification and SVM. Dr. Xin Zhang Linear Classification and SVM Dr. Xin Zhang Email: eexinzhang@scut.edu.cn What is linear classification? Classification is intrinsically non-linear It puts non-identical things in the same class, so a

More information

IDENTIFICATION OF TARGETS FOR COPPER MINERALIZATION USING MULTI-SPECTRAL ANALYSIS OF ASTER SATELLITE IMAGES IN HASHTIJAN 1: SHEET, NW IRAN

IDENTIFICATION OF TARGETS FOR COPPER MINERALIZATION USING MULTI-SPECTRAL ANALYSIS OF ASTER SATELLITE IMAGES IN HASHTIJAN 1: SHEET, NW IRAN IDENTIFICATION OF TARGETS FOR COPPER MINERALIZATION USING MULTI-SPECTRAL ANALYSIS OF ASTER SATELLITE IMAGES IN HASHTIJAN 1:100000 SHEET, NW IRAN Ramin ArameshAsl 1, Peyman Afzal 1,2, Ahmad Adib 1, Naji

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines Nonlinear Support Vector Machines through Iterative Majorization and I-Splines P.J.F. Groenen G. Nalbantov J.C. Bioch July 9, 26 Econometric Institute Report EI 26-25 Abstract To minimize the primal support

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Support Vector Machine Regression for Volatile Stock Market Prediction

Support Vector Machine Regression for Volatile Stock Market Prediction Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,

More information

1. Introduction. S.S. Patil 1, Sachidananda 1, U.B. Angadi 2, and D.K. Prabhuraj 3

1. Introduction. S.S. Patil 1, Sachidananda 1, U.B. Angadi 2, and D.K. Prabhuraj 3 Cloud Publications International Journal of Advanced Remote Sensing and GIS 2014, Volume 3, Issue 1, pp. 525-531, Article ID Tech-249 ISSN 2320-0243 Research Article Open Access Machine Learning Technique

More information

Analysis of Multiclass Support Vector Machines

Analysis of Multiclass Support Vector Machines Analysis of Multiclass Support Vector Machines Shigeo Abe Graduate School of Science and Technology Kobe University Kobe, Japan abe@eedept.kobe-u.ac.jp Abstract Since support vector machines for pattern

More information

Robust Multiple Estimator Systems for the Analysis of Biophysical Parameters From Remotely Sensed Data

Robust Multiple Estimator Systems for the Analysis of Biophysical Parameters From Remotely Sensed Data IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 43, NO. 1, JANUARY 2005 159 Robust Multiple Estimator Systems for the Analysis of Biophysical Parameters From Remotely Sensed Data Lorenzo Bruzzone,

More information

Discriminative Direction for Kernel Classifiers

Discriminative Direction for Kernel Classifiers Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering

More information

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet. CS 189 Spring 013 Introduction to Machine Learning Final You have 3 hours for the exam. The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet. Please

More information

Online publication date: 22 January 2010 PLEASE SCROLL DOWN FOR ARTICLE

Online publication date: 22 January 2010 PLEASE SCROLL DOWN FOR ARTICLE This article was downloaded by: On: 29 January 2010 Access details: Access Details: Free Access Publisher Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered

More information

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane

More information

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights Linear Discriminant Functions and Support Vector Machines Linear, threshold units CSE19, Winter 11 Biometrics CSE 19 Lecture 11 1 X i : inputs W i : weights θ : threshold 3 4 5 1 6 7 Courtesy of University

More information

Support Vector Machine. Industrial AI Lab.

Support Vector Machine. Industrial AI Lab. Support Vector Machine Industrial AI Lab. Classification (Linear) Autonomously figure out which category (or class) an unknown item should be categorized into Number of categories / classes Binary: 2 different

More information

Lênio Soares Galvão Raimundo Almeida Filho Ícaro Vitorello

Lênio Soares Galvão Raimundo Almeida Filho Ícaro Vitorello Use of ASTER short-wave infrared bands for the spectral discrimination of hydrothermally altered-materials: evaluation in a tropical savannah environment Lênio Soares Galvão Raimundo Almeida Filho Ícaro

More information

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring / Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring 2015 http://ce.sharif.edu/courses/93-94/2/ce717-1 / Agenda Combining Classifiers Empirical view Theoretical

More information

Deep Learning. Convolutional Neural Networks Applications

Deep Learning. Convolutional Neural Networks Applications Deep Learning Using a Convolutional Neural Network Dr. Ing. Morris Riedel Adjunct Associated Professor School of Engineering and Natural Sciences, University of Iceland Research Group Leader, Juelich Supercomputing

More information

Multivariate statistical methods and data mining in particle physics Lecture 4 (19 June, 2008)

Multivariate statistical methods and data mining in particle physics Lecture 4 (19 June, 2008) Multivariate statistical methods and data mining in particle physics Lecture 4 (19 June, 2008) RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement

More information

SVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION

SVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION International Journal of Pure and Applied Mathematics Volume 87 No. 6 2013, 741-750 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v87i6.2

More information

Support Vector Machine (continued)

Support Vector Machine (continued) Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need

More information

Testing the Sensitivity of Vegetation Indices for Crop Type Classification Using Rapideye Imagery

Testing the Sensitivity of Vegetation Indices for Crop Type Classification Using Rapideye Imagery Testing the Sensitivity of Vegetation Indices for Crop Type Classification Using Rapideye Imagery Mustafa USTUNER and Fusun BALIK SANLI, Turkey Keywords: Remote Sensing, Agriculture, Geoinformation SUMMARY

More information

A Method to Improve the Accuracy of Remote Sensing Data Classification by Exploiting the Multi-Scale Properties in the Scene

A Method to Improve the Accuracy of Remote Sensing Data Classification by Exploiting the Multi-Scale Properties in the Scene Proceedings of the 8th International Symposium on Spatial Accuracy Assessment in Natural Resources and Environmental Sciences Shanghai, P. R. China, June 25-27, 2008, pp. 183-188 A Method to Improve the

More information

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction

More information

Machine Learning Lecture 7

Machine Learning Lecture 7 Course Outline Machine Learning Lecture 7 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Statistical Learning Theory 23.05.2016 Discriminative Approaches (5 weeks) Linear Discriminant

More information

PATTERN CLASSIFICATION

PATTERN CLASSIFICATION PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Andreas Maletti Technische Universität Dresden Fakultät Informatik June 15, 2006 1 The Problem 2 The Basics 3 The Proposed Solution Learning by Machines Learning

More information

FUSION OF IMAGING SPECTROMETER AND LIDAR DATA USING SUPPORT VECTOR MACHINES FOR LAND COVER CLASSIFICATION IN THE CONTEXT OF FOREST FIRE MANAGEMENT

FUSION OF IMAGING SPECTROMETER AND LIDAR DATA USING SUPPORT VECTOR MACHINES FOR LAND COVER CLASSIFICATION IN THE CONTEXT OF FOREST FIRE MANAGEMENT FUSION OF IMAGING SPECTROMETER AND LIDAR DATA USING SUPPORT VECTOR MACHINES FOR LAND COVER CLASSIFICATION IN THE CONTEXT OF FOREST FIRE MANAGEMENT B. Koetz a,*, F. Morsdorf a, T. Curt b, S. van der Linden

More information

Proceedings - AutoCarto Columbus, Ohio, USA - September 16-18, Dee Shi and Xiaojun Yang

Proceedings - AutoCarto Columbus, Ohio, USA - September 16-18, Dee Shi and Xiaojun Yang Support Vector Machines for Landscape Mapping from Remote Sensor Imagery Dee Shi and Xiaojun Yang ABSTRACT: Landscape mapping is an important activity leading to the generation of various geospatial products

More information

Constrained Optimization and Support Vector Machines

Constrained Optimization and Support Vector Machines Constrained Optimization and Support Vector Machines Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Machine Perceptual Learning and Sensory Summer Augmented 15 Computing Many slides adapted from B. Schiele Machine Learning Lecture 2 Probability Density Estimation 16.04.2015 Bastian Leibe RWTH Aachen

More information

URBAN MAPPING AND CHANGE DETECTION

URBAN MAPPING AND CHANGE DETECTION URBAN MAPPING AND CHANGE DETECTION Sebastian van der Linden with contributions from Akpona Okujeni Humboldt-Unveristät zu Berlin, Germany Introduction Introduction The urban millennium Source: United Nations,

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and

More information

Feature selection and extraction Spectral domain quality estimation Alternatives

Feature selection and extraction Spectral domain quality estimation Alternatives Feature selection and extraction Error estimation Maa-57.3210 Data Classification and Modelling in Remote Sensing Markus Törmä markus.torma@tkk.fi Measurements Preprocessing: Remove random and systematic

More information

Predicting the Probability of Correct Classification

Predicting the Probability of Correct Classification Predicting the Probability of Correct Classification Gregory Z. Grudic Department of Computer Science University of Colorado, Boulder grudic@cs.colorado.edu Abstract We propose a formulation for binary

More information

CLASSIFICATION approaches based on neural networks

CLASSIFICATION approaches based on neural networks IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 37, NO. 3, MAY 1999 1367 Classification of Multisource and Hyperspectral Data Based on Decision Fusion Jon Atli Benediktsson, Member, IEEE, and

More information

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396 Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table

More information

Support Vector Machines Explained

Support Vector Machines Explained December 23, 2008 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),

More information

Support Vector Machine. Industrial AI Lab. Prof. Seungchul Lee

Support Vector Machine. Industrial AI Lab. Prof. Seungchul Lee Support Vector Machine Industrial AI Lab. Prof. Seungchul Lee Classification (Linear) Autonomously figure out which category (or class) an unknown item should be categorized into Number of categories /

More information

Object-based land use/cover extraction from QuickBird image using Decision tree

Object-based land use/cover extraction from QuickBird image using Decision tree Object-based land use/cover extraction from QuickBird image using Decision tree Eltahir. M. Elhadi. 12, Nagi. Zomrawi 2 1-China University of Geosciences Faculty of Resources, Wuhan, 430074, China, 2-Sudan

More information

Resource Management through Machine Learning

Resource Management through Machine Learning Resource Management through Machine Learning Justin Granek Eldad Haber* Elliot Holtham University of British Columbia University of British Columbia NEXT Exploration Inc Vancouver, BC V6T 1Z4 Vancouver,

More information

Selection of Classifiers based on Multiple Classifier Behaviour

Selection of Classifiers based on Multiple Classifier Behaviour Selection of Classifiers based on Multiple Classifier Behaviour Giorgio Giacinto, Fabio Roli, and Giorgio Fumera Dept. of Electrical and Electronic Eng. - University of Cagliari Piazza d Armi, 09123 Cagliari,

More information

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie Computational Biology Program Memorial Sloan-Kettering Cancer Center http://cbio.mskcc.org/leslielab

More information

Virtual Sensors and Large-Scale Gaussian Processes

Virtual Sensors and Large-Scale Gaussian Processes Virtual Sensors and Large-Scale Gaussian Processes Ashok N. Srivastava, Ph.D. Principal Investigator, IVHM Project Group Lead, Intelligent Data Understanding ashok.n.srivastava@nasa.gov Coauthors: Kamalika

More information

Multivariate statistical methods and data mining in particle physics

Multivariate statistical methods and data mining in particle physics Multivariate statistical methods and data mining in particle physics RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement of the problem Some general

More information

Reading, UK 1 2 Abstract

Reading, UK 1 2 Abstract , pp.45-54 http://dx.doi.org/10.14257/ijseia.2013.7.5.05 A Case Study on the Application of Computational Intelligence to Identifying Relationships between Land use Characteristics and Damages caused by

More information

CS534 Machine Learning - Spring Final Exam

CS534 Machine Learning - Spring Final Exam CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the

More information

Learning with multiple models. Boosting.

Learning with multiple models. Boosting. CS 2750 Machine Learning Lecture 21 Learning with multiple models. Boosting. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Learning with multiple models: Approach 2 Approach 2: use multiple models

More information

A GLOBAL ANALYSIS OF URBAN REFLECTANCE. Christopher SMALL

A GLOBAL ANALYSIS OF URBAN REFLECTANCE. Christopher SMALL A GLOBAL ANALYSIS OF URBAN REFLECTANCE Christopher SMALL Lamont Doherty Earth Observatory Columbia University Palisades, NY 10964 USA small@ldeo.columbia.edu ABSTRACT Spectral characterization of urban

More information

An introduction to Support Vector Machines

An introduction to Support Vector Machines 1 An introduction to Support Vector Machines Giorgio Valentini DSI - Dipartimento di Scienze dell Informazione Università degli Studi di Milano e-mail: valenti@dsi.unimi.it 2 Outline Linear classifiers

More information

NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan

NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan In The Proceedings of ICONIP 2002, Singapore, 2002. NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION Haiqin Yang, Irwin King and Laiwan Chan Department

More information

Linear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers)

Linear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers) Support vector machines In a nutshell Linear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers) Solution only depends on a small subset of training

More information

Relevance Vector Machines for Earthquake Response Spectra

Relevance Vector Machines for Earthquake Response Spectra 2012 2011 American American Transactions Transactions on on Engineering Engineering & Applied Applied Sciences Sciences. American Transactions on Engineering & Applied Sciences http://tuengr.com/ateas

More information

ML (cont.): SUPPORT VECTOR MACHINES

ML (cont.): SUPPORT VECTOR MACHINES ML (cont.): SUPPORT VECTOR MACHINES CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 40 Support Vector Machines (SVMs) The No-Math Version

More information

Pattern Recognition 2018 Support Vector Machines

Pattern Recognition 2018 Support Vector Machines Pattern Recognition 2018 Support Vector Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recognition 1 / 48 Support Vector Machines Ad Feelders ( Universiteit Utrecht

More information

URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS

URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS Ivan Lizarazo Universidad Distrital, Department of Cadastral Engineering, Bogota, Colombia; ilizarazo@udistrital.edu.co

More information

LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition LINEAR CLASSIFIERS Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification, the input

More information

Support Vector Ordinal Regression using Privileged Information

Support Vector Ordinal Regression using Privileged Information Support Vector Ordinal Regression using Privileged Information Fengzhen Tang 1, Peter Tiňo 2, Pedro Antonio Gutiérrez 3 and Huanhuan Chen 4 1,2,4- The University of Birmingham, School of Computer Science,

More information

Comparison of MLC and FCM Techniques with Satellite Imagery in A Part of Narmada River Basin of Madhya Pradesh, India

Comparison of MLC and FCM Techniques with Satellite Imagery in A Part of Narmada River Basin of Madhya Pradesh, India Cloud Publications International Journal of Advanced Remote Sensing and GIS 013, Volume, Issue 1, pp. 130-137, Article ID Tech-96 ISS 30-043 Research Article Open Access Comparison of MLC and FCM Techniques

More information

Unsupervised Learning with Permuted Data

Unsupervised Learning with Permuted Data Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University

More information

Introduction to SVM and RVM

Introduction to SVM and RVM Introduction to SVM and RVM Machine Learning Seminar HUS HVL UIB Yushu Li, UIB Overview Support vector machine SVM First introduced by Vapnik, et al. 1992 Several literature and wide applications Relevance

More information

Mineral Mapping with AVIRIS and EO-1 Hyperion **

Mineral Mapping with AVIRIS and EO-1 Hyperion ** Mineral Mapping with AVIRIS and EO-1 Hyperion ** Fred A. Kruse 1 1.0 Introduction Imaging Spectrometry data or Hyperspectral Imagery (HSI) acquired using airborne systems have been used in the geologic

More information

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2018 CS 551, Fall

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table

More information

LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES. Supervised Learning

LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES. Supervised Learning LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES Supervised Learning Linear vs non linear classifiers In K-NN we saw an example of a non-linear classifier: the decision boundary

More information

URBAN VEGETATION COVER EXTRACTION FROM HYPERSPECTRAL REMOTE SENSING IMAGERY AND GIS-BASED SPATIAL ANALYSIS TECHNIQUES: THE CASE OF ATHENS, GREECE

URBAN VEGETATION COVER EXTRACTION FROM HYPERSPECTRAL REMOTE SENSING IMAGERY AND GIS-BASED SPATIAL ANALYSIS TECHNIQUES: THE CASE OF ATHENS, GREECE Proceedings of the 13 th International Conference on Environmental Science and Technology Athens, Greece, 5-7 September 2013 URBAN VEGETATION COVER EXTRACTION FROM HYPERSPECTRAL REMOTE SENSING IMAGERY

More information

Support Vector Machine via Nonlinear Rescaling Method

Support Vector Machine via Nonlinear Rescaling Method Manuscript Click here to download Manuscript: svm-nrm_3.tex Support Vector Machine via Nonlinear Rescaling Method Roman Polyak Department of SEOR and Department of Mathematical Sciences George Mason University

More information

CSCI-567: Machine Learning (Spring 2019)

CSCI-567: Machine Learning (Spring 2019) CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Brief Introduction to Machine Learning

Brief Introduction to Machine Learning Brief Introduction to Machine Learning Yuh-Jye Lee Lab of Data Science and Machine Intelligence Dept. of Applied Math. at NCTU August 29, 2016 1 / 49 1 Introduction 2 Binary Classification 3 Support Vector

More information

Object Based Image Classification for Mapping Urban Land Cover Pattern; A case study of Enugu Urban, Enugu State, Nigeria.

Object Based Image Classification for Mapping Urban Land Cover Pattern; A case study of Enugu Urban, Enugu State, Nigeria. Object Based Image Classification for Mapping Urban Land Cover Pattern; A case study of Enugu Urban, Enugu State, Nigeria. Nnam, Godwin Uchechukwu, u_nnam@yahoo.com a Ndukwu, Raphael I., raphael.ndukwu@unn.edu.ng

More information

p(d θ ) l(θ ) 1.2 x x x

p(d θ ) l(θ ) 1.2 x x x p(d θ ).2 x 0-7 0.8 x 0-7 0.4 x 0-7 l(θ ) -20-40 -60-80 -00 2 3 4 5 6 7 θ ˆ 2 3 4 5 6 7 θ ˆ 2 3 4 5 6 7 θ θ x FIGURE 3.. The top graph shows several training points in one dimension, known or assumed to

More information

Learning Kernel Parameters by using Class Separability Measure

Learning Kernel Parameters by using Class Separability Measure Learning Kernel Parameters by using Class Separability Measure Lei Wang, Kap Luk Chan School of Electrical and Electronic Engineering Nanyang Technological University Singapore, 3979 E-mail: P 3733@ntu.edu.sg,eklchan@ntu.edu.sg

More information

Machine Learning And Applications: Supervised Learning-SVM

Machine Learning And Applications: Supervised Learning-SVM Machine Learning And Applications: Supervised Learning-SVM Raphaël Bournhonesque École Normale Supérieure de Lyon, Lyon, France raphael.bournhonesque@ens-lyon.fr 1 Supervised vs unsupervised learning Machine

More information

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation. CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.

More information

Sparse Representation-based Analysis of Hyperspectral Remote Sensing Data

Sparse Representation-based Analysis of Hyperspectral Remote Sensing Data Sparse Representation-based Analysis of Hyperspectral Remote Sensing Data Ribana Roscher Institute of Geodesy and Geoinformation Remote Sensing Group, University of Bonn 1 Remote Sensing Image Data Remote

More information

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Final Overview. Introduction to ML. Marek Petrik 4/25/2017 Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,

More information