MSc Project Feature Selection using Information Theoretic Techniques. Adam Pocock

Size: px
Start display at page:

Download "MSc Project Feature Selection using Information Theoretic Techniques. Adam Pocock"

Transcription

1 MSc Project Feature Selection using Information Theoretic Techniques Adam Pocock 15/08/2008

2 Abstract This document presents a investigation into 3 different areas of feature selection, using information theoretic methods. The first area of research is an investigation into the selection of the first feature in common feature selection algorithms. This step is often overlooked in the construction of feature selection algorithms, with the assumption that the most informative feature is the best first choice. This can be proven to be untrue, and so an investigation into how to select a better suited feature forms the first part of the research. New methods for selecting the first feature are proposed and empirically tested to see if they offer an improvement over the standard method. The second area of research is an investigation into applying the Rényi extension to information theory to standard feature selection techniques. This requires the development of a Rényi mutual information measure, and two different measures are proposed. The Rényi extension provides a positive real parameter, α, which can be varied. The new Rényi feature selection techniques are empirically tested, varying the measure and value of α used. The third area of research is an investigation into a different method of estimating the Rényi entropy of a variable, by using a function based upon the length of the minimal spanning tree of the variable. This enables a high dimensional estimate of the entropy to be constructed in O(s log s) time and can be used to implement the Max Dependency criterion. This research investigates the use of this estimator as a feature selection criterion and proposes a new genetic algorithm based method for selecting features, in contrast to the traditional forward search used in other algorithms.

3 Declaration No portion of the work referred to in this thesis has been submitted in support of an application for another degree or qualification of this or any other university or other institution of learning. i

4 Copyright Copyright in text of this dissertation rests with the author. Copies (by any process) either in full, or of extracts, may be made only in accordance with instructions given by the author. Details may be obtained from the appropriate Graduate Office. This page must form part of any such copies made. Further copies (by any process) of copies made in accordance with such instructions may not be made without the permission (in writing) of the author. The ownership of any intellectual property rights which may be described in this dissertation is vested in the University of Manchester, subject to any prior agreement to the contrary, and may not be made available for use by third parties without the written permission of the University, which will prescribe the terms and conditions of any such agreement. Further information on the conditions under which disclosures and exploitation may take place is available from the Head of the School of Computer Science. ii

5 Acknowledgements I would like to thank my parents for their continued support, both personal and financial. I would also like to thank my supervisor Gavin Brown for his guidance which has made this thesis possible. iii

6 Contents 1 Project Aims & Contributions Motivation Description of Document Project Contributions Knowledge Contributions Implementation Contributions Background Machine Learning Feature Selection Search Strategies Document Notation Information Theory Mutual Information Information Theory and Feature Selection Feature Selection Algorithms mrmr CMIM DISR Rényi s Information Theory Rényi Entropy Rényi Generalised Divergence Rényi entropy estimation Summary Testing Framework Datasets iv

7 CONTENTS v Public Datasets Adenocarcinoma Dataset Classifiers Support Vector Machine k-nearest Neighbour Feature Selection Algorithms CMIM implementation DISR implementation mrmr implementation Graph implementation Test Construction Comparing CMIM, DISR and mrmr Introduction Results Lymphoma Dataset NCI 9 Dataset Colon Dataset Leukaemia Dataset Lung Cancer Dataset Adenocarcinoma Dataset Analysis Adenocarcinoma Dataset Colon Dataset Leukaemia Dataset Lung Cancer Dataset Lymphoma Dataset NCI 9 Dataset Execution Speed Conclusions Summary Investigating the first feature Introduction Creating New Criteria CMIM Criterion

8 CONTENTS vi DISR Criterion mrmr Criterion Computational complexity Summary Results Lung cancer dataset NCI 9 Dataset Lymphoma Dataset Analysis Lung Cancer Dataset NCI 9 Dataset Lymphoma Dataset Summary Conclusions Summary of the work Investigating the Rényi measures of information Introduction Different Formulations of Rényi Information Constructing the Conditional Mutual Information Reconstructing the Algorithms Constructing CMIM Constructing DISR & mrmr Results CMIM Results DISR Results mrmr-d Results mrmr-q Results Analysis Analysis of CMIM variations Analysis of DISR variations Analysis of mrmr-d variations Analysis of mrmr-q variations Conclusion Summary of the work

9 CONTENTS vii 7 Graph-based Entropy Estimation Introduction Graph-based Entropy Image Registration Graph-based entropy estimation Conclusion A note on the continuous entropy Graph-based Estimation for Feature Selection Creating the Algorithm Usefulness of this measure Rényi Entropy Genetic Algorithm What is a Genetic Algorithm Why use a Genetic Algorithm Constructing the Genetic Algorithm Results Leukaemia Dataset Lung Dataset Execution Speed Analysis Leukaemia Dataset Lung Dataset Summary Conclusions Summary of the work Conclusions Summary of the research First Feature Selection Rényi Mutual Information Graph-based Entropy Estimation Critique Of The Graph-based Entropy Estimation Further Investigation General Observations Investigation into First Feature Selection Investigation into Rényi Mutual Information Investigation into Graph-based Entropy Estimation

10 CONTENTS viii Bibliography 111

11 List of Figures 2.1 How entropy and mutual information relate Difference between H α (X Y ) and H α (XY ) H α (X), using the Lung dataset, and X = feature 1, Y = class Original DISR pseudocode Optimised DISR pseudocode Lymphoma Dataset, 3-NN and SVM Classifiers NCI9 Dataset, 3-NN and SVM Classifiers Colon Dataset, 3-NN and SVM Classifiers Leukaemia Dataset, 3-NN and SVM Classifiers Lung Cancer Dataset, 3-NN and SVM Classifiers Adenocarcinoma Dataset, 3-NN and SVM Classifiers Lung Cancer Dataset, 3-NN Classifier, CMIM and DISR Lung Cancer Dataset, 3-NN Classifier, mrmr-d and mrmr-q Lung Cancer Dataset, Linear SVM Classifier, CMIM and DISR Lung Cancer Dataset, Linear SVM Classifier, mrmr-d and mrmr-q NCI9 Dataset, 3-NN Classifier, CMIM and DISR NCI9 Dataset, 3-NN Classifier, mrmr-d and mrmr-q NCI9 Dataset, Linear SVM Classifier, CMIM and DISR NCI9 Dataset, Linear SVM Classifier, mrmr-d and mrmr-q Lymphoma Dataset, 3-NN Classifier, CMIM and DISR Lymphoma Dataset, 3-NN Classifier, mrmr-d and mrmr-q Lymphoma Dataset, Linear SVM Classifier, CMIM and DISR Lymphoma Dataset, Linear SVM Classifier, mrmr-d and mrmr-q The asymmetry of the conditional formulation ix

12 LIST OF FIGURES x 6.2 Rényi mutual information - lung dataset X = feature 14, Y = class Rényi mutual information - lung dataset X = feature 29, Y = class Rényi mutual information - lung dataset X = feature 49, Y = class The different formulations of conditional mutual information CMIM, Lung Cancer Dataset, 0.6 α CMIM, Lung Cancer Dataset, 1.7 α CMIM, NCI9 Dataset, SVM Classifier, 0.1 α 0.5 & 0.6 α CMIM, NCI9 Dataset, SVM Classifier, 1.2 α 1.6 & 1.7 α DISR Joint and Divergence, NCI9 Dataset, SVM Classifier, 0.1 α DISR Joint and Divergence, NCI9 Dataset, SVM Classifier, 0.6 α DISR Joint and Divergence, NCI9 Dataset, SVM Classifier, 1.7 α mrmr-d Joint and Divergence, Adenocarcinoma Dataset, SVM Classifier, 0.1 α mrmr-d Joint and Divergence, Adenocarcinoma Dataset, 3-NN Classifier, 0.1 α mrmr-d Joint and Divergence, Colon Dataset, SVM Classifier, 1.2 α mrmr-d Joint and Divergence, Lymphoma Dataset, SVM Classifier, 0.1 α mrmr-d Joint and Divergence, Lymphoma Dataset, SVM Classifier, 1.7 α mrmr-q Joint and Divergence, Colon Dataset, SVM Classifier, 1.2 α mrmr-q Joint and Divergence, Colon Dataset, SVM Classifier, 1.7 α mrmr-q Joint and Divergence, Colon Dataset, 3-NN Classifier, 1.7 α mrmr-q Joint and Divergence, Lung Dataset, SVM Classifier, 1.7 α mrmr-q Joint and Divergence, Lymphoma Dataset, SVM Classifier, 0.1 α mrmr-q Joint and Divergence, Lymphoma Dataset, SVM Classifier, 0.6 α The Minimal Spanning Tree of a set of connected points, image sourced from spanning tree.svg, retrieved on: Graph MI Genetic Algorithm Pseudocode Leukaemia Dataset, 3-NN & SVM Classifier, α = Lung Dataset, 3-NN & SVM Classifier, α =

13 List of Tables 3.1 Dataset properties DISR optimisation performance Algorithm execution time (s) XOR Problem Rényi algorithm execution time (s) xi

14 Chapter 1 Project Aims & Contributions 1.1 Motivation Feature selection has become an area of interest due to the increasingly large amount of data that is generated by new scientific methods like genomics. It provides a mechanism for stripping out extraneous and noisy data to avoid masking the informative data. Different mechanisms for feature selection can be found such as linear correlation, and other statistical tests, but an emerging field is the use of information theoretic methods to select informative features using a quantifiable measure of information. The use of such techniques provides a way of comparing the use of different features with different properties, and determining how useful each feature is to the current problem. Further information theoretic techniques enable the combination of features to be evaluated, and this enables informative feature sets to be constructed. Information theory can be applied when there are small numbers of samples in a dataset, which has applications in text-mining and genomics as there is usually a paucity of samples in biological fields. Improvements to feature selection techniques enable biologists to test large numbers of genes and extract ones which are relevant to a particular classification problem, which can lead to the inference of a link between a particular gene and a problem such as a disease. This knowledge can then either be used to aid the diagnosis of difficult cases or to indicate which genes need further investigation to generate more knowledge about a condition. More generally, feature selection techniques provide a way of reducing the complexity and increasing the accuracy of classification and regression tasks as the stripping out of extraneous data enables a more streamlined classifier that is less likely to over-fit on certain data. The reduction of a feature set to its most useful subset enables further data to be collected in a less expensive way, by reducing the amount of data collected, and also to reduce 1

15 CHAPTER 1. PROJECT AIMS & CONTRIBUTIONS 2 the classification time for new data because of the reduction in information that needs to be processed. Research that improves the current set of feature selection techniques therefore has wide ranging applications in many fields of data analysis, and it is the analysis and development of feature selection techniques that forms the basis of this research. 1.2 Description of Document Background The background chapter contains a detailed look into the process of feature selection and why it is becoming increasingly important as the size and complexity of datasets increases. It contains a review of information theory including several extensions to the theory which parametrise it. Then there is a review of the current state of the art feature selection algorithms, and how these use information theoretic techniques to select useful feature sets for further analysis. Testing Framework This chapter details the datasets used for testing the various feature selection algorithms used throughout the research. It describes an optimisation to two of the techniques to bring an improvement in execution time, and a discrepancy found in the reference implementation of mrmr which can alter the outcome of the algorithm. Then the various implementations of algorithms are described and the classifiers used to test them are analysed. Finally the testing framework for the research is constructed, detailing how the selected feature sets are tested. Comparing CMIM, DISR and mrmr An initial study is performed testing the state of the art algorithms against the selected datasets chosen for this work. Additionally a comparison of the execution times of the implementations of the algorithms is performed to give a baseline for the complexity of the algorithms modified for the research. Investigating the first feature This chapter forms an investigation into the properties of the feature selection algorithms and how they select the first feature. The greedy assumption that the feature with the highest mutual information is the best feature to select first is questioned and a number of different methods for selecting the first feature are proposed. These novel methods are then empirically tested against the standard method using all the feature selection algorithms, and across all the datasets. Finally the analysis of the results leads to the conclusion that the greedy assumption is not optimal in the first stage of the feature selection algorithms, and selecting the

16 CHAPTER 1. PROJECT AIMS & CONTRIBUTIONS 3 first feature using a criterion that promotes feature independence provides an increase in classification accuracy. Investigating the Rényi measures of information This chapter forms an investigation into applying the Rényi entropy extension to information theory, and deriving a measure of mutual information. It is found that two methods can be used to measure the mutual information using the Rényi entropy, and this leads to two measures of mutual information parametrised by a positive real number α. The feature selection algorithms analysed previously are then modified to use these different measures of mutual information. The new algorithms are empirically tested using a range of values for the α parameter and the two different measures of the mutual information, across all the datasets and classifiers. Finally conclusions are drawn showing that the use of the Rényi mutual information in these algorithms can provide an increase in classification performance but it adds an extra layer of model inference and so is more prone to over-fitting the data. Graph-based Entropy Estimation This chapter forms an investigation into a different method of estimating the Rényi entropy than the standard method using histogram bins. This method is adapted for use with feature selection in [1] and an extension to that work is developed which, instead of estimating the Shannon entropy, works directly from the Rényi entropy to derive measures of the Rényi mutual information as discussed in chapter 6. Then a genetic algorithm is developed to replace the forward search used in the feature selection algorithm, as the new entropy estimate is less prone to dimensionality problems found in other estimates. This genetic algorithm uses the new entropy estimator as the fitness function for stochastic hill-climbing to search for an optimal solution. The new feature selection techniques developed are empirically tested using the datasets and classifiers used elsewhere in the research. Finally conclusions are drawn about the performance of the new methods, with a functional flaw leading to poor performance of the new entropy estimator. However the performance of the genetic algorithm appears promising and deserves further study. Conclusions This chapter provides a summary of the work performed in this research, and provides a list of further directions for research in this area to take.

17 CHAPTER 1. PROJECT AIMS & CONTRIBUTIONS Project Contributions Knowledge Contributions A comparison of four state of the art feature selection algorithms, using a variety of datasets and two classifiers The creation of several different methods for selecting the initial features used in the 3 selection algorithms, including a thorough empirical investigation An theoretical study into the differing constructions of the mutual information using the Rényi entropy and generalised divergence, including a thorough empirical investigation A thorough investigation of the performance of the different feature selection algorithms when using different values of the Rényi α parameter A critique of the feature selection technique from [1] An investigation into using a genetic algorithm to replace the forward feature selection, with the graph-based entropy estimator A comparison between the graph-based forward feature selection algorithm, the graphbased genetic feature selection algorithm, and the Rényi versions of CMIM, DISR and mrmr, using a variety of datasets Implementation Contributions An optimisation of the DISR and mrmr algorithms to improve execution time The development of a set of Rényi entropy based MATLAB/C++ functions to enable further work The recreation of CMIM, DISR and mrmr to use the Rényi based mutual information The creation of a MATLAB/C++ implementation of the Graph-based entropy feature selection technique from [1] The development of a genetic algorithm in MATLAB/C++ to replace the forward feature selection used in the graph-based feature selection algorithm

18 Chapter 2 Background 2.1 Machine Learning Machine Learning is the use of statistical processes to derive information and rules used to interpret new data. It is designed to automate the process of analysing data and extracting rule-based information from it. This leads to several different areas of work. Machine learning systems can be trained to respond to certain stimulus to provide control frameworks for machinery or to automate tasks that have been demonstrated. Data-based systems can be trained to classify data into different groups based upon the values in each datapoint, or they can be trained to generate real values based upon the values in the datapoint to provide estimates of the value of new unseen datapoints. Both these approaches require labelled samples of data, with the end value, or class defined so the algorithm can learn by example. It is the classification of data that this research will focus upon. Classification systems are being used to analyse large datasets generated in the fields of text analysis and genomics provide preprocessing for human researchers to remove extraneous data, and to perform complex numerical analysis to generate rules and trend information that human researchers would find difficult to extract from the surrounding noise of the dataspace. Simple classification tasks involve the placement of data samples into different categories based upon a function of the data contained in the sample. These tasks can be blurred by large quantities of data, as it becomes more difficult to sift out the informative pieces of data, from the surrounding noise. 5

19 CHAPTER 2. BACKGROUND Feature Selection Feature Selection techniques have become increasingly important within Machine Learning as the increase in processing power of computers has enabled ever more complex datasets to be created, manipulated and analysed, and especially with the recent growth of Genomics as a scientific discipline. Feature Selection is a process where a group of features from a dataset are selected with the aim of improving classification accuracy and decreasing computational complexity [6]. It is closely related to Feature Extraction which is a process where feature vectors are created from the original dataset through manipulations of the data space, which can be considered to be a superset of Feature Selection techniques. In addition to its use as a preprocessor for classification tasks, Feature Selection can also be used to identify features which are informative about different classes, so these features can be further analysed [6]. This is extremely useful in the analysis of gene expression data, which can contain many thousands of features, with each one taken from the expression values of an individual gene. The techniques can then be applied to find genes that appear indicative of certain classes, which can then be used for further experimentation. Feature Selection techniques come in 2 main groups: filters and wrappers [6]. Filter techniques use classifier independent methods to select features from the feature space. These techniques are based on a number of different statistical tests, or on the information theoretic concept of Mutual Information. Wrapper techniques embed the feature selection into a classifier, where the classification performance is used to measure the quality of the currently selected feature set [8]. A further group of embedded techniques are similar to wrappers, but instead of using the classifier as a black box it integrates the classifier into the feature selection algorithm. It is similar to the training of a neural network which can gradually reduce the influence of useless inputs by giving it a lower weight. The different techniques have differing advantages and drawbacks. Filter techniques rely upon measuring information values which have to be defined separately from classification performance, and do not necessarily translate to a set of features which is the optimal set for a particular type of classifier. However they are faster than wrappers, and find a set of features which can be shown to have a high level of information about the target irrespective of the classifier. Wrapper techniques are extremely dependent on the classifier used to test the performance of the feature set. They find feature sets which are highly linked to the choice of classifier, so the features selected do not generalise well to other classifiers. Due to the wrapping process the selected features are well suited for use with the chosen classifier, so may produce a more accurate class prediction. Also, due to the process of constantly retraining and testing a classifier, wrapper techniques have a high computational complexity

20 CHAPTER 2. BACKGROUND 7 when compared to filter techniques [6] Search Strategies Both types of technique require a search strategy to be selected, to determine how the selection space is searched. As the search space for feature sets is exponential in the number of features, the most commonly used strategy is a greedy search, choosing the best option at each current iteration rather than taking a sub-optimal option to increase the solution performance in later iterations. The two main kinds of search strategy are forward search and backward search. Forward search starts from an empty set of selected features, and then proceeds to add features to the set until a specific requirement has been reached. This can be stated as, at each iteration m, the feature x n is found which maximises the condition F, which is then added to the solution set S (equation (2.1)). S = S {arg max X n X \S F (X \S )} (2.1) Backward search starts from a set of all the features then proceeds to eliminate features from the set until a specific requirement has been reached. This can be stated as, at each iteration m, the feature x n is found which minimises the condition F, which is then removed from the solution set S (equation (2.2)). S = S \ {arg min F (S)} (2.2) X n S Backwards selection can be computationally more intensive when selecting a small number of features from a large feature set, as there are more iterations before the search reaches the desired number of features. Search strategies can also be constructed that combine both types, so features are added to the set, and then features are removed, and the process is iterated until the set has met the requirements. The stopping conditions for the algorithms are usually defined as being a set number of selected features, but in the case of wrappers the condition can be the classification performance of the selected features using the internal classifier. This helps stop over-fitting of data by not adding extraneous information.

21 CHAPTER 2. BACKGROUND Document Notation Throughout the rest of this document the following notation will hold. Capital letters will denote random variables, and lower case letter denote a value from that random variable, e.g. X, Y and Z denotes random variables, with x, y and z being a particular value from their respective variables. XY denotes the joint random variable composed from X and Y, and X Y denotes the conditioning of X by Y. p(x) denotes the probability of the event recorded in x occurring. S denotes the selected set of features, and F denotes the set of all features. k denotes the number of desired features output by a feature selection algorithm, n denotes the number of features input into the algorithm (otherwise known as the cardinality of F ). 2.4 Information Theory To have a way of classifying features by information level, first information itself must be quantified in a way that is not subjective. This was proposed by Shannon in 1948 [20], and is known as Shannon s Information Theory. This formulation is the standard method for quantifying information but other methods have been proposed, namely by Rényi in 1961 [18]. Information Theory is constructed from analysis of information flows in systems, and how much data is required to transmit information [20]. A more general explanation of the topic of information theory can be found in [3]. The main concept in Information Theory is Entropy, which measures the degree of uncertainty in a single variable. It gives high values to variables where each possible value has an equal probability and low values where values have an unequal probability distribution (e.g. for the boolean random variable {0,1,1,0} either 0 or 1 is equally likely thus it has a high entropy, for the boolean random variable {0,0,0,1}, 0 is much more likely than 1, thus it has a low entropy). Entropy is valued between 0 and 1 when there are 2 possible values, with the maximum possible entropy increasing depending upon the number of possible values in the variable. Shannon s Information Theory uses the logarithm as the method for deriving the amount of information held in a variable. n H(X) = p(x i ) log 2 (p(x i )) (2.3) i=1 Where X is a random variable with n possible values, H(X) is the entropy of X, and p(x)

22 CHAPTER 2. BACKGROUND 9 is the probability of the value x occurring in X. It sums over the possible values of X taking the probability of each value of X multiplied by the logarithm of the probability. The logarithmic base is usually chosen to be 2, which results in an entropy value measured in bits, though other bases are possible. Other bi-variate entropies are also defined such as the joint entropy and the conditional entropy. The joint entropy is the sum of the uncertainty contained in two variables, and is bounded above by the sum of the two individual entropies, and bounded below by the highest individual entropy. n m H(XY ) = p(x i, y j ) log 2 (p(x i, y j )) (2.4) i=1 j=1 max(h(x), H(Y )) H(XY ) H(X) + H(Y ) (2.5) The lower equality is true when X is completely dependent upon Y, or vice versa. The upper equality is true when X & Y are independent of each other. This is because adding a variable to a system can only increase the uncertainty present, but not decrease it. The conditional entropy is the uncertainty remaining in one variable when another variable has a known value. It is bounded below by 0, and above by the entropy of the original variable. The lower bound is true when X is completely dependent upon Y, and the upper bound is true when X & Y are independent of each other. In general conditioning on entropy reduces the value of the entropy, as it is rare that two variables are completely independent. H(X Y ) = n p(y j )H(X Y = y j ) (2.6) j=1 0 H(X Y ) H(X) (2.7) The interactions between all these values can be modelled in the two variable case by a Venn diagram, showing how the values relate to each other and to the quantity of mutual information. In the diagram (figure 2.1) the term H(XY ) relates to the total area of both H(X) and H(Y ). Also it can be seen that the conditional entropy is also equal to the joint entropy minus the entropy of the conditioning variable. H(X Y ) = H(XY ) H(Y ) (2.8)

23 CHAPTER 2. BACKGROUND 10 Figure 2.1: How entropy and mutual information relate Mutual Information Mutual Information (MI) is defined as a measure of how much information is jointly contained in two variables, or how much knowledge of one variable determines the other variable. It forms the basis of information theoretic feature selection as it provides a function for calculating the relevance of a variable to the target class, irrespective of correlations and other assumptions. Mutual Information can be defined in two different but equivalent ways, either in terms of entropies, or in terms of probability distributions, and it is symmetric with respect to the ordering of X and Y. I(X; Y ) = H(X) H(X Y ) I(X; Y ) = H(Y ) H(Y X) (2.9) I(X; Y ) = H(X) + H(Y ) H(X, Y ) These differing formulations of the mutual information all result in the same value, and are related by figure 2.1. The mutual information is the overlap between the variables, and can thus be obtained as the sum of the two entropies, minus the joint entropy which removes all information that is not contained twice in the variables. This gives rise to an equation for the mutual information. Another equation can be obtained from the entropy minus the conditional entropy as the amount of uncertainty that is lost through the conditioning of the variable by another variable, and lost uncertainty is equivalent to a gain in information. This conditional equation is symmetric, so the ordering of the variables can be changed.

24 CHAPTER 2. BACKGROUND 11 I(X; Y ) = y Y p(x, y) p(x, y) log( p(x)p(y) ) (2.10) x X I(X; Y ) = D KL (p(x, y) p(x)p(y)) (2.11) The probability based definition is defined as the Kullback-Liebler divergence between the joint probability distribution and the sum of the marginal distributions. The Kullback- Liebler divergence is similar to a distance metric between two probability distributions, however it is not a true distance metric as it is not symmetric. It can also be defined in the continuous case, where the discrete probability values become continuous probability densities. I(X; Y ) = Y X p(x, y) p(x, y) log( p(x)p(y) ) (2.12) In general the mutual information when used for feature selection tasks is constructed in the discrete form, or in the continuous form upon a discrete class. This leads to a summation over the discrete class space, and an integral over the continuous feature space. Conditional forms of the mutual information can be formed but unlike with entropies, conditioning the mutual information does not always reduce it. This is because the knowledge of a third variable can make the two original variables dependent upon each other, (e.g. in the case of the XOR problem detailed in Chapter 5, conditioning by the other variable makes the mutual information equal to the entropy of the class). I(X; Y Z) = p(z)i(x; Y Z = z) (2.13) z Z In the multivariate case the mutual information becomes a more complex measure, and can no longer be represented by a simple diagram. This is because the presence of an additional variable can increase the information between the original variables more than the sum of their parts. I(X 1,2 ; Y ) = I(X 1 ; Y ) + I(X 2 ; Y ) + (I(X 1 ; X 2 Y ) I(X 1 ; X 2 )) }{{} increases or decreases information (2.14) In the example above it is possible for I(X 1,2 ; Y ) > I(X 1 ; Y ) + I(X 2 ; Y ) or I(X 1,2 ; Y ) I(X 1 ; Y ) + I(X 2 ; Y ) depending upon the interaction of the variables X 1 and X 2. Therefore an information system cannot be treated as a strict sum of its various components, as they interact in counter-intuitive ways.

25 CHAPTER 2. BACKGROUND Information Theory and Feature Selection Once a metric has been decided for quantifying information then a number of feature selection techniques can be created. The most simple and intuitive one is to rank the features by their mutual information upon the class, and then select the top k features to use for classification. This technique is the simplest to implement, and requires the fewest evaluations of the mutual information (n evaluations, where n is the number of features), but it leads to some problems. Firstly it has the potential to introduce large amounts of redundant information, that is information already held in a selected feature, and secondly it does not take into account the multivariate interactions of a set of features. Therefore more complex techniques have been devised to try and reduce the flaws inherent in this selection algorithm, with an accompanying increase in computational complexity and the number of mutual information evaluations. These techniques try and generate an optimal feature set, but before they can be created the notion of an optimal feature set must be defined. Optimality for feature sets can be defined in numerous ways, based upon the number of features selected, and various techniques for measuring the information content of a feature set. Estimating the information content of a feature set becomes computationally intractable the more features that are added to the set, as the mutual information operator scales badly with dimension. One accepted definition of optimality in feature selection comes from the analysis of Markov blankets [9]. A Markov blanket is defined as a set of features, M, such that all the information held in a particular feature X not in the set is captured. This enables the removal of the feature X from the selected feature set as it is redundant. Koller & Sahami then prove that if a previously removed feature (F i ) had a Markov blanket within the feature set (G), and another feature (F j ) that is about to be removed has a Markov blanket in the set G, then F i has a Markov blanket in G F j. This gives rise to a backward search of the feature space, removing only features whose information is completely contained in the remaining set. The Markov blanket result forms an ideal method for calculating an optimal feature set in terms of minimal redundancy. It does not aim to produce a feature set of an expected size, though the algorithm could easily have such a criterion introduced by modifying it to calculate approximate Markov blankets. This would mean that a feature would be discarded if enough of its information was held in the current feature set, where this value could be modified. However the calculation of Markov blankets is computationally intractable, as it has an exponential complexity. From this notion of a redundant feature several algorithms have been developed which explicitly punish redundancy in the selected feature set. These algorithms are all formulated

26 CHAPTER 2. BACKGROUND 13 using the standard definition of entropy and mutual information. Additionally they are all greedy forward searches, instead of the backward search proposed in [9]. The techniques chosen for study in this research have been detailed below. 2.6 Feature Selection Algorithms mrmr From the work done by Koller & Sahami which provides a formalisation for the redundancy of a feature set, and the use of the mutual information on the class as a measure of relevance, several different ways have been used to combine the two [17, 23]. The accepted method is detailed in [17], as this also provides a useful selection algorithm that compares well with other techniques. The framework given in [17] defines the terms relevance, redundancy, and introduces the maximum dependency criterion, and the minimal-redundancy Maximal-Relevance (mrmr) criterion. It defines the relevance of a feature as its mutual information on the class, and thus the maximal relevance will select features based upon the highest mutual information values on the class. Redundancy is defined as the amount of information held about a feature by the feature set and approximated by averaging the bi-variate mutual informations between the feature and each individual feature in the feature set (equation (2.15)) where X i is the candidate feature and S is the feature set). R = 1 S 2 I(X i ; X j ) (2.15) X j S The maximum dependency criterion is to find the feature set which maximises the mutual information of the feature set upon the class (equation (2.16)) where F is the feature set). D = max I(S; Y ) (2.16) S F Maximum dependency is computationally intractable with the standard method of computing the mutual information, as multi-variate mutual informations are too computationally intensive to calculate, and also it is an exponential search over the feature space [16]. Because of this problem the paper then proposes the mrmr criterion as an acceptable tradeoff between calculating the maximum dependency and the performance of the algorithm. The criterion used in the algorithm is designed to simultaneously minimise the shared information within the feature set, and maximise the information on the class. The mrmr criterion

27 CHAPTER 2. BACKGROUND 14 is developed to approximate the maximum dependency whilst remaining computationally tractable for large datasets. The mrmr criterion is a bi-variate measure of information held in a single feature, penalised by an average of the inter-feature dependence. The criterion is proved to be equivalent to the maximum dependency in the first-order case, where features are added to the set one by one. X mrmr = arg max X n F \S I(X n ; Y ) 1 S 1 X k S I(X n ; X k ) (2.17) The mrmr criterion has no tri-variate estimation, meaning that it is an efficient algorithm when implemented. However this reduces the selection power of the criterion as it has no method for determining any form of variable complementarity, making it ill-suited for feature selection on datasets that have high levels of complementarity CMIM There is a limit to the level of performance that can be extracted from bi-variate feature selection methods. This is due to the issue of feature complementarity, where more information can be found through the use of multiple features at once than can be used by taking features separately. One of the first algorithms that can be classified as using feature complementarity is the Conditional Mutual Information Maximisation (CMIM), presented in [4] (with erratum issued here pdf). The criterion used in the algorithm is designed to minimise the shared information within the selected feature set. This is to prevent redundant information accumulating in the feature set at the expense of more specific information. The CMIM criterion is a tri-variate measure of the information held in a single feature about the class, conditioned upon an already selected feature. It loops over the selected features and assigns each candidate feature a score based upon the lowest conditional mutual information, between the selected features the candidate feature and the class. The feature then selected is the one with the maximum score. X CMIM (1) = arg max I(Y ; X n ) (2.18) n k, 1 k K, X CMIM (k + 1) = arg max min{i(y ; X n ) min I(Y ; X n X n l k CMIM (l))} (2.19) As the algorithm takes a pessimistic view of the data, by taking the minimum value as the

28 CHAPTER 2. BACKGROUND 15 score for a particular feature, this enables a shortcut in the algorithm implementation. At each iteration the score for a particular feature can only decrease so scores that are already below the current best score are not updated in the current run through, as they do not affect the calculation. Also as the scores are maintained through iterations it means there is no need to recalculate the conditional mutual information between already selected features and the remaining features as these values persist throughout the algorithm. The CMIM algorithm is precisely defined with explicit conditions in the initial loop, and so is not ambiguous in how the algorithm proceeds. The fast implementation of CMIM makes the algorithm efficient and quick to execute. The implementation also loses no accuracy over the direct implementation of equation (2.19). The implementation used is detailed in chapter 3. As the CMIM algorithm only considers conditional mutual information it fails to cope well with complementary variables as the pessimistic assumption does not improve the scores of highly complementary variables. This means it may not be well suited to working on datasets with a high complementarity such as gene expression data DISR Feature complementarity can have a large effect upon the information level of the resulting feature set, and so an algorithm was developed by Bontempi & Meyer in [12] to explicitly use complementarity to select features. The resulting criterion, called the Double Input Symmetrical Relevance (DISR) is a tri-variate measure of the information held in two features about the class. It is based on the Symmetrical Relevance (SR) which is defined as the mutual information of the feature on the class, normalised by the joint entropy of the feature and the class (see equation (2.20)). In DISR this is always used in the 3 variable case where it takes the joint mutual information of the two features on the class, normalised by the joint entropy of the 2 features and the class. SR(X; Y ) = I(X, Y ) H(X, Y ) (2.20) At each stage of the algorithm the selected feature is then the feature that maximises the sum of the SR s over all the currently selected features (see equation (2.21)). X DISR = arg max ( SR(X n,k ; Y )) (2.21) X n F \S X k S The DISR criterion works well when variable complementarity is a feature of the dataset, outperforming CMIM in most cases and performing equivalently with mrmr, but its more

29 CHAPTER 2. BACKGROUND 16 computationally intensive than either due to the tri-variate mutual information evaluations. One flaw in the algorithm as presented is that it fails to state explicit initial conditions for the algorithm, implicitly assuming that the equation (2.21) reduces to selecting the feature with the highest mutual information in X. These techniques form the state of the art with respect to information theoretic feature selection, using Shannon s measure of entropy. The DISR criterion takes into account two feature complementarity, and this technique captures the most information about the feature space. Unfortunately none of the techniques described above can perform multi-variate feature analysis in a computationally tractable amount of time. They are all tricks in information theory to approximate the multi-variate estimations. As computation power increases it should be possible to extend DISR to take 3 feature variables at once, but this will produce a significant increase in computation time, as it adds another order of complexity to the algorithm. 2.7 Rényi s Information Theory Rényi Entropy Rényi in a paper analysing the Shannon entropy [18] proposed a new class of entropies which form a generalisation of the Shannon entropy. This definition of entropy has a parameter α and is defined α 0, α H α (X) = 1 1 α log( p(x) α ) (2.22) x X Additionally the Shannon entropy is the limiting case of the Rényi entropy as α tends to lim H α(x) = H(X) (2.23) α 1 From this entropy the joint entropy is easily formed, and holds as the joint probability distribution. However it is not possible to conventionally form the conditional entropy. An expression for the conditional entropy can be created but it does not hold to the relation established in equation (2.8), as demonstrated by figure 2.2. This created conditional entropy does however take the Shannon conditional entropy as its limiting case. H α (X Y ) = y Y 1 p(y) 1 α log( p(x Y = y) α ) (2.24) x X

30 CHAPTER 2. BACKGROUND 17 Figure 2.2: Difference between H α (X Y ) and H α (XY ) H α (X), using the Lung dataset, and X = feature 1, Y = class The implications of this are explained in detail in chapter Rényi Generalised Divergence Additionally Rényi specifies a series of divergence measures that generalise the Kullback- Leibler divergence [11], providing a range of metrics for finding the divergence between two probability distributions. These are also parametrised by α defined α 1. D α (P Q) = 1 n α 1 log( p α k q 1 α k ) (2.25) Again the Kullback-Leibler divergence forms the limiting case of the Rényi generalised divergence as α tends to 1. k=1 lim D α(p Q) = D KL (P Q) (2.26) α 1

31 CHAPTER 2. BACKGROUND 18 These divergences can be used to construct a mutual information like value similar to the Shannon one. This is achieved by substituting into the α divergence the values p(x, y) and p(x)p(y). This forms a divergence between the joint probability density and the sum of the marginals like the mutual information, and is thus called the Rényi mutual information, and will be denoted throughout this paper as I α (X; Y ). D α (p(x, y) p(x)p(y)) = 1 α 1 log( x X p(x, y) α p(x) 1 α p(y) 1 α ) (2.27) This value, like the Rényi entropy, tends to the standard mutual information as α 1, and can thus be used as a basis for analysing the usefulness of the Renyi measure of information with respect to feature selection, and analysing the different values of the α parameter, and how this affects performance. y Y Rényi entropy estimation lim I α(x; Y ) = I(X; Y ) (2.28) α 1 There are various different ways of estimating the entropy of a variable. The standard method used with discrete entropy is to derive the probability for a given event from the number of times the event occurs divided by the number of observed events, as seen in [19]. This can be extended to cope with unseen events, but introduces uncertainty. The standard method for estimating the continuous entropy uses a Parzen Window [15] with a suitable interpolation function to provide an estimate for the probability density function [21]. A new method of estimating the continuous entropy was developed by Hero et al in [7], and was extended in [13]. This graph-based method for estimating the entropy is not subject to the dimensionality constraints of the Parzen Window estimator and enables the entropy of a multidimensional variable to be estimated. The applications of this estimator are explored in detail in chapter Summary This chapter has provided an introduction to the field of Feature Selection and Information theory. It has: Detailed why feature selection is used Explained the basics of information theory

32 CHAPTER 2. BACKGROUND 19 Shown how information theory is used to construct feature selection algorithms Explained how information theory is expanded and complicated by Rényi s work

33 Chapter 3 Testing Framework 3.1 Datasets The datasets used for this research comprise a selection of microarray experiments recording gene expression levels related to various different types of cancer. These datasets are taxing for classifiers due to their small number of samples compared with the high number of features, and is one of the reasons feature selection techniques have been an area of research, as they alleviate some of the problems caused by the high number of features Public Datasets The first 5 datasets used in the research are taken from various sources detailed in [17], which collates and uses them in the analysis of the mrmr-d and mrmr-q feature selection algorithms. All the features in these algorithms are discretised into three states, with a varying number of classes and examples per class. Title Features Classes Examples Colon Leukaemia Lung Lymphoma NCI Table 3.1: Dataset properties 20

34 CHAPTER 3. TESTING FRAMEWORK Adenocarcinoma Dataset A dataset has been provided by Prof. Andrew Brass of the University of Manchester which contains gene expression data taken from cancer patients. The dataset contains continuous features with 2 resultant classes (correlating to two different forms of lung cancer), and 59 samples. Before being used in this research the dataset was discretised along the mean into two states, this reduces some of the problems when trying to generalise from the continuous data. 3.2 Classifiers Classifiers are algorithms that take an amount of training data, and can then be used to predict the class of any new data given. They can be used with a set of known data to provide an estimate for the information contained in the features given by training on a portion of the data, then testing the prediction given against the rest of the data. The correct classification rate can then be used as a measure of how much information the classifier could extract from the chosen features. Different classifiers have different inherent biases. This means that a set of features that works well with one classifier is not guaranteed to work well with another, so to judge the usefulness of different feature selection algorithms then multiple classifiers must be used. There were 2 chosen classifiers used in this work for testing feature selection algorithms, they were a linear Support Vector Machine (SVM), and a 3-Nearest Neighbour classifier Support Vector Machine Support Vector Machines are a way of generating an optimal classification boundary in many dimensional space. The linear variant acts much like a perceptron, and fits a linear separating boundary to the dataset, that correctly classifies the maximum amount of training data. The algorithm can be modified with a kernel to map the boundary into a higher dimensional space, to give a non-linear separating boundary when mapped back into the original space k-nearest Neighbour The k-nn classifier is the simplest way of classifying data. It takes the nearest k neighbours of the new sample and classifies it with the largest group of neighbours. Effectively this creates a non-linear boundary separating the classes depending upon the original training data. An odd value should be chosen for k, as otherwise it introduces ambiguity into the

35 CHAPTER 3. TESTING FRAMEWORK 22 classification process as there may be an equal number of examples in all the classes. In the case where two classes are tied for the highest number of examples, the algorithm generates a random number to select which is the predicted class. This can still be a problem when an odd value of k is chosen if the number of classes is greater than 2 (e.g. in the case where you have a 3 class problem, and a 3-NN classifier, you could have an example of each class as the 3 nearest neighbours, leading to ambiguity). The k-nn classifier, despite being one of the simplest, provides good classification accuracy, and is simple to construct and understand which is why it was chosen for this research. 3.3 Feature Selection Algorithms The three chosen algorithms (CMIM, DISR, mrmr) were all recoded in C++/MEX with a MATLAB wrapper script to improve their performance. This was due to the time constraints and the number of tests that were performed using each algorithm. The basic Shannon entropy and mutual information functions are taken from H. Peng s mutual information toolbox, constructed for the work in [17], and then reverted to C++ and streamlined. The basic Rényi entropy and mutual informations are constructed using the probability functions from the mutual information toolbox, and developed in C++, with MATLAB wrappers to enable their use in other algorithms. The accompanying code will be released with source at a later date CMIM implementation The CMIM implementation is a C++ implementation of the optimised pseudocode provided in [4]. This pseudocode exploits the minimisation property of the CMIM equation to avoid calculating results that can only be worse than the current optimum. The resulting implementation is extremely optimised and was not suitable for further algorithm optimisation DISR implementation The DISR implementation is a modified C++ implementation of the criterion provided in [12]. The paper presents an algorithm based around the DISR criterion. A simple implementation of this algorithm proved to be inefficiently recalculating past values. A decision was made to sacrifice some of the space complexity of the algorithm in exchange for an decrease in computation time. This involves storing SR values between selected features and candidate features, and reloading this value rather than recalculating it from scratch.

36 CHAPTER 3. TESTING FRAMEWORK 23 for i = 2 to k for j = 1 to totalfeatures if not selected j for m = 1 : selectedfeatures currentscore = currentscore + SR(m,j,class); end if currentscore > score score = currentscore; currenthighestfeature = j; end end end answerfeatures(i) = currenthighestfeature; end Figure 3.1: Original DISR pseudocode The pseudocode given in figure 3.1 has k2 n 2 2 calls to the SR function. The optimisation, given in figure 3.2, adds one operation to each iteration of the loop, but only has k n calls to the SR function, as it stores the previously computed values. This leads to a large improvement in computation time, whilst not altering the output of the algorithm in any way. This does result in an increase in the memory used of k n doubles, but due to the performance increase this was found to be acceptable. Table 3.2 shows the increase in performance using the optimised version of the DISR algorithm over the standard naive implementation. The final C++ implementation is included to show the speed increase of MEX implementation over a MATLAB script. The optimised implementation provides a 43 fold improvement on execution time, on average, with the C++ implementation providing a further 9 fold increase, on average. Dataset DISR (MATLAB) DISR Optimised (MATLAB) Optimisation Performance Increase Colon x Leukaemia x Lung x Lymphoma x NCI x Table 3.2: DISR optimisation performance

37 CHAPTER 3. TESTING FRAMEWORK 24 for i = 2 to k for j = 1 to totalfeatures if not selected j for m = 1 to iminus if not calculated SR(m,j,class) calculate SR(m,j,class) store in SRMatrix end currentscore = currentscore + featuresrmatrix(m,j); end if currentscore > score score = currentscore; currenthighestfeature = j; end end end answerfeatures(i) = currenthighestfeature; end mrmr implementation Figure 3.2: Optimised DISR pseudocode The reference implementation of mrmr contains an modification that is designed to speed up the calculation time on large datasets (over 1,000 features). It only labels the top 1000 features ranked by their mutual information as possible candidates for selection even if the dataset is much larger. This modification is not detailed in the paper that details the algorithm [17] and changes the criterion, meaning that other experiments performed where the algorithm has been reconstructed cannot be compared with the original experiments given in [17]. It changes the set F used in equation 3.1 so instead of containing all the features, it contains the top 1000 ranked by the mutual information on the class. X mrmr = arg max X n F \S I(X n ; Y ) 1 S 1 X k S I(X n ; X k ) (3.1) This could lead to inaccuracies when the algorithm is executed on a dataset where the standard mutual information is a bad measure for differentiating between features in the dataset, when selecting a number of features close to 1000, or when selecting features from a dataset with an extremely large number of features. As a consequence of this all results for mrmr use a version of the algorithm where this modification has been removed, due to the

38 CHAPTER 3. TESTING FRAMEWORK 25 possible bias consequences when comparing between different feature selection techniques, and so the algorithm used is a strict implementation of the mrmr criterion. In addition to this flaw with the reference implementation, which was corrected in the modified C++ implementation of mrmr used for this research, a similar optimisation to the one used in the DISR implementation was added. Where the DISR implementation only calculates the SR if it hasn t previously calculated this value then the mrmr implementation only calculates the intra feature mutual information if this value hasn t been previously calculated. This results in a similar complexity decrease to the modified DISR implementation. Both the Quotient and Difference variants of mrmr were coded, and both were optimised and corrected in the same way. It is to be noted that this optimisation does not change the performance of the algorithm when compared to a strict implementation of the mrmr criterion, as compared with the modification in the reference implementation Graph implementation The implementation of the graph-based entropy estimator was done in C++, with a MAT- LAB wrapper. It uses Kruskal s algorithm [10] for finding the minimal spanning tree of a graph, as implemented in the Boost C++ library ver 1.35 ( The remainder of the code was implemented around the minimal spanning tree algorithm in C Test Construction There are several problems in analysing datasets with low numbers of samples and high numbers of features. The high number of features raises the probability that any particular strongly correlated feature has only achieved this correlation through random chance, and is unrelated to the classification of the variable. The method of detection for this problem is to acquire more data, but this is often not feasible. The low number of samples causes different issues, as it leads to an over-fitting of the classifier to the data. This is one of the problems that feature selection tries to mitigate as it reduces the complexity of the task, though it is still possible for the feature selection algorithm to over-fit to the task due to the high number of features. Additionally the paucity of samples leads to problems when splitting the data into training and testing sets for the construction of classifiers, as there is not enough data to ensure a satisfactory level of reliability in the results. To alleviate this problem cross-validation is used, where the data is broken up into several folds, with one folds selected for testing, and the remainder used

39 CHAPTER 3. TESTING FRAMEWORK 26 for training. This is repeated until all the folds have been used for testing, and the resultant errors are averaged to give a measure of the performance. Due to the nature of the datasets used for this research leave one out cross-validation is used, where the size of each fold is reduced to 1. Then the test error is the number of incorrect classifications divided by the number of samples. It is not possible to construct error bars with this technique, as the overall error is only determined after all folds have been tested, so there is no mean calculated. The first area of research, into the selection of the first feature in feature selection algorithms uses the 3 different algorithms constructed to select 50 features, with the first feature being selected by a variety of different methods. These are then tested using the 2 selected classifiers, and graphed. The research related to various investigations of the Rényi entropy select 100 features in each of the tests, which are then tested using the 2 selected classifiers. Each generated feature set is tested using the first k features selected by the algorithm, 1 k m where m is the total number of features selected. This generates a graph of the performance to see which subset of the selected features produces the best classification accuracy. The graphs display the classification error rate, ranging between 0 and 0.5 for the Rényi investigation, and between 0 and 0.6 for the first feature investigation. This however means the performance of the first few features can be missing from the graph of the performance of the Rényi algorithms, but the scale would cause visibility issues with the smaller fluctuations in performance if a larger scale was used.

40 Chapter 4 Comparing CMIM, DISR and mrmr 4.1 Introduction Before the various investigations into different properties of the three feature selection algorithms a comparative test will be performed between the unmodified versions of these algorithms on the datasets detailed in chapter Results Each of the tests were performed according to the strategy given in chapter 3. The discrete jumps in error rate present in the graphs are due to the low sample number and the technique of leave one out cross validation, as the error rate can only take values of the form i where n n is the number of samples, and i is an integer Lymphoma Dataset This dataset has 4026 features, discretised into 3 states. There are 9 classes, and 96 examples. The graphs are shown in figure NCI 9 Dataset This dataset has 9712 features, discretised into 3 states. There are 9 classes and 60 examples. The graphs are shown in figure

41 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 28 Figure 4.1: Lymphoma Dataset, 3-NN and SVM Classifiers

42 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 29 Figure 4.2: NCI9 Dataset, 3-NN and SVM Classifiers

43 CHAPTER 4. COMPARING CMIM, DISR AND MRMR Colon Dataset This dataset has 2000 features discretised as 3 states. There are 2 classes and 62 examples. The graphs are shown in figure Leukaemia Dataset This dataset has 7070 features discretised into 3 states. There are 2 classes and 72 examples. The graphs are shown in figure Lung Cancer Dataset This dataset has 325 features discretised into 3 states. There are 7 classes and 73 examples. The graphs are shown in figure Adenocarcinoma Dataset This dataset has features, discretised into 2 states. There are 2 classes and 59 examples. The graphs are shown in figure Analysis Adenocarcinoma Dataset The mrmr-q algorithm has exceptional performance on this dataset, after 6 features for the 3-NN classifier and 9 features for the SVM it is able to select features which perfectly predict the result. This performance is due to the low number of samples, and the way the dataset is discretised, as it removes large amounts of data which could confuse the classifier. The remaining 3 feature selection techniques converge to accuracies below 5%. With the 3- NN classifier CMIM and mrmr-d converge to the same result with a classification error of 1.6%, and DISR converges to a classification error of 3.3%. With the SVM classifier CMIM and DISR converge to the higher error rate, and mrmr-d tends towards the lower error rate Colon Dataset This dataset has best classification performance with the 3-NN classifier, as the SVM error rates increase as more features are added (within the testing range). The CMIM algorithm

44 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 31 Figure 4.3: Colon Dataset, 3-NN and SVM Classifiers

45 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 32 Figure 4.4: Leukaemia Dataset, 3-NN and SVM Classifiers

46 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 33 Figure 4.5: Lung Cancer Dataset, 3-NN and SVM Classifiers

47 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 34 Figure 4.6: Adenocarcinoma Dataset, 3-NN and SVM Classifiers

48 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 35 provides the best performance with the 3-NN classifier, and in this case the result is deterministic as the colon dataset only contains two classes. The mrmr-d algorithm also reaches the best performance level (with an error rate of 6.4%), with only 5 features, but fails to converge to that level as more features are added. Using the SVM classifier the DISR algorithm converges to the lowest error rate, and the DISR algorithm and mrmr-q reach the lowest error rate of 8.1%, using 2 and 10 features respectively Leukaemia Dataset Using both classifiers the CMIM and mrmr-q algorithms achieve the best performance with this dataset, as they converge to a perfect test error rate. With the 3-NN classifier the perfect classification rate is achieved with 4 features by the mrmr-d algorithm, though as it adds more features to the set its performance decreases. With the SVM classifier the perfect classification rate is achieved with 8 features by the DISR algorithm, though again as it adds more features the performance decreases Lung Cancer Dataset Using the 3-NN classifier the performance of all the feature selection algorithms is approximately equivalent. There is large variation around the mean with no algorithm consistently outperforming the others. The lowest classification error with the 3-NN classifier is achieved by DISR, mrmr-d and CMIM of 5.5%. Using the SVM classifier mrmr-d achieves the lowest classification error, with 17 features, and 4.1% error Lymphoma Dataset Using the 3-NN classifier, mrmr-q performs consistently better than the other algorithms, converging to a lower error rate, and achieving the lowest classification error with 37 features and an error rate of 2.1%. Using the SVM classifier the performance of all the techniques is approximately equivalent, with the mrmr-d algorithm achieving the lowest error rate of 2.1% with the fewest features NCI 9 Dataset Using the 3-NN classifier this dataset is extremely variable, with a wide range in classification performance by adding or removing a single feature. This could be in part due to the construction of the dataset, with the varying numbers of examples per class, and the problem

49 CHAPTER 4. COMPARING CMIM, DISR AND MRMR 36 that the nearest neighbour classifier has with multiclass datasets. However the mrmr-q algorithm consistently outperforms the other algorithms after selecting the first 18 features, achieving the lowest classification error of 10% using the first 44 features. Using the SVM classifier, the dataset is more stable with more separation of the different feature selection algorithms. Again the mrmr-q algorithm outperforms the other algorithms, achieving the lowest classification error of 10% first when using 51 features Execution Speed Table 4.1 lists the execution time for the MATLAB/C++ implementations of the algorithms across the various datasets, when executed on an Athlon XP 3000+, with 1GB of RAM, and selecting the 100 features used in the above test results. Dataset CMIM DISR mrmr-d mrmr-q Adenocarcinoma (41672 features) Colon (2000 features) Leukaemia (7070 features) Lung (325 features) Lymphoma (4026 features) NCI9 (9712 features) Table 4.1: Algorithm execution time (s) The results for mrmr-d and mrmr-q are separated for completeness, as the algorithms only differ by an arithmetic operator, and any differences in performance are due to the task scheduling in the CPU. There is a small indication that in large datasets the costly floating point division operation slightly increases the execution time of the mrmr-q algorithm, but the increase is negligible. The optimised CMIM implementation given in [4] provides extremely fast results, due to the minimisation step in the algorithm specification, causing it to not evaluate the mutual information when it would not affect the current feature choice. This decouples the algorithm s time complexity from the number of features, and thus there is little correlation between the number of features and the execution time of the algorithm. In contrast the DISR algorithm evaluates the Symmetric Relevance, which is a costly operation, and requires far more execution time than the mrmr s simple mutual information evaluation.

50 CHAPTER 4. COMPARING CMIM, DISR AND MRMR Conclusions The mrmr set of algorithms consistently perform the best on these selected datasets, though the CMIM algorithm provides a good solution with a vastly reduced execution time, in comparison with the other algorithms Summary This chapter has provided a brief comparison of the 3 different feature selection algorithms. The test results were obtained using the testing framework developed for the rest of the work, and the algorithm implementations used therein. This gives a baseline performance level and a comparison of how the different feature selection algorithms perform when used with the chosen classifiers and the selected datasets.

51 Chapter 5 Investigating the first feature 5.1 Introduction In this chapter the selection of the first feature in common feature selection algorithms is investigated, and numerous alternatives to the standard method are developed and tested. There have been many different feature selection algorithms [6] created due to the growth in processing power and the availability of ever more complex datasets. These algorithms all use variations on information theoretic values to construct a measure for the usefulness of the feature, and so to determine an optimum feature set. All of these algorithms also start from a basic assumption, that the feature with the highest mutual information is the best feature, and should be chosen to start the algorithm. It can be shown that the feature with the highest mutual information is not automatically a member of the ideal feature set. This was shown in [22], but can also be seen in a simple XOR problem by adding a noise feature (see Table 5.1) as shown in [6]. Strict selection of the highest mutual information feature selects the Noise feature and another feature, whereas the best feature pair is XY and does not contain the Noise feature. X Y Noise Class I(X;Class) 0 I(Y;Class) 0 I(Noise;Class) I(X & Y;Class) 1 I(X & Noise;Class) 0.5 I(Y & Noise;Class) 0.5 Table 5.1: XOR Problem The reason that the highest mutual information feature is selected in the first stage of the algorithm is due to the greedy assumption in the forward search. Each of the criteria for 38

52 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 39 selecting features in CMIM, DISR and mrmr reduce to the selection of the highest mutual information when the selected feature set is empty, but as has been demonstrated this might not be the best solution to the problem of selecting the first feature. 5.2 Creating New Criteria To investigate the effect modifying the first feature has on the performance of a feature selection algorithm requires a consideration of what aspects are useful in a given feature that make it a good choice for forming the initial condition of the algorithm. There are two possibilities investigated in this research. The first is to identify a feature that contains the most independent information from the dataset (i.e. the feature with the most information that is not contained in the other features). The second possibility is to identify a feature that combines with other features to give the greatest possible amount of information. These two strategies can be identified as the two main ideas behind the feature selection algorithms considered in this research. CMIM and mrmr aim to find features that are strongly independent of each other, to capture all the information in the dataset, in as few features as possible. DISR in contrast tries to find the features that combine best with the current set of features it has selected, to provide a set of features that work well together to capture the information in the dataset. These two different approaches will be referred to as the independence and combination approaches. By analysing the different approaches of the algorithms several different methods for determining the initial feature were created and tested CMIM Criterion The CMIM selection criterion is created by applying the CMIM feature selection criterion, and taking the whole feature space as the selected set. It selects the feature that is the most independent of the whole feature set, using individual conditional mutual information calculations. It uses the pessimistic assumption in the original CMIM feature selection criterion, by taking the minimum value for the conditional mutual information over all other features as the score for that particular feature, then selecting the feature that produces the maximum value. This criterion takes the independence approach to selecting the first feature. X CMIMS = arg max (min I(X; Y Z)) (5.1) X Z

53 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 40 A modification was proposed that took the sum over the conditional mutual informations, removing the pessimistic assumption as all features are being used in the calculation, so there is no need for approximating the feature independence. This technique removes the benefit of the minimisation, which is to prioritise features which are strongly independent of the whole dataset. The summation changes the criterion so it prioritises the feature with the highest information when all other features are known. This leads to a bias as one high score can distort the selection process, favouring a feature that combines well with a subset of features, over a feature which has strong general performance. The change to a summation makes this criterion take the combination approach to selecting the first feature. X CMIMS Sum = arg max ( X Z I(X; Y Z)) (5.2) Both of these criteria have complexity of O(N 2 ) with N being the number of features in the dataset DISR Criterion The DISR selection criterion is created by applying the DISR feature selection criterion, and finding the two features that have the maximal information jointly on the class. This performs differently to the original criterion as there is no guarantee that the feature with the highest mutual information will combine well with any other feature in the dataset, whereas the new criterion finds the two features that jointly contain the most information about the class. The DISR criterion thus takes the combination approach to selecting the first feature. X DISR = arg max X,Z ( I(XZ, Y ) H(XZ, Y ) ) (5.3) The DISR selection criterion is symmetrical, and thus has a lower complexity than the other criteria developed, because the algorithm need only search for Z > X as all previous values have been calculated at an earlier stage mrmr Criterion There were two versions of the mrmr selection criterion created, to parallel the two different versions of the mrmr feature selection algorithm. The mrmrd selection criterion is created by applying the mrmrd feature selection criterion, using the whole feature space as the selected set. It is similar to the CMIM selection criterion in that it selects the feature that is most independent of the feature set,

54 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 41 but by a different measure. It penalises the candidate features by their redundancy and selects the least redundant feature, over all the features. The pessimistic assumption was used as this aids in selecting a feature which is not skewed by a low redundancy score with only a small number of features. This takes the independence approach to selecting the first feature. X mrmr = arg max (min (I(X; Y ) I(X; Z))) (5.4) X Ω Z The modification replacing the minimisation with a summation was also tested. This selects for the feature which gives the smallest redundancy score. A similar analysis can be applied to the mrmr sum criterion as applied to the CMIM sum criterion, and as there this uses the combination approach to selecting the first feature. X mrmr = arg max X Ω ( Z (I(X; Y ) I(X; Z))) (5.5) The mrmrq selection criterion is the same as the mrmrd criterion, except using a quotient instead of a subtraction to penalise redundancy. As the quotient is inherently more unstable than the subtraction this gives different results to the mrmrd version, but with similar properties. X mrmr = arg max (min(i(x; Y ) )) (5.6) X Ω Z I(X; Z) X mrmr = arg max X Ω ( Z ( I(X; Y ) )) (5.7) I(X; Z) Computational complexity Each of these proposed new first feature selectors gives an increase in the computational complexity of the first stage of the algorithms. In general these algorithms have complexity order n 2 where n is the number of features, compared with the standard highest mutual information method which has complexity n. This is due to the double loop in each criteria where the whole feature set is compared with the currently selected feature, compared to the single loop for ranking the features by mutual information. In addition, the implementations of the mrmr and CMIM algorithms also require a list of the mutual informations between the features and the class to save recalculating this value at each stage, so the complexity n loop must stay in the algorithm.

55 CHAPTER 5. INVESTIGATING THE FIRST FEATURE Summary The different criteria for selecting the first feature in a feature selection algorithm have been derived through applications of the criteria used for the selection of the remaining features. Each one selects a feature that has the properties that the algorithm is selecting for, with the aim of improving the performance of the algorithm as a whole. As the feature selection algorithms gain more complex information theoretic criteria to select features, the first step remains an area where there is little attention paid. The first step in all of these algorithms has not progressed in a similar manner, as it is still taken from the simple ranking procedure that was first developed to select features. 5.3 Results Each of the first feature selectors was paired with each of the feature selection algorithms, and tested on the variety of datasets listed in chapter 3. The results are presented from the multiclass datasets, as these provide the most variation Lung cancer dataset This dataset has 325 features discretised into 3 states. There are 7 classes and 73 examples. The results are split by classifier and feature selector. Figure 5.1 contains the results for CMIM and DISR using a 3-NN classifier, figure 5.2 contains the results for mrmr using the same classifier. Figures 5.3 & 5.4 contain the results when using a linear SVM NCI 9 Dataset This dataset has 9712 features, discretised into 3 states. There are 9 classes and 60 examples. Figure 5.5 contains the results for CMIM and DISR using a 3-NN classifier, figure 5.6 contains the results for mrmr using the same classifier. Figures 5.7 & 5.8 contain the results when using a linear SVM Lymphoma Dataset This dataset has 4026 features, discretised into 3 states. There are 9 classes, and 96 examples. Figure 5.9 contains the results for CMIM and DISR using a 3-NN classifier, figure 5.10 contains the results for mrmr using the same classifier. Figures 5.11 & 5.12 contain the results when using a linear SVM.

56 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 43 Figure 5.1: Lung Cancer Dataset, 3-NN Classifier, CMIM and DISR

57 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 44 Figure 5.2: Lung Cancer Dataset, 3-NN Classifier, mrmr-d and mrmr-q

58 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 45 Figure 5.3: Lung Cancer Dataset, Linear SVM Classifier, CMIM and DISR

59 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 46 Figure 5.4: Lung Cancer Dataset, Linear SVM Classifier, mrmr-d and mrmr-q

60 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 47 Figure 5.5: NCI9 Dataset, 3-NN Classifier, CMIM and DISR

61 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 48 Figure 5.6: NCI9 Dataset, 3-NN Classifier, mrmr-d and mrmr-q

62 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 49 Figure 5.7: NCI9 Dataset, Linear SVM Classifier, CMIM and DISR

63 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 50 Figure 5.8: NCI9 Dataset, Linear SVM Classifier, mrmr-d and mrmr-q

64 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 51 Figure 5.9: Lymphoma Dataset, 3-NN Classifier, CMIM and DISR

65 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 52 Figure 5.10: Lymphoma Dataset, 3-NN Classifier, mrmr-d and mrmr-q

66 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 53 Figure 5.11: Lymphoma Dataset, Linear SVM Classifier, CMIM and DISR

67 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 54 Figure 5.12: Lymphoma Dataset, Linear SVM Classifier, mrmr-d and mrmr-q

68 CHAPTER 5. INVESTIGATING THE FIRST FEATURE Analysis As can be seen from the results above, the mrmr-q feature selection algorithm is highly sensitive to the choice of the first feature, and produces a wide variation in classification performance when the first feature is changed. Additionally the feature selection algorithms tend towards a similar classification performance, and select a similar set of features Lung Cancer Dataset The best performing new first feature selector over this dataset is the DISR selector, as it has equivalent or better performance to the standard method of selecting the feature with the highest mutual information. The lowest classification error when using the SVM is found by combining the mrmr-q selector with the DISR feature selection algorithm, which results in a 2% classification error, when using the first 15 features selected by the algorithm. The lowest classification error when using the 3-NN classifier is found by combining the mrmr-q selector with the mrmr-d feature selection algorithm, which results in a 4% classification error, when using the first 37 features NCI 9 Dataset This dataset is a good example of the differences between classifiers, with the various different feature sets giving the same SVM performance, whilst the 3-NN classifier has a large variation in classification performance between the different feature sets. Part of this variability could however be due to the nature of the dataset, as it has 9 different classes, with a wide range in the number of examples of each class. This confuses the 3-NN classifier as it will often reach a situation where it has 3 nearest neighbours, each with a different class, at which point the classification is determined by a random number generation. With the SVM only the mrmr-q sum selector generates a sufficiently different feature set to give a different classification performance, but only with the mrmr-d feature selector does it provide a feature set with a significantly better feature set. The lowest classification error with the SVM is found by the CMIM sum selector, the mrmr-d selector, the mrmr- D sum selector and the standard method simultaneously using the mrmr-q feature selector, where it achieves a classification error of 12%. The results for the 3-NN classifier are too variable to draw any conclusions about the performance of the different first feature selectors.

69 CHAPTER 5. INVESTIGATING THE FIRST FEATURE Lymphoma Dataset The best performing first feature selectors over this dataset are the mrmr-d, and the mrmr-d sum selectors, as they perform consistently better than the standard method when using feature selection methods other than DISR. When using the DISR feature selection algorithm their performance is approximately equivalent to the standard method. The lowest classification error with the SVM classifier is found by the mrmr-d, and mrmr-d sum selectors with the mrmr-q feature selection algorithm, and with the mrmr-q and mrmr-q sum selectors when using the mrmr-d feature selection algorithm, with a classification error of 1%. It must be noted that the CMIM selector provided this lowest level of classification error with the mrmr-q feature selection algorithm when using 23 features, before converging with the standard selector as more features were added to the selected set Summary From these results it can be seen that changing the initial choice of feature may provide a worse basis for classification when used alone, if a feature other than the one with the highest mutual information is chosen, however it generally improves the resulting feature set. This is expected because of the way the different first features are selected. The different methods are selecting features which combine well with the rest of the dataset, but do not necessarily provide good classification performance when taken on their own (as the mutual information is a measure of the shared data between the feature and the class, the highest mutual information is the feature with the best classification performance when only using one feature). However when the selected feature sets are tested it can be seen that modifying the first feature so it is the highest ranked by another criteria than the highest mutual information improves or maintains classification performance. 5.5 Conclusions This investigation into the selection of the first feature in common feature selection algorithms has created several new ways of selecting the first feature. The empirical testing of these different selectors has shown that several of them improve classification performance, though the amount of improvement is dataset dependent, and in general only becomes apparent after 10 or more features have been selected. The best performing selectors were those based upon the mrmr feature selection algorithm, though this may in part be due to the datasets involved, as the mrmr algorithm generated the best performing feature sets

70 CHAPTER 5. INVESTIGATING THE FIRST FEATURE 57 suggesting that the datasets have a low amount of intra feature interactions. In general the results indicate that selecting an independent feature is more important than selecting a feature that combines well with the rest of the dataset, though the sum selectors also show an improvement over the standard method Summary of the work This chapter has detailed an investigation into the selection of the first feature in common feature selection algorithms. It has: Stated why the standard method for selecting the first feature is not optimal Constructed a set of different methods for selecting the first feature Empirically tested these methods against the standard method, using a variety of datasets, classifiers and feature selection algorithms Analysed the test results, and concluded that selecting an independent feature on average improves the classification performance

71 Chapter 6 Investigating the Rényi measures of information 6.1 Introduction The Rényi extension to entropy, and the associated extension to the Kullback-Liebler divergence provides a framework for a parametrised measure of information. However various properties of the Shannon entropy do not hold under Rényi s extension, including the creation of conditional entropies and the relationship between the mutual information and entropy. To create a measure of information that is valid over the whole space of the Rényi entropy and generalised divergences means the measure must be parametrised by α, it must measure the amount of shared information between two variables, with a high value when the variables are strongly correlated, and a low value when they are strongly independent, and it must be symmetric with respect to the placement of X & Y. Also the measure should only use the Rényi entropy and generalised divergences, instead of the Shannon entropy and the Kullback-Liebler divergence. This leads to three competing formulations of the mutual information in the Rényi space, which will be compared in detail in this chapter. 6.2 Different Formulations of Rényi Information The Shannon mutual information can be described in 3 different ways. It is the sum of the individual entropies minus the joint entropy, the entropy minus the conditional entropy, and the Kullback-Liebler divergence between the joint probability density and the sum of the marginal (individual) probability densities. All these definitions are equivalent when using 58

72 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 59 the Shannon entropy, and given in equations (6.1) & (6.2). However, when using the Rényi entropy & divergence all these formulations give different results (equation (6.3)). I(X; Y ) = H(X) H(X Y ) I(X; Y ) = H(Y ) H(Y X) (6.1) I(X; Y ) = H(X) + H(Y ) H(X, Y ) I(X; Y ) = D KL (p(x, y) p(x)p(y)) (6.2) H α (X) H α (X Y ) H α (Y ) H α (Y X) H α (X) + H α (Y ) H α (X, Y ) (6.3) The standard formulation of the Rényi mutual information (derived from the information gain measure in [19]) is to use the Rényi generalised divergence between the joint probability density and the sum of the marginal probability densities, as in the Kullback-Liebler definition of Shannon s mutual information. This gives a measure which tends towards the Shannon mutual information as α 1 from both sides. Additionally the measure meets the three requirements set out in the introduction, namely that it is parametrised by α, measures the shared amount of information between two variables, and is constructed solely from Rényi s measures. I αdiv (X; Y ) = D α(p(x, y) p(x)p(y)) (6.4) There is no reason why the mutual information cannot be constructed from the sum of the entropies minus the joint entropy, as this formulation simply returns the uncertainty which is jointly contained in both of the variables, or the information gain when about a variable when another is known. It can be shown that this value also tends towards the Shannon mutual information as α 1, as the individual Rényi entropies tend to the Shannon entropy at this value, so the overall value must approach the Shannon mutual information. Intuitively then this value must be equal to the value derived from the divergence measure, as this is also measuring the information gain about one variable when another is known. However this is not the case with the joint entropy based measure, and the divergence measure both converging to the Shannon mutual information as α 1, but diverging when α 1. I αjoint (X; Y ) = H α(x) + H α (Y ) H α (XY ) (6.5) Creating the conditional formulation of the Rényi entropy is more difficult. As has been shown (in chapter 2), the value derived from simple conditioning of the Rényi entropy by

73 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 60 another variable diverges from the value derived by taking the joint entropy minus the entropy of the conditioning variable (H α (XY ) H α (Y ) H α (X Y )). As the definition of the conditional entropy is the uncertainty remaining in one variable when the conditioning one is known, and this is the value the equation provides, then the conditional entropy has an unknown meaning in Rényi s information theory. I αcond (X; Y ) = H α(x) H α (X Y ) (6.6) I αcond (Y ; X) = H α(y ) H α (Y X) (6.7) I αcond (X; Y ) I αcond (Y ; X) (6.8) As these values are asymmetric it is not suitable for use as a mutual information. This can be seen in figure 6.1, which uses data from the Lung dataset, with X = feature 40, and Y = class, graphed for 0.9 α Both of these equations converge to the Shannon mutual information as α 1 and they also diverge from the other formulations of the Rényi mutual information discussed above. It is clear that the conditional equations and the joint formulation must diverge as the conditional entropy is no longer equal to the joint entropy minus the entropy of the conditioning variable, and it is the substitution of this relationship into either entropy formulation for the Shannon mutual information that relates the two. However as these equations are asymmetric with respect to the positions of X & Y they cannot be used as a valid measure of the mutual information. Of the three different formulations possible for the mutual information the standard divergence method is the accepted one, as it directly takes a measure of the similarity of two probability distributions as the mutual information, and the set of divergences was constructed as an information gain measure by Rényi in [19]. However, as the joint entropy is a valid construct in Rényi s information theory, the use of the joint formulation as a mutual information should be equally valid, as it is still a measure of the information gain when one variable of a pair is known. The conditional formulation is less well founded. The conditioning of a variable should be a property that is independent of the entropy measure used, as it effectively splits the variable into several separate sub-variables depending upon the value of the conditioning variable. However due to the position of the logarithm in the Rényi entropy it is not possible to separate the variables and apply Bayes rule to gain the joint entropy and the entropy of the

74 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 61 Figure 6.1: The asymmetry of the conditional formulation conditional variable. Whilst the conditional equations do appear similar to a mutual information it is due to the lack of symmetry, and the unknown properties of the equation derived for the conditional entropy, which prevents it from being used as a mutual information. Figures 6.2, 6.3 & 6.4 show the varying ways the different formulations of the mutual information diverge. The datapoint at α = 1 is generated with the Shannon mutual information to give continuity to the graph, with the lines of best fit passing through this point when graphed in the region around α = 1 incrementing α in steps of The various different formulations vary in data dependent ways with no formulation having constant properties in relation to the others. 6.3 Constructing the Conditional Mutual Information Due to the problems in finding a conditional entropy that has a valid meaning in Rényi s information theory the conditional mutual information is harder to form. The possible

75 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 62 Figure 6.2: Rényi mutual information - lung dataset X = feature 14, Y = class formulations for the conditional mutual information using Shannon entropy are given below: I(X; Y Z) = H(X Z) H(X Y Z) = H(X Z) + H(Y Z) H(XY Z) (6.9) I(X; Y Z) = z Z p(z)d KL (p(x, y Z = z) p(x Z = z)p(y Z = z)) (6.10) As the first two formulations use the conditional entropy, these are not valid, and have strange behaviour as shown in figure 6.5. The divergence formulation is more stable, and is thus used as the basis for the Rényi conditional mutual information. Therefore the equation for the conditional mutual information using the Rényi generalised divergence is given in equation (6.11). I α (X; Y Z) = z Z p(z)d α (p(x, y Z = z) p(x Z = z)p(y Z = z)) (6.11) Now all the different measures have been constructed it is possible to extend the three feature selection algorithms chosen (CMIM, DISR and mrmr) to work with the Rényi

76 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 63 Figure 6.3: Rényi mutual information - lung dataset X = feature 29, Y = class entropy and information measures. 6.4 Reconstructing the Algorithms Due to the complexities of constructing the algorithms in the Rényi space, there were different ways the algorithms could be constructed. The algorithms based around simple mutual informations with Shannon entropy could have two possible formulations, either using the joint entropy form of mutual information, or the divergence form of mutual information. The CMIM algorithm which is based around the conditional mutual information has only one valid formulation, due to the inconsistencies in the constructed version of the Rényi conditional entropy.

77 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 64 Figure 6.4: Rényi mutual information - lung dataset X = feature 49, Y = class Constructing CMIM As there is only one valid form of the conditional mutual information with the Rényi entropy, then the reconstruction of the CMIM algorithm is straightforward. The conditional mutual information function in the original algorithm can be simply replaced with the conditional mutual information divergence formulation, to lead to a valid algorithm. However the initial state of the CMIM algorithm involves populating the score vector with the mutual information of each feature on the class, and there are two valid forms for this mutual information (namely the divergence and the joint entropy). The divergence formulation was chosen as this is the valid form for the conditional mutual information, and the testing is designed to investigate different formulations of the Rényi mutual information Constructing DISR & mrmr Both of these algorithms use standard mutual informations, which can be replaced with two different formulations when working with the Rényi entropy. Two versions of the algorithms

78 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 65 Figure 6.5: The different formulations of conditional mutual information were constructed, with the first version using the divergence formulation for the Rényi mutual information, and the second using the joint formulation for the Rényi mutual information. Additionally the joint entropy used in the calculation of the Symmetric Relevance (equation (2.20)) in the DISR algorithm was replaced with the Rényi joint entropy. In both cases all the mutual informations, including the initial feature ranking step to determine the first feature, were replaced with Rényi mutual informations. 6.5 Results The tests were ran according to the plan detailed in chapter 3. As the α parameter in the Rényi Entropy and the Rényi generalised divergence is a continuous parameter and the performance of the algorithms cannot be defined by an equation it is impossible to fully capture the behaviour of the algorithms when the alpha parameter is varied. An

79 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 66 approximation is gained by sampling the results at fixed values of α. It was decided to sample the α parameter between 0.1 and 2.0 in steps of 0.1, to provide an estimate of the selection power of different values of α. Each test run provides 19 selected feature sets, in addition to the standard Shannon algorithms, which are provided for comparison. This data is separated out over 4 different graphs, to preserve the readability of the results, with the Shannon algorithms provided on each graph for comparison, and the SVM and 3-NN classifier results provided on separate graphs. This leads to 8 graphs per valid formulation of each algorithm with the Rényi mutual information, per dataset. A subset of the graphs with interesting properties are given below CMIM Results There is one valid form for the CMIM algorithm when working with the Rényi mutual information. The results are shown for the lung dataset, and the NCI9 dataset. Lung Dataset Two results are presented, 0.6 α 1.1 and 1.6 α 2.0, in figures 6.6 and 6.7. In figure 6.6, the algorithm using α = 0.6 outperforms the rest of the α values, by being consistently lower than the traditional Shannon version of CMIM, when using the 3-NN classifier, and after the first 15 features have been selected. When selecting the last 15 features a noticeable divide opens between α < 0.9 and 0.9 α 1.1 including the Shannon algorithm. In general after selecting the 15th feature, the algorithms with α < 1 provide an increase in classification performance over the standard Shannon algorithm. When using the SVM classifier, the results are less clear. The values either side of 1 follow the Shannon classification error quite closely until 50 features have been selected, where the differences in the features chosen cause divergence in the classification performance. The α values below 0.9, have an erratic classification performance until after the 40th feature has been selected, where they start to improve upon the classification performance of the other values. In figure 6.7, the higher values of α are outperformed by the Shannon algorithm when using the SVM classifier, only at the end producing a set of features which is an improvement on the one generated by the Shannon algorithm. Using the 3-NN classifier the picture is less clear, as the classification errors are more variable, with the high α values appearing to provide a performance increase when using selecting up to 30 features, before the performance becomes similar to the standard Shannon algorithm.

80 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 67 Figure 6.6: CMIM, Lung Cancer Dataset, 0.6 α 1.1

81 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 68 Figure 6.7: CMIM, Lung Cancer Dataset, 1.7 α 2.0

82 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 69 NCI9 Dataset Four results are presented, but only using the SVM classifier, as due to the number of classes in the NCI9 dataset and the properties of the 3-NN classifier the results are extremely variable. The results cover the whole range of α tested, with 0.1 α 1.1 in figure 6.8 and 1.2 α 2.0 in figure 6.9. In figure 6.8, α < 0.6 reaches a low classification error faster than the standard Shannon algorithm, but plateaus at that level and doesn t provide a decrease in classification error as the number of features increases. For 0.6 α 1.1 the performance is approximately the same as the Shannon algorithm. In figure 6.9, where 1.2 α 1.6 the performance is equivalent to the Shannon algorithm except when selecting the features between 40 and 50, where the performance improves over the standard Shannon algorithm. Additionally when α = 1.4 in this region it has a classification error 17.5 percentage points lower than the Shannon algorithm DISR Results There are two valid forms for the DISR algorithm when working with the Rényi mutual information, the divergence formulation, and the joint entropy formulation. The results presented form a comparison between the two different formulations, and what advantages, if any, they provide. NCI 9 Dataset Three results are presented, using the SVM classifier. These results show the contrast between the classification power of the divergence formulation and the joint entropy formulation. In figure 6.10 the variation between the two formulations of the Rényi mutual information at lower values of α can be seen, with the joint entropy formulation generating classification errors greater than 50%, where the divergence formulation generates classification errors in the region of 25-30%. This shows that for low values of α the joint entropy formulation is unable to select features according to their information. For 0.4 α 0.5 the two formulations perform equivalently, though the joint formulation still varies more than the divergence formulation for the value α = 0.4. In figure 6.11 the variation between the two formulations is minimal, with the joint formulation being closer centred on the standard Shannon algorithm.

83 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 70 Figure 6.8: CMIM, NCI9 Dataset, SVM Classifier, 0.1 α 0.5 & 0.6 α 1.1

84 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 71 Figure 6.9: CMIM, NCI9 Dataset, SVM Classifier, 1.2 α 1.6 & 1.7 α 2.0

85 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 72 Figure 6.10: DISR Joint and Divergence, NCI9 Dataset, SVM Classifier, 0.1 α 0.5

86 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 73 Figure 6.11: DISR Joint and Divergence, NCI9 Dataset, SVM Classifier, 0.6 α 1.1

87 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 74 Figure 6.12: DISR Joint and Divergence, NCI9 Dataset, SVM Classifier, 1.7 α 2.0

88 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 75 In figure 6.12, the variation between the two formulations increases, with the joint formulation losing classification accuracy as the α value increases mrmr-d Results There are two valid forms for the mrmr-d algorithm when working with the Rényi mutual information, the divergence formulation, and the joint entropy formulation. The results presented form a comparison between the two different formulations, and what advantages, if any, they provide. Adenocarcinoma Dataset The adenocarcinoma dataset has two classes, and thus the 3-NN classifier is not subject to the problems found with multiclass datasets. These results show the difference in classification performance between the joint formulation and the divergence formulation. Figure 6.13 shows the results from the SVM classifier for both formulations. The divergence formulation consistently performs better for 0.2 α 0.5, compared with both the joint formulation and the standard Shannon algorithm. The joint formulation outperforms the Shannon algorithm whilst selecting features 23 through 62, but converges to the Shannon result for higher numbers of features. Figure 6.14 shows the results from the 3-NN classifier for both formulations. With this classifier both formulations are outperformed by the Shannon algorithm, with the joint formulation initially performing as well as the Shannon algorithm before strongly diverging to give a higher classification error. Colon Dataset The colon dataset also has two classes, but the results presented are from the SVM classifier. These results show how the joint formulation can improve upon the classification performance of the divergence formulation for given α values. Figure 6.15 shows the results from the SVM classifier with 1.2 α 1.6. The joint formulation consistently outperforms the divergence formulation for all values except when α = 1.6 where the performance is mixed when compared with the Shannon version, but results in a much lower classification error. The minimum error achieved by the divergence formulation is 3.2%, selecting 61 to 65 features, with α = 1.5, 1.4 and briefly α = 1.2 (though this value gives an highly varying performance in the set of features selected). After this low point the classification error for the values of α shown rise until they exceed the error for the standard Shannon algorithm.

89 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 76 igure 6.13: mrmr-d Joint and Divergence, Adenocarcinoma Dataset, SVM Classifier, 0.1 α 0.5

90 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 77 igure 6.14: mrmr-d Joint and Divergence, Adenocarcinoma Dataset, 3-NN Classifier, 0.1 α 0.5

91 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 78 With the joint formulation this low point of classification error is reached consistently when α = 1.6 selecting 89 features, though all other solutions remain below the error rate for the Shannon algorithm and do not rise in a similar manner to the divergence formulation. Lymphoma Dataset The lymphoma dataset has 9 classes, but generates lower classification errors than expected as can be seen from the results in chapter 4. Figure 6.16 shows the results from the SVM classifier with 0.1 α 0.5 for both formulations. The joint formulation outperforms the divergence formulation consistently after selecting the first 10 features and its performance is consistently lower than the Shannon algorithm after selecting the first 40 features. In contrast the divergence formulation only begins to outperform the Shannon algorithm after 76 features have been selected. Figure 6.17 shows the results from the SVM classifier with 1.7 α 2.0 for both formulations. In contrast to the previous results, here the divergence formulation performs equivalently to the Shannon algorithm, whilst the joint formulation is outperformed by more than 5 percentage points on average. This shows how the changing α parameter modifies the properties of the different formulations in different ways therefore results derived from one formulation cannot be transferred to another mrmr-q Results There are two valid forms for the mrmr-q algorithm when working with the Rényi mutual information, the divergence formulation, and the joint entropy formulation. The results presented form a comparison between the two different formulations, and what advantages they provide. Colon Dataset The results presented for the colon dataset using the mrmr-q algorithm are similar to those presented for the mrmr-d algorithm, with the joint formulation outperforming the divergence formulation with 1.2 α 1.6 (presented in figure However the joint formulation has one more noticeable feature using the mrmr-q algorithm, which is the unusual performance of the value α = 1.5 when selecting the 28th and 29th features, where the classification error drops substantially to give only one incorrect classification under the leave-one-out cross validation testing. This result is not repeated in other tests, and is the lowest error estimate gained for this dataset using the SVM classifier.

92 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 79 igure 6.15: mrmr-d Joint and Divergence, Colon Dataset, SVM Classifier, 1.2 α 1.6

93 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 80 igure 6.16: mrmr-d Joint and Divergence, Lymphoma Dataset, SVM Classifier, 0.1 α 0.5

94 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 81 igure 6.17: mrmr-d Joint and Divergence, Lymphoma Dataset, SVM Classifier, 1.7 α 2.0

95 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 82 igure 6.18: mrmr-q Joint and Divergence, Colon Dataset, SVM Classifier, 1.2 α 1.6

96 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 83 Figure 6.19 presents the results using the SVM classifier with 1.7 α 2.0. In this result it can be seen how the joint formulation has greater variance depending upon the α value, compared to the divergence estimate, which is highly convergent over this range. Additionally the performance of the joint formulation can be seen to decrease in this range as the α increases, whilst it remains an improvement upon the Shannon algorithm for α = 1.7, 1.8 the performance is not consistently better over the selection of 100 features, with the end of the testing range providing the most stable performance and the best result for the Rényi mutual information. Figure 6.20 presents the results using the 3-NN classifier with 1.7 α 2.0. The performance of the Rényi mutual information in the algorithms is lower than the Shannon algorithm, only occasionally improving upon the classification error. This is representative of several results where the SVM performance is not replicated under the 3-NN classifier, when using different versions of the Rényi feature selection algorithms. Lung Dataset Figure 6.21 presents the results from the SVM classifier with 1.7 α 2.0. In the results the divergence formulation outperforms the joint formulation and the Shannon algorithm after 50 features have been selected, though this happens at the higher end of the α range, which is uncommon in the rest of the results. The joint formulation outperforms the Shannon algorithm when using features, but only equals the best classification performance of the Shannon algorithm when it does so, in contrast to the divergence formulation outperforming the Shannon algorithm s best by 3 percentage points. Lymphoma Dataset Figure 6.22 presents the results from the SVM classifier with 0.1 α 0.5. In the results the divergence formulation has less variance around the Shannon algorithm, with performance approximately the same, and achieving the same best classification error. The joint formulation has much more variance in the results with α = 0.1, 0.3 providing the best classification results, and achieving a minimal classification error. Figure 6.23 presents the results from the SVM classifier with 0.6 α 1.1. In the results both formulations perform equivalently to the Shannon algorithm, with a slightly greater variance in the results of the joint formulation. This leads to the joint formulation gaining the optimal classification accuracy before converging with the Shannon algorithm.

97 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 84 igure 6.19: mrmr-q Joint and Divergence, Colon Dataset, SVM Classifier, 1.7 α 2.0

98 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 85 igure 6.20: mrmr-q Joint and Divergence, Colon Dataset, 3-NN Classifier, 1.7 α 2.0

99 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 86 Figure 6.21: mrmr-q Joint and Divergence, Lung Dataset, SVM Classifier, 1.7 α 2.0

100 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 87 igure 6.22: mrmr-q Joint and Divergence, Lymphoma Dataset, SVM Classifier, 0.1 α 0.5

101 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 88 igure 6.23: mrmr-q Joint and Divergence, Lymphoma Dataset, SVM Classifier, 0.6 α 1.1

102 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION Analysis Analysis of CMIM variations The Rényi version of the CMIM algorithm can only use the divergence formulation as the others use the invalid Rényi conditional entropy. Therefore the only parameter varied in the CMIM testing was the α parameter. The varying of this parameter leads to some improvement over the standard Shannon method of measuring the mutual information, but it is dataset dependent and does not provide a measurable improvement over all of the tested datasets. Over all the datasets the extremes of the α parameter provided little improvement over the Shannon algorithm, though the region around α = 1 provides similar performance to the Shannon algorithm with occasional improvements to the classification error Analysis of DISR variations The Rényi version of the DISR algorithm can use both the joint and divergence formulations of the Rényi mutual information. Therefore there are two parameters to be varied in the DISR testing, both the α parameter and the formulation of the mutual information. The change to the Rényi mutual information does not result in any noticeable improvements to the classification performance of the feature sets produced by the DISR algorithm, though the region around α = 1 still provides similar and occasionally better performance compared with the Shannon version of the algorithm. The joint formulation of the mutual information is more variable at either end of the tested range using this algorithm, with classification errors 25 percentage points greater than the Shannon algorithm for α < 0.3, and up to 10 percentage points greater for 1.7 α Analysis of mrmr-d variations The Rényi version of the mrmr-d algorithm can use both the joint and divergence formulations of the Rényi mutual information. Therefore there are two parameters varied in the testing of the Rényi mrmr-d algorithm. The results using this algorithm show the joint formulation has more variance than the divergence formulation, ranging above and below the classification errors of the Shannon algorithm, and thus can be an improvement upon the standard algorithm. For α < 0.5 the joint formulation performs better than the Shannon algorithm on the datasets though this is only for the SVM classifier. The feature sets generated by the Rényi mrmr-d algorithm using the joint formulation do not perform well with the 3-NN classifier, and only have a performance increase when using the SVM.

103 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 90 The region where 1.2 α 1.6 gives a better performance from both the formulations, but α > 1.6 degrades the performance of the joint formulation Analysis of mrmr-q variations The Rényi version of the mrmr-q algorithm is fundamentally the same as the mrmr-d algorithm and thus the same testing constraints apply. Due to the similarity between the mrmr-d and mrmr-q criteria the variation of the α parameter and changing the mutual information formulation provides similar results to the mrmr-d criterion. 6.7 Conclusion The results from the empirical testing of the Rényi mutual information, and the algorithms constructed using it show that in general there is a value of α which can improve the classification result, with both formulations of the mutual information. The lack of a clear value or formulation that consistently provides the best result means that the value of α and the formulation must be added as a layer of model inference to the system, or an arbitrary value chosen for comparison of the work. The problem with adding a layer of inference (as must be done if a highly performing feature set is to be chosen by picking from among several different algorithms) is that it causes more complexity in the model of the data which leads to over-fitting of the problem, especially in low sample fields such as genomics. In general the performance of the divergence formulation mirrors or slightly exceeds the performance of the Shannon algorithm, and the joint formulation provides a more variable level of performance around the Shannon-based algorithms, dictated by the α value, the dataset and the classifier used. A finer grained search of the region 0 < α 2.0 could be performed to aid the estimation of which values of the α parameter improve upon the Shannon algorithms, but this would still be a dataset dependent problem, unless a wider ranging study of more data was performed Summary of the work This chapter has detailed an investigation the properties of the Rényi entropy, and the various Rényi mutual informations that can be constructed. It has: Constructed 2 different methods for measuring the mutual information with the Rényi entropies and divergences.

104 CHAPTER 6. INVESTIGATING THE RÉNYI MEASURES OF INFORMATION 91 Investigated why the conditional formulation of the mutual information is invalid with the Rényi entropy. Modified the feature selection algorithms investigated previously to use the Rényi mutual informations. Empirically tested these algorithms against the standard method, using a variety of datasets and classifiers, and varying the α parameter. Analysed the test results, and concluded that the Rényi mutual informations can provide an improvement to classification accuracy when using selected values of the α parameter. Concluded that of the two different formulations of the Rényi mutual information, the divergence formulation provides an improvement in classification accuracy over a wider range of α values than the joint formulation.

105 Chapter 7 Graph-based Entropy Estimation 7.1 Introduction 7.2 Graph-based Entropy Image Registration Image registration is the process of identifying a transform that maps similar areas of two different images onto each other. It is used for creating a standard co-ordinate system for working with multiple images of the same area taken from similar non-identical viewpoints. The work by Neemuchwala et al. [14] is concerned with image registration problems, rather than strict feature selection. However it develops a series of interesting information theoretic constructs based around the Rényi entropy, which are used to determine the precise transformation for registering images. The paper [14] provides a review of the previous work carried out by the authors in the image registration field, using several novel information theoretic techniques Graph-based entropy estimation This paper presents a method for calculating the Rényi entropy of a dataset as a whole, in one calculation that scales with the number of samples in the dataset, rather than the dimensionality of the dataspace. It does this through work on entropic graphs, that is graphs which are related to the entropy level of the dataset. It shows that depending on the graph construction, an accurate estimation of the continuous Rényi entropy can be found, through calculation on the lengths of the edges of the graph. These graphs are mathematical 92

106 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 93 constructs, which are un-directed and exist as a set of points and edges which connect the points, and where the edges are weighted according to their length raised to a power γ. Several different kinds of graph are proposed for this, but the type of graph used is the Minimal Spanning Tree (MST). The MST is a graph with a minimal summed edge weight which connects all the points in the space, and contains no closed cycles (so the graph can be represented as a tree with a root node), see figure 7.1, in which the black edges form the minimal spanning tree of the set of points, with the gray edges being unused edges from the set of available edges. In previous work the authors have proved that this forms an accurate estimate of the Rényi entropy [7]. Figure 7.1: The Minimal Spanning Tree of a set of connected points, image sourced from spanning tree.svg, retrieved on: In equation (7.1) Ω is the set of all graphs with the required properties (in this case all spanning trees), s is the euclidean distance between the datapoints in X, i is an iterator over all the edges in E, and γ is the edge exponent, ranging 0 < γ < d where d is the dimension of the feature space. This equation specifies the construction of the minimal graph, and provides an equation for its power-weighted sum L γ (X). L γ (X) is thus a function with value equal to the summation of the edges raised to the power γ of the minimal graph that fits the criterion, i.e. L γ (X) equals the summation of the MST s edge lengths (in figure 7.1 this is equivalent to the sum of the lengths of the bold black lines).

107 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 94 L γ (X) = min E Ω s γ i (7.1) In equation (7.2) α = 1 γ and β d d,γ is a constant independent of the data, depending only on the type of graph and n is the number of datapoints in X. This gives a value for α ranging 0 < α < 1. L γ (X) lim n n α = β d,γ i E f α (x)dx (7.2) In equation (7.3) the equation for the Rényi entropy has been integrated with equation (7.2) to form an expression for the Rényi entropy in terms of the power-weighted sum of the edge lengths, with Ĥα(f) denoting the estimation for the Rényi entropy. This expression is proved in [7] to be a strongly consistent estimator for the Rényi entropy of order α. This framework then provides a Rényi entropy estimator for 0 < α < 1 with the parameter γ being adjustable to generate different α values. Ĥ α (f) α log(β d,γ) = 1 1 α log(l γ(x) n α ) (7.3) This can be intuited by thinking of the Rényi continuous entropy, and what it measures. The Rényi continuous entropy is a measure of the variability of a variable, it takes high values for a wide variation in the data, and a low value for a low variation in the data. The graph linking all the points in the variable will have a similarly high value if there is a large variation in the data as there will be a number of long links between distant points, and will have a low value if there is a small variation in the data as most of the points will be close together. The equations derived give a formalisation of this link Conclusion With the new method of estimating entropy, the paper goes on to detail a method for constructing and estimating the mutual information between two continuous spaces. This definition is not useful for the classification task as it uses an extremely large continuous data sample to provide the information and classification tasks are mainly mapping discrete data onto a discrete class. Then they apply the two derived constructs to the task of image registration. This formulation of the Rényi entropy estimation has a computational complexity which scales with the number of samples in the dataset, rather than the dimensionality of the dataset, and thus provides a framework for doing large scale analyses in a computationally tractable amount of time, similar to the Maximum Dependency feature selection concept in

108 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 95 [17] A note on the continuous entropy The equations give a value for the Rényi continuous entropy, however all the previous work in this paper has been using the discrete entropies. The continuous entropy can be used in a similar way to the discrete entropy, forming mutual informations and other values (in the Shannon space). However it does give much larger values when it is used to calculate a discrete variable. Also the continuous entropy is distorted when the discrete values have non-sequential values, e.g. if a variable takes the values {1, 2, 3} then it will have a lower continuous entropy than if the variable took the values {1, 4, 9} though this would give the same discrete entropy. This issue only causes a problem if the feature vectors have different possible values, however in all the test data they take the same range of values. 7.3 Graph-based Estimation for Feature Selection Once above framework had been created, then it was applied to feature selection by Bonev et al. in [1]. It details an implementation of the graph-based Rényi entropy estimator, used to calculate the Maximum Dependency criterion from [17]. It goes on to state the comparative computational complexities between mrmr (O(n 3 ) where n is the dimension of the feature space) and the new algorithm (O(n 2 )) though it must be noted that the entropy estimation has its own computational complexity which scales as O(s log(s)) where s is the number of samples. The algorithm reuses the Maximum Dependency criterion from [17] (equation (2.16)). It is made computationally tractable by the different method of estimating the Rényi entropy which is used to replace the Shannon entropy given in the original formulation. As the Rényi entropy tends towards the Shannon entropy as α 1, a series of graph-based estimations are performed to capture the behaviour of the function in the region around 1. This is then used to construct an estimate for the Shannon entropy, which is then used to construct the mutual information. X Graph = arg max X n X /S I(X S Xn ; Y ) (7.4) The mutual information is constructed by using the conditional formulation, (i.e. I(X; Y ) = H(X) H(X Y )), so an accurate method for constructing the conditional Rényi entropy is required. Simple conditioning of the Rényi entropy will provide a value that tends towards

109 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 96 the Shannon conditional entropy as α 1, so this should give a valid estimate for the Shannon mutual information, though as been previously seen this value has little meaning when working solely with Rényi entropies. 7.4 Creating the Algorithm For this work it was decided to investigate the properties of the Graph-based entropy estimator when using it solely with the Rényi entropy rather than using it to approximate the Shannon entropy. This changes the formulation as it is based upon the Rényi entropy rather than the Shannon entropy, and the Rényi entropy has weaker additivity properties, though the previous research (along with [5]) indicates that the Rényi entropy is a better measure for quantifying useful information in some cases. For forward selection this creates additional problems, as the estimate of the bias factor which constrains the graph value into an entropy is only valid when using in a high dimensional feature space. This means that the initial stages of forward selection are subject to a much greater bias, and are thus less reliable choices than the later stages. Other approximations can be found by generating graphs in a [0, 1] d space and calculating their lengths, then using this value to estimate the bias. The bias estimator was chosen to be the large dimension approximation used by Neemuchwala et al. for reasons of simplicity, despite an approximation problem with the initial stages of forward selection. The equation is given in (7.5). β γ,d = γ 2 ln( d 2πe ) (7.5) As the graph-based entropy estimator is only valid when the dimension of the space is greater than 1, a different function must be used to select the initial feature. In this work it was chosen to be the divergence formulation of the Rényi mutual information, with the same α parameter as the graph-based estimator. This was decided before the investigation into the different formulations of the Rényi mutual information, and the divergence method was chosen as it is the simplest derivation from the original work by Rényi in [18]. Due to the inability to accurately construct the divergence measure used by Neemuchwala et al. in a discrete space, the mutual information has to be constructed from the various entropies that can be calculated using the graph estimator. As the conditional formulation is ill-founded with the Rényi entropy, the joint formulation (equation (7.6)) was chosen.

110 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 97 I α (X; Y ) = H α (X) + H α (Y ) H α (XY ) (7.6) This raises further problems, as the entropy estimator used for the earlier Rényi work was a discrete estimator, and the graph-based estimator treats it as a continuous space thus providing a much higher estimate for the entropy of the tristate variables used in testing. This causes the H α (X) and H α (XY ) terms which are estimated using the graph-based method to dominate the H α (Y ) term, which is a one dimensional vector of the classes, and is calculated using a discrete Rényi entropy estimator. This leads to a bad feature selection algorithm, as it doesn t prioritise information on the class. Instead it chooses the highest entropy features, leading to poor classification performance. Once the problems with the joint formulation were noticed, the conditional formulation (equation (7.7)) was used for the testing of the algorithm, as it provides an approximation of the mutual information, even though it is ill-founded, as both terms can be calculated using the graph-based estimator. The conditional entropy was calculated by simple conditioning on the graph-based estimator. This formulation was found to select features that were correlated with the class and increase classification performance over the joint formulation. The resultant feature selection algorithm is termed Graph-Based Forward Search (GBFS) for the remainder of this document. I α (X; Y ) = H α (X) H α (X Y ) (7.7) Usefulness of this measure There are two main problems with the use of the graph-based mutual information that has been constructed. Firstly, the bias is a bad approximation for low dimensional spaces, so in the early stages of forward selection the mutual information estimate is wrong. As the bias is constant throughout a particular iteration it doesn t affect the performance of the algorithm, though it does cause negative mutual information values which have to be compensated for in the algorithm. Secondly, the use of the conditional formulation of the mutual information is ill founded with the Rényi entropy. It varies with the correlation between two variables, but it is unknown what factors influence this value as it is not a true mutual information.

111 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION Rényi Entropy Genetic Algorithm A further extension to the graph-based feature selector is proposed, as an alternative to a simple forward search. The majority of feature selection techniques use variations on greedy search strategies, generally using a greedy forward search as the basis for the algorithm. The reason for this is twofold: firstly the algorithm is simple to construct, and secondly, there is no accurate way of estimating the total mutual information of a set of variables upon a class. The graph-based entropy estimator provides a solution to the second problem, as now the mutual information of a set of variables can be calculated in an efficient and accurate manner. As the aim of feature selection is to provide a set of features which contain the maximum information about a class, this function can be maximised over the whole feature space to provide an optimal feature set. The search space for such a maximisation is exponential with respect to the number of features, and thus an exhaustive search is not possible in most cases. One such algorithm for searching such a wide search space is a genetic algorithm, which is based upon evolutionary principles applied to a population of competing solutions What is a Genetic Algorithm Genetic algorithms consist of a population of solutions, with each solution described as a string of numbers or letters and an fitness function which can measure the performance of a particular string [2]. Additionally there are two operations that can be performed on the strings, mutation and crossover. Mutation replaces a single value from the string, with another possible value selected at random. Crossover takes two members of the population and produces a new member of the population with a percentage of its string copied from one string, and the remainder copied from the other string. The algorithm then iterates over a number of generations. At each generation each member of the population is evaluated, with a percentage of the low scoring members being removed. New members of the population are then generated using crossover from the existing high scoring members. Then a percentage of the population is mutated, to generate new (hopefully improved) solutions. This process is a form of stochastic hill climber, with each generation producing an improvement in the average performance of the population.

112 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION Why use a Genetic Algorithm Greedy forward searches have a problem with highly interactive systems, such as feature selection. They assume that the addition of the current optimum to the solution only improves the solution, whereas in feature selection adding a feature to a set can lower the information contained in the set (see chapter 2). Also the three algorithms chosen for study proceed incrementally, adding one feature at a time to the solution set. This doesn t take into account complementarity of features in a useful way, as it can only cope with at best two way complementarity (i.e. two features combining to increase the information). There is no method for coping with adding together useful sets of features without using techniques such as boosting, or ensemble methods. The genetic algorithm doesn t have this problem. By mutating and crossing over solutions it can generate solution sets where many features have changed in one generation, and can so test for interactions between large groups of features. As the process is optimising with respect to the evaluation function this process will slowly provide feature sets with a higher mutual information, which greedy forward search cannot find as they involve making nonoptimal choices for a long term benefit. The levels of crossover and mutation have to be carefully chosen to ensure the system doesn t converge to a local optima whilst ignoring all other optimum solutions, as the populations converge. An increase in the level of mutation helps to increase the variability of solutions, to enable the algorithm to cover more of the search space Constructing the Genetic Algorithm The version of the genetic algorithm used in the testing implements a subset of the features described above due to time constraints. It implements mutation, and the culling & repopulation parts of the algorithm, but does not contain a crossover operator. The algorithm also has a fixed mutation rate, where each feature in the solution has a 5% chance of being mutated into another feature. Additionally each solution has the same fixed length, there is no provision for growing or shrinking the number of features in a given solution. Figure 7.2 contains the pseudocode algorithm for the Graph entropy genetic algorithm. The implementation given for the genetic algorithm is relatively simple to test the validity of the graph-based mutual information as a measure of the performance of the set of selected features. This algorithm will be referred to as the Graph-Based Genetic Algorithm (GBGA) to differentiate it from the forward search method constructed above. It performs g p mutual information calculations, where g is the number of generations and p is the population size,

113 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 100 randomly generate population of solutions, of size popsize test the initial fitness of the solutions for i = 0 : numberofgenerations { copy out the top 5 solutions replace the population with a new population, taken from the top 5 solutions mutate each solution { for j = 0 : solutionlength { generate random number between 0 and 1 if number < 0.05 { swap that feature for an unused one } } - loop over solution length } - mutate function recheck the fitness of the solutions } - loop over generations return the top 5 solutions Figure 7.2: Graph MI Genetic Algorithm Pseudocode in comparison to the forward search, which performs k n mutual information calculations where k is the number of desired features and n is the total number of features. It enables a better coverage of the search space than the forward selector with a smaller number of mutual information calculations, as each calculation in the forward search differs by only 1 feature, whereas in the genetic algorithm each calculation could differ by the string length. 7.6 Results The results for the two graph-based feature selection algorithms have been combined, along with comparison results taken from the Rényi mutual information based versions of the

114 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION feature selection algorithms used in the remainder of this research. The Rényi mutual information chosen was the divergence formulation as this is valid over all the different algorithms. All results were generated with α = 0.9, for a more detailed investigation into the properties of the α parameter see chapter 6. The way of analysing the results used throughout the paper is not optimal for analysing the performance of the genetic algorithm, as it does not do a forward search of the space, and is thus not searching for incrementally better feature sets. Therefore testing the performance of progressively larger feature sets does not relate to the performance of the whole feature set, as the ordering of the features (in the selected set) is unrelated to their performance. The only valid testing result for the genetic algorithm is therefore the final result, i.e. the classification error when using all 100 selected features, as this is the value that is being optimised in the genetic algorithm Leukaemia Dataset Figure 7.3 contains the results of the 6 algorithms compared when using the 3-NN and SVM classifiers Lung Dataset Figure 7.4 contains the results of the 6 algorithms compared when using the 3-NN and SVM classifiers Execution Speed Table 7.1 lists the execution time for the MATLAB/C++ implementations of the algorithms several datasets, when executed on an Athlon XP 3000+, with 1GB of RAM, and selecting the 100 features used in the above test results. The GBGA would not terminate on the Adenocarcinoma dataset, and so no time is recorded for that algorithm. Dataset CMIM DISR mrmr-d GBFS GBGA Adenocarcinoma (41672 features, 59 samples) x Lung (325 features, 73 samples) NCI9 (9712 features, 60 samples) Table 7.1: Rényi algorithm execution time (s) The results given in the table combine both the mrmr variants into one score, as they provide almost the same execution time (see Table 4.1), and the time taken for the mrmr-q algorithm was thus not recorded.

115 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 102 Figure 7.3: Leukaemia Dataset, 3-NN & SVM Classifier, α = 0.9

116 CHAPTER 7. GRAPH-BASED ENTROPY ESTIMATION 103 Figure 7.4: Lung Dataset, 3-NN & SVM Classifier, α = 0.9

INFORMATION-THEORETIC FEATURE SELECTION IN MICROARRAY DATA USING VARIABLE COMPLEMENTARITY

INFORMATION-THEORETIC FEATURE SELECTION IN MICROARRAY DATA USING VARIABLE COMPLEMENTARITY INFORMATION-THEORETIC FEATURE SELECTION IN MICROARRAY DATA USING VARIABLE COMPLEMENTARITY PATRICK E. MEYER, COLAS SCHRETTER AND GIANLUCA BONTEMPI {pmeyer,cschrett,gbonte}@ulb.ac.be http://www.ulb.ac.be/di/mlg/

More information

Information Theory and Feature Selection (Joint Informativeness and Tractability)

Information Theory and Feature Selection (Joint Informativeness and Tractability) Information Theory and Feature Selection (Joint Informativeness and Tractability) Leonidas Lefakis Zalando Research Labs 1 / 66 Dimensionality Reduction Feature Construction Construction X 1,..., X D f

More information

Conditional Likelihood Maximization: A Unifying Framework for Information Theoretic Feature Selection

Conditional Likelihood Maximization: A Unifying Framework for Information Theoretic Feature Selection Conditional Likelihood Maximization: A Unifying Framework for Information Theoretic Feature Selection Gavin Brown, Adam Pocock, Mingjie Zhao and Mikel Lujan School of Computer Science University of Manchester

More information

Feature selection. c Victor Kitov August Summer school on Machine Learning in High Energy Physics in partnership with

Feature selection. c Victor Kitov August Summer school on Machine Learning in High Energy Physics in partnership with Feature selection c Victor Kitov v.v.kitov@yandex.ru Summer school on Machine Learning in High Energy Physics in partnership with August 2015 1/38 Feature selection Feature selection is a process of selecting

More information

Sparse Approximation and Variable Selection

Sparse Approximation and Variable Selection Sparse Approximation and Variable Selection Lorenzo Rosasco 9.520 Class 07 February 26, 2007 About this class Goal To introduce the problem of variable selection, discuss its connection to sparse approximation

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

Filter Methods. Part I : Basic Principles and Methods

Filter Methods. Part I : Basic Principles and Methods Filter Methods Part I : Basic Principles and Methods Feature Selection: Wrappers Input: large feature set Ω 10 Identify candidate subset S Ω 20 While!stop criterion() Evaluate error of a classifier using

More information

CS6375: Machine Learning Gautam Kunapuli. Decision Trees

CS6375: Machine Learning Gautam Kunapuli. Decision Trees Gautam Kunapuli Example: Restaurant Recommendation Example: Develop a model to recommend restaurants to users depending on their past dining experiences. Here, the features are cost (x ) and the user s

More information

3. If a choice is broken down into two successive choices, the original H should be the weighted sum of the individual values of H.

3. If a choice is broken down into two successive choices, the original H should be the weighted sum of the individual values of H. Appendix A Information Theory A.1 Entropy Shannon (Shanon, 1948) developed the concept of entropy to measure the uncertainty of a discrete random variable. Suppose X is a discrete random variable that

More information

Probabilistic Graphical Models for Image Analysis - Lecture 1

Probabilistic Graphical Models for Image Analysis - Lecture 1 Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.

More information

Perspectives of feature selection in bioinformatics: from relevance to causal inference

Perspectives of feature selection in bioinformatics: from relevance to causal inference Perspectives of feature selection in bioinformatics: from relevance to causal inference Gianluca Bontempi Machine Learning Group, Computer Science Department ULB, Université Libre de Bruxelles Boulevard

More information

Active Learning for Sparse Bayesian Multilabel Classification

Active Learning for Sparse Bayesian Multilabel Classification Active Learning for Sparse Bayesian Multilabel Classification Deepak Vasisht, MIT & IIT Delhi Andreas Domianou, University of Sheffield Manik Varma, MSR, India Ashish Kapoor, MSR, Redmond Multilabel Classification

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

Comparison of Shannon, Renyi and Tsallis Entropy used in Decision Trees

Comparison of Shannon, Renyi and Tsallis Entropy used in Decision Trees Comparison of Shannon, Renyi and Tsallis Entropy used in Decision Trees Tomasz Maszczyk and W lodzis law Duch Department of Informatics, Nicolaus Copernicus University Grudzi adzka 5, 87-100 Toruń, Poland

More information

Noisy channel communication

Noisy channel communication Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 6 Communication channels and Information Some notes on the noisy channel setup: Iain Murray, 2012 School of Informatics, University

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability

More information

Degrees of Freedom in Regression Ensembles

Degrees of Freedom in Regression Ensembles Degrees of Freedom in Regression Ensembles Henry WJ Reeve Gavin Brown University of Manchester - School of Computer Science Kilburn Building, University of Manchester, Oxford Rd, Manchester M13 9PL Abstract.

More information

Medical Imaging. Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases

Medical Imaging. Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases Uses of Information Theory in Medical Imaging Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases Norbert.schuff@ucsf.edu With contributions from Dr. Wang Zhang Medical Imaging Informatics,

More information

CS Homework 3. October 15, 2009

CS Homework 3. October 15, 2009 CS 294 - Homework 3 October 15, 2009 If you have questions, contact Alexandre Bouchard (bouchard@cs.berkeley.edu) for part 1 and Alex Simma (asimma@eecs.berkeley.edu) for part 2. Also check the class website

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Feature Engineering. Knowledge Discovery and Data Mining 1. Roman Kern. ISDS, TU Graz

Feature Engineering. Knowledge Discovery and Data Mining 1. Roman Kern. ISDS, TU Graz Feature Engineering Knowledge Discovery and Data Mining 1 Roman Kern ISDS, TU Graz 2017-11-09 Roman Kern (ISDS, TU Graz) Feature Engineering 2017-11-09 1 / 66 Big picture: KDDM Probability Theory Linear

More information

Machine Learning. Lecture 9: Learning Theory. Feng Li.

Machine Learning. Lecture 9: Learning Theory. Feng Li. Machine Learning Lecture 9: Learning Theory Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Why Learning Theory How can we tell

More information

Generative MaxEnt Learning for Multiclass Classification

Generative MaxEnt Learning for Multiclass Classification Generative Maximum Entropy Learning for Multiclass Classification A. Dukkipati, G. Pandey, D. Ghoshdastidar, P. Koley, D. M. V. S. Sriram Dept. of Computer Science and Automation Indian Institute of Science,

More information

Quantitative Biology Lecture 3

Quantitative Biology Lecture 3 23 nd Sep 2015 Quantitative Biology Lecture 3 Gurinder Singh Mickey Atwal Center for Quantitative Biology Summary Covariance, Correlation Confounding variables (Batch Effects) Information Theory Covariance

More information

Proteomics and Variable Selection

Proteomics and Variable Selection Proteomics and Variable Selection p. 1/55 Proteomics and Variable Selection Alex Lewin With thanks to Paul Kirk for some graphs Department of Epidemiology and Biostatistics, School of Public Health, Imperial

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

Lecture 1: Introduction, Entropy and ML estimation

Lecture 1: Introduction, Entropy and ML estimation 0-704: Information Processing and Learning Spring 202 Lecture : Introduction, Entropy and ML estimation Lecturer: Aarti Singh Scribes: Min Xu Disclaimer: These notes have not been subjected to the usual

More information

Machine Learning and Bayesian Inference. Unsupervised learning. Can we find regularity in data without the aid of labels?

Machine Learning and Bayesian Inference. Unsupervised learning. Can we find regularity in data without the aid of labels? Machine Learning and Bayesian Inference Dr Sean Holden Computer Laboratory, Room FC6 Telephone extension 6372 Email: sbh11@cl.cam.ac.uk www.cl.cam.ac.uk/ sbh11/ Unsupervised learning Can we find regularity

More information

Gene Expression Data Classification with Revised Kernel Partial Least Squares Algorithm

Gene Expression Data Classification with Revised Kernel Partial Least Squares Algorithm Gene Expression Data Classification with Revised Kernel Partial Least Squares Algorithm Zhenqiu Liu, Dechang Chen 2 Department of Computer Science Wayne State University, Market Street, Frederick, MD 273,

More information

Fantope Regularization in Metric Learning

Fantope Regularization in Metric Learning Fantope Regularization in Metric Learning CVPR 2014 Marc T. Law (LIP6, UPMC), Nicolas Thome (LIP6 - UPMC Sorbonne Universités), Matthieu Cord (LIP6 - UPMC Sorbonne Universités), Paris, France Introduction

More information

Information in Biology

Information in Biology Lecture 3: Information in Biology Tsvi Tlusty, tsvi@unist.ac.kr Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve

More information

How to Quantitate a Markov Chain? Stochostic project 1

How to Quantitate a Markov Chain? Stochostic project 1 How to Quantitate a Markov Chain? Stochostic project 1 Chi-Ning,Chou Wei-chang,Lee PROFESSOR RAOUL NORMAND April 18, 2015 Abstract In this project, we want to quantitatively evaluate a Markov chain. In

More information

SUPPORT VECTOR MACHINE

SUPPORT VECTOR MACHINE SUPPORT VECTOR MACHINE Mainly based on https://nlp.stanford.edu/ir-book/pdf/15svm.pdf 1 Overview SVM is a huge topic Integration of MMDS, IIR, and Andrew Moore s slides here Our foci: Geometric intuition

More information

Variable Selection in Data Mining Project

Variable Selection in Data Mining Project Variable Selection Variable Selection in Data Mining Project Gilles Godbout IFT 6266 - Algorithmes d Apprentissage Session Project Dept. Informatique et Recherche Opérationnelle Université de Montréal

More information

CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs

CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs Professor Erik Sudderth Brown University Computer Science October 6, 2016 Some figures and materials courtesy

More information

Machine Learning, Fall 2009: Midterm

Machine Learning, Fall 2009: Midterm 10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

Information Theory Primer:

Information Theory Primer: Information Theory Primer: Entropy, KL Divergence, Mutual Information, Jensen s inequality Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,

More information

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality

More information

Information in Biology

Information in Biology Information in Biology CRI - Centre de Recherches Interdisciplinaires, Paris May 2012 Information processing is an essential part of Life. Thinking about it in quantitative terms may is useful. 1 Living

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Classification and Regression Trees

Classification and Regression Trees Classification and Regression Trees Ryan P Adams So far, we have primarily examined linear classifiers and regressors, and considered several different ways to train them When we ve found the linearity

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

Time: 1 hour 30 minutes

Time: 1 hour 30 minutes Paper Reference(s) 6663/0 Edexcel GCE Core Mathematics C Gold Level G5 Time: hour 30 minutes Materials required for examination Mathematical Formulae (Green) Items included with question papers Nil Candidates

More information

Lecture 4 October 18th

Lecture 4 October 18th Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations

More information

CSCE 478/878 Lecture 6: Bayesian Learning

CSCE 478/878 Lecture 6: Bayesian Learning Bayesian Methods Not all hypotheses are created equal (even if they are all consistent with the training data) Outline CSCE 478/878 Lecture 6: Bayesian Learning Stephen D. Scott (Adapted from Tom Mitchell

More information

Algorithm-Independent Learning Issues

Algorithm-Independent Learning Issues Algorithm-Independent Learning Issues Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007, Selim Aksoy Introduction We have seen many learning

More information

Conditional probabilities and graphical models

Conditional probabilities and graphical models Conditional probabilities and graphical models Thomas Mailund Bioinformatics Research Centre (BiRC), Aarhus University Probability theory allows us to describe uncertainty in the processes we model within

More information

Bayesian Machine Learning - Lecture 7

Bayesian Machine Learning - Lecture 7 Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1

More information

Introduction to Bayesian Learning

Introduction to Bayesian Learning Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline

More information

13 : Variational Inference: Loopy Belief Propagation and Mean Field

13 : Variational Inference: Loopy Belief Propagation and Mean Field 10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction

More information

5 Mutual Information and Channel Capacity

5 Mutual Information and Channel Capacity 5 Mutual Information and Channel Capacity In Section 2, we have seen the use of a quantity called entropy to measure the amount of randomness in a random variable. In this section, we introduce several

More information

CMU-Q Lecture 24:

CMU-Q Lecture 24: CMU-Q 15-381 Lecture 24: Supervised Learning 2 Teacher: Gianni A. Di Caro SUPERVISED LEARNING Hypotheses space Hypothesis function Labeled Given Errors Performance criteria Given a collection of input

More information

ECE662: Pattern Recognition and Decision Making Processes: HW TWO

ECE662: Pattern Recognition and Decision Making Processes: HW TWO ECE662: Pattern Recognition and Decision Making Processes: HW TWO Purdue University Department of Electrical and Computer Engineering West Lafayette, INDIANA, USA Abstract. In this report experiments are

More information

Information theory and decision tree

Information theory and decision tree Information theory and decision tree Jianxin Wu LAMDA Group National Key Lab for Novel Software Technology Nanjing University, China wujx2001@gmail.com June 14, 2018 Contents 1 Prefix code and Huffman

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!

More information

the tree till a class assignment is reached

the tree till a class assignment is reached Decision Trees Decision Tree for Playing Tennis Prediction is done by sending the example down Prediction is done by sending the example down the tree till a class assignment is reached Definitions Internal

More information

Introduction of Recruit

Introduction of Recruit Apr. 11, 2018 Introduction of Recruit We provide various kinds of online services from job search to hotel reservations across the world. Housing Beauty Travel Life & Local O2O Education Automobile Bridal

More information

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie Computational Biology Program Memorial Sloan-Kettering Cancer Center http://cbio.mskcc.org/leslielab

More information

Kyle Reing University of Southern California April 18, 2018

Kyle Reing University of Southern California April 18, 2018 Renormalization Group and Information Theory Kyle Reing University of Southern California April 18, 2018 Overview Renormalization Group Overview Information Theoretic Preliminaries Real Space Mutual Information

More information

Classification Part 2. Overview

Classification Part 2. Overview Classification Part 2 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville Overview Rule based Classifiers Nearest-neighbor Classifiers Data Mining

More information

Lecture 1: March 7, 2018

Lecture 1: March 7, 2018 Reinforcement Learning Spring Semester, 2017/8 Lecture 1: March 7, 2018 Lecturer: Yishay Mansour Scribe: ym DISCLAIMER: Based on Learning and Planning in Dynamical Systems by Shie Mannor c, all rights

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

Feature selection and extraction Spectral domain quality estimation Alternatives

Feature selection and extraction Spectral domain quality estimation Alternatives Feature selection and extraction Error estimation Maa-57.3210 Data Classification and Modelling in Remote Sensing Markus Törmä markus.torma@tkk.fi Measurements Preprocessing: Remove random and systematic

More information

Introduction to Information Theory

Introduction to Information Theory Introduction to Information Theory Gurinder Singh Mickey Atwal atwal@cshl.edu Center for Quantitative Biology Kullback-Leibler Divergence Summary Shannon s coding theorems Entropy Mutual Information Multi-information

More information

Machine Learning 3. week

Machine Learning 3. week Machine Learning 3. week Entropy Decision Trees ID3 C4.5 Classification and Regression Trees (CART) 1 What is Decision Tree As a short description, decision tree is a data classification procedure which

More information

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health

More information

Hands-On Learning Theory Fall 2016, Lecture 3

Hands-On Learning Theory Fall 2016, Lecture 3 Hands-On Learning Theory Fall 016, Lecture 3 Jean Honorio jhonorio@purdue.edu 1 Information Theory First, we provide some information theory background. Definition 3.1 (Entropy). The entropy of a discrete

More information

CS 6375 Machine Learning

CS 6375 Machine Learning CS 6375 Machine Learning Decision Trees Instructor: Yang Liu 1 Supervised Classifier X 1 X 2. X M Ref class label 2 1 Three variables: Attribute 1: Hair = {blond, dark} Attribute 2: Height = {tall, short}

More information

AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013

AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013 AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013 Lecturer: Dr. Mark Tame Introduction With the emergence of new types of information, in this case

More information

Information Theory in Intelligent Decision Making

Information Theory in Intelligent Decision Making Information Theory in Intelligent Decision Making Adaptive Systems and Algorithms Research Groups School of Computer Science University of Hertfordshire, United Kingdom June 7, 2015 Information Theory

More information

BAYESIAN DECISION THEORY

BAYESIAN DECISION THEORY Last updated: September 17, 2012 BAYESIAN DECISION THEORY Problems 2 The following problems from the textbook are relevant: 2.1 2.9, 2.11, 2.17 For this week, please at least solve Problem 2.3. We will

More information

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection Instructor: Herke van Hoof (herke.vanhoof@cs.mcgill.ca) Based on slides by:, Jackie Chi Kit Cheung Class web page:

More information

Recent Advances in Bayesian Inference Techniques

Recent Advances in Bayesian Inference Techniques Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian

More information

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation. ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Previous lectures: Sparse vectors recap How to represent

More information

ANLP Lecture 22 Lexical Semantics with Dense Vectors

ANLP Lecture 22 Lexical Semantics with Dense Vectors ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Henry S. Thompson ANLP Lecture 22 5 November 2018 Previous

More information

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the

More information

Lab 12: Structured Prediction

Lab 12: Structured Prediction December 4, 2014 Lecture plan structured perceptron application: confused messages application: dependency parsing structured SVM Class review: from modelization to classification What does learning mean?

More information

Lecture 2: August 31

Lecture 2: August 31 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy

More information

Regret Analysis for Performance Metrics in Multi-Label Classification The Case of Hamming and Subset Zero-One Loss

Regret Analysis for Performance Metrics in Multi-Label Classification The Case of Hamming and Subset Zero-One Loss Regret Analysis for Performance Metrics in Multi-Label Classification The Case of Hamming and Subset Zero-One Loss Krzysztof Dembczyński 1, Willem Waegeman 2, Weiwei Cheng 1, and Eyke Hüllermeier 1 1 Knowledge

More information

An Introduction to Bayesian Machine Learning

An Introduction to Bayesian Machine Learning 1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems

More information

DATA MINING LECTURE 9. Minimum Description Length Information Theory Co-Clustering

DATA MINING LECTURE 9. Minimum Description Length Information Theory Co-Clustering DATA MINING LECTURE 9 Minimum Description Length Information Theory Co-Clustering MINIMUM DESCRIPTION LENGTH Occam s razor Most data mining tasks can be described as creating a model for the data E.g.,

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Support Vector Machine & Its Applications

Support Vector Machine & Its Applications Support Vector Machine & Its Applications A portion (1/3) of the slides are taken from Prof. Andrew Moore s SVM tutorial at http://www.cs.cmu.edu/~awm/tutorials Mingyue Tan The University of British Columbia

More information

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make

More information

The connection of dropout and Bayesian statistics

The connection of dropout and Bayesian statistics The connection of dropout and Bayesian statistics Interpretation of dropout as approximate Bayesian modelling of NN http://mlg.eng.cam.ac.uk/yarin/thesis/thesis.pdf Dropout Geoffrey Hinton Google, University

More information

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395 Data Mining Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 42 Outline 1 Introduction 2 Feature selection

More information

Decision trees COMS 4771

Decision trees COMS 4771 Decision trees COMS 4771 1. Prediction functions (again) Learning prediction functions IID model for supervised learning: (X 1, Y 1),..., (X n, Y n), (X, Y ) are iid random pairs (i.e., labeled examples).

More information

CS540 ANSWER SHEET

CS540 ANSWER SHEET CS540 ANSWER SHEET Name Email 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 1 2 Final Examination CS540-1: Introduction to Artificial Intelligence Fall 2016 20 questions, 5 points

More information

Variational algorithms for marginal MAP

Variational algorithms for marginal MAP Variational algorithms for marginal MAP Alexander Ihler UC Irvine CIOG Workshop November 2011 Variational algorithms for marginal MAP Alexander Ihler UC Irvine CIOG Workshop November 2011 Work with Qiang

More information

Computational Systems Biology: Biology X

Computational Systems Biology: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA L#8:(November-08-2010) Cancer and Signals Outline 1 Bayesian Interpretation of Probabilities Information Theory Outline Bayesian

More information

Qualifier: CS 6375 Machine Learning Spring 2015

Qualifier: CS 6375 Machine Learning Spring 2015 Qualifier: CS 6375 Machine Learning Spring 2015 The exam is closed book. You are allowed to use two double-sided cheat sheets and a calculator. If you run out of room for an answer, use an additional sheet

More information

Support'Vector'Machines. Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan

Support'Vector'Machines. Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan Support'Vector'Machines Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan kasthuri.kannan@nyumc.org Overview Support Vector Machines for Classification Linear Discrimination Nonlinear Discrimination

More information

Max-Planck-Institut für Mathematik in den Naturwissenschaften Leipzig

Max-Planck-Institut für Mathematik in den Naturwissenschaften Leipzig Max-Planck-Institut für Mathematik in den Naturwissenschaften Leipzig Hierarchical Quantification of Synergy in Channels by Paolo Perrone and Nihat Ay Preprint no.: 86 2015 Hierarchical Quantification

More information