Iron Ore Rock Recognition Trials

Size: px
Start display at page:

Download "Iron Ore Rock Recognition Trials"

Transcription

1 Australian Centre for Field Robotics A Key Centre of Teaching and Research The Rose Street Building J04 The University of Sydney 2006 NSW, Australia Iron Ore Rock Recognition Trials Fabio Ramos, Peter Hatherly & Sildomar Monteiro T: E: f.ramos@acfr.usyd.edu.au Technical Report ACFR-TR June 2008 Released 12 March 2009

2 Table of Contents 1 Introduction The Learning Procedure Boosting Experiments Trial Area Drill Sensors Geophysical Interpretation of Trial Site Results Conclusions References...9 Appendix 1 Selected Drill Monitoring Data for Row Appendix 2 Geophysical Logging Results for Row Appendix 3 Core Photographs and Geotechnical Logs...25 ii

3 1 Introduction With the availability of monitoring data relating to drill performance and operation, the possibility arises for using that data to estimate the properties of the rocks being drilled. These properties could include rock type and strength. ACFR has previously developed theory and algorithms for the analysis of multi-parameter data for other automation projects requiring fusion of sensor information to provide a most likely estimate of a required situation. It was felt that these techniques could be applied to the problem of estimating ground conditions using the monitoring data gathered during the drilling of blast holes. For a trial in November 2006, a comprehensive set of drill monitoring data were obtained together with geotechnical, geological and geophysical support data aimed at providing an insight into the ground conditions. This report concerns the analysis of the drill monitoring data. In keeping with the terminology of the petroleum industry, the drill monitoring data is termed measurement-while-drilling (MWD) data. The analysis of the MWD data to determine subsurface geological conditions is termed rock recognition. 2 The Learning Procedure 2.1 Boosting A machine learning approach was used to combine the sensors inputs for rock recognition. Among the techniques tested (logistic regression, K-nearest neighbour and boosting), boosting performed the best. A classical supervised learning approach consists of two phases. In the learning phase, the algorithm is presented with input data and corresponding labels (rock recognition assignments). An optimisation procedure is then performed to learn a function that maps the inputs to the labels. Once this function is learned, new input data can be evaluated and a label obtained. This second phase is known as the testing phase. The final implementation of rock recognition will consist of an off-line algorithm for learning and an online (embedded) algorithm for testing. Importantly, the machine learning procedure employed in this work can easily accommodate new information. There is no need for a specialist on the method, the only requirement is a training set consisting of input data and corresponding labels. This allows extreme flexibility for using the method in different drills and different geologies. Boosting is a machine learning technique for supervised classification that employs a combination of weak simple classifiers to produce a powerful committee [2][1]. It has a sound theoretical foundation and provides probability estimates for each class. Boosting has become very popular due to many empirical studies showing that it tends to yield smaller classification error rates and be more robust to overfitting than competing methods such as Support Vector Machines or Neural Networks. Other advantages of boosting are implementation simplicity and computational efficiency. The main disadvantage is the difficulty in interpreting the classification boundary due to the high degree of nonlinearity. There are many different variations of boosting algorithms. The most commonly used version is AdaBoost (from adaptive boosting), developed by Freund and Schapire [2]. The idea of AdaBoost is to train many weak learners on various distributions (or set of weights) of the input data and then combine the classifiers produced into a single committee. A weak learner can be any classifier whose performance is guaranteed to be better than a random guess. A very common weak learner is the decision stump, which produces an axis-orthogonal hyperplane and can be viewed as a one-level binary decision tree. Initially, the weights of all training examples are set equally, but after each round of the AdaBoost

4 algorithm, the weights of incorrectly classified examples are increased. The final committee or ensemble is a weighted majority combination of the weak classifiers. A formal description of the AdaBoost algorithm for binary classification is shown below. Empirical observations have shown that, although AdaBoost presents good generalization, it may overfit in some cases and its performance is degraded in the presence of noise. AdaBoost Given a training set of n labelled examples ( 1 1) where y { 1, + 1}. i Initialize the weighting coefficients: w For m= 1,, M a. Normalize the weighting coefficients w i 1 = for i = 1,, n. n w m m i i = n m w j = 1 j x, y,,( x, y ) b. Fit a classifier g m to the training data that minimizes the weighted error n m function m = wi I( yi gm( xi) ) ε i= 1 c. Evaluate the log-odds 1 1 ε log m αm = 2 ε m m 1 m w w exp α g ( x ) y d. Update the weighting coefficients i i { m m i i} M α ( ) Return predictions of the final ensemble 1 mg m m x Y ( x = ) M 1 if > 0 = 1 otherwise. n n Several methods have been proposed to extend boosting to handle the multiclass case. The typical strategy, called one-vs-all approach, is to reduce the multiclass problem to a set of independent binary problems. However, this approach has scalability issues, i.e., increase in computational cost as the number of classes grows, due to the independent training and run required by each binary classifier. For rock recognition purposes we use a version of boosting named LogitBoost. LogitBoost fits additive logistic regression models by stagewise optimization of the maximum likelihood. The algorithm below shows the multiclass version of LogitBoost which optimizes the maximum likelihood through adaptive Newton steps [1]. 2

5 LogitBoost (J classes) 1. Start with weights w ij = 1/N, i = 1,, N, j = 1,,J, F j (x) = 0 and p j (x) = 1/J j. 2. Repeat for m = 1, 2,, M: (a) Repeat for j = 1,, J: (i) Compute working responses and weights in the jth class, p j ( ) y * z ij = ij p j x i p j ( x i )1 ( p j ( x i ), w ij = p j (x i )(1 p j (x i )). (ii) Fit the function f mj (x) by a weighted least-squares regression of z ij to x i with weights w ij. (b) Set f mj () x J 1 f mj () x 1 J f mk () x, and F j ( x) F j ()+ x f mj (). x J J k=1 (c) Update p j (x) using, ()= x ef j ( x) J, F k ()= x Output the classifier argmax j F j ( x). J e F k () x k=1 k=1 3 Experiments 3.1 Trial Area Figure 1 shows the arrangement of the drill holes. The site was chosen because a variety of down-hole conditions were expected to vary. Three rows of holes were drilled with the drill operating in percussion mode (down-hole hammer) with and without shock absorber, and rotary mode with and without shock absorber. Of the three rows of holes, row 1 was the most comprehensive. In this row, 28 holes 12 m deep were drilled at 3 m spacings with the drill in percussion mode and with the use of the shock absorber. In addition to the test holes drilled with the blast hole rig, 7 diamond drill holes were drilled to a depth of 12 m at the locations shown in Figure 1. The cores were recovered for logging and testing. At the completion of the drilling all blast holes and the diamond drill holes were geophysically logged using caliper, natural gamma, magnetic susceptibility and density (gamma gamma) logging tools. The detailed geology was determined by site geologists using a combination of cone logging, core logging and the geophysical logging results. As shown by the geological section for this set of holes (Figure 2), the geological conditions ranged from shale to ore to banded iron formation (BIF). All cores were photographed and geotechnical logs were prepared. Appendix 2 contains the geophysical logging data for row 1. Core photographs and the geotechnical logs are provided in Appendix 3. 3

6 Blast hole Resource hole Rock recognition row 1 Rock recognition row 2 Rock recognition row 3 Rock recognition core hole 5240 Local North Local East Figure 1. Pattern of resource holes, test holes and subsequent blast holes drilled at the test site. Figure 2. Geological section through row 1 of the test holes. 3.2 Drill Sensors The MWD sensors used for the rock recognition were: 1. Bit air pressure; 2. Pull-down pressure; 3. Rotation pressure; 4. Pull-down rate; 5. Head speed; 6. Pressure transducers (7 in total); 7. Accelerometers mounted on the mast (5 in total); 4

7 A Programmable Logic Controller (PLC) was used to log the data from sensors (1-5) of the above list. Although the sampling frequency was variable, the average was 10 readings (per sensor) per second. The pressure transducers were logged at 500Hz. The accelerometers were logged at a much higher sampling rate of about 10 khz. As the data was considerably noisy, a feature extraction procedure was performed before the data were input into the classifier. For sensors 1-6, the feature extraction procedure consisted of grouping the measurements into sections of 0.1 m by using the sensors time stamp and the corresponding head position of the drill. From this, an average value was obtained. This process worked well for the sensor data for row 1 but unfortunately there was serious sensor failure for the subsequent rows and these data could not be analysed. For the accelerometers, the measurements were grouped into sections of 0.1m as before and the standard deviation was computed. The idea was to quantify the vibration in a simple manner. Different analyses of the accelerometer data were performed at University of Queensland in the frequency domain. However, no conclusive results were obtained and the simple feature extraction procedure above was used for the results reported. During the data acquisition, the drill operated in percussion and rotary mode, which caused large vibrations in the machine. This fact makes accelerometer data challenging to interpret. 4 Geophysical Interpretation of Trial Site Detailed analysis of MWD data for rock recognition purposes requires independent data for the purposes of training the learning algorithm and the evaluation of the results. While the geological data from the cone and core logging provided a general geological section, the data were not at the 0.1 m increments utilised for the MWD data. The geotechnical data was also not available at this scale. Fortunately the geophysical borehole logs provided the quantitative down-hole measurements that could be correlated with the geology and also allowed direct comparison with the MWD data at 0.1 m increments. The analysis of the geophysical logging data was undertaken as follows. The natural gamma logs are ideally suited for discriminating shale from ore and Banded Iron Formation (BIF). Figure 3 shows the raw data for row 1. If a cut-off of 15 API is used to mark the transition from the ore and BIF, the results shown in Figure 4 are obtained. These compare favourably with the geological section shown in Figure 2. Figure 3. Smoothed natural gamma data for row 1 shown at 0.1 m depth intervals (units are API) 5

8 Figure 4. The occurrence of shale (bright red) defined by natural gamma data being > 15 API. To discriminate ore from BIF, the magnetic susceptibility logs were used. Figure 5 shows the magnetic susceptibility data. The criterion applied to distinguish BIF from the ore types (the shale had already been identified) was that for the BIF, somewhere within a preset distance (less than 1 m) the magnetic susceptibility needed to be above a preset threshold (about 12). Figure 6 shows the results of applying this rule. Figure 5. Smoothed magnetic susceptibility data for row 1. (Dimensionless susceptibility values are by 10-5.) Figure 6. Interpreted geological section showing shale (bright red), BIF (blue) and ore (brown). Finally, an attempt was made to estimate the hardness of the ore types on the basis of the density logs. Figure 7 shows the smoothed density. It can be seen that there are higher densities present within the ore zones in holes 11 to 15. It was assumed that these higher densities represented harder ore. The distribution of the densities within the ore zones is shown in more detail in Figure 8. On the basis of this distribution, hard, medium and soft ore types were assumed to be present. Figure 9 shows the final geophysical interpretation 6

9 Figure 7. Smoothed density data for row 1 (units of t/m3). Figure 8. Distribution of densities within the ore zone. Interpretations of hard, medium and soft ore were made on the basis of this distribution. Figure 9. Final classification of the geological section for row 1 based on the analysis of the geophysical logs. 5 Results As mentioned above, there were problems in the MWD data for rows 2 and 3. It was therefore only possible to analyse the results from row 1. Furthermore the results shown below, accelerometer data has not been used. In the first experiment, the objective was to distinguish shale from the other rock types. The natural gamma-ray data was used to label the training data. Figure 10 shows results for leave-one-out cross 7

10 validation; i.e., each 0.1m section is evaluated using a model trained with all the other sections. A direct comparison between Figure 10 and the basic shale model in Figure 4 shows that the boosting classifier is able to predict shale bands quite accurately from the MWD data. This is not totally a surprise given that shale is the softest rock type considered and, in general, possesses very different mechanical properties than the others. Figure 10. Shale detection results using boosting. The red colour indicates shale. The second experiment was aimed at recognising four different rock types: shale, zone A iron ore, zone B iron ore and BIF. The same evaluation procedure as before was used, i.e., leave-one-out cross validation. Figure 11 depicts the results, which can be compared to the geological interpretation of Figure 2. As can be observed, the shale bands are detected accurately except in the top of hole 28 where a sensor failure occurred. The BIF zone is on the right part of the section, separated from Zone B by a shale band. In this particular experiment, it was possible to distinguish between iron ore zones (Zones A and B), however, we believe this should be difficult in the general case since iron zones might have very similar mechanical properties. Figure11. Geological recognition using boosting. In the final experiment, the goal was to combine geology and hardness in the classification process. Labelling of the training data was obtained using the procedure described in the previous section. In summary, the system was trained to identify shale, BIF, and iron ore zones. Within the iron ore zones, the system classifies sections according to the hardness as soft, medium and hard ores. In this particular experiment, the cross validation was performed differently. Instead of the leave-onesection out as before, the whole testing hole was left aside, i.e., the classifier was trained on 27 holes and tested on the remaining hole. This process was then repeated 28 times. Compared to the other cross validation procedure where only one section was left out of the training, this procedure is much more difficult and requires better generalisation properties therefore previous results are not directly comparable. However, comparison of Figure 12 with the geophysical interpretation in Figure 9 shows a favourable correlation. 8

11 Figure 12. Combined geological recognition and hardness results using boosting. A qualitative assessment of these results was obtained by visually inspecting the bench face at the trial position and the diamond drill core next to hole 12. The drill core in hole had large sections of relatively strong, competent ore with few fractures. By observing the size of the rocks on the floor formed after blasting, it was possible to note that larger rocks were formed around holes 12 and 13. This gave a good indication that this area had harder rocks which were not correctly blasted resulting in larger post blasting fragments. 6 Conclusions While the failure of the sensor logging for the drilling of the second and third rows was unfortunate, the results for row 1 which utilised percussion drilling are most encouraging given the degree of dampening from the shock sub. Indeed, this may be one of the first successful instances of rock recognition being undertaken on in the hole hammer percussion drilling anywhere. With the good geological control provided by the geophysical logging, the rock recognition tests utilising the MWD data were clearly able to delineate the ore zones from shale and BIF. If the variations in density shown by geophysical density logging can be taken to indicate the differences in the hardness of the ore, then it would seem that hardness can also be recognised from the MWD data. Provision of good geological control will continue to be an issue for further rock recognition trials. Geophysical logging will be essential. Coring and good geological logging are also desirable. Means of calibrating results against blast performance are needed. The approach to supervised learning that has been utilised here has been successful. It is hoped that a machine learning procedure that can easily accommodate new information will be the final outcome, without the need for specialist involvement. The main requirement will be for a training set consisting of input data and corresponding labels. This will allow extreme flexibility for using the method with different drills and different geological conditions. 7 References [1] Friedman, J., Hastie, T. and Tibshirani, R. (2000). Additive logistic regression: A statistical view of boosting. Annals of Statistics, vol. 28, no. 2, pp [2] Freund, Y. and Schapire, R.E. (1996). Experiments with a new boosting algorithm. In Proceedings of the Thirteenth International Conference on Machine Learning, pp [3] Torralba, A., Murphy, K.P. and Freeman, W.T. (2007). Sharing visual features for multiclass and multiview object detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 5, pp

12 Appendix 1 Selected Drill Monitoring Data for Row 1 Of the following data, the PLC fitted to the drill provided the bit pressure, the rotation pressure, the pulldown pressure, the pull-down rate and the head speed. In addition, pressure transducers were installed to monitor the following drill functions: Feed down Transducer 1 Feed up Transducer 2 Reverse rotation Transducer 3 Forward rotation Transducer 4 Rotation relief Transducer 5 Feed relief Transducer 6 Hold back Transducer 7 10

13 11

14 12

15 13

16 14

17 15

18 16

19 Appendix 2 Geophysical Logging Results for Row 1 17

20 18

21 19

22 20

23 21

24 22

25 23

26 24

27 Appendix 3 Core Photographs and Geotechnical Logs 25

28 26

29 27

30 28

31 29

32 30

33 31

34 32

Analysis of the Performance of AdaBoost.M2 for the Simulated Digit-Recognition-Example

Analysis of the Performance of AdaBoost.M2 for the Simulated Digit-Recognition-Example Analysis of the Performance of AdaBoost.M2 for the Simulated Digit-Recognition-Example Günther Eibl and Karl Peter Pfeiffer Institute of Biostatistics, Innsbruck, Austria guenther.eibl@uibk.ac.at Abstract.

More information

CS229 Supplemental Lecture notes

CS229 Supplemental Lecture notes CS229 Supplemental Lecture notes John Duchi 1 Boosting We have seen so far how to solve classification (and other) problems when we have a data representation already chosen. We now talk about a procedure,

More information

Learning with multiple models. Boosting.

Learning with multiple models. Boosting. CS 2750 Machine Learning Lecture 21 Learning with multiple models. Boosting. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Learning with multiple models: Approach 2 Approach 2: use multiple models

More information

A Brief Introduction to Adaboost

A Brief Introduction to Adaboost A Brief Introduction to Adaboost Hongbo Deng 6 Feb, 2007 Some of the slides are borrowed from Derek Hoiem & Jan ˇSochman. 1 Outline Background Adaboost Algorithm Theory/Interpretations 2 What s So Good

More information

AdaBoost. Lecturer: Authors: Center for Machine Perception Czech Technical University, Prague

AdaBoost. Lecturer: Authors: Center for Machine Perception Czech Technical University, Prague AdaBoost Lecturer: Jan Šochman Authors: Jan Šochman, Jiří Matas Center for Machine Perception Czech Technical University, Prague http://cmp.felk.cvut.cz Motivation Presentation 2/17 AdaBoost with trees

More information

Introduction to Machine Learning Lecture 11. Mehryar Mohri Courant Institute and Google Research

Introduction to Machine Learning Lecture 11. Mehryar Mohri Courant Institute and Google Research Introduction to Machine Learning Lecture 11 Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Boosting Mehryar Mohri - Introduction to Machine Learning page 2 Boosting Ideas Main idea:

More information

Boosting. CAP5610: Machine Learning Instructor: Guo-Jun Qi

Boosting. CAP5610: Machine Learning Instructor: Guo-Jun Qi Boosting CAP5610: Machine Learning Instructor: Guo-Jun Qi Weak classifiers Weak classifiers Decision stump one layer decision tree Naive Bayes A classifier without feature correlations Linear classifier

More information

Data Mining und Maschinelles Lernen

Data Mining und Maschinelles Lernen Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting

More information

ECE 5984: Introduction to Machine Learning

ECE 5984: Introduction to Machine Learning ECE 5984: Introduction to Machine Learning Topics: Ensemble Methods: Bagging, Boosting Readings: Murphy 16.4; Hastie 16 Dhruv Batra Virginia Tech Administrativia HW3 Due: April 14, 11:55pm You will implement

More information

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted

More information

COMS 4771 Lecture Boosting 1 / 16

COMS 4771 Lecture Boosting 1 / 16 COMS 4771 Lecture 12 1. Boosting 1 / 16 Boosting What is boosting? Boosting: Using a learning algorithm that provides rough rules-of-thumb to construct a very accurate predictor. 3 / 16 What is boosting?

More information

ECE 5424: Introduction to Machine Learning

ECE 5424: Introduction to Machine Learning ECE 5424: Introduction to Machine Learning Topics: Ensemble Methods: Bagging, Boosting PAC Learning Readings: Murphy 16.4;; Hastie 16 Stefan Lee Virginia Tech Fighting the bias-variance tradeoff Simple

More information

The AdaBoost algorithm =1/n for i =1,...,n 1) At the m th iteration we find (any) classifier h(x; ˆθ m ) for which the weighted classification error m

The AdaBoost algorithm =1/n for i =1,...,n 1) At the m th iteration we find (any) classifier h(x; ˆθ m ) for which the weighted classification error m ) Set W () i The AdaBoost algorithm =1/n for i =1,...,n 1) At the m th iteration we find (any) classifier h(x; ˆθ m ) for which the weighted classification error m m =.5 1 n W (m 1) i y i h(x i ; 2 ˆθ

More information

Statistics and learning: Big Data

Statistics and learning: Big Data Statistics and learning: Big Data Learning Decision Trees and an Introduction to Boosting Sébastien Gadat Toulouse School of Economics February 2017 S. Gadat (TSE) SAD 2013 1 / 30 Keywords Decision trees

More information

Learning Ensembles. 293S T. Yang. UCSB, 2017.

Learning Ensembles. 293S T. Yang. UCSB, 2017. Learning Ensembles 293S T. Yang. UCSB, 2017. Outlines Learning Assembles Random Forest Adaboost Training data: Restaurant example Examples described by attribute values (Boolean, discrete, continuous)

More information

Stochastic Gradient Descent

Stochastic Gradient Descent Stochastic Gradient Descent Machine Learning CSE546 Carlos Guestrin University of Washington October 9, 2013 1 Logistic Regression Logistic function (or Sigmoid): Learn P(Y X) directly Assume a particular

More information

CS7267 MACHINE LEARNING

CS7267 MACHINE LEARNING CS7267 MACHINE LEARNING ENSEMBLE LEARNING Ref: Dr. Ricardo Gutierrez-Osuna at TAMU, and Aarti Singh at CMU Mingon Kang, Ph.D. Computer Science, Kennesaw State University Definition of Ensemble Learning

More information

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring / Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring 2015 http://ce.sharif.edu/courses/93-94/2/ce717-1 / Agenda Combining Classifiers Empirical view Theoretical

More information

Boosting: Foundations and Algorithms. Rob Schapire

Boosting: Foundations and Algorithms. Rob Schapire Boosting: Foundations and Algorithms Rob Schapire Example: Spam Filtering problem: filter out spam (junk email) gather large collection of examples of spam and non-spam: From: yoav@ucsd.edu Rob, can you

More information

Machine Learning Lecture 5

Machine Learning Lecture 5 Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

More information

Machine Learning Lecture 7

Machine Learning Lecture 7 Course Outline Machine Learning Lecture 7 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Statistical Learning Theory 23.05.2016 Discriminative Approaches (5 weeks) Linear Discriminant

More information

Ordinal Classification with Decision Rules

Ordinal Classification with Decision Rules Ordinal Classification with Decision Rules Krzysztof Dembczyński 1, Wojciech Kotłowski 1, and Roman Słowiński 1,2 1 Institute of Computing Science, Poznań University of Technology, 60-965 Poznań, Poland

More information

Neural Networks and Ensemble Methods for Classification

Neural Networks and Ensemble Methods for Classification Neural Networks and Ensemble Methods for Classification NEURAL NETWORKS 2 Neural Networks A neural network is a set of connected input/output units (neurons) where each connection has a weight associated

More information

Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers

Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers Erin Allwein, Robert Schapire and Yoram Singer Journal of Machine Learning Research, 1:113-141, 000 CSE 54: Seminar on Learning

More information

Voting (Ensemble Methods)

Voting (Ensemble Methods) 1 2 Voting (Ensemble Methods) Instead of learning a single classifier, learn many weak classifiers that are good at different parts of the data Output class: (Weighted) vote of each classifier Classifiers

More information

Ensembles of Classifiers.

Ensembles of Classifiers. Ensembles of Classifiers www.biostat.wisc.edu/~dpage/cs760/ 1 Goals for the lecture you should understand the following concepts ensemble bootstrap sample bagging boosting random forests error correcting

More information

A Decision Stump. Decision Trees, cont. Boosting. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University. October 1 st, 2007

A Decision Stump. Decision Trees, cont. Boosting. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University. October 1 st, 2007 Decision Trees, cont. Boosting Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University October 1 st, 2007 1 A Decision Stump 2 1 The final tree 3 Basic Decision Tree Building Summarized

More information

Logistic Regression and Boosting for Labeled Bags of Instances

Logistic Regression and Boosting for Labeled Bags of Instances Logistic Regression and Boosting for Labeled Bags of Instances Xin Xu and Eibe Frank Department of Computer Science University of Waikato Hamilton, New Zealand {xx5, eibe}@cs.waikato.ac.nz Abstract. In

More information

Boosting & Deep Learning

Boosting & Deep Learning Boosting & Deep Learning Ensemble Learning n So far learning methods that learn a single hypothesis, chosen form a hypothesis space that is used to make predictions n Ensemble learning à select a collection

More information

Boosting. Ryan Tibshirani Data Mining: / April Optional reading: ISL 8.2, ESL , 10.7, 10.13

Boosting. Ryan Tibshirani Data Mining: / April Optional reading: ISL 8.2, ESL , 10.7, 10.13 Boosting Ryan Tibshirani Data Mining: 36-462/36-662 April 25 2013 Optional reading: ISL 8.2, ESL 10.1 10.4, 10.7, 10.13 1 Reminder: classification trees Suppose that we are given training data (x i, y

More information

Frank C Porter and Ilya Narsky: Statistical Analysis Techniques in Particle Physics Chap. c /9/9 page 331 le-tex

Frank C Porter and Ilya Narsky: Statistical Analysis Techniques in Particle Physics Chap. c /9/9 page 331 le-tex Frank C Porter and Ilya Narsky: Statistical Analysis Techniques in Particle Physics Chap. c15 2013/9/9 page 331 le-tex 331 15 Ensemble Learning The expression ensemble learning refers to a broad class

More information

Lecture 8. Instructor: Haipeng Luo

Lecture 8. Instructor: Haipeng Luo Lecture 8 Instructor: Haipeng Luo Boosting and AdaBoost In this lecture we discuss the connection between boosting and online learning. Boosting is not only one of the most fundamental theories in machine

More information

Boosting with decision stumps and binary features

Boosting with decision stumps and binary features Boosting with decision stumps and binary features Jason Rennie jrennie@ai.mit.edu April 10, 2003 1 Introduction A special case of boosting is when features are binary and the base learner is a decision

More information

Application and Challenges of Artificial Intelligence in Exploration

Application and Challenges of Artificial Intelligence in Exploration Application and Challenges of Artificial Intelligence in Exploration XPLOR 2017 John McGaughey 2017 Mira Geoscience Ltd. Artificial Intelligence Artificial Intelligence is colossally hyped these days,

More information

Boosting Methods: Why They Can Be Useful for High-Dimensional Data

Boosting Methods: Why They Can Be Useful for High-Dimensional Data New URL: http://www.r-project.org/conferences/dsc-2003/ Proceedings of the 3rd International Workshop on Distributed Statistical Computing (DSC 2003) March 20 22, Vienna, Austria ISSN 1609-395X Kurt Hornik,

More information

Ensembles. Léon Bottou COS 424 4/8/2010

Ensembles. Léon Bottou COS 424 4/8/2010 Ensembles Léon Bottou COS 424 4/8/2010 Readings T. G. Dietterich (2000) Ensemble Methods in Machine Learning. R. E. Schapire (2003): The Boosting Approach to Machine Learning. Sections 1,2,3,4,6. Léon

More information

Machine Learning and Data Mining. Linear classification. Kalev Kask

Machine Learning and Data Mining. Linear classification. Kalev Kask Machine Learning and Data Mining Linear classification Kalev Kask Supervised learning Notation Features x Targets y Predictions ŷ = f(x ; q) Parameters q Program ( Learner ) Learning algorithm Change q

More information

ABC-LogitBoost for Multi-Class Classification

ABC-LogitBoost for Multi-Class Classification Ping Li, Cornell University ABC-Boost BTRY 6520 Fall 2012 1 ABC-LogitBoost for Multi-Class Classification Ping Li Department of Statistical Science Cornell University 2 4 6 8 10 12 14 16 2 4 6 8 10 12

More information

Learning theory. Ensemble methods. Boosting. Boosting: history

Learning theory. Ensemble methods. Boosting. Boosting: history Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over

More information

VBM683 Machine Learning

VBM683 Machine Learning VBM683 Machine Learning Pinar Duygulu Slides are adapted from Dhruv Batra Bias is the algorithm's tendency to consistently learn the wrong thing by not taking into account all the information in the data

More information

COMS 4721: Machine Learning for Data Science Lecture 13, 3/2/2017

COMS 4721: Machine Learning for Data Science Lecture 13, 3/2/2017 COMS 4721: Machine Learning for Data Science Lecture 13, 3/2/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University BOOSTING Robert E. Schapire and Yoav

More information

A New Perspective on Boosting in Linear Regression via Subgradient Optimization and Relatives

A New Perspective on Boosting in Linear Regression via Subgradient Optimization and Relatives A New Perspective on Boosting in Linear Regression via Subgradient Optimization and Relatives Paul Grigas May 25, 2016 1 Boosting Algorithms in Linear Regression Boosting [6, 9, 12, 15, 16] is an extremely

More information

Hierarchical Boosting and Filter Generation

Hierarchical Boosting and Filter Generation January 29, 2007 Plan Combining Classifiers Boosting Neural Network Structure of AdaBoost Image processing Hierarchical Boosting Hierarchical Structure Filters Combining Classifiers Combining Classifiers

More information

TDT4173 Machine Learning

TDT4173 Machine Learning TDT4173 Machine Learning Lecture 3 Bagging & Boosting + SVMs Norwegian University of Science and Technology Helge Langseth IT-VEST 310 helgel@idi.ntnu.no 1 TDT4173 Machine Learning Outline 1 Ensemble-methods

More information

A Gentle Introduction to Gradient Boosting. Cheng Li College of Computer and Information Science Northeastern University

A Gentle Introduction to Gradient Boosting. Cheng Li College of Computer and Information Science Northeastern University A Gentle Introduction to Gradient Boosting Cheng Li chengli@ccs.neu.edu College of Computer and Information Science Northeastern University Gradient Boosting a powerful machine learning algorithm it can

More information

Boos$ng Can we make dumb learners smart?

Boos$ng Can we make dumb learners smart? Boos$ng Can we make dumb learners smart? Aarti Singh Machine Learning 10-601 Nov 29, 2011 Slides Courtesy: Carlos Guestrin, Freund & Schapire 1 Why boost weak learners? Goal: Automa'cally categorize type

More information

Robotics 2 AdaBoost for People and Place Detection

Robotics 2 AdaBoost for People and Place Detection Robotics 2 AdaBoost for People and Place Detection Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard v.1.0, Kai Arras, Oct 09, including material by Luciano Spinello and Oscar Martinez Mozos

More information

Boosting Based Conditional Quantile Estimation for Regression and Binary Classification

Boosting Based Conditional Quantile Estimation for Regression and Binary Classification Boosting Based Conditional Quantile Estimation for Regression and Binary Classification Songfeng Zheng Department of Mathematics Missouri State University, Springfield, MO 65897, USA SongfengZheng@MissouriState.edu

More information

Evidence Contrary to the Statistical View of Boosting

Evidence Contrary to the Statistical View of Boosting Journal of Machine Learning Research 9 (2008) 131-156 Submitted 10/05; Revised 7/07; Published 2/08 Evidence Contrary to the Statistical View of Boosting David Mease Department of Marketing and Decision

More information

Machine Learning Basics Lecture 7: Multiclass Classification. Princeton University COS 495 Instructor: Yingyu Liang

Machine Learning Basics Lecture 7: Multiclass Classification. Princeton University COS 495 Instructor: Yingyu Liang Machine Learning Basics Lecture 7: Multiclass Classification Princeton University COS 495 Instructor: Yingyu Liang Example: image classification indoor Indoor outdoor Example: image classification (multiclass)

More information

Open Problem: A (missing) boosting-type convergence result for ADABOOST.MH with factorized multi-class classifiers

Open Problem: A (missing) boosting-type convergence result for ADABOOST.MH with factorized multi-class classifiers JMLR: Workshop and Conference Proceedings vol 35:1 8, 014 Open Problem: A (missing) boosting-type convergence result for ADABOOST.MH with factorized multi-class classifiers Balázs Kégl LAL/LRI, University

More information

Evidence Contrary to the Statistical View of Boosting

Evidence Contrary to the Statistical View of Boosting University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 2-2008 Evidence Contrary to the Statistical View of Boosting David Mease San Jose State University Abraham J. Wyner

More information

SPECIAL INVITED PAPER

SPECIAL INVITED PAPER The Annals of Statistics 2000, Vol. 28, No. 2, 337 407 SPECIAL INVITED PAPER ADDITIVE LOGISTIC REGRESSION: A STATISTICAL VIEW OF BOOSTING By Jerome Friedman, 1 Trevor Hastie 2 3 and Robert Tibshirani 2

More information

Self Supervised Boosting

Self Supervised Boosting Self Supervised Boosting Max Welling, Richard S. Zemel, and Geoffrey E. Hinton Department of omputer Science University of Toronto 1 King s ollege Road Toronto, M5S 3G5 anada Abstract Boosting algorithms

More information

CS 231A Section 1: Linear Algebra & Probability Review

CS 231A Section 1: Linear Algebra & Probability Review CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability

More information

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang CS 231A Section 1: Linear Algebra & Probability Review Kevin Tang Kevin Tang Section 1-1 9/30/2011 Topics Support Vector Machines Boosting Viola Jones face detector Linear Algebra Review Notation Operations

More information

LogitBoost with Trees Applied to the WCCI 2006 Performance Prediction Challenge Datasets

LogitBoost with Trees Applied to the WCCI 2006 Performance Prediction Challenge Datasets Applied to the WCCI 2006 Performance Prediction Challenge Datasets Roman Lutz Seminar für Statistik, ETH Zürich, Switzerland 18. July 2006 Overview Motivation 1 Motivation 2 The logistic framework The

More information

Chapter 14 Combining Models

Chapter 14 Combining Models Chapter 14 Combining Models T-61.62 Special Course II: Pattern Recognition and Machine Learning Spring 27 Laboratory of Computer and Information Science TKK April 3th 27 Outline Independent Mixing Coefficients

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction

More information

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer. University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x

More information

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington Linear Classification CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Example of Linear Classification Red points: patterns belonging

More information

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

MIDTERM SOLUTIONS: FALL 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE

MIDTERM SOLUTIONS: FALL 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE MIDTERM SOLUTIONS: FALL 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE March 28, 2012 The exam is closed book. You are allowed a double sided one page cheat sheet. Answer the questions in the spaces provided on

More information

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information

Support Vector Machines for Classification: A Statistical Portrait

Support Vector Machines for Classification: A Statistical Portrait Support Vector Machines for Classification: A Statistical Portrait Yoonkyung Lee Department of Statistics The Ohio State University May 27, 2011 The Spring Conference of Korean Statistical Society KAIST,

More information

Announcements Kevin Jamieson

Announcements Kevin Jamieson Announcements My office hours TODAY 3:30 pm - 4:30 pm CSE 666 Poster Session - Pick one First poster session TODAY 4:30 pm - 7:30 pm CSE Atrium Second poster session December 12 4:30 pm - 7:30 pm CSE Atrium

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Thomas G. Dietterich tgd@eecs.oregonstate.edu 1 Outline What is Machine Learning? Introduction to Supervised Learning: Linear Methods Overfitting, Regularization, and the

More information

10701/15781 Machine Learning, Spring 2007: Homework 2

10701/15781 Machine Learning, Spring 2007: Homework 2 070/578 Machine Learning, Spring 2007: Homework 2 Due: Wednesday, February 2, beginning of the class Instructions There are 4 questions on this assignment The second question involves coding Do not attach

More information

Recap from previous lecture

Recap from previous lecture Recap from previous lecture Learning is using past experience to improve future performance. Different types of learning: supervised unsupervised reinforcement active online... For a machine, experience

More information

Classification Ensemble That Maximizes the Area Under Receiver Operating Characteristic Curve (AUC)

Classification Ensemble That Maximizes the Area Under Receiver Operating Characteristic Curve (AUC) Classification Ensemble That Maximizes the Area Under Receiver Operating Characteristic Curve (AUC) Eunsik Park 1 and Y-c Ivan Chang 2 1 Chonnam National University, Gwangju, Korea 2 Academia Sinica, Taipei,

More information

Infinite Ensemble Learning with Support Vector Machinery

Infinite Ensemble Learning with Support Vector Machinery Infinite Ensemble Learning with Support Vector Machinery Hsuan-Tien Lin and Ling Li Learning Systems Group, California Institute of Technology ECML/PKDD, October 4, 2005 H.-T. Lin and L. Li (Learning Systems

More information

TDT4173 Machine Learning

TDT4173 Machine Learning TDT4173 Machine Learning Lecture 9 Learning Classifiers: Bagging & Boosting Norwegian University of Science and Technology Helge Langseth IT-VEST 310 helgel@idi.ntnu.no 1 TDT4173 Machine Learning Outline

More information

Gradient Boosting (Continued)

Gradient Boosting (Continued) Gradient Boosting (Continued) David Rosenberg New York University April 4, 2016 David Rosenberg (New York University) DS-GA 1003 April 4, 2016 1 / 31 Boosting Fits an Additive Model Boosting Fits an Additive

More information

18.9 SUPPORT VECTOR MACHINES

18.9 SUPPORT VECTOR MACHINES 744 Chapter 8. Learning from Examples is the fact that each regression problem will be easier to solve, because it involves only the examples with nonzero weight the examples whose kernels overlap the

More information

The Boosting Approach to. Machine Learning. Maria-Florina Balcan 10/31/2016

The Boosting Approach to. Machine Learning. Maria-Florina Balcan 10/31/2016 The Boosting Approach to Machine Learning Maria-Florina Balcan 10/31/2016 Boosting General method for improving the accuracy of any given learning algorithm. Works by creating a series of challenge datasets

More information

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function

More information

Boosting. Jiahui Shen. October 27th, / 44

Boosting. Jiahui Shen. October 27th, / 44 Boosting Jiahui Shen October 27th, 2017 1 / 44 Target of Boosting Figure: Weak learners Figure: Combined learner 2 / 44 Boosting introduction and notation Boosting: combines weak learners into a strong

More information

Artificial Intelligence Roman Barták

Artificial Intelligence Roman Barták Artificial Intelligence Roman Barták Department of Theoretical Computer Science and Mathematical Logic Introduction We will describe agents that can improve their behavior through diligent study of their

More information

Self Organizing Maps. We are drowning in information and starving for knowledge. A New Approach for Integrated Analysis of Geological Data.

Self Organizing Maps. We are drowning in information and starving for knowledge. A New Approach for Integrated Analysis of Geological Data. Radiometrics Processed Landsat TM Self Organizing Maps A New Approach for Integrated Analysis of Geological Data. We are drowning in information and starving for knowledge. Rutherford D. Roger Stephen.Fraser@csiro.au

More information

FINAL: CS 6375 (Machine Learning) Fall 2014

FINAL: CS 6375 (Machine Learning) Fall 2014 FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for

More information

CSE 151 Machine Learning. Instructor: Kamalika Chaudhuri

CSE 151 Machine Learning. Instructor: Kamalika Chaudhuri CSE 151 Machine Learning Instructor: Kamalika Chaudhuri Ensemble Learning How to combine multiple classifiers into a single one Works well if the classifiers are complementary This class: two types of

More information

ABC-Boost: Adaptive Base Class Boost for Multi-class Classification

ABC-Boost: Adaptive Base Class Boost for Multi-class Classification ABC-Boost: Adaptive Base Class Boost for Multi-class Classification Ping Li Department of Statistical Science, Cornell University, Ithaca, NY 14853 USA pingli@cornell.edu Abstract We propose -boost (adaptive

More information

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run

More information

Harrison B. Prosper. Bari Lectures

Harrison B. Prosper. Bari Lectures Harrison B. Prosper Florida State University Bari Lectures 30, 31 May, 1 June 2016 Lectures on Multivariate Methods Harrison B. Prosper Bari, 2016 1 h Lecture 1 h Introduction h Classification h Grid Searches

More information

Geotechnical modelling based on geophysical logging data

Geotechnical modelling based on geophysical logging data University of Wollongong Research Online Coal Operators' Conference Faculty of Engineering and Information Sciences 2013 Geotechnical modelling based on geophysical logging data Peter Hatherly Coalbed

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Article from Predictive Analytics and Futurism July 2016 Issue 13 Regression and Classification: A Deeper Look By Jeff Heaton Classification and regression are the two most common forms of models fitted

More information

Introduction to Machine Learning Lecture 13. Mehryar Mohri Courant Institute and Google Research

Introduction to Machine Learning Lecture 13. Mehryar Mohri Courant Institute and Google Research Introduction to Machine Learning Lecture 13 Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Multi-Class Classification Mehryar Mohri - Introduction to Machine Learning page 2 Motivation

More information

Man, Machine and Data: A Mineral Exploration Perspective

Man, Machine and Data: A Mineral Exploration Perspective Man, Machine and Data: A Mineral Exploration Perspective Eun-Jung Holden, Geodata Algorithms Team The Centre for Exploration Targeting, School of Earth Sciences The University of Western Australia Team

More information

Boosting. Acknowledgment Slides are based on tutorials from Robert Schapire and Gunnar Raetsch

Boosting. Acknowledgment Slides are based on tutorials from Robert Schapire and Gunnar Raetsch . Machine Learning Boosting Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg riedmiller@informatik.uni-freiburg.de

More information

A TWO-STAGE COMMITTEE MACHINE OF NEURAL NETWORKS

A TWO-STAGE COMMITTEE MACHINE OF NEURAL NETWORKS Journal of the Chinese Institute of Engineers, Vol. 32, No. 2, pp. 169-178 (2009) 169 A TWO-STAGE COMMITTEE MACHINE OF NEURAL NETWORKS Jen-Feng Wang, Chinson Yeh, Chen-Wen Yen*, and Mark L. Nagurka ABSTRACT

More information

Nonlinear Classification

Nonlinear Classification Nonlinear Classification INFO-4604, Applied Machine Learning University of Colorado Boulder October 5-10, 2017 Prof. Michael Paul Linear Classification Most classifiers we ve seen use linear functions

More information

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,

More information

Online Learning and Sequential Decision Making

Online Learning and Sequential Decision Making Online Learning and Sequential Decision Making Emilie Kaufmann CNRS & CRIStAL, Inria SequeL, emilie.kaufmann@univ-lille.fr Research School, ENS Lyon, Novembre 12-13th 2018 Emilie Kaufmann Online Learning

More information

Machine Learning Lecture 10

Machine Learning Lecture 10 Machine Learning Lecture 10 Neural Networks 26.11.2018 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Today s Topic Deep Learning 2 Course Outline Fundamentals Bayes

More information

Learning and Memory in Neural Networks

Learning and Memory in Neural Networks Learning and Memory in Neural Networks Guy Billings, Neuroinformatics Doctoral Training Centre, The School of Informatics, The University of Edinburgh, UK. Neural networks consist of computational units

More information

Classification objectives COMS 4771

Classification objectives COMS 4771 Classification objectives COMS 4771 1. Recap: binary classification Scoring functions Consider binary classification problems with Y = { 1, +1}. 1 / 22 Scoring functions Consider binary classification

More information