Context-based Reasoning in Ambient Intelligence - CoReAmI -

Size: px
Start display at page:

Download "Context-based Reasoning in Ambient Intelligence - CoReAmI -"

Transcription

1 Context-based in Ambient Intelligence - CoReAmI - Hristijan Gjoreski Department of Intelligent Systems, Jožef Stefan Institute Supervisor: Prof. Dr. Matjaž Gams Co-supervisor: Dr. Mitja Luštrek

2 Background What is Ambient Intelligence (AmI)? Refers environments with embedded technologies that are sensitive and responsive to the presence of people What is context? [A. Dey, 00] Context is any information that characterizes the circumstances in which an event occurs J. C. Augusto et al. (007) Context-based reasoning To reason about something using contextual information

3 Motivation () Context Numerous systems that provide services and reason about a user: health-care systems, elderly support systems, fitness applications, etc. The various sensors used in these systems provide rich contextual information: location, activity, etc. Illustrative example, which shows the importance of the context User sits and has high heart rate an alarming situation? Not always depends on the context, e.g., running as a previous activity Context-based reasoning approach CoReAmI

4 Motivation () Multi-view from multiple points of view (Multi-view reasoning) An intuitive example: sensing/reasoning about food Sight Smell Taste Complete picture of the food CoReAmI includes the idea of using multiple points of view Each view is created by using each source of information as a context individually In accordance with the: Principle of multiple knowledge Gams Ensemble learning (combining multiple ML models) Dietterich Gams M. Weak intelligence: through the principle and paradox of multiple knowledge. New York: Nova Science Publishers, 00. Dietterich TG. Ensemble methods in Machine Learning. In Proc. of the first International Workshop on Multiple Classifier Systems

5 Outline Background and Motivation Hypothesis CoReAmI Approach CoReAmI Case Studies Scientific Contributions Conclusions and Future Work 5

6 Hypothesis Extracting and combining multiple sources of information by using a context-based approach (that is, using each source of information as a context) can lead to better reasoning performance compared to conventional approaches in an AmI domain. To validate the hypothesis we developed the CoReAmI approach and tested it on three AmI problem domains: o Activity recognition o Energy expenditure estimation o Fall detection 6

7 Outline Introduction and Motivation Hypothesis CoReAmI Approach CoReAmI Case Studies Scientific Contributions Conclusions and Future Work 7

8 Outline Introduction and Motivation Hypothesis CoReAmI Approach Context Extraction Context Modeling Context Aggregation CoReAmI Case Studies Scientific Contributions Conclusions and Future Work 8

9 Outline Introduction and Motivation Hypothesis CoReAmI Approach Context Extraction Context Modeling Context Aggregation CoReAmI Case Studies Scientific Contributions Conclusions and Future Work 9

10 CoReAmI Approach s s s s m Sensor Data c c n A: Context Extraction v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c v c Aggregation Final decision C: Context Aggregation 0

11 CoReAmI Approach s s s s m Sensor Data c c n A: Context Extraction v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c v c Aggregation Final decision C: Context Aggregation

12 CoReAmI Approach s s s s m Sensor Data Accelerometer Heart rate sensor Location sensor c c n A: Context Extraction v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c v c Aggregation Final decision C: Context Aggregation

13 CoReAmI Approach A: Context Extraction s s s m Each context is a nominal variable (feature) For example: Activity (standing, sitting) Context Extraction segmentation, filtering, synchronization, classification, Heart rate (low, medium, ) Breath rate (low, medium, ) c c c c n

14 CoReAmI Approach s s s s m Sensor Data Accelerometer Heart rate sensor Location sensor c c n A: Context Extraction Location Activity Breath rate v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c v c Aggregation Final decision C: Context Aggregation 4

15 CoReAmI Approach s s s s m Breath rate Low c c n Sensor Data Accelerometer Heart rate sensor Location sensor A: Context Extraction Location Activity Breath rate v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c ML Expert rules v c Aggregation C: Context Aggregation Final decision 5

16 CoReAmI Approach B: Context Modeling Activity Heart rate Breath rate Decision Standing Low: 50 Low: 0 Yes Sitting Medium: 85 Low: No Sitting Low: 60 High: 9 No Standing Medium: 9 High: No Sitting Medium: 9 Low: No Running High: 5 Medium: No Sitting Medium: 0 Low: Yes Running High: 8 Medium: Yes Sitting Low: 59 Low: 6 Yes Standing High: 45 Medium: 9 Yes Running High: 44 High: 6 Yes The fundamental phase of CoReAmI Provides the multiple views Performs context-based partitioning Constructs an ensemble of models Breath rate Low Medium High Activity Heart rate Breath rate Decision Standing 50 0 Yes Sitting 85 No Sitting 9 No Sitting 99 Yes Sitting 59 6 Yes for the low breath rate Activity Heart rate Breath rate Decision Running 5 No Running 8 Yes Standing 45 9 Yes for the medium breath rate Activity Heart rate (min - ) Breath rate (min - ) Decision Sitting 60 9 No Standing 9 No Running 44 6 Yes for the high breath rate The same for: heart rate and activity 6 m breath rate = low m breath rate = medium m breath rate = high

17 CoReAmI Approach s s s s m Sensor Data Accelerometer Heart rate sensor Location sensor c c n A: Context Extraction Activity Heart rate Location v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c v c Aggregation Final decision C: Context Aggregation 7

18 CoReAmI Approach s s s s m Sensor Data Accelerometer Heart rate sensor Location sensor c c n A: Context Extraction Activity Heart rate Location v c v c v c Rv c Rv c Rv c R R R B: Context Modeling m c m c m c v c Aggregation Final decision C: Context Aggregation 8

19 CoReAmI Approach C: Context Aggregation Executed while evaluating a new instance Models that correspond to the instance s context values are invoked and their decisions are aggregated. m A=Standing m A=Sitting m A=Running m HR=Low m HR=Medium m HR=High Instance Activity (A) Standing Sitting Running Running Aggregation Heart rate (HR) Low Medium High Medium Decision Aggregation techniques: average, weighted average, median, voting, stacking Assumption: it is better to find a good aggregation function instead of choosing the best single model (Gams, Dietterich) Gams M. Weak intelligence: through the principle and paradox of multiple knowledge. New York: Nova Science Publishers, 00. Dietterich TG. Ensemble methods in Machine Learning. In Proc. of the first International Workshop on Multiple Classifier Systems

20 CoReAmI Summary CoReAmI General approach for context-based reasoning in Ambient Intelligence Contexts Multiple-views s s s s 4 s m c c c n c c c n Three phases: Context extraction Context modeling Context aggregation c c c c c cn v c v v x v v v y v v Rv c Rv c Rv c Rv c Rv c Rv c Rv c Rv c Rv c x y z m c m c m c m c m c m c v c v c cn v z cn Aggregation Final decision 0

21 Outline Introduction and Motivation Hypothesis CoReAmI Approach CoReAmI Case Studies Scientific Contributions Conclusions and Future Work

22 Outline Introduction and Motivation Hypothesis CoReAmI Approach CoReAmI Case Studies Energy Expenditure Estimation Activity Recognition Fall Detection Scientific Contributions Conclusions and Future Work

23 Case Study: EE Estimation Problem Description Human Energy Expenditure (EE) estimation Directly reflects the level of physical activity For sports training, weight control, management of metabolic disorders (e.g., diabetes), etc. True EE is difficult to measure by traditional means Solution: Automatic, unobtrusive and accurate EE estimation using various sensors (accelerometers, heart rate, temperature)

24 Case Study: EE Estimation Experimental Setup Sensor equipment Zephyr chest strap Shimmer accelerometers BodyMedia arm band Raw sensor used as input for CoReAmI EE estimation output used for comparison Cosmed ground truth (MET) Experimental scenario 90-minute Supervised by an expert Including 5 activities: lying, walking, running, cycling, shoveling snow, washing dishes, etc. Recorded by 0 volunteers 4

25 Case Study: EE Estimation CoReAmI for EE Estimation Context Extraction Activity Acceleration peaks count Heart rate Breath rate Chest skin temperature Galvanic skin response Arm skin temperature Near-body temperature Context modeling Discretization Decision Tree splitting criteria adapted for regression by Yong et al. 46 regression models 5 regression learning methods Context Aggregation Average Median 5

26 Case Study: EE Estimation Results RMSE Root Mean Square Error (RMSE) for CoReAmI's MET estimation compared to : single regression, Random Subspace and Bagging using 5 base learners: o Artificial Neural Network (ANN) o Support Vector machine for Regression (SVR) o Multiple Linear Regression (MLR) o Gaussian Processes for Regression (GPR) o Model tree (M5P) Base learner Single regression Random Subspace Bagging CoReAmI (average) CoReAmI (median) Improvement Improvement ANN % % SVR % % MLR % % GPR % % M5P % % CoReAmI achieves the lowest RMSE regardless of the baseline learner Statistically significant, confirmed by T-Test p = (0.05) 6

27 RMSE Case Study: EE Estimation Results CoReAmI's MET estimation compared to each of the contexts used individually (only the context models learned for the particular context). Individual contexts CoReAmI Activity Peak count Breath rate Heart rate Chest temp GSR Ambient temp Arm temp CoreAmI 7

28 Estimeted MET Case Study: EE Estimation Results Measured and estimated METs for different types of activities using: CoReAmI BodyMedia (SOTA commercial EE estimation device) ANN trained on chest-accelerometer only (ANN-Acc) CoReAmI BodyMedia ANN-Acc Vigorous stationary cycling Light stationary cycling Running Method RMSE overall CoReAmI Moderate to vigorous household activities BodyMedia.6 Walking Light household activities & exercise Sedentary ANN-Acc True MET 8

29 Case Study: EE Estimation Summary EE Estimation A novel approach to EE estimation using CoReAmI Outperformed the competing approaches The main advantages of CoReAmI: Exploits the complementarity of multiple context-based regression models. Semantically split the set (by using contexts) and not by sampling it with replacement (Bagging) or randomly choosing features (Random Subspace). The results confirmed the hypothesis. 9

30 Case Study Activity Recognition The RaReFall system (Real-time Activity Recognition and Fall detection) Evaluated as the best performing at the EvAAL-AR competition* Uses two accelerometers Problem: distinguishing standing and sitting with single torso accelerometer CoReAmI significantly improved the recognition of the two activities, by using their context The results confirm the hypothesis 80% 70% F-measure 65% 68% 78% 60% 55% 50% Baseline J48 with context features RF with context features CoReAmI *Gjoreski H, Kozina S, Gams M, Luštrek M, Álvarez JA, Hong JH, Ramos J, Dey AK, Bocca M, Patwari N. Competitive Live Evaluation of Activity-recognition Systems. IEEE Pervasive Computing, accepted, (04). JCR:.0 0

31 Case Study Fall Detection Goal: Accurate fall detection with as few sensors as possible Combining contexts extracted from inertial and location sensors by using expert rules CoReAmI significantly improved the detection performance achieving 96.6% F-measure with minimal sensor configuration inertial and location sensor The results confirm the hypothesis F-measure in % AFP MLA CoReAmI Falls Non-fall (fal-like) Non-fall Overall

32 Outline Introduction and Motivation Hypothesis CoReAmI Approach CoReAmI Case Studies Energy Expenditure Estimation Activity Recognition Fall Detection Scientific Contributions Conclusions and Future Work

33 Scientific Contributions. Development of a novel, general, context-based reasoning approach in AmI, called CoReAmI. The approach extracts multiple contexts from sensor and performs reasoning about the user using multiple models constructed for each of the contexts individually.. Application/Adaptation of CoReAmI to three AmI problem domains, which resulted in novel approach in each domain (outperforming the conventional approaches):. CoReAmI for activity recognition. CoReAmI for energy expenditure estimation. CoReAmI for fall detection. Development of a novel method for context-based partitioning of a set into multiple subsets and this way creating multiple views on the by using each feature as a context. Also applicable outside of CoReAmI.

34 Conclusions The thesis addressed the problem of combining multiple sources of information by using contextual information. The hypothesis was confirmed on all three AmI problem domains. The key idea is to partition the set (problem) using each source of information as a context individually. The CoReAmI approach can be adapted and applied to different problem domains, for which the available information can be presented by multiple contexts. 4

35 Future Work Optimizing the hyper-parameters. (Auto Weka) Context-based ensembles. Ensemble-learning algorithm for general purpose ML, (similar to Random forest). Part of WEKA. Directions of how to use CoReAmI on a new problem domain. Release the code, its documentation and appropriate sample applications. Dealing with missing context values. Dealing with redundant or similar context information. (Feature selection algorithms) Deep learning and CoReAmI 5

An Empirical Study of Building Compact Ensembles

An Empirical Study of Building Compact Ensembles An Empirical Study of Building Compact Ensembles Huan Liu, Amit Mandvikar, and Jigar Mody Computer Science & Engineering Arizona State University Tempe, AZ 85281 {huan.liu,amitm,jigar.mody}@asu.edu Abstract.

More information

Calculating Land Values by Using Advanced Statistical Approaches in Pendik

Calculating Land Values by Using Advanced Statistical Approaches in Pendik Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey Calculating Land Values by Using Advanced Statistical Approaches in Pendik Prof. Dr. Arif Cagdas AYDINOGLU Ress. Asst. Rabia BOVKIR

More information

Ensemble Methods. NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan

Ensemble Methods. NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan Ensemble Methods NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan How do you make a decision? What do you want for lunch today?! What did you have last night?! What are your favorite

More information

Ensembles. Léon Bottou COS 424 4/8/2010

Ensembles. Léon Bottou COS 424 4/8/2010 Ensembles Léon Bottou COS 424 4/8/2010 Readings T. G. Dietterich (2000) Ensemble Methods in Machine Learning. R. E. Schapire (2003): The Boosting Approach to Machine Learning. Sections 1,2,3,4,6. Léon

More information

Variance Reduction and Ensemble Methods

Variance Reduction and Ensemble Methods Variance Reduction and Ensemble Methods Nicholas Ruozzi University of Texas at Dallas Based on the slides of Vibhav Gogate and David Sontag Last Time PAC learning Bias/variance tradeoff small hypothesis

More information

Bagging and Boosting for the Nearest Mean Classifier: Effects of Sample Size on Diversity and Accuracy

Bagging and Boosting for the Nearest Mean Classifier: Effects of Sample Size on Diversity and Accuracy and for the Nearest Mean Classifier: Effects of Sample Size on Diversity and Accuracy Marina Skurichina, Liudmila I. Kuncheva 2 and Robert P.W. Duin Pattern Recognition Group, Department of Applied Physics,

More information

FEATURE SELECTION COMBINED WITH RANDOM SUBSPACE ENSEMBLE FOR GENE EXPRESSION BASED DIAGNOSIS OF MALIGNANCIES

FEATURE SELECTION COMBINED WITH RANDOM SUBSPACE ENSEMBLE FOR GENE EXPRESSION BASED DIAGNOSIS OF MALIGNANCIES FEATURE SELECTION COMBINED WITH RANDOM SUBSPACE ENSEMBLE FOR GENE EXPRESSION BASED DIAGNOSIS OF MALIGNANCIES Alberto Bertoni, 1 Raffaella Folgieri, 1 Giorgio Valentini, 1 1 DSI, Dipartimento di Scienze

More information

VBM683 Machine Learning

VBM683 Machine Learning VBM683 Machine Learning Pinar Duygulu Slides are adapted from Dhruv Batra Bias is the algorithm's tendency to consistently learn the wrong thing by not taking into account all the information in the data

More information

ECE 5424: Introduction to Machine Learning

ECE 5424: Introduction to Machine Learning ECE 5424: Introduction to Machine Learning Topics: Ensemble Methods: Bagging, Boosting PAC Learning Readings: Murphy 16.4;; Hastie 16 Stefan Lee Virginia Tech Fighting the bias-variance tradeoff Simple

More information

Ensemble Methods and Random Forests

Ensemble Methods and Random Forests Ensemble Methods and Random Forests Vaishnavi S May 2017 1 Introduction We have seen various analysis for classification and regression in the course. One of the common methods to reduce the generalization

More information

Learning with multiple models. Boosting.

Learning with multiple models. Boosting. CS 2750 Machine Learning Lecture 21 Learning with multiple models. Boosting. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Learning with multiple models: Approach 2 Approach 2: use multiple models

More information

Robotics 2 AdaBoost for People and Place Detection

Robotics 2 AdaBoost for People and Place Detection Robotics 2 AdaBoost for People and Place Detection Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard v.1.0, Kai Arras, Oct 09, including material by Luciano Spinello and Oscar Martinez Mozos

More information

Empirical Risk Minimization, Model Selection, and Model Assessment

Empirical Risk Minimization, Model Selection, and Model Assessment Empirical Risk Minimization, Model Selection, and Model Assessment CS6780 Advanced Machine Learning Spring 2015 Thorsten Joachims Cornell University Reading: Murphy 5.7-5.7.2.4, 6.5-6.5.3.1 Dietterich,

More information

Multimodal context analysis and prediction

Multimodal context analysis and prediction Multimodal context analysis and prediction Valeria Tomaselli (valeria.tomaselli@st.com) Sebastiano Battiato Giovanni Maria Farinella Tiziana Rotondo (PhD student) Outline 2 Context analysis vs prediction

More information

Probabilistic Graphical Models for Image Analysis - Lecture 1

Probabilistic Graphical Models for Image Analysis - Lecture 1 Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.

More information

Will it rain tomorrow?

Will it rain tomorrow? Will it rain tomorrow? Bilal Ahmed - 561539 Department of Computing and Information Systems, The University of Melbourne, Victoria, Australia bahmad@student.unimelb.edu.au Abstract With the availability

More information

Data Mining und Maschinelles Lernen

Data Mining und Maschinelles Lernen Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting

More information

Anomaly Detection for the CERN Large Hadron Collider injection magnets

Anomaly Detection for the CERN Large Hadron Collider injection magnets Anomaly Detection for the CERN Large Hadron Collider injection magnets Armin Halilovic KU Leuven - Department of Computer Science In cooperation with CERN 2018-07-27 0 Outline 1 Context 2 Data 3 Preprocessing

More information

day month year documentname/initials 1

day month year documentname/initials 1 ECE471-571 Pattern Recognition Lecture 13 Decision Tree Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi

More information

Ensembles of Classifiers.

Ensembles of Classifiers. Ensembles of Classifiers www.biostat.wisc.edu/~dpage/cs760/ 1 Goals for the lecture you should understand the following concepts ensemble bootstrap sample bagging boosting random forests error correcting

More information

Human Activity Recognition and Estimation of Calorie Expenditure - A Data Analytic Approach

Human Activity Recognition and Estimation of Calorie Expenditure - A Data Analytic Approach Human Activity Recognition and Estimation of Calorie Expenditure - A Data Analytic Approach Mushtaque Ahamed A Dr. Snehanshu Saha PESIT-Bangalore South Campus December 4, 2015 Mushtaque Ahamed A Dr. Snehanshu

More information

Decision Trees: Overfitting

Decision Trees: Overfitting Decision Trees: Overfitting Emily Fox University of Washington January 30, 2017 Decision tree recap Loan status: Root 22 18 poor 4 14 Credit? Income? excellent 9 0 3 years 0 4 Fair 9 4 Term? 5 years 9

More information

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Hujia Yu, Jiafu Wu [hujiay, jiafuwu]@stanford.edu 1. Introduction Housing prices are an important

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Ensembles Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique Fédérale de Lausanne

More information

Hierarchical Boosting and Filter Generation

Hierarchical Boosting and Filter Generation January 29, 2007 Plan Combining Classifiers Boosting Neural Network Structure of AdaBoost Image processing Hierarchical Boosting Hierarchical Structure Filters Combining Classifiers Combining Classifiers

More information

Decision Trees. Tirgul 5

Decision Trees. Tirgul 5 Decision Trees Tirgul 5 Using Decision Trees It could be difficult to decide which pet is right for you. We ll find a nice algorithm to help us decide what to choose without having to think about it. 2

More information

Ensemble Methods: Jay Hyer

Ensemble Methods: Jay Hyer Ensemble Methods: committee-based learning Jay Hyer linkedin.com/in/jayhyer @adatahead Overview Why Ensemble Learning? What is learning? How is ensemble learning different? Boosting Weak and Strong Learners

More information

Ensemble Methods. Charles Sutton Data Mining and Exploration Spring Friday, 27 January 12

Ensemble Methods. Charles Sutton Data Mining and Exploration Spring Friday, 27 January 12 Ensemble Methods Charles Sutton Data Mining and Exploration Spring 2012 Bias and Variance Consider a regression problem Y = f(x)+ N(0, 2 ) With an estimate regression function ˆf, e.g., ˆf(x) =w > x Suppose

More information

Active Sonar Target Classification Using Classifier Ensembles

Active Sonar Target Classification Using Classifier Ensembles International Journal of Engineering Research and Technology. ISSN 0974-3154 Volume 11, Number 12 (2018), pp. 2125-2133 International Research Publication House http://www.irphouse.com Active Sonar Target

More information

Machine Learning Recitation 8 Oct 21, Oznur Tastan

Machine Learning Recitation 8 Oct 21, Oznur Tastan Machine Learning 10601 Recitation 8 Oct 21, 2009 Oznur Tastan Outline Tree representation Brief information theory Learning decision trees Bagging Random forests Decision trees Non linear classifier Easy

More information

Molinas. June 15, 2018

Molinas. June 15, 2018 ITT8 SAMBa Presentation June 15, 2018 ling Data The data we have include: Approx 30,000 questionnaire responses each with 234 questions during 1998-2017 A data set of 60 questions asked to 500,000 households

More information

Click Prediction and Preference Ranking of RSS Feeds

Click Prediction and Preference Ranking of RSS Feeds Click Prediction and Preference Ranking of RSS Feeds 1 Introduction December 11, 2009 Steven Wu RSS (Really Simple Syndication) is a family of data formats used to publish frequently updated works. RSS

More information

Inertial Navigation and Various Applications of Inertial Data. Yongcai Wang. 9 November 2016

Inertial Navigation and Various Applications of Inertial Data. Yongcai Wang. 9 November 2016 Inertial Navigation and Various Applications of Inertial Data Yongcai Wang 9 November 2016 Types of Gyroscope Mechanical Gyroscope Laser Gyroscope Sagnac Effect Stable Platform IMU and Strapdown IMU In

More information

Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll

Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll Maarten Bieshaar, Günther Reitberger, Stefan Zernetsch, Prof. Dr. Bernhard Sick, Dr. Erich Fuchs, Prof. Dr.-Ing. Konrad Doll 08.02.2017 By 2030 road traffic deaths will be the fifth leading cause of death

More information

TDT4173 Machine Learning

TDT4173 Machine Learning TDT4173 Machine Learning Lecture 9 Learning Classifiers: Bagging & Boosting Norwegian University of Science and Technology Helge Langseth IT-VEST 310 helgel@idi.ntnu.no 1 TDT4173 Machine Learning Outline

More information

Algorithms for Classification: The Basic Methods

Algorithms for Classification: The Basic Methods Algorithms for Classification: The Basic Methods Outline Simplicity first: 1R Naïve Bayes 2 Classification Task: Given a set of pre-classified examples, build a model or classifier to classify new cases.

More information

CS7267 MACHINE LEARNING

CS7267 MACHINE LEARNING CS7267 MACHINE LEARNING ENSEMBLE LEARNING Ref: Dr. Ricardo Gutierrez-Osuna at TAMU, and Aarti Singh at CMU Mingon Kang, Ph.D. Computer Science, Kennesaw State University Definition of Ensemble Learning

More information

Applying Machine Learning for Gravitational-wave Burst Data Analysis

Applying Machine Learning for Gravitational-wave Burst Data Analysis Applying Machine Learning for Gravitational-wave Burst Data Analysis Junwei Cao LIGO Scientific Collaboration Research Group Research Institute of Information Technology Tsinghua University June 29, 2016

More information

What makes good ensemble? CS789: Machine Learning and Neural Network. Introduction. More on diversity

What makes good ensemble? CS789: Machine Learning and Neural Network. Introduction. More on diversity What makes good ensemble? CS789: Machine Learning and Neural Network Ensemble methods Jakramate Bootkrajang Department of Computer Science Chiang Mai University 1. A member of the ensemble is accurate.

More information

Ensemble Methods for Machine Learning

Ensemble Methods for Machine Learning Ensemble Methods for Machine Learning COMBINING CLASSIFIERS: ENSEMBLE APPROACHES Common Ensemble classifiers Bagging/Random Forests Bucket of models Stacking Boosting Ensemble classifiers we ve studied

More information

Hypothesis Evaluation

Hypothesis Evaluation Hypothesis Evaluation Machine Learning Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Hypothesis Evaluation Fall 1395 1 / 31 Table of contents 1 Introduction

More information

Machine Learning and Deep Learning! Vincent Lepetit!

Machine Learning and Deep Learning! Vincent Lepetit! Machine Learning and Deep Learning!! Vincent Lepetit! 1! What is Machine Learning?! 2! Hand-Written Digit Recognition! 2 9 3! Hand-Written Digit Recognition! Formalization! 0 1 x = @ A Images are 28x28

More information

Random Forests. These notes rely heavily on Biau and Scornet (2016) as well as the other references at the end of the notes.

Random Forests. These notes rely heavily on Biau and Scornet (2016) as well as the other references at the end of the notes. Random Forests One of the best known classifiers is the random forest. It is very simple and effective but there is still a large gap between theory and practice. Basically, a random forest is an average

More information

Introduction to Machine Learning Midterm Exam

Introduction to Machine Learning Midterm Exam 10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but

More information

Infinite Ensemble Learning with Support Vector Machinery

Infinite Ensemble Learning with Support Vector Machinery Infinite Ensemble Learning with Support Vector Machinery Hsuan-Tien Lin and Ling Li Learning Systems Group, California Institute of Technology ECML/PKDD, October 4, 2005 H.-T. Lin and L. Li (Learning Systems

More information

Big Data Analytics. Special Topics for Computer Science CSE CSE Feb 24

Big Data Analytics. Special Topics for Computer Science CSE CSE Feb 24 Big Data Analytics Special Topics for Computer Science CSE 4095-001 CSE 5095-005 Feb 24 Fei Wang Associate Professor Department of Computer Science and Engineering fei_wang@uconn.edu Prediction III Goal

More information

Exam Machine Learning for the Quantified Self with answers :00-14:45

Exam Machine Learning for the Quantified Self with answers :00-14:45 Exam Machine Learning for the Quantified Self with answers 21. 06. 2017 12:00-14:45 NOTES: 1. YOUR NAME MUST BE WRITTEN ON EACH SHEET IN CAPITALS. 2. Answer the questions in Dutch or English. 3. Points

More information

UVA CS 4501: Machine Learning

UVA CS 4501: Machine Learning UVA CS 4501: Machine Learning Lecture 21: Decision Tree / Random Forest / Ensemble Dr. Yanjun Qi University of Virginia Department of Computer Science Where are we? è Five major sections of this course

More information

A Support Vector Regression Model for Forecasting Rainfall

A Support Vector Regression Model for Forecasting Rainfall A Support Vector Regression for Forecasting Nasimul Hasan 1, Nayan Chandra Nath 1, Risul Islam Rasel 2 Department of Computer Science and Engineering, International Islamic University Chittagong, Bangladesh

More information

Hierarchical Signal Segmentation and Classification for Accurate Activity Recognition

Hierarchical Signal Segmentation and Classification for Accurate Activity Recognition Hierarchical Signal Segmentation and Classification for Accurate Activity Recognition Ali Akbari Department of Biomedical Engineering Texas A&M University aliakbari@tamu.edu Jian Wu Department of Computer

More information

TDT4173 Machine Learning

TDT4173 Machine Learning TDT4173 Machine Learning Lecture 3 Bagging & Boosting + SVMs Norwegian University of Science and Technology Helge Langseth IT-VEST 310 helgel@idi.ntnu.no 1 TDT4173 Machine Learning Outline 1 Ensemble-methods

More information

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition Data Mining Classification: Basic Concepts and Techniques Lecture Notes for Chapter 3 by Tan, Steinbach, Karpatne, Kumar 1 Classification: Definition Given a collection of records (training set ) Each

More information

Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers

Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers Erin Allwein, Robert Schapire and Yoram Singer Journal of Machine Learning Research, 1:113-141, 000 CSE 54: Seminar on Learning

More information

Hierarchical models for the rainfall forecast DATA MINING APPROACH

Hierarchical models for the rainfall forecast DATA MINING APPROACH Hierarchical models for the rainfall forecast DATA MINING APPROACH Thanh-Nghi Do dtnghi@cit.ctu.edu.vn June - 2014 Introduction Problem large scale GCM small scale models Aim Statistical downscaling local

More information

A Novel Activity Detection Method

A Novel Activity Detection Method A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of

More information

Linear Classifiers: Expressiveness

Linear Classifiers: Expressiveness Linear Classifiers: Expressiveness Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 Lecture outline Linear classifiers: Introduction What functions do linear classifiers express?

More information

Chapter 14 Combining Models

Chapter 14 Combining Models Chapter 14 Combining Models T-61.62 Special Course II: Pattern Recognition and Machine Learning Spring 27 Laboratory of Computer and Information Science TKK April 3th 27 Outline Independent Mixing Coefficients

More information

Analyzing dynamic ensemble selection techniques using dissimilarity analysis

Analyzing dynamic ensemble selection techniques using dissimilarity analysis Analyzing dynamic ensemble selection techniques using dissimilarity analysis George D. C. Cavalcanti 1 1 Centro de Informática - Universidade Federal de Pernambuco (UFPE), Brazil www.cin.ufpe.br/~gdcc

More information

Lecture 3: Decision Trees

Lecture 3: Decision Trees Lecture 3: Decision Trees Cognitive Systems - Machine Learning Part I: Basic Approaches of Concept Learning ID3, Information Gain, Overfitting, Pruning last change November 26, 2014 Ute Schmid (CogSys,

More information

Holdout and Cross-Validation Methods Overfitting Avoidance

Holdout and Cross-Validation Methods Overfitting Avoidance Holdout and Cross-Validation Methods Overfitting Avoidance Decision Trees Reduce error pruning Cost-complexity pruning Neural Networks Early stopping Adjusting Regularizers via Cross-Validation Nearest

More information

Classification using stochastic ensembles

Classification using stochastic ensembles July 31, 2014 Topics Introduction Topics Classification Application and classfication Classification and Regression Trees Stochastic ensemble methods Our application: USAID Poverty Assessment Tools Topics

More information

Explaining Results of Neural Networks by Contextual Importance and Utility

Explaining Results of Neural Networks by Contextual Importance and Utility Explaining Results of Neural Networks by Contextual Importance and Utility Kary FRÄMLING Dep. SIMADE, Ecole des Mines, 158 cours Fauriel, 42023 Saint-Etienne Cedex 2, FRANCE framling@emse.fr, tel.: +33-77.42.66.09

More information

EECS 349:Machine Learning Bryan Pardo

EECS 349:Machine Learning Bryan Pardo EECS 349:Machine Learning Bryan Pardo Topic 2: Decision Trees (Includes content provided by: Russel & Norvig, D. Downie, P. Domingos) 1 General Learning Task There is a set of possible examples Each example

More information

Predicting New Search-Query Cluster Volume

Predicting New Search-Query Cluster Volume Predicting New Search-Query Cluster Volume Jacob Sisk, Cory Barr December 14, 2007 1 Problem Statement Search engines allow people to find information important to them, and search engine companies derive

More information

Outline: Ensemble Learning. Ensemble Learning. The Wisdom of Crowds. The Wisdom of Crowds - Really? Crowd wiser than any individual

Outline: Ensemble Learning. Ensemble Learning. The Wisdom of Crowds. The Wisdom of Crowds - Really? Crowd wiser than any individual Outline: Ensemble Learning We will describe and investigate algorithms to Ensemble Learning Lecture 10, DD2431 Machine Learning A. Maki, J. Sullivan October 2014 train weak classifiers/regressors and how

More information

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring / Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring 2015 http://ce.sharif.edu/courses/93-94/2/ce717-1 / Agenda Combining Classifiers Empirical view Theoretical

More information

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann Machine Learning! in just a few minutes Jan Peters Gerhard Neumann 1 Purpose of this Lecture Foundations of machine learning tools for robotics We focus on regression methods and general principles Often

More information

On the Problem of Error Propagation in Classifier Chains for Multi-Label Classification

On the Problem of Error Propagation in Classifier Chains for Multi-Label Classification On the Problem of Error Propagation in Classifier Chains for Multi-Label Classification Robin Senge, Juan José del Coz and Eyke Hüllermeier Draft version of a paper to appear in: L. Schmidt-Thieme and

More information

ABC random forest for parameter estimation. Jean-Michel Marin

ABC random forest for parameter estimation. Jean-Michel Marin ABC random forest for parameter estimation Jean-Michel Marin Université de Montpellier Institut Montpelliérain Alexander Grothendieck (IMAG) Institut de Biologie Computationnelle (IBC) Labex Numev! joint

More information

From statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu

From statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu From statistics to data science BAE 815 (Fall 2017) Dr. Zifei Liu Zifeiliu@ksu.edu Why? How? What? How much? How many? Individual facts (quantities, characters, or symbols) The Data-Information-Knowledge-Wisdom

More information

Machine Learning. Ensemble Methods. Manfred Huber

Machine Learning. Ensemble Methods. Manfred Huber Machine Learning Ensemble Methods Manfred Huber 2015 1 Bias, Variance, Noise Classification errors have different sources Choice of hypothesis space and algorithm Training set Noise in the data The expected

More information

Algorithm-Independent Learning Issues

Algorithm-Independent Learning Issues Algorithm-Independent Learning Issues Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007, Selim Aksoy Introduction We have seen many learning

More information

B555 - Machine Learning - Homework 4. Enrique Areyan April 28, 2015

B555 - Machine Learning - Homework 4. Enrique Areyan April 28, 2015 - Machine Learning - Homework Enrique Areyan April 8, 01 Problem 1: Give decision trees to represent the following oolean functions a) A b) A C c) Ā d) A C D e) A C D where Ā is a negation of A and is

More information

ECE 5984: Introduction to Machine Learning

ECE 5984: Introduction to Machine Learning ECE 5984: Introduction to Machine Learning Topics: Ensemble Methods: Bagging, Boosting Readings: Murphy 16.4; Hastie 16 Dhruv Batra Virginia Tech Administrativia HW3 Due: April 14, 11:55pm You will implement

More information

Neural Networks and Ensemble Methods for Classification

Neural Networks and Ensemble Methods for Classification Neural Networks and Ensemble Methods for Classification NEURAL NETWORKS 2 Neural Networks A neural network is a set of connected input/output units (neurons) where each connection has a weight associated

More information

Forecasting of Solar Photovoltaic System Power Generation using Wavelet Decomposition and Biascompensated

Forecasting of Solar Photovoltaic System Power Generation using Wavelet Decomposition and Biascompensated Forecasting of Solar Photovoltaic System Power Generation using Wavelet Decomposition and Biascompensated Random Forest Po-Han Chiang, Siva Prasad Varma Chiluvuri, Sujit Dey, Truong Q. Nguyen Dept. of

More information

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning Mark Schmidt University of British Columbia, May 2016 www.cs.ubc.ca/~schmidtm/svan16 Some images from this lecture are

More information

The Deep Forest and its Modi cations

The Deep Forest and its Modi cations The Deep Forest and its Modi cations Lev V. Utkin pml.spbstu.ru Munich, November 4, 2017 In memory of Kurt Weichselberger Peter the Great Saint-Petersburg Polytechnic University Agenda 1 Decision Trees

More information

Linear and Logistic Regression. Dr. Xiaowei Huang

Linear and Logistic Regression. Dr. Xiaowei Huang Linear and Logistic Regression Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Two Classical Machine Learning Algorithms Decision tree learning K-nearest neighbor Model Evaluation Metrics

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University October 11, 2012 Today: Computational Learning Theory Probably Approximately Coorrect (PAC) learning theorem

More information

Gradient Boosting, Continued

Gradient Boosting, Continued Gradient Boosting, Continued David Rosenberg New York University December 26, 2016 David Rosenberg (New York University) DS-GA 1003 December 26, 2016 1 / 16 Review: Gradient Boosting Review: Gradient Boosting

More information

Boosting & Deep Learning

Boosting & Deep Learning Boosting & Deep Learning Ensemble Learning n So far learning methods that learn a single hypothesis, chosen form a hypothesis space that is used to make predictions n Ensemble learning à select a collection

More information

Recitation 9. Gradient Boosting. Brett Bernstein. March 30, CDS at NYU. Brett Bernstein (CDS at NYU) Recitation 9 March 30, / 14

Recitation 9. Gradient Boosting. Brett Bernstein. March 30, CDS at NYU. Brett Bernstein (CDS at NYU) Recitation 9 March 30, / 14 Brett Bernstein CDS at NYU March 30, 2017 Brett Bernstein (CDS at NYU) Recitation 9 March 30, 2017 1 / 14 Initial Question Intro Question Question Suppose 10 different meteorologists have produced functions

More information

Fuzzy Systems. Introduction

Fuzzy Systems. Introduction Fuzzy Systems Introduction Prof. Dr. Rudolf Kruse Christian Moewes {kruse,cmoewes}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge

More information

Evaluation of Two Level Classifier for Predicting Compressor Failures in Heavy Duty Vehicles. Slawomir Nowaczyk SAIS 2017 workshop, May

Evaluation of Two Level Classifier for Predicting Compressor Failures in Heavy Duty Vehicles. Slawomir Nowaczyk SAIS 2017 workshop, May Evaluation of Two Level Classifier for Predicting Compressor Failures in Heavy Duty Vehicles Yuantao Fan, Pablo De Moral & Slawomir Nowaczyk SAIS 2017 workshop, 15-16 May Objective & Motivation Predictive

More information

OVERLAPPING ANIMAL SOUND CLASSIFICATION USING SPARSE REPRESENTATION

OVERLAPPING ANIMAL SOUND CLASSIFICATION USING SPARSE REPRESENTATION OVERLAPPING ANIMAL SOUND CLASSIFICATION USING SPARSE REPRESENTATION Na Lin, Haixin Sun Xiamen University Key Laboratory of Underwater Acoustic Communication and Marine Information Technology, Ministry

More information

Summary and discussion of: Dropout Training as Adaptive Regularization

Summary and discussion of: Dropout Training as Adaptive Regularization Summary and discussion of: Dropout Training as Adaptive Regularization Statistics Journal Club, 36-825 Kirstin Early and Calvin Murdock November 21, 2014 1 Introduction Multi-layered (i.e. deep) artificial

More information

A Comparison among various Classification Algorithms for Travel Mode Detection using Sensors data collected by Smartphones

A Comparison among various Classification Algorithms for Travel Mode Detection using Sensors data collected by Smartphones CUPUM 2015 175-Paper A Comparison among various Classification Algorithms for Travel Mode Detection using Sensors data collected by Smartphones Muhammad Awais Shafique and Eiji Hato Abstract Nowadays,

More information

Voting (Ensemble Methods)

Voting (Ensemble Methods) 1 2 Voting (Ensemble Methods) Instead of learning a single classifier, learn many weak classifiers that are good at different parts of the data Output class: (Weighted) vote of each classifier Classifiers

More information

Fuzzy Systems. Introduction

Fuzzy Systems. Introduction Fuzzy Systems Introduction Prof. Dr. Rudolf Kruse Christoph Doell {kruse,doell}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge Processing

More information

Math 6330: Statistical Consulting Class 5

Math 6330: Statistical Consulting Class 5 Math 6330: Statistical Consulting Class 5 Tony Cox tcoxdenver@aol.com University of Colorado at Denver Course web site: http://cox-associates.com/6330/ What is a predictive model? The probability that

More information

Analysis of Data Mining Techniques for Weather Prediction

Analysis of Data Mining Techniques for Weather Prediction ISSN (Print) : 0974-6846 ISSN (Online) : 0974-5645 Indian Journal of Science and Technology, Vol 9(38), DOI: 10.17485/ijst/2016/v9i38/101962, October 2016 Analysis of Data Mining Techniques for Weather

More information

Chapter 18. Decision Trees and Ensemble Learning. Recall: Learning Decision Trees

Chapter 18. Decision Trees and Ensemble Learning. Recall: Learning Decision Trees CSE 473 Chapter 18 Decision Trees and Ensemble Learning Recall: Learning Decision Trees Example: When should I wait for a table at a restaurant? Attributes (features) relevant to Wait? decision: 1. Alternate:

More information

CS 229 Final report A Study Of Ensemble Methods In Machine Learning

CS 229 Final report A Study Of Ensemble Methods In Machine Learning A Study Of Ensemble Methods In Machine Learning Abstract The idea of ensemble methodology is to build a predictive model by integrating multiple models. It is well-known that ensemble methods can be used

More information

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Speaker Representation and Verification Part II. by Vasileios Vasilakakis Speaker Representation and Verification Part II by Vasileios Vasilakakis Outline -Approaches of Neural Networks in Speaker/Speech Recognition -Feed-Forward Neural Networks -Training with Back-propagation

More information

Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns

Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns Aly Kane alykane@stanford.edu Ariel Sagalovsky asagalov@stanford.edu Abstract Equipped with an understanding of the factors that influence

More information

ECE521 Lecture 7/8. Logistic Regression

ECE521 Lecture 7/8. Logistic Regression ECE521 Lecture 7/8 Logistic Regression Outline Logistic regression (Continue) A single neuron Learning neural networks Multi-class classification 2 Logistic regression The output of a logistic regression

More information

Predicting Individual Thermal Comfort using Machine Learning Algorithms

Predicting Individual Thermal Comfort using Machine Learning Algorithms Predicting Individual Thermal Comfort using Machine Learning Algorithms Asma Ahmad Farhan 1, Krishna Pattipati 2, Bing Wang 1, and Peter Luh 2 Abstract thermal sensation in an environment may be delayed,

More information

Machine Learning Approaches to Crop Yield Prediction and Climate Change Impact Assessment

Machine Learning Approaches to Crop Yield Prediction and Climate Change Impact Assessment Machine Learning Approaches to Crop Yield Prediction and Climate Change Impact Assessment Andrew Crane-Droesch FCSM, March 2018 The views expressed are those of the authors and should not be attributed

More information

Maximum Direction to Geometric Mean Spectral Response Ratios using the Relevance Vector Machine

Maximum Direction to Geometric Mean Spectral Response Ratios using the Relevance Vector Machine Maximum Direction to Geometric Mean Spectral Response Ratios using the Relevance Vector Machine Y. Dak Hazirbaba, J. Tezcan, Q. Cheng Southern Illinois University Carbondale, IL, USA SUMMARY: The 2009

More information