Are You The Beatles Or The Rolling Stones?
|
|
- Alexis Mosley
- 5 years ago
- Views:
Transcription
1 N. Ortelli, G. Lederrey, A. Alahi, M. Bierlaire Transport and Mobility Laboratory School of Architecture, Civil, and Environmental Engineering ENAC École Polytechnique Fédérale de Lausanne 13th Workshop on Discrete Choice Models June 21-23, 2018
2
3 The new and good ideas come from having a very broad and multidisciplinary range of interests. Robin Chase
4 Discrete choice models Understand and predict individual choice behavior using mathematical models 4 / 51
5 Discrete choice models Understand and predict individual choice behavior using mathematical models 4 / 51
6 Discrete choice models U in = V in + ε in 5 / 51
7 Discrete choice models U in = V in + ε in P in = e V in j Cn ev jn 5 / 51
8 Discrete choice models U in = V in + ε in P in = e V in j Cn ev jn V in =... + β c Cost in + β t Time in / 51
9 Discrete choice models U in = V in + ε in P in = e V in j Cn ev jn V in =... + β c Cost in + β t Time in +... }{{} linear in parameters 5 / 51
10 Discrete choice models U in = V in + ε in P in = e V in j Cn ev jn V in =... + β c Cost in + β t Time in +... }{{} linear in parameters Prediction accuracy + Interpretability 5 / 51
11 Machine learning Neural networks 6 / 51
12 Machine learning Neural networks Very good prediction accuracy 6 / 51
13 Machine learning Neural networks Very good prediction accuracy 6 / 51
14 Machine learning 7 / 51
15 Machine learning 7 / 51
16 Enhancing Discrete Choice Models with Neural Networks
17 General idea Bringing the predictive strength of Neural Networks, a powerful machine learning-based technique, to the field of Discrete Choice Models without compromising the interpretability of these models 9 / 51
18 Multinomial Logit as Convolution NN 10 / 51
19 Multinomial Logit as Convolution NN Activation Function: Softmax Loss: Categorical Cross-Entropy 11 / 51
20 Multinomial Logit as Convolution NN A single convolution gives us the utility functions 12 / 51
21 Model and Validation on Swissmetro Stated Preference Survey Simple utilities for validation purposes: V car = ASC car + β cost Cost car + β time Time car V train = ASC train + β cost Cost train + β time Time train V SM = ASC SM + β cost Cost SM + β time Time SM Biogeme betas Convolution Kernel weights 13 / 51
22 Neural Network enhanced DCM Case 1: same inputs The new term from the NN overruns the MNL linear parameters 14 / 51
23 Neural Network enhanced DCM Case 2: different inputs New term in the utility specification coming from a NN using all unused features 15 / 51
24 Neural Network enhanced DCM Case 2: different inputs U in = ASC i + β cost Cost in + β time Time in + β NN NN in + ε in Interpretation: NN in = uncaptured information of MNL model MNL estimates keep their significance Increase in log-likelihood 16 / 51
25 Swissmetro case Utility Functions 17 / 51
26 Swissmetro case Unused Features 18 / 51
27 Results Benchmark - standard MNL 19 / 51
28 Results Benchmark - standard MNL 19 / 51
29 Results Hybrid model ++ Increased Likelihood 20 / 51
30 Results Hybrid model ++ Increased likelihood 20 / 51
31 Results Hybrid model ++ Keeps important parameters significant 20 / 51
32 Results Hybrid model Some parameters loose significance 20 / 51
33 Results Simple hybrid model - only key parameters ++ Keeps important parameters significant 21 / 51
34 Results Simple hybrid model - only key parameters Increased likelihood + significant parameters 21 / 51
35 Results Comparison 22 / 51
36 Conclusions Enhancing DCM greatly increases likelihood Same input or correlated input - Breaks original parameter significance Independent Input - Best performances for likelihood - All parameters have good sign and significance 23 / 51
37 Future work Interpretability: what is really happening? 24 / 51
38 What do you think?
39 Automatic Utility Specification Using Machine Learning Techniques
40 Credit Nicola Ortelli 27 / 51
41 Are You The Beatles Or The Rolling Stones? Motivation Determining the appropriate utility specification for a particular application is a difficult task 28 / 51
42 Motivation Determining the appropriate utility specification for a particular application is a difficult task Expertise + Intuition 28 / 51
43 Motivation Determining the appropriate utility specification for a particular application is a difficult task Inspiration + Experience 28 / 51
44 Motivation Determining the appropriate utility specification for a particular application is a difficult task Trial-and-error 28 / 51
45 Motivation Determining the appropriate utility specification for a particular application is a difficult task Time consuming 28 / 51
46 Objective Build a procedure that automatically finds a good utility specification based on the data Define neighborhood relations between specifications and use classical local search algorithms 29 / 51
47 Automatic utility specification framework 30 / 51
48 Data partition 31 / 51
49 Data partition Data set separated into train set and test set Train set separated into K folds 32 / 51
50 Specifications and neighborhood structure 33 / 51
51 Specifications and neighborhood structure Current assumptions: Only continuous and binary variables Each continuous variable is included either on its own or in interaction with one binary variable All parameters are alternative specific Example: V =... + β x x +... V =... + β x,b1 =0 x (1 b 1) + β x,b1 =1 x b V =... + β x,b2 =0 x (1 b 2) + β x,b2 =1 x b V =... + β x,b1 =0,b 2 =0 x (1 b 1) (1 b 2) + β x,b1 =1,b 2 =0 x b 1 (1 b 2) + β x,b1 =0,b 2 =1 x (1 b 1) b 2 + β x,b1 =1,b 2 =1 x b 1 b / 51
52 Specifications and neighborhood structure Neighbors of a particular specification = all specifications that are a single change away from it Four possible changes: add, remove, interact, un-interact 35 / 51
53 Measure of performance 36 / 51
54 Measure of performance Current assumptions: Performance of a specification is measured as the log-likelihood it yields on new data Cross-validation is used to avoid overfitting or favorizing models with a large number of parameters The global performance P m of model m is defined as the sum of the log-likelihoods obtained on each fold after estimation on the K-1 others: K P m = k=1 L mk 37 / 51
55 Neighbor validation 38 / 51
56 Neighbor validation Conditions of acceptance of a model m : A neighbor that performs better than the baseline is always accepted P(accept m) = 1 if P m P mb One that performs worse might still get accepted, depending on the difference of performance P m P mb and the current iteration number P(accept m) = exp(pm Pm b ) T t if P m < P mb The algorithm ends when all neighbors of a model are rejected 39 / 51
57 Case study 1: Optima
58 Case study: Optima Revealed preference survey conducted in Switzerland for CarPostal: Three alternatives: public transports, car and soft modes observations left after discarding incomplete ones Two binary and six continuous variables are considered different specifications can be obtained 41 / 51
59 Best specification 42 / 51
60 Comparison with an existing model 43 / 51
61 Case study 2: SwissMetro
62 Case study: SwissMetro Stated choice survey to analyze the impact of the Swissmetro: Three alternatives: train, Swissmetro and car Nine different situations for each of the respondents Two binary and eight continuous variables are considered different specifications can be obtained 45 / 51
63 Best specification 46 / 51
64 Comparison with an existing model 47 / 51
65 Conclusions
66 Conclusions Despite a certain number of restrictions, results show that this topic is worth further investigation The procedure reaches very good specifications in a matter of minutes In some cases, the obtained model outperforms the benchmark with less variables under consideration 49 / 51
67 Future work VNS > neighborhood structures + relax assumptions What is a good utility specification? Predictability Parameters significance Behavioral interpretation / 51
68 Thank you! The fact is I ve always loved both bands :-)
Binary choice. Michel Bierlaire
Binary choice Michel Bierlaire Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique Fédérale de Lausanne M. Bierlaire (TRANSP-OR ENAC EPFL)
More informationConvolution and Pooling as an Infinitely Strong Prior
Convolution and Pooling as an Infinitely Strong Prior Sargur Srihari srihari@buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Convolutional
More informationDiscrete panel data. Michel Bierlaire
Discrete panel data Michel Bierlaire Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique Fédérale de Lausanne M. Bierlaire (TRANSP-OR ENAC
More informationStatistical Tests. Matthieu de Lapparent
Statistical Tests Matthieu de Lapparent matthieu.delapparent@epfl.ch Transport and Mobility Laboratory, School of Architecture, Civil and Environmental Engineering, Ecole Polytechnique Fédérale de Lausanne
More informationRegression Adjustment with Artificial Neural Networks
Regression Adjustment with Artificial Neural Networks Age of Big Data: data comes in a rate and in a variety of types that exceed our ability to analyse it Texts, image, speech, video Real motivation:
More informationIntegrating advanced discrete choice models in mixed integer linear optimization
Integrating advanced discrete choice models in mixed integer linear optimization Meritxell Pacheco Shadi Sharif Azadeh, Michel Bierlaire, Bernard Gendron Transport and Mobility Laboratory (TRANSP-OR) École
More informationCalculating indicators with Biogeme
Calculating indicators with Biogeme Michel Bierlaire Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique Fédérale de Lausanne September 25,
More informationSwissmetro. March 3, Model Specification with Generic Attributes
Swissmetro March 3, 2015 Model Specification with Generic Attributes Model file: MNL_SM_generic.mod The dataset consists of survey data collected on the trains between St. Gallen and Geneva in Switzerland.
More informationNormalization and correlation of cross-nested logit models
Normalization and correlation of cross-nested logit models E. Abbe, M. Bierlaire, T. Toledo Lab. for Information and Decision Systems, Massachusetts Institute of Technology Inst. of Mathematics, Ecole
More informationDeep Learning Srihari. Deep Belief Nets. Sargur N. Srihari
Deep Belief Nets Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines for continuous
More informationMachine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group
Machine Learning for Computer Vision 8. Neural Networks and Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group INTRODUCTION Nonlinear Coordinate Transformation http://cs.stanford.edu/people/karpathy/convnetjs/
More informationOptimization and Simulation
Optimization and Simulation Variance reduction Michel Bierlaire Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique Fédérale de Lausanne M.
More informationClick Prediction and Preference Ranking of RSS Feeds
Click Prediction and Preference Ranking of RSS Feeds 1 Introduction December 11, 2009 Steven Wu RSS (Really Simple Syndication) is a family of data formats used to publish frequently updated works. RSS
More informationCSC321 Lecture 4: Learning a Classifier
CSC321 Lecture 4: Learning a Classifier Roger Grosse Roger Grosse CSC321 Lecture 4: Learning a Classifier 1 / 31 Overview Last time: binary classification, perceptron algorithm Limitations of the perceptron
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationECE 5424: Introduction to Machine Learning
ECE 5424: Introduction to Machine Learning Topics: Ensemble Methods: Bagging, Boosting PAC Learning Readings: Murphy 16.4;; Hastie 16 Stefan Lee Virginia Tech Fighting the bias-variance tradeoff Simple
More informationHoldout and Cross-Validation Methods Overfitting Avoidance
Holdout and Cross-Validation Methods Overfitting Avoidance Decision Trees Reduce error pruning Cost-complexity pruning Neural Networks Early stopping Adjusting Regularizers via Cross-Validation Nearest
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Ensembles Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique Fédérale de Lausanne
More informationarxiv:math/ v1 [math.st] 20 Oct 2005
Technical Report No. 0510, Department of Statistics, University of Toronto arxiv:math/0510449v1 [math.st] 20 Oct 2005 Improving Classification When a Class Hierarchy is Available Using a Hierarchy-Based
More informationConditional distributions
Conditional distributions Will Monroe July 6, 017 with materials by Mehran Sahami and Chris Piech Independence of discrete random variables Two random variables are independent if knowing the value of
More informationSupporting Information
Supporting Information Convolutional Embedding of Attributed Molecular Graphs for Physical Property Prediction Connor W. Coley a, Regina Barzilay b, William H. Green a, Tommi S. Jaakkola b, Klavs F. Jensen
More informationECE521 Lectures 9 Fully Connected Neural Networks
ECE521 Lectures 9 Fully Connected Neural Networks Outline Multi-class classification Learning multi-layer neural networks 2 Measuring distance in probability space We learnt that the squared L2 distance
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationCOMP 551 Applied Machine Learning Lecture 21: Bayesian optimisation
COMP 55 Applied Machine Learning Lecture 2: Bayesian optimisation Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp55 Unless otherwise noted, all material posted
More informationCSC321 Lecture 4: Learning a Classifier
CSC321 Lecture 4: Learning a Classifier Roger Grosse Roger Grosse CSC321 Lecture 4: Learning a Classifier 1 / 28 Overview Last time: binary classification, perceptron algorithm Limitations of the perceptron
More informationCSC321 Lecture 5: Multilayer Perceptrons
CSC321 Lecture 5: Multilayer Perceptrons Roger Grosse Roger Grosse CSC321 Lecture 5: Multilayer Perceptrons 1 / 21 Overview Recall the simple neuron-like unit: y output output bias i'th weight w 1 w2 w3
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationA Lagrangian relaxation method for solving choice-based mixed linear optimization models that integrate supply and demand interactions
A Lagrangian relaxation method for solving choice-based mixed linear optimization models that integrate supply and demand interactions Meritxell Pacheco Shadi Sharif Azadeh, Michel Bierlaire, Bernard Gendron
More informationCOMS 4771 Regression. Nakul Verma
COMS 4771 Regression Nakul Verma Last time Support Vector Machines Maximum Margin formulation Constrained Optimization Lagrange Duality Theory Convex Optimization SVM dual and Interpretation How get the
More informationPredicting New Search-Query Cluster Volume
Predicting New Search-Query Cluster Volume Jacob Sisk, Cory Barr December 14, 2007 1 Problem Statement Search engines allow people to find information important to them, and search engine companies derive
More informationApprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning
Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning Nicolas Thome Prenom.Nom@cnam.fr http://cedric.cnam.fr/vertigo/cours/ml2/ Département Informatique Conservatoire
More informationPAC-learning, VC Dimension and Margin-based Bounds
More details: General: http://www.learning-with-kernels.org/ Example of more complex bounds: http://www.research.ibm.com/people/t/tzhang/papers/jmlr02_cover.ps.gz PAC-learning, VC Dimension and Margin-based
More informationCSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18
CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$
More informationLogistic Regression Introduction to Machine Learning. Matt Gormley Lecture 9 Sep. 26, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Logistic Regression Matt Gormley Lecture 9 Sep. 26, 2018 1 Reminders Homework 3:
More informationThe estimation of multivariate extreme value models from choice-based samples
The estimation of multivariate extreme value models from choice-based samples M. Bierlaire, D. Bolduc and D. McFadden Transport and Mobility Laboratory, EPFL Département d économique, Université Laval,
More informationDoes the Wake-sleep Algorithm Produce Good Density Estimators?
Does the Wake-sleep Algorithm Produce Good Density Estimators? Brendan J. Frey, Geoffrey E. Hinton Peter Dayan Department of Computer Science Department of Brain and Cognitive Sciences University of Toronto
More informationEXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING
EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING DATE AND TIME: June 9, 2018, 09.00 14.00 RESPONSIBLE TEACHER: Andreas Svensson NUMBER OF PROBLEMS: 5 AIDING MATERIAL: Calculator, mathematical
More informationNeural Network Training
Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification
More informationThe estimation of Generalized Extreme Value models from choice-based samples
The estimation of Generalized Extreme Value models from choice-based samples M. Bierlaire D. Bolduc D. McFadden August 10, 2006 Abstract We consider an estimation procedure for discrete choice models in
More informationStatistical Machine Learning
Statistical Machine Learning Lecture 9 Numerical optimization and deep learning Niklas Wahlström Division of Systems and Control Department of Information Technology Uppsala University niklas.wahlstrom@it.uu.se
More informationGradient-Based Learning. Sargur N. Srihari
Gradient-Based Learning Sargur N. srihari@cedar.buffalo.edu 1 Topics Overview 1. Example: Learning XOR 2. Gradient-Based Learning 3. Hidden Units 4. Architecture Design 5. Backpropagation and Other Differentiation
More informationCOMS 4771 Introduction to Machine Learning. James McInerney Adapted from slides by Nakul Verma
COMS 4771 Introduction to Machine Learning James McInerney Adapted from slides by Nakul Verma Announcements HW1: Please submit as a group Watch out for zero variance features (Q5) HW2 will be released
More informationDestination Choice Model including panel data using WiFi localization in a pedestrian facility
Destination Choice Model including panel data using WiFi localization in a pedestrian facility Loïc Tinguely, Antonin Danalet, Matthieu de Lapparent & Michel Bierlaire Transport and Mobility Laboratory
More informationDark knowledge. Geoffrey Hinton, Oriol Vinyals & Jeff Dean) Google
Dark knowledge Geoffrey Hinton, Oriol Vinyals & Jeff Dean) Google An analogy butterfly gold small production model caterpillar paydirt big ensemble of learned models leaves planet training data The conflicting
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)
More informationHOMEWORK #4: LOGISTIC REGRESSION
HOMEWORK #4: LOGISTIC REGRESSION Probabilistic Learning: Theory and Algorithms CS 274A, Winter 2018 Due: Friday, February 23rd, 2018, 11:55 PM Submit code and report via EEE Dropbox You should submit a
More informationCOMS 4771 Introduction to Machine Learning. Nakul Verma
COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative
More informationStatistical Machine Learning from Data
January 17, 2006 Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Multi-Layer Perceptrons Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole
More informationLecture 4: Training a Classifier
Lecture 4: Training a Classifier Roger Grosse 1 Introduction Now that we ve defined what binary classification is, let s actually train a classifier. We ll approach this problem in much the same way as
More informationInterpreting Deep Classifiers
Ruprecht-Karls-University Heidelberg Faculty of Mathematics and Computer Science Seminar: Explainable Machine Learning Interpreting Deep Classifiers by Visual Distillation of Dark Knowledge Author: Daniela
More informationMulticlass Logistic Regression
Multiclass Logistic Regression Sargur. Srihari University at Buffalo, State University of ew York USA Machine Learning Srihari Topics in Linear Classification using Probabilistic Discriminative Models
More informationComparison of Log-Linear Models and Weighted Dissimilarity Measures
Comparison of Log-Linear Models and Weighted Dissimilarity Measures Daniel Keysers 1, Roberto Paredes 2, Enrique Vidal 2, and Hermann Ney 1 1 Lehrstuhl für Informatik VI, Computer Science Department RWTH
More informationCS 6375 Machine Learning
CS 6375 Machine Learning Nicholas Ruozzi University of Texas at Dallas Slides adapted from David Sontag and Vibhav Gogate Course Info. Instructor: Nicholas Ruozzi Office: ECSS 3.409 Office hours: Tues.
More informationPAC-learning, VC Dimension and Margin-based Bounds
More details: General: http://www.learning-with-kernels.org/ Example of more complex bounds: http://www.research.ibm.com/people/t/tzhang/papers/jmlr02_cover.ps.gz PAC-learning, VC Dimension and Margin-based
More informationWhat Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1
What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single
More informationLinear and Logistic Regression. Dr. Xiaowei Huang
Linear and Logistic Regression Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Two Classical Machine Learning Algorithms Decision tree learning K-nearest neighbor Model Evaluation Metrics
More informationDeep Learning for Natural Language Processing. Sidharth Mudgal April 4, 2017
Deep Learning for Natural Language Processing Sidharth Mudgal April 4, 2017 Table of contents 1. Intro 2. Word Vectors 3. Word2Vec 4. Char Level Word Embeddings 5. Application: Entity Matching 6. Conclusion
More informationFrom perceptrons to word embeddings. Simon Šuster University of Groningen
From perceptrons to word embeddings Simon Šuster University of Groningen Outline A basic computational unit Weighting some input to produce an output: classification Perceptron Classify tweets Written
More informationHomework 1 Solutions Probability, Maximum Likelihood Estimation (MLE), Bayes Rule, knn
Homework 1 Solutions Probability, Maximum Likelihood Estimation (MLE), Bayes Rule, knn CMU 10-701: Machine Learning (Fall 2016) https://piazza.com/class/is95mzbrvpn63d OUT: September 13th DUE: September
More informationHow do we compare the relative performance among competing models?
How do we compare the relative performance among competing models? 1 Comparing Data Mining Methods Frequent problem: we want to know which of the two learning techniques is better How to reliably say Model
More informationIntroduction to Machine Learning
Introduction to Machine Learning 3. Instance Based Learning Alex Smola Carnegie Mellon University http://alex.smola.org/teaching/cmu2013-10-701 10-701 Outline Parzen Windows Kernels, algorithm Model selection
More informationKernel Logistic Regression and the Import Vector Machine
Kernel Logistic Regression and the Import Vector Machine Ji Zhu and Trevor Hastie Journal of Computational and Graphical Statistics, 2005 Presented by Mingtao Ding Duke University December 8, 2011 Mingtao
More informationLecture 4: Training a Classifier
Lecture 4: Training a Classifier Roger Grosse 1 Introduction Now that we ve defined what binary classification is, let s actually train a classifier. We ll approach this problem in much the same way as
More informationIntroduction to mixtures in discrete choice models
Introduction to mixtures in discrete choice models p. 1/42 Introduction to mixtures in discrete choice models Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Introduction to
More informationDeep Feedforward Networks. Sargur N. Srihari
Deep Feedforward Networks Sargur N. srihari@cedar.buffalo.edu 1 Topics Overview 1. Example: Learning XOR 2. Gradient-Based Learning 3. Hidden Units 4. Architecture Design 5. Backpropagation and Other Differentiation
More informationLearning from the Wisdom of Crowds by Minimax Entropy. Denny Zhou, John Platt, Sumit Basu and Yi Mao Microsoft Research, Redmond, WA
Learning from the Wisdom of Crowds by Minimax Entropy Denny Zhou, John Platt, Sumit Basu and Yi Mao Microsoft Research, Redmond, WA Outline 1. Introduction 2. Minimax entropy principle 3. Future work and
More informationReal Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report
Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Hujia Yu, Jiafu Wu [hujiay, jiafuwu]@stanford.edu 1. Introduction Housing prices are an important
More informationCalculating indicators with PythonBiogeme
Calculating indicators with PythonBiogeme Michel Bierlaire May 17, 2017 Report TRANSP-OR 170517 Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique
More informationHomework 2 Solutions Kernel SVM and Perceptron
Homework 2 Solutions Kernel SVM and Perceptron CMU 1-71: Machine Learning (Fall 21) https://piazza.com/cmu/fall21/17115781/home OUT: Sept 25, 21 DUE: Oct 8, 11:59 PM Problem 1: SVM decision boundaries
More informationTHE POTENTIAL OF APPLYING MACHINE LEARNING FOR PREDICTING CUT-IN BEHAVIOUR OF SURROUNDING TRAFFIC FOR TRUCK-PLATOONING SAFETY
THE POTENTIAL OF APPLYING MACHINE LEARNING FOR PREDICTING CUT-IN BEHAVIOUR OF SURROUNDING TRAFFIC FOR TRUCK-PLATOONING SAFETY Irene Cara Jan-Pieter Paardekooper TNO Helmond The Netherlands Paper Number
More informationThe Origin of Deep Learning. Lili Mou Jan, 2015
The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets
More informationChoice Theory. Matthieu de Lapparent
Choice Theory Matthieu de Lapparent matthieu.delapparent@epfl.ch Transport and Mobility Laboratory, School of Architecture, Civil and Environmental Engineering, Ecole Polytechnique Fédérale de Lausanne
More informationThe Network GEV model
The Network GEV model Michel Bierlaire January, 2002 Discrete choice models play a major role in transportation demand analysis. Their nice and strong theoretical properties, and their flexibility to capture
More informationQuasi-Random Simulation of Discrete Choice Models
Quasi-Random Simulation of Discrete Choice Models by Zsolt Sándor and Kenneth Train Erasmus University Rotterdam and University of California, Berkeley June 12, 2002 Abstract We describe the properties
More informationMulticlass Classification-1
CS 446 Machine Learning Fall 2016 Oct 27, 2016 Multiclass Classification Professor: Dan Roth Scribe: C. Cheng Overview Binary to multiclass Multiclass SVM Constraint classification 1 Introduction Multiclass
More informationLecture 10. Announcement. Mixture Models II. Topics of This Lecture. This Lecture: Advanced Machine Learning. Recap: GMMs as Latent Variable Models
Advanced Machine Learning Lecture 10 Mixture Models II 30.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ Announcement Exercise sheet 2 online Sampling Rejection Sampling Importance
More informationIntroduction to Machine Learning. Introduction to ML - TAU 2016/7 1
Introduction to Machine Learning Introduction to ML - TAU 2016/7 1 Course Administration Lecturers: Amir Globerson (gamir@post.tau.ac.il) Yishay Mansour (Mansour@tau.ac.il) Teaching Assistance: Regev Schweiger
More informationLearning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text
Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text Yi Zhang Machine Learning Department Carnegie Mellon University yizhang1@cs.cmu.edu Jeff Schneider The Robotics Institute
More informationLatent Dirichlet Allocation Introduction/Overview
Latent Dirichlet Allocation Introduction/Overview David Meyer 03.10.2016 David Meyer http://www.1-4-5.net/~dmm/ml/lda_intro.pdf 03.10.2016 Agenda What is Topic Modeling? Parametric vs. Non-Parametric Models
More informationLecture 2. Judging the Performance of Classifiers. Nitin R. Patel
Lecture 2 Judging the Performance of Classifiers Nitin R. Patel 1 In this note we will examine the question of how to udge the usefulness of a classifier and how to compare different classifiers. Not only
More informationIntroduction to Machine Learning Midterm Exam Solutions
10-701 Introduction to Machine Learning Midterm Exam Solutions Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes,
More informationBackpropagation Introduction to Machine Learning. Matt Gormley Lecture 12 Feb 23, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Backpropagation Matt Gormley Lecture 12 Feb 23, 2018 1 Neural Networks Outline
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationStable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems
Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems Thore Graepel and Nicol N. Schraudolph Institute of Computational Science ETH Zürich, Switzerland {graepel,schraudo}@inf.ethz.ch
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationRoute choice models: Introduction and recent developments
Route choice models: Introduction and recent developments Michel Bierlaire transp-or.epfl.ch Transport and Mobility Laboratory, EPFL Route choice models: Introduction and recent developments p.1/40 Route
More informationVBM683 Machine Learning
VBM683 Machine Learning Pinar Duygulu Slides are adapted from Dhruv Batra Bias is the algorithm's tendency to consistently learn the wrong thing by not taking into account all the information in the data
More informationCOMPARING FIXED AND ADAPTIVE COMPUTATION TIME FOR RE-
Workshop track - ICLR COMPARING FIXED AND ADAPTIVE COMPUTATION TIME FOR RE- CURRENT NEURAL NETWORKS Daniel Fojo, Víctor Campos, Xavier Giró-i-Nieto Universitat Politècnica de Catalunya, Barcelona Supercomputing
More informationSegmentation of Cell Membrane and Nucleus using Branches with Different Roles in Deep Neural Network
Segmentation of Cell Membrane and Nucleus using Branches with Different Roles in Deep Neural Network Tomokazu Murata 1, Kazuhiro Hotta 1, Ayako Imanishi 2, Michiyuki Matsuda 2 and Kenta Terai 2 1 Meijo
More informationLogistic Regression & Neural Networks
Logistic Regression & Neural Networks CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Logistic Regression Perceptron & Probabilities What if we want a probability
More informationNeural networks and optimization
Neural networks and optimization Nicolas Le Roux Criteo 18/05/15 Nicolas Le Roux (Criteo) Neural networks and optimization 18/05/15 1 / 85 1 Introduction 2 Deep networks 3 Optimization 4 Convolutional
More informationCS534 Machine Learning - Spring Final Exam
CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the
More informationNeed for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels
Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)
More informationScuola di Calcolo Scientifico con MATLAB (SCSM) 2017 Palermo 31 Luglio - 4 Agosto 2017
Scuola di Calcolo Scientifico con MATLAB (SCSM) 2017 Palermo 31 Luglio - 4 Agosto 2017 www.u4learn.it Ing. Giuseppe La Tona Sommario Machine Learning definition Machine Learning Problems Artificial Neural
More informationNeural Networks 2. 2 Receptive fields and dealing with image inputs
CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There
More informationMachine Learning, Fall 2009: Midterm
10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all
More informationECE521 Lecture 7/8. Logistic Regression
ECE521 Lecture 7/8 Logistic Regression Outline Logistic regression (Continue) A single neuron Learning neural networks Multi-class classification 2 Logistic regression The output of a logistic regression
More informationINF 5860 Machine learning for image classification. Lecture 14: Reinforcement learning May 9, 2018
Machine learning for image classification Lecture 14: Reinforcement learning May 9, 2018 Page 3 Outline Motivation Introduction to reinforcement learning (RL) Value function based methods (Q-learning)
More informationKernel Density Topic Models: Visual Topics Without Visual Words
Kernel Density Topic Models: Visual Topics Without Visual Words Konstantinos Rematas K.U. Leuven ESAT-iMinds krematas@esat.kuleuven.be Mario Fritz Max Planck Institute for Informatics mfrtiz@mpi-inf.mpg.de
More informationOptimization and Simulation
Optimization and Simulation Simulating events: the Michel Bierlaire Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique Fédérale de Lausanne
More information