Gaussian Mixture Modelling for Grouping Traffic Patterns Using Automatic Traffic Recorder Data

Size: px
Start display at page:

Download "Gaussian Mixture Modelling for Grouping Traffic Patterns Using Automatic Traffic Recorder Data"

Transcription

1 0 0 Gaussian Mixture Modelling for Grouping Traffic Patterns Using Automatic Traffic Recorder Data Sunil K. Madanu Graduate Research Assistant Department of Civil Engineering, the University of Texas at Arlington Nedderman Hall B, Yates Street, Arlington, TX 0 sunil.madanu@mavs.uta.edu Stephen P. Mattingly, Corresponding Author Associate Professor Department of Civil Engineering, the University of Texas at Arlington Nedderman Hall, Yates Street, Arlington, TX 0 Tel: --; Fax: --0; mattingly@uta.edu Christina A. McDaniel-Wilson, P.E. Oregon Department of Transportation Planning Analysis Unit th St. NE, Ste, Salem, OR 0 Tel: Christina.A.MCDANIEL-WILSON@odot.state.or.us Word count:,00 words for text + tables/figures x 0 words (each) =,00 words

2 Madanu, Mattingly, and McDaniel-Wilson 0 0 ABSTRACT The grouping of similar traffic patterns plays an important role while estimating Annual Average Daily Traffic (AADT) values from Short Term Traffic Counts (STTCs). Incorrect grouping often becomes a significant source of AADT estimation errors. For instance, grouping a commuter traffic trend pattern into a recreational traffic trend may produce an erroneous AADT value. The traditional knowledge based methods, often aided with visual interpretation, introduce subjective bias while grouping traffic patterns. In addition, the grouping requires additional time to process large amounts of data, and remains inefficient with unapparent traffic patterns. Under limited resources and constraints, better methods and techniques may group sites with similar characteristics. The paper presents a refined Gaussian Mixture Modelling (GMM) for clustering continuous count data. The number of mixture components identified using Bayesian Information Criteria (BIC) from mixture modelling leads to overestimation if the number of components is interpreted as the number of clusters. Hence, the authors adopt a hierarchical combination of the initial mixture components using an entropy criterion to identify the number of clusters. The researchers compare this modified clustering solution with model based agglomerative hierarchical clustering, volume factor grouping and a traditional approach. The study uses Automatic Traffic Recorder (ATR) data from the Oregon Department of Transportation (ODOT) as a comparative case study. Overall, the GMM clustering solution shows a lower average monthly and average daily Mean Absolute Percent Error (MAPE) than other methods; however, for summer months, the ODOT method shows less MAPE than the GMM solution. Keywords: Clustering Analysis, Gaussian Mixture Model, AADT, Grouping, Model-based Clustering, Short Term Traffic Count, Seasonal Traffic Trend Grouping, ATR, Automatic Traffic Recorder

3 Madanu, Mattingly, and McDaniel-Wilson 0 0 INTRODUCTION The state Departments of Transportation (DOTs) allocate significant resources on collecting historic traffic data throughout the state-wide networks. The estimation of Annual Average Daily Traffic (AADT) has many practical applications in planning, operations, maintenance, and decision making. The literature proposes a wide range of techniques like regression analysis, geographically weighted regression, artificial neural networks, time-series analysis, genetic algorithms, and kriging-based methods (see references () and () for review). According to the Federal Highway Administration (FHWA) Traffic Monitoring Guide (TMG), monitoring traffic volume trends represents a key task for continuous traffic count program (). The deployment of Automatic Traffic Recorders (ATR) on a state-wide network helps state DOTs to collect and monitor traffic patterns. State agencies face a tough decision on how many ATRs to deploy, where to deploy, and how frequently to collect the data with limited resources. In addition, decisions on the type of traffic patterns to monitor, either by vehicle type, monthly, day of week, or hourly distribution complicates the monitoring process. The FHWA and state agencies have developed some guidelines for the Traffic Monitoring Analysis System (TMAS) to evaluate volume trends over a specified time period. Moreover, monitoring the AADT and its trends requires continuous data from only at a limited number of locations. In order to cover a specific location of interest, Short Term Traffic Counts (STTC) are taken and seasonally adjusted using factor groups (). Factor groups have reasonably homogeneous patterns, usually but not necessarily, calculated based on monthly traffic patterns. In addition, the TMG also suggests day of week, hourly patterns, and patterns by vehicle type (passenger or trucks) or geographic region. Many state DOTs group traffic patterns from ATR sites and calculate AADT values using seasonal factors (). In fact, the TMG outlines three types of analysis for grouping: Traditional Approach, Cluster Analysis, and Volume Factor Grouping. The traditional approach uses general knowledge of the road system with visual interpretation to identify groups. Cluster analysis is a procedure to group the patterns, often, using monthly factors (ratio of AADT to MADT-Monthly Average Daily Traffic) at continuous count stations. Volume factor grouping maintains separate volume factor groups by highway functional category. Finding groups through knowledge and functional class seems neither practical nor likely to produce better results because of large amount of continuous count data, and dynamic changes in travel activity patterns (irrespective of highway functional class). In addition, bias due to subjectivity, difficulty in analyzing these large datasets, and significant time resource requirements elevate the problems of the conventional methods (). A few alternative methods, generally labeled as clustering techniques, have evolved for the automatic grouping of traffic patterns. The paper aims to provide a modified clustering process using Gaussian Mixture Modelling (GMM). The study adopts an entropy difference criterion to systematically select clustering solutions using a fitted piecewise linear regression model. A hierarchical cluster combination method is introduced to select the number of clusters. In addition, the GMM solution is compared with model based agglomerative hierarchical clustering, traditional approach and volume factor grouping to assess its relative performance. After the background information, the paper outlines Gaussian mixture modelling and the cluster selection process. The following section describes the classification of test data according to the

4 Madanu, Mattingly, and McDaniel-Wilson 0 0 cluster solutions. Finally, the paper presents a comparative case study of clustering solutions for estimating AADT values. BACKGROUND Clustering techniques try to determine the structure of data when no information is available except observational data. Cluster analysis partitions the data into meaningful subgroups without knowing its components and structure. Cluster analysis is broadly divided into heuristic methods and statistical models, which follow either hierarchical or relocation strategies (). Hierarchical methods produce clusters in stages either using an agglomerative (merging) or divisive (splitting) method at each stage. The agglomerative hierarchical clustering treats each observation as a cluster by itself at the beginning and merges observations at each stage using a variety of methods, like single link (nearest neighbor), complete link (farthest neighbor), and Ward s distance. Model based hierarchical agglomerative clustering, on other hand, uses maximum-likelihood criteria for merging observations into groups (). A drawback of agglomerative methods is that they do not address the issue of determining the number of clusters as a result the criteria for merging and selecting clusters will produce different classifications from agglomerative clustering (). Relocation based methods assign the observations iteratively among the groups. The number of clusters or groups has to be specified in advance, and they do not change during the course of iterations. However, at each iteration, observations move from one group to other groups, usually, using some form of distance criteria. K-means clustering and finite mixture modelling represent two examples of relocation based methods. K-means clustering selects the number of clusters that minimize within-group variance (). Mixture modelling involves probability based cluster analysis, where observational data is assumed to come from a mixture of probability distributions. The Bayesian Information Criterion (BIC) is used to determine the number of components and model form for clustering. Often, mixture components are modelled to follow a Gaussian distribution where its maximum likelihood parameters are found using an Expected- Maximization (EM) algorithm (). Past research has categorized ATR data using a variety of techniques: agglomerative hierarchical grouping (,), k-means clustering (,), model-based clustering (), fuzzy C-means method (-), regression models (), Bayesian statistics (), mixture of regression models (), neural networks (), genetic algorithms (), quantum-frequency algorithm for automated identification of traffic patterns (), fuzzy logic (), Support Vector Machines (SVM) (), and wavelets (). These methods present a variety of drawbacks, which include inconsistent clustering results for different locations and time periods, clustering solution sensitivity to missing data, inability to provide sematic meaning (e.g. summer, commuter, recreational, etc...), theoretical nature of methods, and ill-suitability for wider implementation (). The groups formed from cluster analysis, unlike groupings based on expert judgment, avoid biasness because they are chosen by their data-driven similarity measure (). However, cluster analysis lacks guidelines on establishing the optimal number of clusters for a given data set. Priori information on the number of clusters, cluster initializations, and cluster evaluation criteria play an important role in cluster analysis outputs. Unfortunately, no standard procedure exists for selecting a priori

5 Madanu, Mattingly, and McDaniel-Wilson 0 information. Cluster analysis output groups, often, appear relatively unidentifiable on a given state-wide network due to their pure mathematical nature (). Bad clustering or grouping is the chief source of error while estimating AADT using a seasonal factoring approach (). In addition, day to day traffic variation, missing or bad data due to ATR malfunctions, and incorrect assignment of STTCs to seasonal factor groups make significant contributions to AADT estimation errors (). Cluster analysis may compute a different number of clusters across years and ATR sites may change cluster year by year (). These drawbacks highlight the challenges of cluster analysis and the difficulty of identifying clusters for practical situations. Despite the previous criticism, clustering analysis with evolution of machine learning algorithms provides an opportunity to improve the clustering analysis outputs. The objective of this paper is to provide a modified clustering process using Gaussian Mixture Modelling. The study introduces a hierarchical cluster combination method to systematically select a clustering solution using an entropy difference criterion. GAUSSIAN MIXTURE MODELLING In mixture modelling, a distribution f is a mixture of K component distributions like f, f f K if f(x) = K k= P k f k (x) () Where P k is called mixing weight or mixing proportion with P k > 0 and k P k =. The component distributions can be completely arbitrary and may come with different distributions. Most cases select parametric mixture models where component distributions come from the same family (all Gaussian, all Gamma, etc ) (). In general, Gaussian mixture models use different parameters (mean, variances, and mixing weights). Let x = {x, x x n } in R n d is a d-dimensional data patterns with a sample size of n. The likelihood or probability density f (x i K,θ K ) of a data point x i belongs to a sample from a finite K mixture density equal to k= P k f k (x i a k ), where θ K are the parameters (P P K-, a a K- ). The component densities are Gaussian densities with parameter a k = (µ k, Σ k ), µ k being the mean and Σ k the variance matrix of component k. The group or class, z = (z z n ), to which a data point belongs is unknown or to be determined. The group of a data point x i is expressed as z i = (z i z ik ) where z ik takes a value of one if x i belong to group k. n Assuming x to be independent samples, then the likelihood is expressed as f(x i K, θ K ) and Log-likelihood is n n K l(k, θ K ) = i= log f(x i K, θ K ) = log( P k f k (x i a k ) i=, i= k= ) () The parameters of mixture models are found using an Expected-Maximization (EM) algorithm (). Data generated from mixture models are characterized by either groups, clusters or classes. Cluster geometrical features (shape, volume, and orientation) are determined by the covariance matrices (Σ k ). The Eigenvalue decomposition of Σ k provides a general framework for clustering analysis, which is often called model based clustering (). The covariance matrix ( Σ k ) is decomposed into: Σ k = λ k D k A k D k T ()

6 Madanu, Mattingly, and McDaniel-Wilson 0 0 Where D k is an orthogonal matrix of Eigenvectors (gives cluster orientation), A k is a diagonal matrix whose elements are proportional to the eigenvalues (gives cluster shape), and λ k governs the volume of the cluster. Restricting the cluster orientation, shape and volume yields different models for clustering (). These models are broadly categorized into three main model families: spherical, diagonal, and general. A three-letter code describes the volume, shape, and orientation of the cluster groups. Each letter in the code may take a value of E for equal, V for variable, or I for identity (). For instance, model VEV stands for a cluster model with variable volume, equal shape and variable orientation. Selection of mixture models is usually done using Bayesian Information Criteria (BIC). In model-based clustering, a decisive first local maximum of BIC indicates strong evidence for a good model (). Usually, the number of mixture components is taken as the number of clusters (). However, this can lead to an overestimation, because BIC selects the number of mixture components only to provide a good approximation to the density rather than the number of clusters (). Also, a mixture of Gaussian components may better represent a cluster than a single Gaussian component. Thus, a best approximating Gaussian mixture model may not represent the correct number of clusters in the data (). Biernacki et al. () proposed the Integrated Complete Likelihood (ICL) criteria to assess the number of clusters. Often, the number of clusters identified by ICL criteria is smaller than the number of components selected by BIC because ICL is approximately equal to BIC with a penalty term of mean entropy (). The ICL usage for estimating mixture components may give an underestimated number of components (). In order solve the dilemma of cluster selection, Baudry et al. () proposed a method to achieve best clustering using entropy criterion. The mixture components produced by a BIC solution are hierarchically combined using entropy criterion. The best number of clusters are selected based on a piecewise linear regression fitted to the rescaled entropy plot. The entropy of K mixture components is defined as (): K n k= i= () Ent(K) = t ik (θ K ) log t ik (θ K ) 0 Where t ik is the conditional probability of x i from the kth mixture component, which is given by t ik (θ K ) = p k f k (x i a k ) K j= p j f j (x i a j ) STUDY METHODOLOGY The study methodology has three stages. In the first stage, the GMM clustering technique is applied to Oregon Department of Transportation (ODOT) ATR data. The initial clustering solution provides the best model form and number of components using BIC criteria. The initial clustering solution (based on BIC criteria) is hierarchically combined until obtaining a single cluster. At any step, during the combination process, only two components are added based on the minimization of entropy criterion. Once the clustering solution is found, each data point is assigned to a group number or class. The next stage designs a data classifier using the group labels assigned during the first stage. In the final stage, the designed classifier helps in assigning a group number to the test data. The relative merits of clustering solutions are also assessed by comparing estimated and actual AADT values. In particular, the GMM solution is compared ()

7 Madanu, Mattingly, and McDaniel-Wilson 0 with model based agglomerative hierarchical clustering, and ODOT s seasonal trend grouping method and functional class grouping method. CASE STUDY Data The data sets are developed from ATR counters throughout the State of Oregon for the years and. For the purpose of analysis, a Short Term Traffic Count (STTC) is defined as a complete hour count on a given day. At each ATR station, STTCs are sampled and used for clustering analysis. Table provides a summary of the ATR data characteristics. Regions, and share more ATRs and also have higher travel activity. The researchers try to generalize the grouping methodology by considering all possible variations in the traffic trends. Thus, the analysis does not subset the data by vehicle type, weekdays, weekends, and seasonal patterns when establishing groupings. The data is used only for clustering and design of a classifier. The clustering solution and its relative performance is tested with the data. The next section presents the clustering analysis outputs and evaluates the solution. Year Table Study Data Characteristics ATRs per Region (#) Mean ATR AADT per Region (veh/year) Total ATRs Annual ADT Number of STTCs per year 0,,,,,, 0,,0,0,,,,, First Stage - Clustering The clustering analysis uses the data with a sample size of 0,. Each sample has continuous hourly traffic volume (veh/hour), daily traffic volume (DT in veh/day), and the ratio of AADT to DT. At the beginning of the first stage, GMM clustering is fit to the data and refined to obtain a final cluster solution. The researchers, then, present model based Gaussian agglomerative hierarchical clustering analysis, and other clustering solutions for comparative analysis. The study uses the mclust package available in the R programming language to analyze clustering (). The following sections present these steps. GMM - Initial Clustering Solution Sample hourly traffic data (with dimensionality d = ) represents the input data for the GMM clustering. The model based clustering chooses mixture components using both BIC and ICL criteria. An unconstrained model VVV (variable shape, volume and orientation of covariance matrix) is chosen in both cases. The mix proportions of the fifteen components are not uniform; five components have a proportion between and percent, three components between and percent, four components between and percent, and the proportion varies for the other three components. GMM Solution Refinement Starting with the initial BIC solution (having components), hierarchically, components are added until obtaining a single cluster (clusters,, to ). At each step, two mixture

8 Madanu, Mattingly, and McDaniel-Wilson 0 0 components are combined based on the minimization of entropy criterion calculated at each step. The same steps are repeated for the ICL criteria solution. As the model form and number of mixture components are the same for the BIC and ICL criteria, a hierarchical combination of the initial solution is done only once. The hierarchical combination does not change the likelihood and distribution of the observations from the initial cluster solution. Hence, no formal statistical inference is available to choose between different numbers of clusters produced at every consecutive combination step. However, the decrease in entropy at each step guides the selection of a cluster solution. Figure (a) shows the entropies of the combined solutions, and Figure (b) shows the difference in entropy. Reducing the clusters from to appears beneficial because the decrease in entropy is large. Similarly, reducing from to clusters provides a large decrease in entropy. Each of the five subsequent reduction steps (from down to clusters) provide an entropy reduction roughly half as large as the initial change from to. The next cluster reduction steps (from down to ) each decrease the entropy a little less than one third of the first change from to. The change in entropy alone does not provide sufficient guidance for the number of clusters. To obtain more insight, Figure (c) shows the decrease in entropy with respect to the number of observations merged at each step. On the abscissas, the difference between two successive points shows the number of observations used for merging at a given step in the hierarchical combination process. Using the difference in observations and its corresponding entropy difference, a normalized difference in entropy is computed (Figure (d)). The ratio of entropy difference (from K+ clusters to K clusters) to number of merged observations gives a normalized entropy difference. At the first merging, a large entropy difference over a small number of observations leads to a large normalized difference in entropy. Byers and Raftery () proposed fitting a piecewise linear regression model to the values in the entropy plot, and selecting the number of clusters corresponding to an estimated breakpoint (elbow) in the plot. The regression line is shown as the dotted lines for both the entropy and normalized entropy plots in Figure (a) and Figure (c). The fit of piecewise linear regression plot shows one elbow and chooses eight clusters. Model based Gaussian agglomerative Hierarchical Clustering Analysis (MHCA) The GMM clustering solution is compared with the Model based Gaussian agglomerative Hierarchical Clustering Analysis (MHCA). Since the MHCA is more efficient than conventional agglomerative hierarchical clustering and model based clustering shows improved estimates over K-means and other types of clustering while estimating AADT (), the study focuses on a comparison of the GMM and MHCA methods. The MHCA uses maximum-likelihood criteria for merging observations into groups (). Even, the MHCA requires a prioi information on the number of clusters. Unfortunately, such information is often not available before hand; instead the algorithm is tested for a number of clusters varying from to. The best number of groups is selected using one of commonly used cluster evaluation criteria. Calinski-Harabasz index (0), Davies-Bouldin index (), Dunn index () and Silhouette values () are some of the frequently used clustering evaluation criteria. The study selects six cluster solutions for MHCA as the optimum number of clusters using the Dunn Index.

9 Madanu, Mattingly, and McDaniel-Wilson (a) (b) (c) (d) FIGURE Entropy and normalized entropy plots for case study data. Other Clustering In addition, the clustering solutions are compared with the ODOT seasonal trend grouping method (TMG labeled traditional approach) and Functional Class (FC) grouping method (TMG labeled volume factor grouping method). The ODOT seasonal trend grouping method uses knowledge of the road system with aid of visual interpretation. The ODOT, historically, identifies ten seasonal trends: Summer < 00 AADT, Recreational Summer, Interstate Non- Urbanized, Interstate Urbanized, Agricultural, Commuter, Recreational Summer /Winter, Coastal Destination, Summer, and Coastal Destination Route. Functional class grouping bundles the data patterns based on ODOT highway functional class: Rural Interstate, Urban Interstate,

10 Madanu, Mattingly, and McDaniel-Wilson 0 0 Freeways and Expressways, Principal Arterial, Principal Arterial Urban, Minor Arterial, and Major Collector. Next, the research team tests the relative merit of the cluster solutions for estimating the AADT. The study adopts the GMM eight cluster solution and MHCA six cluster solution for estimation. First, the cluster solution classifies each data pattern into one of the groups (either one of eight or one of six depending on the solution). However, the test data needs a classifier that assigns data patterns into one of the groups. In the past, researchers have developed numerous methods for assigning traffic data to factor groups (see reference ()). Both tasks, reasonably grouping the traffic data (current study focus) and assigning traffic data to correct groups (classification) play an important role when computing AADT values. Second Stage - Classification The study uses the commonly adopted Quadratic Discriminant Analysis (QDA) technique for classifying test data patterns. The QDA statistically assigns a given data pattern to the groups of known characteristics. The error rate or misclassification rate (percent of data patterns whose groups are misclassified) shows the predictability of the QDA. The data with known group labels from GMM clustering solution is divided into two parts, namely, training data and test data using 0 and 0 proportion rule. Next, the QDA is performed to classify the training data patterns. Trained QDA classifier has a -fold cross validation error rate of. percent for the GMM clustering solution. When tested for QDA predictability, a test error rate of. percent is observed. A separate QDA is performed for the MHCA solution using the same training and testing data. Final Stage - AADT Estimation After the clustering step, each cluster group is again sub grouped by month to compute the average ratio of AADT to DT. For instance, the GMM eight cluster solution produces average ratios of AADT to DT ( groups months) after sub-grouping. Subgrouping helps to address monthly and seasonal variation of traffic data. Depending upon the analysis type, subgrouping continues only on a part of the test data. For instance, while studying weekday patterns, subgrouping is performed only for the weekdays of the test data. In total, three subgrouping tables namely weekday, weekend, and daily or total are developed. The subgrouping of each cluster by month continues for GMM, MHCA, ODOT seasonal trend grouping (ODOT), and highway functional class (FC) grouping methods. The ATR data set with a sample of, data patterns is used for testing. The dataset has hourly traffic data showing time of day variation and the groupings according to the ODOT seasonal trend grouping method and highway functional class. The QDA classify the test data and assign the group number according to the GMM and MHCA solution. The AADT is calculated by matching group number and month of the test patterns (using data). The product of the matched average ratio of AADT to DT (corresponding to a matched group number and month) and sum of -hour traffic volume (daily traffic or DT) estimates the AADT. The computed AADT value is compared with the actual AADT value to obtain an error. The Mean Absolute Percent Error (MPAE) given in Equation () is used to compare the estimates from clustering methods (, ):

11 Madanu, Mattingly, and McDaniel-Wilson MAPE = N ( AADT Estimated AADT Actual n= 0) () AADT Actual n RESULTS First, the study examines MAPEs among the clustering solutions by day of week patterns (see Table ). The research team follows the ODOT definition of weekdays (Monday to Thursday) and weekends (Friday to Sunday). The daily pattern type refers to any day of week that combines both weekday and weekend patterns. The GMM clustering solution gives less percent error for the weekday pattern type compared to weekends. The percent error for daily patterns lies between the weekday and weekend types. Compared to the MHCA solution, the GMM solution yields a lower error rate for any pattern type. Grouping by functional class shows a higher error rate than any other method. The ODOT method shows almost two percent less error than the next best solution by the GMM method for the weekday pattern. The ODOT method considers seasonal trends based on a dedicated weekday and weekend traffic patterns. This allows the ODOT method to show less error than the GMM method, which is based on daily traffic patterns. For other patterns, the GMM and ODOT solutions appear similar, while the GMM holds a little advantage when comparing daily patterns. Table does not give a clear picture on whether the GMM solution performs better than the conventional ODOT method. The team develops Figure, which shows monthly variation of MAPEs for different clustering solutions. Table Mean Absolute Percent Error (MPAE) among Clustering Solutions Mean Absolute Percent Error (MPAE) in % Pattern ODOT Seasonal Trend Grouping Functional Class Grouping GMM MHCA (ODOT) (FC) Weekday....0 Weekend.... Daily Both GMM and ODOT methods have more MAPEs for winter months compared to summer months. Overall, the monthly variation of the MAPE is lower for the GMM method compared to the ODOT method except for summer months. The summer month trends typically represent recreational activities that are mostly based on weekend patterns. The daily traffic pattern based GMM solution does not capture the AADT variation compared to the ODOT method for those months. However, application of the monthly subgrouping from the GMM cluster solution seems to produce less AADT estimation error. The monthly error variation remains high for the MHCA and the functional class grouping methods.

12 MAPE (%) Madanu, Mattingly, and McDaniel-Wilson Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec GMM Month MHCA ODOT Seasonal Trend Grouping Functional Class Grouping FIGURE Monthly variation of mean absolute percent error among clustering methods. CONCLUDING REMARKS Summary and Conclusions Cluster analysis partitions the data into meaningful subgroups without knowing its components and structure. However, choosing the number of clusters or patterns in the traffic data represents a challenging task. The study introduces a statistical clustering method, Gaussian Mixture Modelling, for grouping the ATR traffic data patterns. The study methodology has three stages: clustering, classification, and AADT estimation. In the clustering stage, first, the GMM is fit to the data and the number of components is selected using BIC criteria. Matching the number of components, from the BIC criteria, with the number of clusters overestimates the number of necessary clusters. Hence, the research team introduces a strategy to refine the initial GMM cluster solution using entropy criterion minimization. Next, the authors design a quadratic discriminant classifier to classify the data with known groups from the GMM clustering solution. Finally, The GMM solution is tested for estimating AADT values using a data set. The GMM clustering solution shows lower MAPEs for most months compared to other clustering solutions; however, the current study has the following limitations and possible improvements. Practical Implication The study uses a classifier designed based on a GMM clustering solution to obtain a group number for the field-collected STTCs. The AADT is computed by taking a product of the corresponding group s ratio of AADT to DT and daily traffic of that test pattern (see AADT Estimation section). Usually, STTCs span multiple days across multiple months. The average of AADT values estimated for all those days will become the AADT for the given study facility.

13 Madanu, Mattingly, and McDaniel-Wilson 0 0 Study Limitations and Future Directions The GMM clustering solution is constructed using daily traffic patterns. The clustering solution may be enhanced by further dividing the data set into subgroups by weekdays, weekends, season, and vehicle type. However, the current methodological framework needs only modified input for performing clustering analysis on these data sets. As the traffic patterns change by geographical location on a state-wide transportation network, performing a clustering analysis within each region may provide consistent results for different locations. Similarly, clustering with historic data, instead of a single year of data, may capture the traffic variations more effectively. The entropy criterion guides the choice of the number of clusters, but it does not have a statistical interpretation. An inferential means of identifying the number of clusters, for instance using a gap statistic, could provide a robust choice of clustering. The consideration of land-use and socio-economic characteristics while grouping the traffic patterns may provide more significant groups, too. Like GMM, the mixture of either linear or non-linear models with AADT as a dependent variable and land-use and socio-economic characteristics as independent variables may help reduce the AADT estimation errors. ACKNOWLEDGMENTS The authors would like to thank Mr. Brian Dunn, Mr. Doug Norval, Mr. Peter Schuytema, and Mr. Sam Ayash of the Oregon Department of Transportation (ODOT) for their valuable advice and help. The contents of this paper reflect the views of the authors, who are responsible for the facts and the accuracy of the data, opinions, findings, and conclusions presented herein. The contents do not necessarily reflect the official view or policies of the ODOT, and the university that the authors represent. REFERENCES. Duddu, V. R., and S. S. Pulugurtha. Principle of Demographic Gravitation to Estimate Annual Average Daily Traffic: Comparison of Statistical and Neural Network Models. Journal of Transportation Engineering, Vol., No.,, pp... Gastaldi, M., G. Gecchele, and R. Rossi. Estimation of Annual Average Daily Traffic from One-week Traffic Counts - A Combined ANN-Fuzzy Approach. Transportation Research: Part C, vol.,, PP. -.. Traffic Monitoring Guide. Federal Highway Administration (FHWA). U.S. Department of Transportation,. Accessed July,.. Fraley, C., and A. E. Raftery. How Many Clusters? Answers via Model-Based Cluster Analysis. The Computer Journal, vol.,, pp... Sharma, S. C., and A. Werner. Improved Method of Grouping Province Wide Permanent Traffic Counters. In Transport Research Record: Journal of the Transportation Research Board, No., Transportation Research Board of the National Academies, Washington, D.C.,, pp... Faghri, A., and J. Hua. Roadway Seasonal Classification using Neural Networks. Journal of Computing in Civil Engineering, Vol., No.,, pp... Flaherty, J. Cluster Analysis of Arizona Automatic Traffic Record Data. In Transport Research Record: Journal of the Transportation Research Board, No., Transportation Research Board of the National Academies, Washington, D.C.,, pp..

14 Madanu, Mattingly, and McDaniel-Wilson 0 0. Tsapakis, I., W. H. Schneider IV, and A. P. Nichols. Improving the Estimation of Total and Direction-based Heavy-Duty Vehicle Annual Average Daily Traffic, Transportation Planning and Technology, Vol., Issue.,, pp. -.. Rossi, R., M. Gastaldi, and G. Gecchele. Comparison of Clustering Methods for Road Group Identification in FHWA Traffic Monitoring Approach: Effects on AADT Estimates. Journal of Transportation Engineering, Vol. 0, No.,, Gecchele, G., S. Kikuchi, R. Rossi, and M. Gastaldi. Advances in Uncertainty Treatment in FHWA Procedure for Estimating Annual Average Daily Traffic Volume. Presented at st Annual Meeting of the Transportation Research Board, Washington, D.C.,.. Rossi, R., M. Gastaldi, G. Gecchele, and S. Kikuchi. Estimation of Annual Average Daily Truck Traffic Volume. Uncertainty Treatment and Data Collection Requirements. Procedia Social and Behavioral Sciences, vol.,, pp... Gastaldi, M., R. Rossi, G. Gecchele, and L. Della Lucia. Annual Average Daily Traffic Estimation from Seasonal Traffic Counts. Procedia Social and Behavioral Sciences, vol.,, pp... Faghri, A., and P. Chakroborty. Development and Evaluation of a Statistically Reliable Traffic Counting Program, Transportation Planning and Technology, Vol., No.,, pp. -.. Davis, G. A. Accuracy of Estimates of Mean Daily Traffic: A Review. In Transport Research Record: Journal of the Transportation Research Board, No., Transportation Research Board of the National Academies, Washington, D.C.,, pp... Chen, P., K. Yin, and J. Sun. Application of Finite Mixture of Regression Model with Varying Mixing Probabilities to Estimation of Urban Arterial Travel Times. In Transportation Research Record: Journal of the Transportation Research Board, No., Transportation Research Board of the National Academies, Washington, D.C.,, pp... Lingras, P. Traffic Parameter Estimation and Highway Classification: Rough Patterns Using A Neural Networks Approach, Transportation Planning and Technology, Vol., No.,, pp. -.. Lingras, P. Statistical and Genetic Algorithms Classification of Highways, Journal of Transportation Engineering, Vol., No., 0, pp. -.. Venkatanarayana, R., B. L. Smith, and M. J. Demetsky. Quantum-Frequency Algorithm for Automated Identification of Traffic Patterns. In Transportation Research Record: Journal of the Transportation Research Board, No., Transportation Research Board of the National Academies, Washington, D.C., 0, pp... Kirshfink, H., and C. Chadenas. Traffic Situation Prediction Applying Pattern Matching and Fuzzy Classification. European Symposium on Intelligent Techniques, Orthodox Academy of Crete, Greece, June,.. Ren, J. T., X. L. Ou, Y. Zhang, and D. C. Hu. Research on Network-Level Traffic Pattern Recognition. IEEE th International Conference on Intelligent Transportation Systems, Singapore, Sept., 0.. Venkatanarayana, R., B. L. Smith, and M. J. Demetsky. Traffic Pattern Identification Using Wavelet Transforms. Presented at th Annual Meeting of the Transportation Research Board, Washington, D.C., 0.. Ritchie, S. A Statistical Approach to Statewide Traffic Counting. In Transportation Research Record: Journal of the Transportation Research Board, No. 0, Transportation Research Board of the National Academies, Washington, D.C.,, pp..

15 Madanu, Mattingly, and McDaniel-Wilson. Advanced Data Analysis from an Elementary Point of View. Cosma R. Shalizi. Accessed on July,.. Fraley, C., and A. E. Raftery. Model-based Clustering, Discriminant Analysis, and Density Estimation. Journal of the American Statistical Association, Vol., 0, pp. :.. Baudry, J. P., A. E. Raftery, G. Celeux, K. Lo, and R. Gottardo. Combining Mixture Components for Clustering. Journal of Computational and Graphical Statistics, Vol., No.,, pp. -.. Biernacki, C., G. Celeux, and G. Govaert. Assessing a Mixture Model for Clustering with the Integrated Completed Likelihood. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol., 00, pp... Fraley, C., A. E. Raftery, T. B. Murphy and L. Scrucca. mclust Version for R: Normal Mixture Modeling for Model-Based Clustering, Classification, and Density Estimation. Technical Report No., Department of Statistics, University of Washington,.. Byers, S. D., and A. E. Raftery. Nearest Neighbor Clutter Removal for Estimating Features in Spatial Point Processes. Journal of the American Statistical Association, Vol.,, pp... Fraley, C. Algorithms for Model-Based Gaussian Hierarchical Clustering. Technical Report No., Department of Statistics, University of Washington,. 0. Calinski, T., and J. Harabasz. A Dendrite Method for Cluster Analysis. Communications in Statistics. Vol., No.,, pp... Davies, D. L., and D. W. Bouldin. A Cluster Separation Measure. IEEE Transactions on Pattern Analysis and Machine Intelligence. Vol. PAMI-, No.,, pp... Dunn, J. C. A Fuzzy Relative of the ISODATA Process and Its Use in Detecting Compact Well- Separated Clusters. Journal of Cybernetics, Vol., Issue.,, pp... Rousseeuw, P. J. Silhouettes: a Graphical Aid to the Interpretation and Validation of Cluster Analysis. Computational and Applied Mathematics. Vol.,, pp..

Optimization of Short-Term Traffic Count Plan to Improve AADT Estimation Error

Optimization of Short-Term Traffic Count Plan to Improve AADT Estimation Error International Journal Of Engineering Research And Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 13, Issue 10 (October 2017), PP.71-79 Optimization of Short-Term Traffic Count Plan

More information

QUANTIFICATION OF THE NATURAL VARIATION IN TRAFFIC FLOW ON SELECTED NATIONAL ROADS IN SOUTH AFRICA

QUANTIFICATION OF THE NATURAL VARIATION IN TRAFFIC FLOW ON SELECTED NATIONAL ROADS IN SOUTH AFRICA QUANTIFICATION OF THE NATURAL VARIATION IN TRAFFIC FLOW ON SELECTED NATIONAL ROADS IN SOUTH AFRICA F DE JONGH and M BRUWER* AECOM, Waterside Place, Tygerwaterfront, Carl Cronje Drive, Cape Town, South

More information

Project Appraisal Guidelines

Project Appraisal Guidelines Project Appraisal Guidelines Unit 16.2 Expansion Factors for Short Period Traffic Counts August 2012 Project Appraisal Guidelines Unit 16.2 Expansion Factors for Short Period Traffic Counts Version Date

More information

Appendix BAL Baltimore, Maryland 2003 Annual Report on Freeway Mobility and Reliability

Appendix BAL Baltimore, Maryland 2003 Annual Report on Freeway Mobility and Reliability (http://mobility.tamu.edu/mmp) Office of Operations, Federal Highway Administration Appendix BAL Baltimore, Maryland 2003 Annual Report on Freeway Mobility and Reliability This report is a supplement to:

More information

Model-based cluster analysis: a Defence. Gilles Celeux Inria Futurs

Model-based cluster analysis: a Defence. Gilles Celeux Inria Futurs Model-based cluster analysis: a Defence Gilles Celeux Inria Futurs Model-based cluster analysis Model-based clustering (MBC) consists of assuming that the data come from a source with several subpopulations.

More information

Seasonal Factor Assignment Based on the Similarity of Hourly Traffic Patterns and Influential Variables

Seasonal Factor Assignment Based on the Similarity of Hourly Traffic Patterns and Influential Variables 868 JOURNAL OF SOFTWARE, VOL. 8, NO. 4, APRIL 2013 Seasonal Factor Assignment Based on the Similarity of Hourly Traffic Patterns and Influential Variables Chenxi Lu * and Renfa Yang Ningbo University of

More information

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Song Li 1, Peng Wang 1 and Lalit Goel 1 1 School of Electrical and Electronic Engineering Nanyang Technological University

More information

Cipra D. Revised Submittal 1

Cipra D. Revised Submittal 1 Cipra D. Revised Submittal 1 Enhancing MPO Travel Models with Statewide Model Inputs: An Application from Wisconsin David Cipra, PhD * Wisconsin Department of Transportation PO Box 7913 Madison, Wisconsin

More information

Development of Improved Models for Imputing Missing Traffic Counts

Development of Improved Models for Imputing Missing Traffic Counts The Open Transportation Journal, 2009, 3, 35-48 35 Open Access Development of Improved Models for Imputing Missing Traffic Counts Ming Zhong *,1 and Satish Sharma 2 1 Transportation Group, Department of

More information

Traffic Volume Time-Series Analysis According to the Type of Road Use

Traffic Volume Time-Series Analysis According to the Type of Road Use Computer-Aided Civil and Infrastructure Engineering 15 (2000) 365 373 Traffic Volume Time-Series Analysis According to the Type of Road Use Pawan Lingras* Department of Mathematics and Computing Science,

More information

Brief Introduction of Machine Learning Techniques for Content Analysis

Brief Introduction of Machine Learning Techniques for Content Analysis 1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview

More information

Responsive Traffic Management Through Short-Term Weather and Collision Prediction

Responsive Traffic Management Through Short-Term Weather and Collision Prediction Responsive Traffic Management Through Short-Term Weather and Collision Prediction Presenter: Stevanus A. Tjandra, Ph.D. City of Edmonton Office of Traffic Safety (OTS) Co-authors: Yongsheng Chen, Ph.D.,

More information

Short-term traffic volume prediction using neural networks

Short-term traffic volume prediction using neural networks Short-term traffic volume prediction using neural networks Nassim Sohaee Hiranya Garbha Kumar Aswin Sreeram Abstract There are many modeling techniques that can predict the behavior of complex systems,

More information

DEVELOPING DECISION SUPPORT TOOLS FOR THE IMPLEMENTATION OF BICYCLE AND PEDESTRIAN SAFETY STRATEGIES

DEVELOPING DECISION SUPPORT TOOLS FOR THE IMPLEMENTATION OF BICYCLE AND PEDESTRIAN SAFETY STRATEGIES DEVELOPING DECISION SUPPORT TOOLS FOR THE IMPLEMENTATION OF BICYCLE AND PEDESTRIAN SAFETY STRATEGIES Deo Chimba, PhD., P.E., PTOE Associate Professor Civil Engineering Department Tennessee State University

More information

The Assessment of Applying Chaos Theory for Daily Traffic Estimation

The Assessment of Applying Chaos Theory for Daily Traffic Estimation Proceedings of the 2014 International Conference on Industrial Engineering and Operations Management Bali, Indonesia, January 7 9, 2014 The Assessment of Applying Chaos Theory for Daily Traffic Estimation

More information

Freeway rear-end collision risk for Italian freeways. An extreme value theory approach

Freeway rear-end collision risk for Italian freeways. An extreme value theory approach XXII SIDT National Scientific Seminar Politecnico di Bari 14 15 SETTEMBRE 2017 Freeway rear-end collision risk for Italian freeways. An extreme value theory approach Gregorio Gecchele Federico Orsini University

More information

STATISTICAL ANALYSIS OF LAW ENFORCEMENT SURVEILLANCE IMPACT ON SAMPLE CONSTRUCTION ZONES IN MISSISSIPPI (Part 1: DESCRIPTIVE)

STATISTICAL ANALYSIS OF LAW ENFORCEMENT SURVEILLANCE IMPACT ON SAMPLE CONSTRUCTION ZONES IN MISSISSIPPI (Part 1: DESCRIPTIVE) STATISTICAL ANALYSIS OF LAW ENFORCEMENT SURVEILLANCE IMPACT ON SAMPLE CONSTRUCTION ZONES IN MISSISSIPPI (Part 1: DESCRIPTIVE) Tulio Sulbaran, Ph.D 1, David Marchman 2 Abstract It is estimated that every

More information

Estimation of Average Daily Traffic (ADT) Using Short - Duration Volume Counts

Estimation of Average Daily Traffic (ADT) Using Short - Duration Volume Counts Estimation of Average Daily Traffic (ADT) Using Short - Duration Volume Counts M.D.D.V. Meegahapola 1, R.D.S.M. Ranamuka 1 and I.M.S. Sathyaprasad 1 1 Department of Civil Engineering Faculty of Engineering

More information

More on Unsupervised Learning

More on Unsupervised Learning More on Unsupervised Learning Two types of problems are to find association rules for occurrences in common in observations (market basket analysis), and finding the groups of values of observational data

More information

CS534 Machine Learning - Spring Final Exam

CS534 Machine Learning - Spring Final Exam CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the

More information

Real-Time Travel Time Prediction Using Multi-level k-nearest Neighbor Algorithm and Data Fusion Method

Real-Time Travel Time Prediction Using Multi-level k-nearest Neighbor Algorithm and Data Fusion Method 1861 Real-Time Travel Time Prediction Using Multi-level k-nearest Neighbor Algorithm and Data Fusion Method Sehyun Tak 1, Sunghoon Kim 2, Kiate Jang 3 and Hwasoo Yeo 4 1 Smart Transportation System Laboratory,

More information

Forecasting Wind Ramps

Forecasting Wind Ramps Forecasting Wind Ramps Erin Summers and Anand Subramanian Jan 5, 20 Introduction The recent increase in the number of wind power producers has necessitated changes in the methods power system operators

More information

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1 EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle

More information

PATTERN CLASSIFICATION

PATTERN CLASSIFICATION PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS

More information

Different points of view for selecting a latent structure model

Different points of view for selecting a latent structure model Different points of view for selecting a latent structure model Gilles Celeux Inria Saclay-Île-de-France, Université Paris-Sud Latent structure models: two different point of views Density estimation LSM

More information

Classification: The rest of the story

Classification: The rest of the story U NIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN CS598 Machine Learning for Signal Processing Classification: The rest of the story 3 October 2017 Today s lecture Important things we haven t covered yet Fisher

More information

Encapsulating Urban Traffic Rhythms into Road Networks

Encapsulating Urban Traffic Rhythms into Road Networks Encapsulating Urban Traffic Rhythms into Road Networks Junjie Wang +, Dong Wei +, Kun He, Hang Gong, Pu Wang * School of Traffic and Transportation Engineering, Central South University, Changsha, Hunan,

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 6: Cluster Analysis Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2017/2018 Master in Mathematical Engineering

More information

A data mining approach to creating fundamental traffic flow diagram

A data mining approach to creating fundamental traffic flow diagram Available online at www.sciencedirect.com ScienceDirect Procedia - Social and Behavioral Scien ce s 104 ( 2013 ) 430 439 2 nd Conference of Transportation Research Group of India (2nd CTRG) A data mining

More information

ALTERNATIVES FOR ESTIMATING SEASONAL FACTORS ON RURAL AND URBAN ROADS IN FLORIDA

ALTERNATIVES FOR ESTIMATING SEASONAL FACTORS ON RURAL AND URBAN ROADS IN FLORIDA ALTERNATIVES FOR ESTIMATING SEASONAL FACTORS ON RURAL AND URBAN ROADS IN FLORIDA Final Report Contract No. BD015-03 Prepared for Research Office Florida Department of Transportation Prepared by Lehman

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Unsupervised machine learning

Unsupervised machine learning Chapter 9 Unsupervised machine learning Unsupervised machine learning (a.k.a. cluster analysis) is a set of methods to assign objects into clusters under a predefined distance measure when class labels

More information

Classification: Linear Discriminant Analysis

Classification: Linear Discriminant Analysis Classification: Linear Discriminant Analysis Discriminant analysis uses sample information about individuals that are known to belong to one of several populations for the purposes of classification. Based

More information

A clustering view on ESS measures of political interest:

A clustering view on ESS measures of political interest: A clustering view on ESS measures of political interest: An EM-MML approach Cláudia Silvestre Margarida Cardoso Mário Figueiredo Escola Superior de Comunicação Social - IPL BRU_UNIDE, ISCTE-IUL Instituto

More information

CSCI-567: Machine Learning (Spring 2019)

CSCI-567: Machine Learning (Spring 2019) CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March

More information

High Dimensional Discriminant Analysis

High Dimensional Discriminant Analysis High Dimensional Discriminant Analysis Charles Bouveyron LMC-IMAG & INRIA Rhône-Alpes Joint work with S. Girard and C. Schmid High Dimensional Discriminant Analysis - Lear seminar p.1/43 Introduction High

More information

Predicting Long-term Exposures for Health Effect Studies

Predicting Long-term Exposures for Health Effect Studies Predicting Long-term Exposures for Health Effect Studies Lianne Sheppard Adam A. Szpiro, Johan Lindström, Paul D. Sampson and the MESA Air team University of Washington CMAS Special Session, October 13,

More information

Lesson 8: Variability in a Data Distribution

Lesson 8: Variability in a Data Distribution Classwork Example 1: Comparing Two Distributions Robert s family is planning to move to either New York City or San Francisco. Robert has a cousin in San Francisco and asked her how she likes living in

More information

Introduction to Statistical modeling: handout for Math 489/583

Introduction to Statistical modeling: handout for Math 489/583 Introduction to Statistical modeling: handout for Math 489/583 Statistical modeling occurs when we are trying to model some data using statistical tools. From the start, we recognize that no model is perfect

More information

Urban Link Travel Time Estimation Using Large-scale Taxi Data with Partial Information

Urban Link Travel Time Estimation Using Large-scale Taxi Data with Partial Information Urban Link Travel Time Estimation Using Large-scale Taxi Data with Partial Information * Satish V. Ukkusuri * * Civil Engineering, Purdue University 24/04/2014 Outline Introduction Study Region Link Travel

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota,

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Trip Generation Characteristics of Super Convenience Market Gasoline Pump Stores

Trip Generation Characteristics of Super Convenience Market Gasoline Pump Stores Trip Generation Characteristics of Super Convenience Market Gasoline Pump Stores This article presents the findings of a study that investigated trip generation characteristics of a particular chain of

More information

Variable selection for model-based clustering

Variable selection for model-based clustering Variable selection for model-based clustering Matthieu Marbac (Ensai - Crest) Joint works with: M. Sedki (Univ. Paris-sud) and V. Vandewalle (Univ. Lille 2) The problem Objective: Estimation of a partition

More information

Multivariate Regression Model Results

Multivariate Regression Model Results Updated: August, 0 Page of Multivariate Regression Model Results 4 5 6 7 8 This exhibit provides the results of the load model forecast discussed in Schedule. Included is the forecast of short term system

More information

Time Series Analysis

Time Series Analysis Time Series Analysis A time series is a sequence of observations made: 1) over a continuous time interval, 2) of successive measurements across that interval, 3) using equal spacing between consecutive

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and

More information

Predicting Future Energy Consumption CS229 Project Report

Predicting Future Energy Consumption CS229 Project Report Predicting Future Energy Consumption CS229 Project Report Adrien Boiron, Stephane Lo, Antoine Marot Abstract Load forecasting for electric utilities is a crucial step in planning and operations, especially

More information

Data Preprocessing. Cluster Similarity

Data Preprocessing. Cluster Similarity 1 Cluster Similarity Similarity is most often measured with the help of a distance function. The smaller the distance, the more similar the data objects (points). A function d: M M R is a distance on M

More information

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata

Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata Principles of Pattern Recognition C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata e-mail: murthy@isical.ac.in Pattern Recognition Measurement Space > Feature Space >Decision

More information

Course in Data Science

Course in Data Science Course in Data Science About the Course: In this course you will get an introduction to the main tools and ideas which are required for Data Scientist/Business Analyst/Data Analyst. The course gives an

More information

Machine Learning, Fall 2009: Midterm

Machine Learning, Fall 2009: Midterm 10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all

More information

Accessing and Using Indiana Traffic Data

Accessing and Using Indiana Traffic Data Accessing and Using Indiana Traffic Data Purdue Road School March 6, 2013 Presented by: Gregory Katter, PE Asset Planning Indiana Traffic Data 1. Overview of Traffic Data Collection Program 2. Accessing

More information

EVALUATION OF ALGORITHM PERFORMANCE 2012/13 GAS YEAR SCALING FACTOR AND WEATHER CORRECTION FACTOR

EVALUATION OF ALGORITHM PERFORMANCE 2012/13 GAS YEAR SCALING FACTOR AND WEATHER CORRECTION FACTOR EVALUATION OF ALGORITHM PERFORMANCE /3 GAS YEAR SCALING FACTOR AND WEATHER CORRECTION FACTOR. Background The annual gas year algorithm performance evaluation normally considers three sources of information

More information

Model selection criteria in Classification contexts. Gilles Celeux INRIA Futurs (orsay)

Model selection criteria in Classification contexts. Gilles Celeux INRIA Futurs (orsay) Model selection criteria in Classification contexts Gilles Celeux INRIA Futurs (orsay) Cluster analysis Exploratory data analysis tools which aim is to find clusters in a large set of data (many observations

More information

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II) Contents Lecture Lecture Linear Discriminant Analysis Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University Email: fredriklindsten@ituuse Summary of lecture

More information

Minimum message length estimation of mixtures of multivariate Gaussian and von Mises-Fisher distributions

Minimum message length estimation of mixtures of multivariate Gaussian and von Mises-Fisher distributions Minimum message length estimation of mixtures of multivariate Gaussian and von Mises-Fisher distributions Parthan Kasarapu & Lloyd Allison Monash University, Australia September 8, 25 Parthan Kasarapu

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Andrew O. Finley 1 and Sudipto Banerjee 2 1 Department of Forestry & Department of Geography, Michigan

More information

AREP GAW. AQ Forecasting

AREP GAW. AQ Forecasting AQ Forecasting What Are We Forecasting Averaging Time (3 of 3) PM10 Daily Maximum Values, 2001 Santiago, Chile (MACAM stations) 300 Level 2 Pre-Emergency Level 1 Alert 200 Air Quality Standard 150 100

More information

Predicting freeway traffic in the Bay Area

Predicting freeway traffic in the Bay Area Predicting freeway traffic in the Bay Area Jacob Baldwin Email: jtb5np@stanford.edu Chen-Hsuan Sun Email: chsun@stanford.edu Ya-Ting Wang Email: yatingw@stanford.edu Abstract The hourly occupancy rate

More information

Global Climates. Name Date

Global Climates. Name Date Global Climates Name Date No investigation of the atmosphere is complete without examining the global distribution of the major atmospheric elements and the impact that humans have on weather and climate.

More information

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Final Overview. Introduction to ML. Marek Petrik 4/25/2017 Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,

More information

= observed volume on day l for bin j = base volume in jth bin, and = residual error, assumed independent with mean zero.

= observed volume on day l for bin j = base volume in jth bin, and = residual error, assumed independent with mean zero. QB research September 4, 06 Page -Minute Bin Volume Forecast Model Overview In response to strong client demand, Quantitative Brokers (QB) has developed a new algorithm called Closer that specifically

More information

Technical note on seasonal adjustment for M0

Technical note on seasonal adjustment for M0 Technical note on seasonal adjustment for M0 July 1, 2013 Contents 1 M0 2 2 Steps in the seasonal adjustment procedure 3 2.1 Pre-adjustment analysis............................... 3 2.2 Seasonal adjustment.................................

More information

Qualifying Exam in Machine Learning

Qualifying Exam in Machine Learning Qualifying Exam in Machine Learning October 20, 2009 Instructions: Answer two out of the three questions in Part 1. In addition, answer two out of three questions in two additional parts (choose two parts

More information

MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No.

MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No. MODELING OF 85 TH PERCENTILE SPEED FOR RURAL HIGHWAYS FOR ENHANCED TRAFFIC SAFETY ANNUAL REPORT FOR FY 2009 (ODOT SPR ITEM No. 2211) Submitted to: Ginger McGovern, P.E. Planning and Research Division Engineer

More information

Submitted for Presentation at the 2006 TRB Annual Meeting of the Transportation Research Board

Submitted for Presentation at the 2006 TRB Annual Meeting of the Transportation Research Board Investigation of Hot Mix Asphalt Surfaced Pavements Skid Resistance in Maryland State Highway Network System By Wenbing Song, Xin Chen, Timothy Smith, and Adel Hedfi Wenbing Song MDSHA, Office of Material

More information

Clustering VS Classification

Clustering VS Classification MCQ Clustering VS Classification 1. What is the relation between the distance between clusters and the corresponding class discriminability? a. proportional b. inversely-proportional c. no-relation Ans:

More information

A FUZZY TIME SERIES-MARKOV CHAIN MODEL WITH AN APPLICATION TO FORECAST THE EXCHANGE RATE BETWEEN THE TAIWAN AND US DOLLAR.

A FUZZY TIME SERIES-MARKOV CHAIN MODEL WITH AN APPLICATION TO FORECAST THE EXCHANGE RATE BETWEEN THE TAIWAN AND US DOLLAR. International Journal of Innovative Computing, Information and Control ICIC International c 2012 ISSN 1349-4198 Volume 8, Number 7(B), July 2012 pp. 4931 4942 A FUZZY TIME SERIES-MARKOV CHAIN MODEL WITH

More information

PRELIMINARY DRAFT FOR DISCUSSION PURPOSES

PRELIMINARY DRAFT FOR DISCUSSION PURPOSES Memorandum To: David Thompson From: John Haapala CC: Dan McDonald Bob Montgomery Date: February 24, 2003 File #: 1003551 Re: Lake Wenatchee Historic Water Levels, Operation Model, and Flood Operation This

More information

Unsupervised Learning with Permuted Data

Unsupervised Learning with Permuted Data Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University

More information

Choosing a model in a Classification purpose. Guillaume Bouchard, Gilles Celeux

Choosing a model in a Classification purpose. Guillaume Bouchard, Gilles Celeux Choosing a model in a Classification purpose Guillaume Bouchard, Gilles Celeux Abstract: We advocate the usefulness of taking into account the modelling purpose when selecting a model. Two situations are

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Andrew O. Finley Department of Forestry & Department of Geography, Michigan State University, Lansing

More information

The Scope and Growth of Spatial Analysis in the Social Sciences

The Scope and Growth of Spatial Analysis in the Social Sciences context. 2 We applied these search terms to six online bibliographic indexes of social science Completed as part of the CSISS literature search initiative on November 18, 2003 The Scope and Growth of Spatial

More information

Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size

Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size Berkman Sahiner, a) Heang-Ping Chan, Nicholas Petrick, Robert F. Wagner, b) and Lubomir Hadjiiski

More information

U.S. - Canadian Border Traffic Prediction

U.S. - Canadian Border Traffic Prediction Western Washington University Western CEDAR WWU Honors Program Senior Projects WWU Graduate and Undergraduate Scholarship 12-14-2017 U.S. - Canadian Border Traffic Prediction Colin Middleton Western Washington

More information

Expanding the GSATS Model Area into

Expanding the GSATS Model Area into Appendix A Expanding the GSATS Model Area into North Carolina Jluy, 2011 Table of Contents LONG-RANGE TRANSPORTATION PLAN UPDATE 1. Introduction... 1 1.1 Background... 1 1.2 Existing Northern Extent of

More information

Machine Learning Overview

Machine Learning Overview Machine Learning Overview Sargur N. Srihari University at Buffalo, State University of New York USA 1 Outline 1. What is Machine Learning (ML)? 2. Types of Information Processing Problems Solved 1. Regression

More information

Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion

Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion Journal of Advances in Information Technology Vol. 8, No. 1, February 2017 Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion Guizhou Wang Institute of Remote Sensing

More information

Daily Rainfall Disaggregation Using HYETOS Model for Peninsular Malaysia

Daily Rainfall Disaggregation Using HYETOS Model for Peninsular Malaysia Daily Rainfall Disaggregation Using HYETOS Model for Peninsular Malaysia Ibrahim Suliman Hanaish, Kamarulzaman Ibrahim, Abdul Aziz Jemain Abstract In this paper, we have examined the applicability of single

More information

DM-Group Meeting. Subhodip Biswas 10/16/2014

DM-Group Meeting. Subhodip Biswas 10/16/2014 DM-Group Meeting Subhodip Biswas 10/16/2014 Papers to be discussed 1. Crowdsourcing Land Use Maps via Twitter Vanessa Frias-Martinez and Enrique Frias-Martinez in KDD 2014 2. Tracking Climate Change Opinions

More information

Transport Data Analysis and Modeling Methodologies

Transport Data Analysis and Modeling Methodologies 1 Transport Data Analysis and Modeling Methodologies Lab Session #12 (Random Parameters Count-Data Models) You are given accident, evirnomental, traffic, and roadway geometric data from 275 segments of

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms

VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms 03/Feb/2010 VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms Presented by Andriy Temko Department of Electrical and Electronic Engineering Page 2 of

More information

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Lecture 4 Discriminant Analysis, k-nearest Neighbors Lecture 4 Discriminant Analysis, k-nearest Neighbors Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se fredrik.lindsten@it.uu.se

More information

FHWA Planning Data Resources: Census Data Planning Products (CTPP) HEPGIS Interactive Mapping Portal

FHWA Planning Data Resources: Census Data Planning Products (CTPP) HEPGIS Interactive Mapping Portal FHWA Planning Data Resources: Census Data Planning Products (CTPP) HEPGIS Interactive Mapping Portal Jeremy Raw, P.E. FHWA, Office of Planning, Systems Planning and Analysis August 2017 Outline Census

More information

A Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems

A Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems A Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems Hiroyuki Mori Dept. of Electrical & Electronics Engineering Meiji University Tama-ku, Kawasaki

More information

GAMINGRE 8/1/ of 7

GAMINGRE 8/1/ of 7 FYE 09/30/92 JULY 92 0.00 254,550.00 0.00 0 0 0 0 0 0 0 0 0 254,550.00 0.00 0.00 0.00 0.00 254,550.00 AUG 10,616,710.31 5,299.95 845,656.83 84,565.68 61,084.86 23,480.82 339,734.73 135,893.89 67,946.95

More information

Machine Learning for Signal Processing Bayes Classification and Regression

Machine Learning for Signal Processing Bayes Classification and Regression Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/18797 1 Recap: KNN A very effective and simple way of performing classification Simple model: For

More information

Chapter 14 Combining Models

Chapter 14 Combining Models Chapter 14 Combining Models T-61.62 Special Course II: Pattern Recognition and Machine Learning Spring 27 Laboratory of Computer and Information Science TKK April 3th 27 Outline Independent Mixing Coefficients

More information

Load Forecasting Using Artificial Neural Networks and Support Vector Regression

Load Forecasting Using Artificial Neural Networks and Support Vector Regression Proceedings of the 7th WSEAS International Conference on Power Systems, Beijing, China, September -7, 2007 3 Load Forecasting Using Artificial Neural Networks and Support Vector Regression SILVIO MICHEL

More information

Regression Clustering

Regression Clustering Regression Clustering In regression clustering, we assume a model of the form y = f g (x, θ g ) + ɛ g for observations y and x in the g th group. Usually, of course, we assume linear models of the form

More information

Gaussian and Linear Discriminant Analysis; Multiclass Classification

Gaussian and Linear Discriminant Analysis; Multiclass Classification Gaussian and Linear Discriminant Analysis; Multiclass Classification Professor Ameet Talwalkar Slide Credit: Professor Fei Sha Professor Ameet Talwalkar CS260 Machine Learning Algorithms October 13, 2015

More information

Exploring Human Mobility with Multi-Source Data at Extremely Large Metropolitan Scales. ACM MobiCom 2014, Maui, HI

Exploring Human Mobility with Multi-Source Data at Extremely Large Metropolitan Scales. ACM MobiCom 2014, Maui, HI Exploring Human Mobility with Multi-Source Data at Extremely Large Metropolitan Scales Desheng Zhang & Tian He University of Minnesota, USA Jun Huang, Ye Li, Fan Zhang, Chengzhong Xu Shenzhen Institute

More information

Introduction to Course

Introduction to Course .. Introduction to Course Oran Kittithreerapronchai 1 1 Department of Industrial Engineering, Chulalongkorn University Bangkok 10330 THAILAND last updated: September 17, 2016 COMP METH v2.00: intro 1/

More information

Discriminative Direction for Kernel Classifiers

Discriminative Direction for Kernel Classifiers Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information