Fast Adaptive Algorithm for Robust Evaluation of Quality of Experience

Size: px
Start display at page:

Download "Fast Adaptive Algorithm for Robust Evaluation of Quality of Experience"

Transcription

1 Fast Adaptive Algorithm for Robust Evaluation of Quality of Experience Qianqian Xu, Ming Yan, Yuan Yao October 2014

2 1 Motivation Mean Opinion Score vs. Paired Comparisons Crowdsourcing Ranking on Internet 2 Outlier Detection HodgeRank on Graphs LASSO for Outlier Detection Adaptive Least Trimmed Squares 3 Numerical Experiments Simulated Study Real-world Data 4 Conclusions

3 Outline Motivation Outlier Detection Numerical Experiments Subjective Image Quality Assessment Reference Qianqian Xu, Fast-fading Ming Yan, Yuan Yao White-noise iterative Least Trimmed Squares Conclusions

4 Mean Opinion Score Mean opinion score is widely used for evaluation of images, videos, as well books and movies, etc., but Unable to concretely define the concept of scale; Ambiguity interpretations of the scale among users; Difficult to verify whether a participant gives false ratings either intentionally or carelessly.

5 Outline Motivation Outlier Detection Numerical Experiments Paired Comparisons Simpler design with binary choice; Robust decision (invariant up to a monotone transform on personal scaling functions). Which one looks better to you (fastfading vs. white noise)? Qianqian Xu, Ming Yan, Yuan Yao iterative Least Trimmed Squares Conclusions

6 Crowdsourcing Ranking on Internet Start from a movie The Social Network

7 Crowdsourcing Definition The term crowdsourcing is a portmanteau of crowd and outsourcing. It is the act of outsourcing tasks, traditionally performed by an employee or contractor, to an undefined, large group of people or community (a crowd ), through an open call. random participants from Internet; random item pairs in comparison; so, data is incomplete; is imbalanced (heterogeneously distributed); is dynamic; contains outliers.

8 Automatic Outlier Detection in QoE Evaluation iterative Least Trimmed Squares (ilts) is proposed for outlier detection in QoE evaluation. ilts is fast achieves up to 190 times faster than LASSO; ilts is adaptive purifies data automatically without a prior knowledge on the amount of outliers.

9 Least Squares Assume Y α ij = sign(s i s j + Z α ij ) sign( ) = ±1 measures the sign of the value; s = {s 1,, s n} R n is the true scaling score on n items; z α ij is noise. For independent and identically distributed noise zij α with zero mean, the Gauss-Markov Theorem claims Least Square (LS)-rank is the unbiased estimator with minimal variance: 1 minimize s R n 2 (s i s j Yij α ) 2. i,j,α

10 Sparse Outliers If Zij α = Eij α + Nij α, where perturbation E α ij is sparse outliers, LS becomes unstable and may give bad estimation. May due to different test conditions; human errors; abnormal variations in context. How to detect and remove them to achieve a robust estimation against sparse outliers?

11 LASSO 1 minimize s R n,e 2 (s i s j Yij α + Eij α ) 2 + λ E 1. i,j,α Algorithm 1 LASSO for Outlier-Detection and Global Rating Estimation (1) Find the solution path of the Lasso problem; (2) Tuning parameter. Determine an optimal λ by cross-validation with random projections or by looking at the path directly; (3) Rule out outliers and perform least squares to get an unbiased score estimation.

12 Parameter Tuning Remark. Cross-validation may fail when outliers become dense and small in magnitudes. We can look at the solution path directly. An example of the solution path of simulated data with those corresponding to outliers are plotted in red, which mostly lie outside the majority of the paths.

13 Drawbacks of LASSO LASSO is expensive; LASSO needs prior knowledge (i.e., number of outliers in the dataset). Call for an efficient and automatic method for outlier detection!

14 Least Trimmed Squares Use l 0 in the constraint instead of l 1 in the objective. { minimize 1 s R n,e 2 (s i s j Yij α + E α ij )2, i,j,α subject to E 0 K. It is equivalent to minimize s R n,λ subject to Λ α ij (s i s j Yij α)2, (1 Λ α ij ) K, Λα ij {0, 1}. i,j,α i,j,α (1) Here Λ α ij is used to denote the outliers as follows: Λ α ij = { 0, if Y α ij is a outlier, 1, otherwise. (2)

15 Alternating Minimization 1) Fix Λ and update s. We need to solve a least squares problem with the comparisons that Λ α ij = 1. 2) Fix s and update Λ. This time we are solving minimize Λ subject to Λ α ij (s i s j Yij α)2, (1 Λ α ij ) K, Λα ij {0, 1}. i,j,α i,j,α This problem is to choose K elements with largest summation from the set {(s i s j Yij α)2 }. We can choose any Λ such that (1 Λ α ij ) K, Λα ij {0, 1} and i,j,α min i,j,α,λ α=0(s i s j Yij α ) 2 max ij i,j,α,λ α=1(s i s j Yij α ) 2. (4) ij (3)

16 iterative Least Trimmed Squares Algorithm 2 iterative Least Trimmed Squares with K Input: {Yij α }, K 0. Initialization: k = 0, Λ α ij = 1. for k = 1, 2, do Update s k by solving the least squares problem using only the comparisons with Λ α ij = 1. Update Λ k from (4) with one different from previous ones. end for return s.

17 Adaptive Least Trimmed Squares Algorithm 3 Adaptive Least Trimmed Squares Input: {Yij α }, Miter > 0, β 1 < 1, β 2 > 1. Initialization: k = 0, Λ α i,j(k) = 1, K k = 0. for k = 1,, Miter do Update s k with least squares using only the comparisons with Λ α ij (k 1) = 1. Let K k be the total number of comparisons with wrong directions, i.e., Yij α has different sign with si k sj k. { β 1 Kk, if k = 1; K k = min( β 2K k 1, K (5) k ), otherwise. If K k = K k, break. Update Λ(k) using (4) with K = K k. end for Find ŝ with least squares using only the samples with Λ α ij (k) = 1. return ŝ, ˆK = Kk.

18 Remarks β 1 < 1 is small to make sure that the first estimation is underestimated. e.g. β 1 =0.75. β 2 > 1 is small to make sure that it will not overshoot the estimation too much. e.g. β 2 =1.03. Add one step to just compare every pair of two successive items and make the correction on the detection, i.e., if item i is ranking above j but the number of people choosing item i over j is less than the number of people choosing j over i, we can remove those choosing j over i and keep those choosing i over j.

19 Data Description Create a random total order on n candidates as the ground-truth order. Add paired comparison edges (i, j) randomly with preference directions following the ground-truth order. Randomly choose a portion of the comparison edges and reverse them in preference directions. Notation: SN (Sample Number): total number of paired comparisons. ON (Outlier Number): number of outliers. OP (Outlier Percentage): ON/SN.

20 ilts vs. LASSO (Precisions) avg (sd) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= (0.022) 0.993(0.023) 0.993(0.015) 0.942(0.037) 0.670(0.078) 0.505(0.097) SN= (0.000) 1.000(0.000) 0.998(0.009) 0.976(0.023) 0.751(0.067) 0.503(0.089) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.991(0.013) 0.811(0.060) 0.502(0.090) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.995(0.010) 0.829(0.059) 0.498(0.098) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.998(0.006) 0.847(0.052) 0.499(0.101) Precisions for simulated data via ilts, 100 times repeat. avg (sd) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= (0.033) 0.962(0.030) 0.958(0.025) 0.905(0.032) 0.698(0.067) 0.513(0.085) SN= (0.011) 0.990(0.014) 0.984(0.016) 0.942(0.022) 0.750(0.056) 0.516(0.084) SN= (0.005) 0.997(0.008) 0.992(0.012) 0.957(0.020) 0.796(0.050) 0.523(0.083) SN= (0.001) 0.999(0.005) 0.996(0.009) 0.970(0.016) 0.818(0.048) 0.518(0.093) SN= (0.002) 1.000(0.000) 0.998(0.006) 0.972(0.016) 0.837(0.038) 0.525(0.088) Precisions for simulated data via LASSO, 100 times repeat.

21 ilts vs. LASSO (Recalls) avg (sd) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= (0.000) 0.994(0.015) 0.994(0.010) 0.943(0.036) 0.653(0.080) 0.438(0.093) SN= (0.000) 1.000(0.000) 0.999(0.006) 0.978(0.019) 0.727(0.071) 0.456(0.087) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.991(0.012) 0.797(0.062) 0.464(0.089) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.996(0.007) 0.821(0.060) 0.466(0.098) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.998(0.006) 0.842(0.052) 0.470(0.100) Recalls for simulated data via ilts, 100 times repeat. avg (sd) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= (0.033) 0.962(0.030) 0.958(0.025) 0.905(0.032) 0.698(0.067) 0.513(0.085) SN= (0.011) 0.990(0.014) 0.984(0.016) 0.942(0.022) 0.750(0.056) 0.518(0.084) SN= (0.005) 0.997(0.008) 0.992(0.012) 0.957(0.020) 0.796(0.050) 0.523(0.083) SN= (0.001) 0.999(0.005) 0.996(0.009) 0.970(0.016) 0.818(0.048) 0.518(0.093) SN= (0.002) 1.000(0.000) 0.998(0.006) 0.972(0.016) 0.837(0.038) 0.525(0.088) Recalls for simulated data via LASSO, 100 times repeat.

22 ilts vs. LASSO (F1 scores) avg (sd) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= (0.012) 0.994(0.019) 0.994(0.012) 0.943(0.036) 0.675(0.079) 0.469(0.095) SN= (0.000) 1.000(0.000) 0.999(0.007) 0.977(0.021) 0.739(0.069) 0.478(0.088) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.991(0.012) 0.804(0.061) 0.482(0.089) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.996(0.009) 0.825(0.059) 0.482(0.098) SN= (0.000) 1.000(0.000) 1.000(0.000) 0.998(0.006) 0.845(0.052) 0.484(0.101) F1 scores for simulated data via ilts, 100 times repeat. avg (sd) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= (0.033) 0.962(0.030) 0.958(0.025) 0.905(0.032) 0.698(0.067) 0.513(0.085) SN= (0.011) 0.990(0.014) 0.984(0.016) 0.942(0.022) 0.750(0.056) 0.516(0.084) SN= (0.005) 0.997(0.008) 0.992(0.012) 0.957(0.020) 0.796(0.050) 0.523(0.083) SN= (0.001) 0.999(0.005) 0.996(0.009) 0.970(0.016) 0.818(0.048) 0.518(0.093) SN= (0.002) 1.000(0.000) 0.998(0.006) 0.972(0.016) 0.837(0.038) 0.525(0.088) F1 scores for simulated data via LASSO, 100 times repeat.

23 ilts is Fast! time (second) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= SN= SN= SN= SN= Computing time for 100 runs in total via ilts. time (second) OP=5% OP=10% OP=15% OP=30% OP=45% OP=50% SN= SN= SN= SN= SN= Computing time for 100 runs in total via LASSO. ilts can achieve up to about 190 times faster than LASSO!!!

24 Outline Motivation Outlier Detection Numerical Experiments Conclusions Data Description Figure: Left: PC-VQA dataset; right: PC-IQA dataset. PC-VQA (complete and balanced dataset): 38,400 paired comparisons for 10 reference videos, collected from 209 random observers. PC-IQA (incomplete and imbalanced dataset): 23,097 paired comparisons for 15 reference images, collected from 187 random observers. Qianqian Xu, Ming Yan, Yuan Yao iterative Least Trimmed Squares

25 Experimental Results ID Paired comparison matrixes of reference (a) in PC-VQA dataset. Red pairs are outliers obtained by both ilts and LASSO. Open blue circles are outliers obtained by LASSO but not ilts. Filled blue circles are outliers obtained by ilts but not LASSO.

26 Different rankings after outlier removal Video ID L2 LASSO IRLS 1 1 ( ) 1 ( ) 1 ( ) 9 2 ( ) 2 ( ) 2 ( ) 10 3 ( ) 3 ( ) 3 ( ) 13 4 ( ) 4 ( ) 4 ( ) 7 5 ( ) 5 ( ) 5 ( ) 8 6 ( ) 6 ( ) 6 ( ) 11 7 ( ) 7 ( ) 7 ( ) 14 8 ( ) 8 ( ) 8 ( ) 15 9 ( ) 9 ( ) 9 ( ) 3 10 ( ) 11 ( ) 12 ( ) ( ) 10 ( ) 10 ( ) 4 12 ( ) 12 ( ) 11 ( ) ( ) 13 ( ) 13 ( ) 5 14 ( ) 14 ( ) 14 ( ) 6 15 ( ) 15 ( ) 15 ( ) 2 16 ( ) 16 ( ) 16 ( ) Different rankings for reference (a) in the PC-VQA dataset. The integer represents the ranking position. The number in parentheses represents the global ranking score.

27 Experimental Results ID Paired comparison matrixes of reference (c) in PC-IQA dataset. Red pairs are outliers obtained by both ilts and LASSO. Open blue circles are outliers obtained by LASSO but not ilts. Filled blue circles are outliers obtained by ilts but not LASSO.

28 Different rankings after outliers removal for reference (c) in the PC-IQA dataset Image ID L2 LASSO IRLS 1 1 ( ) 1 ( ) 1 ( ) 8 2 ( ) 2 ( ) 2 ( ) 16 3 ( ) 3 ( ) 3 ( ) 2 4 ( ) 4 ( ) 4 ( ) 3 5 ( ) 5 ( ) 5 ( ) 11 6 ( ) 6 ( ) 7 ( ) 6 7 ( ) 7 ( ) 6 ( ) 12 8 ( ) 8 ( ) 8 ( ) 9 9 ( ) 9 ( ) 9 ( ) ( ) 10 ( ) 10 ( ) 5 11 ( ) 11 ( ) 11 ( ) ( ) 12 ( ) 12 ( ) 7 13 ( ) 13 ( ) 13 ( ) ( ) 14 ( ) 15 ( ) ( ) 15 ( ) 14 ( ) 4 16 ( ) 16 ( ) 16 ( ) Remark: ilts prefers to choose minority votings as outliers, while the LASSO selects the large deviations from the gradient of global ranking score as outliers even when they are in majority votings. Such a small difference only leads to a local order change of nearby ranked items, so both of these two ranking algorithms are stable.

29 Summary ilts is a surprisingly simple, efficient, and automatic algorithm for outlier detection in QoE evaluation. ilts achieves up to 190 times faster than LASSO. ilts can automatically estimate the number of outliers and detect them without any prior information about the number of outliers exist in the dataset. Open resources: report, slides, and codes available on

Fast Adaptive Algorithm for Robust Evaluation of Quality of Experience

Fast Adaptive Algorithm for Robust Evaluation of Quality of Experience 1 Fast Adaptive Algorithm for Robust Evaluation of Quality of Experience Qianqian Xu, Ming Yan, and Yuan Yao arxiv:1407.7636v2 [cs.mm] 22 Oct 2014 Abstract Outlier detection is an integral part of robust

More information

HodgeRank on Random Graphs

HodgeRank on Random Graphs Outline Motivation Application Discussions School of Mathematical Sciences Peking University July 18, 2011 Outline Motivation Application Discussions 1 Motivation Crowdsourcing Ranking on Internet 2 HodgeRank

More information

Active Sampling for Subjective Image Quality Assessment

Active Sampling for Subjective Image Quality Assessment Active Sampling for Subjective Image Quality Assessment Peng Ye and David Doermann Institute for Advanced Computer Studies University of Maryland, College Park, MD, USA {pengye,doermann}@umiacs.umd.edu

More information

Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems

Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems Sewoong Oh Massachusetts Institute of Technology joint work with David R. Karger and Devavrat Shah September 28, 2011 1 / 13 Crowdsourcing

More information

Two-Sample Inferential Statistics

Two-Sample Inferential Statistics The t Test for Two Independent Samples 1 Two-Sample Inferential Statistics In an experiment there are two or more conditions One condition is often called the control condition in which the treatment is

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification

More information

CSE 546 Final Exam, Autumn 2013

CSE 546 Final Exam, Autumn 2013 CSE 546 Final Exam, Autumn 0. Personal info: Name: Student ID: E-mail address:. There should be 5 numbered pages in this exam (including this cover sheet).. You can use any material you brought: any book,

More information

Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning

Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning Mingyan Liu (Joint work with Yang Liu) Department of Electrical Engineering and Computer Science University of Michigan,

More information

Linear classifiers: Overfitting and regularization

Linear classifiers: Overfitting and regularization Linear classifiers: Overfitting and regularization Emily Fox University of Washington January 25, 2017 Logistic regression recap 1 . Thus far, we focused on decision boundaries Score(x i ) = w 0 h 0 (x

More information

Recommender systems, matrix factorization, variable selection and social graph data

Recommender systems, matrix factorization, variable selection and social graph data Recommender systems, matrix factorization, variable selection and social graph data Julien Delporte & Stéphane Canu stephane.canu@litislab.eu StatLearn, april 205, Grenoble Road map Model selection for

More information

Whom to Ask? Jury Selection for Decision Making Tasks on Micro-blog Services

Whom to Ask? Jury Selection for Decision Making Tasks on Micro-blog Services Whom to Ask? Jury Selection for Decision Making Tasks on Micro-blog Services Caleb Chen CAO, Jieying SHE, Yongxin TONG, Lei CHEN The Hong Kong University of Science and Technology Is Istanbul the capital

More information

Graph Helmholtzian and Rank Learning

Graph Helmholtzian and Rank Learning Graph Helmholtzian and Rank Learning Lek-Heng Lim NIPS Workshop on Algebraic Methods in Machine Learning December 2, 2008 (Joint work with Xiaoye Jiang, Yuan Yao, and Yinyu Ye) L.-H. Lim (NIPS 2008) Graph

More information

Binary Principal Component Analysis in the Netflix Collaborative Filtering Task

Binary Principal Component Analysis in the Netflix Collaborative Filtering Task Binary Principal Component Analysis in the Netflix Collaborative Filtering Task László Kozma, Alexander Ilin, Tapani Raiko first.last@tkk.fi Helsinki University of Technology Adaptive Informatics Research

More information

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine CS 277: Data Mining Mining Web Link Structure Class Presentations In-class, Tuesday and Thursday next week 2-person teams: 6 minutes, up to 6 slides, 3 minutes/slides each person 1-person teams 4 minutes,

More information

Complex Social System, Elections. Introduction to Network Analysis 1

Complex Social System, Elections. Introduction to Network Analysis 1 Complex Social System, Elections Introduction to Network Analysis 1 Complex Social System, Network I person A voted for B A is more central than B if more people voted for A In-degree centrality index

More information

CSC 411: Lecture 03: Linear Classification

CSC 411: Lecture 03: Linear Classification CSC 411: Lecture 03: Linear Classification Richard Zemel, Raquel Urtasun and Sanja Fidler University of Toronto Zemel, Urtasun, Fidler (UofT) CSC 411: 03-Classification 1 / 24 Examples of Problems What

More information

CPSC 340: Machine Learning and Data Mining. Stochastic Gradient Fall 2017

CPSC 340: Machine Learning and Data Mining. Stochastic Gradient Fall 2017 CPSC 340: Machine Learning and Data Mining Stochastic Gradient Fall 2017 Assignment 3: Admin Check update thread on Piazza for correct definition of trainndx. This could make your cross-validation code

More information

When Dictionary Learning Meets Classification

When Dictionary Learning Meets Classification When Dictionary Learning Meets Classification Bufford, Teresa 1 Chen, Yuxin 2 Horning, Mitchell 3 Shee, Liberty 1 Mentor: Professor Yohann Tendero 1 UCLA 2 Dalhousie University 3 Harvey Mudd College August

More information

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model 204 Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model S. A. Ibrahim 1 ; W. B. Yahya 2 1 Department of Physical Sciences, Al-Hikmah University, Ilorin, Nigeria. e-mail:

More information

Large-Margin Thresholded Ensembles for Ordinal Regression

Large-Margin Thresholded Ensembles for Ordinal Regression Large-Margin Thresholded Ensembles for Ordinal Regression Hsuan-Tien Lin and Ling Li Learning Systems Group, California Institute of Technology, U.S.A. Conf. on Algorithmic Learning Theory, October 9,

More information

SAT, NP, NP-Completeness

SAT, NP, NP-Completeness CS 473: Algorithms, Spring 2018 SAT, NP, NP-Completeness Lecture 22 April 13, 2018 Most slides are courtesy Prof. Chekuri Ruta (UIUC) CS473 1 Spring 2018 1 / 57 Part I Reductions Continued Ruta (UIUC)

More information

FINAL: CS 6375 (Machine Learning) Fall 2014

FINAL: CS 6375 (Machine Learning) Fall 2014 FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for

More information

Crowdsourcing Pareto-Optimal Object Finding by Pairwise Comparisons

Crowdsourcing Pareto-Optimal Object Finding by Pairwise Comparisons 2015 The University of Texas at Arlington. All Rights Reserved. Crowdsourcing Pareto-Optimal Object Finding by Pairwise Comparisons Abolfazl Asudeh, Gensheng Zhang, Naeemul Hassan, Chengkai Li, Gergely

More information

Ad Placement Strategies

Ad Placement Strategies Case Study 1: Estimating Click Probabilities Tackling an Unknown Number of Features with Sketching Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox 2014 Emily Fox January

More information

Large-scale Collaborative Ranking in Near-Linear Time

Large-scale Collaborative Ranking in Near-Linear Time Large-scale Collaborative Ranking in Near-Linear Time Liwei Wu Depts of Statistics and Computer Science UC Davis KDD 17, Halifax, Canada August 13-17, 2017 Joint work with Cho-Jui Hsieh and James Sharpnack

More information

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit

More information

Andriy Mnih and Ruslan Salakhutdinov

Andriy Mnih and Ruslan Salakhutdinov MATRIX FACTORIZATION METHODS FOR COLLABORATIVE FILTERING Andriy Mnih and Ruslan Salakhutdinov University of Toronto, Machine Learning Group 1 What is collaborative filtering? The goal of collaborative

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Introduction to Statistics for Traffic Crash Reconstruction

Introduction to Statistics for Traffic Crash Reconstruction Introduction to Statistics for Traffic Crash Reconstruction Jeremy Daily Jackson Hole Scientific Investigations, Inc. c 2003 www.jhscientific.com Why Use and Learn Statistics? 1. We already do when ranging

More information

Infinite Ensemble Learning with Support Vector Machinery

Infinite Ensemble Learning with Support Vector Machinery Infinite Ensemble Learning with Support Vector Machinery Hsuan-Tien Lin and Ling Li Learning Systems Group, California Institute of Technology ECML/PKDD, October 4, 2005 H.-T. Lin and L. Li (Learning Systems

More information

10-701/ Machine Learning - Midterm Exam, Fall 2010

10-701/ Machine Learning - Midterm Exam, Fall 2010 10-701/15-781 Machine Learning - Midterm Exam, Fall 2010 Aarti Singh Carnegie Mellon University 1. Personal info: Name: Andrew account: E-mail address: 2. There should be 15 numbered pages in this exam

More information

arxiv: v1 [math.co] 13 Dec 2014

arxiv: v1 [math.co] 13 Dec 2014 SEARCHING FOR KNIGHTS AND SPIES: A MAJORITY/MINORITY GAME MARK WILDON arxiv:1412.4247v1 [math.co] 13 Dec 2014 Abstract. There are n people, each of whom is either a knight or a spy. It is known that at

More information

On Markov chain Monte Carlo methods for tall data

On Markov chain Monte Carlo methods for tall data On Markov chain Monte Carlo methods for tall data Remi Bardenet, Arnaud Doucet, Chris Holmes Paper review by: David Carlson October 29, 2016 Introduction Many data sets in machine learning and computational

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

Appendix: Modeling Approach

Appendix: Modeling Approach AFFECTIVE PRIMACY IN INTRAORGANIZATIONAL TASK NETWORKS Appendix: Modeling Approach There is now a significant and developing literature on Bayesian methods in social network analysis. See, for instance,

More information

A Wisdom of the Crowd Approach to Forecasting

A Wisdom of the Crowd Approach to Forecasting A Wisdom of the Crowd Approach to Forecasting Funded by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number D11PC20059 Brandon

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Statistical Ranking Problem

Statistical Ranking Problem Statistical Ranking Problem Tong Zhang Statistics Department, Rutgers University Ranking Problems Rank a set of items and display to users in corresponding order. Two issues: performance on top and dealing

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

COSC 341 Human Computer Interaction. Dr. Bowen Hui University of British Columbia Okanagan

COSC 341 Human Computer Interaction. Dr. Bowen Hui University of British Columbia Okanagan COSC 341 Human Computer Interaction Dr. Bowen Hui University of British Columbia Okanagan 1 Last Topic Distribution of means When it is needed How to build one (from scratch) Determining the characteristics

More information

CS246 Final Exam, Winter 2011

CS246 Final Exam, Winter 2011 CS246 Final Exam, Winter 2011 1. Your name and student ID. Name:... Student ID:... 2. I agree to comply with Stanford Honor Code. Signature:... 3. There should be 17 numbered pages in this exam (including

More information

Maximum Margin Matrix Factorization for Collaborative Ranking

Maximum Margin Matrix Factorization for Collaborative Ranking Maximum Margin Matrix Factorization for Collaborative Ranking Joint work with Quoc Le, Alexandros Karatzoglou and Markus Weimer Alexander J. Smola sml.nicta.com.au Statistical Machine Learning Program

More information

The Perceptron Algorithm

The Perceptron Algorithm The Perceptron Algorithm Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 Outline The Perceptron Algorithm Perceptron Mistake Bound Variants of Perceptron 2 Where are we? The Perceptron

More information

1 [15 points] Frequent Itemsets Generation With Map-Reduce

1 [15 points] Frequent Itemsets Generation With Map-Reduce Data Mining Learning from Large Data Sets Final Exam Date: 15 August 2013 Time limit: 120 minutes Number of pages: 11 Maximum score: 100 points You can use the back of the pages if you run out of space.

More information

CS570 Data Mining. Anomaly Detection. Li Xiong. Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber.

CS570 Data Mining. Anomaly Detection. Li Xiong. Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber. CS570 Data Mining Anomaly Detection Li Xiong Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber April 3, 2011 1 Anomaly Detection Anomaly is a pattern in the data that does not conform

More information

arxiv: v1 [stat.ml] 16 Nov 2017

arxiv: v1 [stat.ml] 16 Nov 2017 arxiv:1711.05957v1 [stat.ml] 16 Nov 2017 HodgeRank with Information Maximization for Crowdsourced Pairwise Ranking Aggregation Qianqian Xu 1,3, Jiechao Xiong 2,3, Xi Chen 4, Qingming Huang 5,6, Yuan Yao

More information

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints By I. Necoara, Y. Nesterov, and F. Glineur Lijun Xu Optimization Group Meeting November 27, 2012 Outline

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2014 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

CS 188: Artificial Intelligence. Outline

CS 188: Artificial Intelligence. Outline CS 188: Artificial Intelligence Lecture 21: Perceptrons Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. Outline Generative vs. Discriminative Binary Linear Classifiers Perceptron Multi-class

More information

Recommendation Systems

Recommendation Systems Recommendation Systems Pawan Goyal CSE, IITKGP October 21, 2014 Pawan Goyal (IIT Kharagpur) Recommendation Systems October 21, 2014 1 / 52 Recommendation System? Pawan Goyal (IIT Kharagpur) Recommendation

More information

Structural Learning and Integrative Decomposition of Multi-View Data

Structural Learning and Integrative Decomposition of Multi-View Data Structural Learning and Integrative Decomposition of Multi-View Data, Department of Statistics, Texas A&M University JSM 2018, Vancouver, Canada July 31st, 2018 Dr. Gen Li, Columbia University, Mailman

More information

CS246 Final Exam. March 16, :30AM - 11:30AM

CS246 Final Exam. March 16, :30AM - 11:30AM CS246 Final Exam March 16, 2016 8:30AM - 11:30AM Name : SUID : I acknowledge and accept the Stanford Honor Code. I have neither given nor received unpermitted help on this examination. (signed) Directions

More information

CS Homework 3. October 15, 2009

CS Homework 3. October 15, 2009 CS 294 - Homework 3 October 15, 2009 If you have questions, contact Alexandre Bouchard (bouchard@cs.berkeley.edu) for part 1 and Alex Simma (asimma@eecs.berkeley.edu) for part 2. Also check the class website

More information

Some graph optimization problems in data mining. P. Van Dooren, CESAME, Univ. catholique Louvain based on work in collaboration with the group on

Some graph optimization problems in data mining. P. Van Dooren, CESAME, Univ. catholique Louvain based on work in collaboration with the group on Some graph optimization problems in data mining P. Van Dooren, CESAME, Univ. catholique Louvain based on work in collaboration with the group on University of Chicago, October 16, 2012 Leuven Lambiotte

More information

arxiv: v1 [stat.me] 30 Dec 2017

arxiv: v1 [stat.me] 30 Dec 2017 arxiv:1801.00105v1 [stat.me] 30 Dec 2017 An ISIS screening approach involving threshold/partition for variable selection in linear regression 1. Introduction Yu-Hsiang Cheng e-mail: 96354501@nccu.edu.tw

More information

Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent

Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent KDD 2011 Rainer Gemulla, Peter J. Haas, Erik Nijkamp and Yannis Sismanis Presenter: Jiawen Yao Dept. CSE, UT Arlington 1 1

More information

Lecture 9: September 28

Lecture 9: September 28 0-725/36-725: Convex Optimization Fall 206 Lecturer: Ryan Tibshirani Lecture 9: September 28 Scribes: Yiming Wu, Ye Yuan, Zhihao Li Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These

More information

Introduction to Logistic Regression

Introduction to Logistic Regression Introduction to Logistic Regression Guy Lebanon Binary Classification Binary classification is the most basic task in machine learning, and yet the most frequent. Binary classifiers often serve as the

More information

Delta Boosting Machine and its application in Actuarial Modeling Simon CK Lee, Sheldon XS Lin KU Leuven, University of Toronto

Delta Boosting Machine and its application in Actuarial Modeling Simon CK Lee, Sheldon XS Lin KU Leuven, University of Toronto Delta Boosting Machine and its application in Actuarial Modeling Simon CK Lee, Sheldon XS Lin KU Leuven, University of Toronto This presentation has been prepared for the Actuaries Institute 2015 ASTIN

More information

Spatial Decision Tree: A Novel Approach to Land-Cover Classification

Spatial Decision Tree: A Novel Approach to Land-Cover Classification Spatial Decision Tree: A Novel Approach to Land-Cover Classification Zhe Jiang 1, Shashi Shekhar 1, Xun Zhou 1, Joseph Knight 2, Jennifer Corcoran 2 1 Department of Computer Science & Engineering 2 Department

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014 Exam policy: This exam allows two one-page, two-sided cheat sheets (i.e. 4 sides); No other materials. Time: 2 hours. Be sure to write

More information

Anomaly Detection for the CERN Large Hadron Collider injection magnets

Anomaly Detection for the CERN Large Hadron Collider injection magnets Anomaly Detection for the CERN Large Hadron Collider injection magnets Armin Halilovic KU Leuven - Department of Computer Science In cooperation with CERN 2018-07-27 0 Outline 1 Context 2 Data 3 Preprocessing

More information

Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

More information

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology

More information

Ridge Regression 1. to which some random noise is added. So that the training labels can be represented as:

Ridge Regression 1. to which some random noise is added. So that the training labels can be represented as: CS 1: Machine Learning Spring 15 College of Computer and Information Science Northeastern University Lecture 3 February, 3 Instructor: Bilal Ahmed Scribe: Bilal Ahmed & Virgil Pavlu 1 Introduction Ridge

More information

Deep Learning & Artificial Intelligence WS 2018/2019

Deep Learning & Artificial Intelligence WS 2018/2019 Deep Learning & Artificial Intelligence WS 2018/2019 Linear Regression Model Model Error Function: Squared Error Has no special meaning except it makes gradients look nicer Prediction Ground truth / target

More information

Logistic Regression with the Nonnegative Garrote

Logistic Regression with the Nonnegative Garrote Logistic Regression with the Nonnegative Garrote Enes Makalic Daniel F. Schmidt Centre for MEGA Epidemiology The University of Melbourne 24th Australasian Joint Conference on Artificial Intelligence 2011

More information

Voting (Ensemble Methods)

Voting (Ensemble Methods) 1 2 Voting (Ensemble Methods) Instead of learning a single classifier, learn many weak classifiers that are good at different parts of the data Output class: (Weighted) vote of each classifier Classifiers

More information

Proximity-Based Anomaly Detection using Sparse Structure Learning

Proximity-Based Anomaly Detection using Sparse Structure Learning Proximity-Based Anomaly Detection using Sparse Structure Learning Tsuyoshi Idé (IBM Tokyo Research Lab) Aurelie C. Lozano, Naoki Abe, and Yan Liu (IBM T. J. Watson Research Center) 2009/04/ SDM 2009 /

More information

Bias-free Sparse Regression with Guaranteed Consistency

Bias-free Sparse Regression with Guaranteed Consistency Bias-free Sparse Regression with Guaranteed Consistency Wotao Yin (UCLA Math) joint with: Stanley Osher, Ming Yan (UCLA) Feng Ruan, Jiechao Xiong, Yuan Yao (Peking U) UC Riverside, STATS Department March

More information

Stat 705: Completely randomized and complete block designs

Stat 705: Completely randomized and complete block designs Stat 705: Completely randomized and complete block designs Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 16 Experimental design Our department offers

More information

Method 1: Geometric Error Optimization

Method 1: Geometric Error Optimization Method 1: Geometric Error Optimization we need to encode the constraints ŷ i F ˆx i = 0, rank F = 2 idea: reconstruct 3D point via equivalent projection matrices and use reprojection error equivalent projection

More information

Generic Text Summarization

Generic Text Summarization June 27, 2012 Outline Introduction 1 Introduction Notation and Terminology 2 3 4 5 6 Text Summarization Introduction Notation and Terminology Two Types of Text Summarization Query-Relevant Summarization:

More information

CSE 546 Midterm Exam, Fall 2014

CSE 546 Midterm Exam, Fall 2014 CSE 546 Midterm Eam, Fall 2014 1. Personal info: Name: UW NetID: Student ID: 2. There should be 14 numbered pages in this eam (including this cover sheet). 3. You can use an material ou brought: an book,

More information

Gaussian and Linear Discriminant Analysis; Multiclass Classification

Gaussian and Linear Discriminant Analysis; Multiclass Classification Gaussian and Linear Discriminant Analysis; Multiclass Classification Professor Ameet Talwalkar Slide Credit: Professor Fei Sha Professor Ameet Talwalkar CS260 Machine Learning Algorithms October 13, 2015

More information

Extending a two-variable mean to a multi-variable mean

Extending a two-variable mean to a multi-variable mean Extending a two-variable mean to a multi-variable mean Estelle M. Massart, Julien M. Hendrickx, P.-A. Absil Universit e catholique de Louvain - ICTEAM Institute B-348 Louvain-la-Neuve - Belgium Abstract.

More information

Dimensionality Reduction

Dimensionality Reduction 394 Chapter 11 Dimensionality Reduction There are many sources of data that can be viewed as a large matrix. We saw in Chapter 5 how the Web can be represented as a transition matrix. In Chapter 9, the

More information

Machine Learning - MT Clustering

Machine Learning - MT Clustering Machine Learning - MT 2016 15. Clustering Varun Kanade University of Oxford November 28, 2016 Announcements No new practical this week All practicals must be signed off in sessions this week Firm Deadline:

More information

Shape Outlier Detection Using Pose Preserving Dynamic Shape Models

Shape Outlier Detection Using Pose Preserving Dynamic Shape Models Shape Outlier Detection Using Pose Preserving Dynamic Shape Models Chan-Su Lee and Ahmed Elgammal Rutgers, The State University of New Jersey Department of Computer Science Outline Introduction Shape Outlier

More information

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20. 10-601 Machine Learning, Midterm Exam: Spring 2008 Please put your name on this cover sheet If you need more room to work out your answer to a question, use the back of the page and clearly mark on the

More information

CMPS 6630: Introduction to Computational Biology and Bioinformatics. Structure Comparison

CMPS 6630: Introduction to Computational Biology and Bioinformatics. Structure Comparison CMPS 6630: Introduction to Computational Biology and Bioinformatics Structure Comparison Protein Structure Comparison Motivation Understand sequence and structure variability Understand Domain architecture

More information

Information Retrieval

Information Retrieval Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices

More information

Learning to Query, Reason, and Answer Questions On Ambiguous Texts

Learning to Query, Reason, and Answer Questions On Ambiguous Texts Learning to Query, Reason, and Answer Questions On Ambiguous Texts Xiaoxiao Guo, Tim Klinger, Clemens Rosenbaum, Joseph P. Bigus, Murray Campbell, Ban Kawas, Kartik Talamadupula, Gerald Tesauro, Satinder

More information

1 Overview. 2 Learning from Experts. 2.1 Defining a meaningful benchmark. AM 221: Advanced Optimization Spring 2016

1 Overview. 2 Learning from Experts. 2.1 Defining a meaningful benchmark. AM 221: Advanced Optimization Spring 2016 AM 1: Advanced Optimization Spring 016 Prof. Yaron Singer Lecture 11 March 3rd 1 Overview In this lecture we will introduce the notion of online convex optimization. This is an extremely useful framework

More information

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run

More information

Data Mining. CS57300 Purdue University. Jan 11, Bruno Ribeiro

Data Mining. CS57300 Purdue University. Jan 11, Bruno Ribeiro Data Mining CS57300 Purdue University Jan, 208 Bruno Ribeiro Regression Posteriors Working with Data 2 Linear Regression: Review 3 Linear Regression (use A)? Interpolation (something is missing) (x,, x

More information

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation. CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.

More information

Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information:

Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information: Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information: aanum@ug.edu.gh College of Education School of Continuing and Distance Education 2014/2015 2016/2017 Session Overview In this Session

More information

cxx ab.ec Warm up OH 2 ax 16 0 axtb Fix any a, b, c > What is the x 2 R that minimizes ax 2 + bx + c

cxx ab.ec Warm up OH 2 ax 16 0 axtb Fix any a, b, c > What is the x 2 R that minimizes ax 2 + bx + c Warm up D cai.yo.ie p IExrL9CxsYD Sglx.Ddl f E Luo fhlexi.si dbll Fix any a, b, c > 0. 1. What is the x 2 R that minimizes ax 2 + bx + c x a b Ta OH 2 ax 16 0 x 1 Za fhkxiiso3ii draulx.h dp.d 2. What is

More information

Motivation Subgradient Method Stochastic Subgradient Method. Convex Optimization. Lecture 15 - Gradient Descent in Machine Learning

Motivation Subgradient Method Stochastic Subgradient Method. Convex Optimization. Lecture 15 - Gradient Descent in Machine Learning Convex Optimization Lecture 15 - Gradient Descent in Machine Learning Instructor: Yuanzhang Xiao University of Hawaii at Manoa Fall 2017 1 / 21 Today s Lecture 1 Motivation 2 Subgradient Method 3 Stochastic

More information

Lesson 15: Solution Sets of Two or More Equations (or Inequalities) Joined by And or Or

Lesson 15: Solution Sets of Two or More Equations (or Inequalities) Joined by And or Or Solution Sets of Two or More Equations (or Inequalities) Joined by And or Or Classwork 1. Determine whether each claim given below is true or false. a. Right now, I am in math class and English class.

More information

Lecture 11. Linear Soft Margin Support Vector Machines

Lecture 11. Linear Soft Margin Support Vector Machines CS142: Machine Learning Spring 2017 Lecture 11 Instructor: Pedro Felzenszwalb Scribes: Dan Xiang, Tyler Dae Devlin Linear Soft Margin Support Vector Machines We continue our discussion of linear soft margin

More information

Interpreting Deep Classifiers

Interpreting Deep Classifiers Ruprecht-Karls-University Heidelberg Faculty of Mathematics and Computer Science Seminar: Explainable Machine Learning Interpreting Deep Classifiers by Visual Distillation of Dark Knowledge Author: Daniela

More information

CS5314 Randomized Algorithms. Lecture 15: Balls, Bins, Random Graphs (Hashing)

CS5314 Randomized Algorithms. Lecture 15: Balls, Bins, Random Graphs (Hashing) CS5314 Randomized Algorithms Lecture 15: Balls, Bins, Random Graphs (Hashing) 1 Objectives Study various hashing schemes Apply balls-and-bins model to analyze their performances 2 Chain Hashing Suppose

More information

Outline: Ensemble Learning. Ensemble Learning. The Wisdom of Crowds. The Wisdom of Crowds - Really? Crowd wiser than any individual

Outline: Ensemble Learning. Ensemble Learning. The Wisdom of Crowds. The Wisdom of Crowds - Really? Crowd wiser than any individual Outline: Ensemble Learning We will describe and investigate algorithms to Ensemble Learning Lecture 10, DD2431 Machine Learning A. Maki, J. Sullivan October 2014 train weak classifiers/regressors and how

More information

A Probabilistic Model for Canonicalizing Named Entity Mentions. Dani Yogatama Yanchuan Sim Noah A. Smith

A Probabilistic Model for Canonicalizing Named Entity Mentions. Dani Yogatama Yanchuan Sim Noah A. Smith A Probabilistic Model for Canonicalizing Named Entity Mentions Dani Yogatama Yanchuan Sim Noah A. Smith Introduction Model Experiments Conclusions Outline Introduction Model Experiments Conclusions Outline

More information

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1 Week 2 Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Part I Other datatypes, preprocessing 2 / 1 Other datatypes Document data You might start with a collection of

More information