Part 3 Introduction to statistical classification techniques
|
|
- Marylou Edwards
- 5 years ago
- Views:
Transcription
1 Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli
2 Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms P(ω i ) p( / ω i ) And we knw the lss matri Λ Nte: in statistics D is ften called the sample f size n drawn frm the distributin p(). In pattern recgnitin the term sample is usually used fr the single pattern i. ØThe minimum risk thery allws us t design the ptimal classifier (that minimizes the classificatin risk) fr the task at hand ØHwever, in practical cases, we never knw all this infrmatin ØThe nly infrmatin that we usually have is a data set D (called design r training data set) D = [,,., n ] i = ( i, i,., id ) i=,..,n i belnging t ne f the c classes ( i ε ω j j=,,c) ØPatterns i are drawn independently accrding t p( i / ω j ) Machine Learning, Part 3, March 07 Fabi Rli
3 Classificatin techniques ØIf we knw the classes t which the patterns i f the design/training set belng t, we speak f Supervised Classificatin Further infrmatin, beynd the data sert D, that we can have are: Ø We can knw the parametric mdel ( parametric frm ) f the distributin p(/ω i ), s that we can use the Parametric Techniques Ø If we knw nthing abut the distributin p(/ω i ), we are bliged t use the s called Nn-Parametric Techniques ØParametric Techniques: we knw the parametric frm f the distributin p(/ω i ), fr eample, we knw that the distributin is Gaussian ØNn-Parametric Techniques: we knw nthing abut the distributin, and we are nt able t get any infrmatin with an unsupervised analysis. Nte that we are assuming that estimating prirs P(ω i ) is an easy prblem, assumptin that is ften but nt always true. ØHere we are disregarding the csts f classificatin. The reasn is that the chice f cst values is a prblem-dependent issue, very little can be said in general abut this chice. Machine Learning, Part 3, March 07 Fabi Rli 3
4 Classificatin Parametric Techniques ØWe knw, r we assume, a parametric frm f the distributins p(/ω i ). ØThe main prblem is then t estimate the parameters f the mdel (e.g., the mean value and the variance f the Guassian mdel) ØWe discuss in detail these techniques in Part 4. ØThe estimate f the parameters is dne using the data set D, r a subset f it mre ften (t avid a prblem called ver-fitting ). ØHw can we assume a gd parametric mdel f the distributins p(/ω i )? In the practical applicatins we have tw pssibilities t d that: We assume different parametric mdels, we cmpute the parameters fr each mdel, then we cmpare the errrs f the mdels and select the best We use Unsupervised Classificatin Techniques (we see basic cncepts later in Part 9) t gain sme knwledge f the parametric frm f the p(/ ω i ). Unsupervised classificatin Using the data set D we try t gain sme knwledge abut p(/ω i ) (e.g., we discver that it is made up f tw clusters, i.e., it is the sum f tw Gaussian distributins) Machine Learning, Part 3, March 07 Fabi Rli 4
5 Classificatin Nn-Parametric Techniques ØWe knw nthing f the distributin p(/ω i ), and we are nt able t gain knwledge with an unsupervised analysis. ØWe use techniques (Part 5) that allw t estimate the densities p(/ω i ), r the prsterir prbabilities P(ω i /), using the data set D. ØNn-parametric techniques are aimed t estimate directly the density functins p() Machine Learning, Part 3, March 07 Fabi Rli 5
6 Eample f parametric techniques in bimetrics In bimetric recgnitin parametric techniques can be used t mdel genuine and impstr distributins Parametric techniques smetimes prvides perfrmances lwer than the nes f nn parametric techniques Machine Learning, Part 3, March 07 Fabi Rli 6
7 Linear discriminant functins ØIn sme cases it can be mre effective t assume a parametric frm f the discriminant functins g i (), i=,..,c, instead f a parametric frm f the p(/ω i ) (We discuss this in Part 6). ØFr eample, t assume a linear frm f the discriminant functins g i () In sme cases linear functins allw t disciminate well classes that wuld be difficult t mdel by cmputing the distributins p(/ω i ). It is wrth nting that, in the end, what we want t d in many cases is just t classify, nt mdelling the p(/ω i )! Even if a linear discriminat functin des nt prvide the ptimal slutins, hwever the errr rate can be acceptable fr the task at hand! Machine Learning, Part 3, March 07 Fabi Rli 7
8 Design f classifier: basic design cycle We have just a design set D = [,,., n ] Unsupervised analysis D yu knw the frm f p()? NO Nn Parametric techniques YES Parametric techniques Split D int 3 sets: training, validatin, and test set Use training+validatin sets t estimate parameters Split D int 3 sets: training, validatin, and test set Use the validatin set t estimate parameters, and training set t train classifier Use test set t estimate errr prbability We see later that nnparametric techniques have sme parameters t be estimated as well! Machine Learning, Part 3, March 07 Fabi Rli 8
9 Sme ntable cncepts: feature (re)scaling ØFeatures used t characterize patterns are usually linked t physical measurements which have different scales. Given samples in D, feature scales can be very different (e.g, height in meters and weight in kg). This is due t nn-hmgenus physical measurements r the intrinsic scales f different features. ØSlutin: nrmalizatin, (re)scaling f features. The nrmalizatin peratin can be regarded as a functin h j applied t feature that takes as input the riginal feature value ij, and utputs the rescaled(nrmalized) feature value ij = h j ( ij ), with h j being the nrmalizatin functin (j =,,..., d). Machine Learning, Part 3, March 07 Fabi Rli 9
10 Sme nrmalizatin functins Given D = [,,., n ], i = ( i, i,., id ) i=,..,n, nrmalizatin functins h j widely used are the fllwings: We divide the feature ij by maimum value (ver D): ij ij =, j,ma = ma kj k=,,..., n j,ma Divide by maimum range: ij = ij j,ma j,min j,min [0,], j,ma j,min Divide by standard deviatin f feature ij : { } kj ( m ) = ij m m j E j k ij =, σ σ j = E j kj j k { } Machine Learning, Part 3, March 07 Fabi Rli 0 = = ma k=,,..., n min k=,,..., n mˆ j σˆ j = = kj kj n N k= N kj n ( ) kj mˆ j k=
11 Remarks n nrmalizatin The third nrmalizatin methd (divisin by standard deviatin) is useful, fr eample, when feature distributin is Gaussian. If feature ij has a Gaussian distributin, the nrmalized feature ij has a nrmalized Gaussian distributin. Nrmalizatin must be dne using all the patterns available in D and fr each feature separately. Hereafter, we assume that all the features used have been prperly nrmalized, and therefre we mit the ape in ij. Machine Learning, Part 3, March 07 Fabi Rli
12 Sme ntable cncepts: Separatin f classes Definitin f separated class: In a bi-dimensinal feature space (d=), a class is called separated if a curve (clsed r pen) eits s that all the samples f that class lies n the same side f the curve. In a d-dimensinal feature space we have hyper-curves. Tw separated classes can be: Linearly separable, if the curve that separates the tw classes is a linear functin (fr d =, the curve is a straight line); Nn linearly separable, the separatin needs nn-linear curves. Nte that the separatin demands that tw patterns belnging t different classes d nt have the same feature values! S we are speaking f deterministic separatin! Machine Learning, Part 3, March 07 Fabi Rli
13 Ntable cncepts: Multi-mdal classes Ø A data class is multimdal if it cntains clusters f patterns which are linearly separable r it has different peaks f the density functin. Esempi ω (a) (b) (c) ω ω ω ω ω (a) (b) (c) ω (a) tw linearly separable classes, (b) e (c) tw classes nn linearly separable. The class ω in (c) is bimdal. In (a) and (c) statistical methds wrk well, the case (b) is much mre difficult. Machine Learning, Part 3, March 07 Fabi Rli 3
14 A ntable cncept: gemetrical cmpleity f classes Characteristics f a class als dipends n the gemetrical features f the data distributin in the feature space. In particular, if classes have elngated distributins and/r are much verlapped, sme techniques wrk prly. Eample it is difficult t discriminate samples in regins where the tw classes are very verlapped. Each class in the figure have a privileged directin in the feature space. Features have a very high crrelatin (cnditinal crrelatin given the class). Machine Learning, Part 3, March 07 Fabi Rli 4
15 Crrelatin Cefficient Crrelatin between tw features i ed j can be measured by the cefficient f crrelatin ρ ij (i, j =,,..., d). It is linked t the variance σ ij = E{( i i )( j j )} and the feature variances σ ii and σ jj by: ρ ij = σ jj σ ii If d is the feature number, [ρ ij ] is a squared matri d d, cn ρ ij i, j =,,..., d e ρ ii = (main diagnal) i =,,..., d. feature i and j are crrelated if ρ ij has a high value (e.g. > 0.8). σ ij The analysis f crrelatin can be dne fr each class and fr the whle data set. Machine Learning, Part 3, March 07 Fabi Rli 5
16 Ntable cncepts: Gemetrical vs. Prbabilistic cmpleity Square 44 Square 00 Prbabilistic cmpleity I must recgnize ne pattern ut f ne millin! Tw very unbalanced classes! The prblem has simple gemetrical features, but it is very hard! Eample f Gemetrical Cmpleity Machine Learning, Part 3, March 07 Fabi Rli 6
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised
More informationPattern Recognition 2014 Support Vector Machines
Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft
More informationLecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised
More informationx 1 Outline IAML: Logistic Regression Decision Boundaries Example Data
Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares
More informationSimple Linear Regression (single variable)
Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins
More informationCOMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification
COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551
More informationThe Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition
The Kullback-Leibler Kernel as a Framewrk fr Discriminant and Lcalized Representatins fr Visual Recgnitin Nun Vascncels Purdy H Pedr Mren ECE Department University f Califrnia, San Dieg HP Labs Cambridge
More informationk-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels
Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t
More informationComputational modeling techniques
Cmputatinal mdeling techniques Lecture 4: Mdel checing fr ODE mdels In Petre Department f IT, Åb Aademi http://www.users.ab.fi/ipetre/cmpmd/ Cntent Stichimetric matrix Calculating the mass cnservatin relatins
More informationResampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017
Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with
More informationCOMP 551 Applied Machine Learning Lecture 4: Linear classification
COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted
More informationTree Structured Classifier
Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients
More informationThe blessing of dimensionality for kernel methods
fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented
More informationSupport-Vector Machines
Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material
More informationMidwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter
Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline
More informationBootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >
Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);
More informationDistributions, spatial statistics and a Bayesian perspective
Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics
More informationReinforcement Learning" CMPSCI 383 Nov 29, 2011!
Reinfrcement Learning" CMPSCI 383 Nv 29, 2011! 1 Tdayʼs lecture" Review f Chapter 17: Making Cmple Decisins! Sequential decisin prblems! The mtivatin and advantages f reinfrcement learning.! Passive learning!
More informationChecking the resolved resonance region in EXFOR database
Checking the reslved resnance regin in EXFOR database Gttfried Bertn Sciété de Calcul Mathématique (SCM) Oscar Cabells OECD/NEA Data Bank JEFF Meetings - Sessin JEFF Experiments Nvember 0-4, 017 Bulgne-Billancurt,
More informationIAML: Support Vector Machines
1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int
More information1 The limitations of Hartree Fock approximation
Chapter: Pst-Hartree Fck Methds - I The limitatins f Hartree Fck apprximatin The n electrn single determinant Hartree Fck wave functin is the variatinal best amng all pssible n electrn single determinants
More informationCOMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)
COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise
More informationThe general linear model and Statistical Parametric Mapping I: Introduction to the GLM
The general linear mdel and Statistical Parametric Mapping I: Intrductin t the GLM Alexa Mrcm and Stefan Kiebel, Rik Hensn, Andrew Hlmes & J-B J Pline Overview Intrductin Essential cncepts Mdelling Design
More information3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression
3.3.4 Prstate Cancer Data Example (Cntinued) 3.4 Shrinkage Methds 61 Table 3.3 shws the cefficients frm a number f different selectin and shrinkage methds. They are best-subset selectin using an all-subsets
More informationModelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA
Mdelling f Clck Behaviur Dn Percival Applied Physics Labratry University f Washingtn Seattle, Washingtn, USA verheads and paper fr talk available at http://faculty.washingtn.edu/dbp/talks.html 1 Overview
More informationChapter 3: Cluster Analysis
Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA
More information, which yields. where z1. and z2
The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin
More informationThe standards are taught in the following sequence.
B L U E V A L L E Y D I S T R I C T C U R R I C U L U M MATHEMATICS Third Grade In grade 3, instructinal time shuld fcus n fur critical areas: (1) develping understanding f multiplicatin and divisin and
More informationComparison of hybrid ensemble-4dvar with EnKF and 4DVar for regional-scale data assimilation
Cmparisn f hybrid ensemble-4dvar with EnKF and 4DVar fr reginal-scale data assimilatin Jn Pterjy and Fuqing Zhang Department f Meterlgy The Pennsylvania State University Wednesday 18 th December, 2013
More informationMODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:
MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use
More informationMODULE 1. e x + c. [You can t separate a demominator, but you can divide a single denominator into each numerator term] a + b a(a + b)+1 = a + b
. REVIEW OF SOME BASIC ALGEBRA MODULE () Slving Equatins Yu shuld be able t slve fr x: a + b = c a d + e x + c and get x = e(ba +) b(c a) d(ba +) c Cmmn mistakes and strategies:. a b + c a b + a c, but
More informationDifferentiation Applications 1: Related Rates
Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm
More informationA Matrix Representation of Panel Data
web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins
More informationStatistical classifiers: Bayesian decision theory and density estimation
3 rd NOSE Shrt Curse Alpbach, st 6 th Mar 004 Statistical classifiers: Bayesian decisin thery and density estimatin Ricard Gutierrez- Department f Cmputer Science rgutier@cs.tamu.edu http://research.cs.tamu.edu/prism
More informationIn SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:
In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin
More informationHypothesis Tests for One Population Mean
Hypthesis Tests fr One Ppulatin Mean Chapter 9 Ala Abdelbaki Objective Objective: T estimate the value f ne ppulatin mean Inferential statistics using statistics in rder t estimate parameters We will be
More information5 th grade Common Core Standards
5 th grade Cmmn Cre Standards In Grade 5, instructinal time shuld fcus n three critical areas: (1) develping fluency with additin and subtractin f fractins, and develping understanding f the multiplicatin
More informationSUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis
SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm
More informationWhat is Statistical Learning?
What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,
More informationInternal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.
Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.
More informationSlide04 (supplemental) Haykin Chapter 4 (both 2nd and 3rd ed): Multi-Layer Perceptrons
Slide04 supplemental) Haykin Chapter 4 bth 2nd and 3rd ed): Multi-Layer Perceptrns CPSC 636-600 Instructr: Ynsuck Che Heuristic fr Making Backprp Perfrm Better 1. Sequential vs. batch update: fr large
More informationExam #1. A. Answer any 1 of the following 2 questions. CEE 371 March 10, Please grade the following questions: 1 or 2
CEE 371 March 10, 2009 Exam #1 Clsed Bk, ne sheet f ntes allwed Please answer ne questin frm the first tw, ne frm the secnd tw and ne frm the last three. The ttal ptential number f pints is 100. Shw all
More informationExam #1. A. Answer any 1 of the following 2 questions. CEE 371 October 8, Please grade the following questions: 1 or 2
CEE 371 Octber 8, 2009 Exam #1 Clsed Bk, ne sheet f ntes allwed Please answer ne questin frm the first tw, ne frm the secnd tw and ne frm the last three. The ttal ptential number f pints is 100. Shw all
More informationLecture 8: Multiclass Classification (I)
Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Lecture 8: Multiclass Classificatin (I) Ha Helen Zhang Fall 07 Ha Helen Zhang Lecture 8: Multiclass Classificatin
More informationLinear Classification
Linear Classificatin CS 54: Machine Learning Slides adapted frm Lee Cper, Jydeep Ghsh, and Sham Kakade Review: Linear Regressin CS 54 [Spring 07] - H Regressin Given an input vectr x T = (x, x,, xp), we
More informationFall 2013 Physics 172 Recitation 3 Momentum and Springs
Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.
More informationB. Definition of an exponential
Expnents and Lgarithms Chapter IV - Expnents and Lgarithms A. Intrductin Starting with additin and defining the ntatins fr subtractin, multiplicatin and divisin, we discvered negative numbers and fractins.
More information7 TH GRADE MATH STANDARDS
ALGEBRA STANDARDS Gal 1: Students will use the language f algebra t explre, describe, represent, and analyze number expressins and relatins 7 TH GRADE MATH STANDARDS 7.M.1.1: (Cmprehensin) Select, use,
More informationDrought damaged area
ESTIMATE OF THE AMOUNT OF GRAVEL CO~TENT IN THE SOIL BY A I R B O'RN EMS S D A T A Y. GOMI, H. YAMAMOTO, AND S. SATO ASIA AIR SURVEY CO., l d. KANAGAWA,JAPAN S.ISHIGURO HOKKAIDO TOKACHI UBPREFECTRAl OffICE
More informationComparing Several Means: ANOVA. Group Means and Grand Mean
STAT 511 ANOVA and Regressin 1 Cmparing Several Means: ANOVA Slide 1 Blue Lake snap beans were grwn in 12 pen-tp chambers which are subject t 4 treatments 3 each with O 3 and SO 2 present/absent. The ttal
More informationCS 109 Lecture 23 May 18th, 2016
CS 109 Lecture 23 May 18th, 2016 New Datasets Heart Ancestry Netflix Our Path Parameter Estimatin Machine Learning: Frmally Many different frms f Machine Learning We fcus n the prblem f predictin Want
More informationCAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank
CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal
More informationNUMBERS, MATHEMATICS AND EQUATIONS
AUSTRALIAN CURRICULUM PHYSICS GETTING STARTED WITH PHYSICS NUMBERS, MATHEMATICS AND EQUATIONS An integral part t the understanding f ur physical wrld is the use f mathematical mdels which can be used t
More information3. Classify the following Numbers (Counting (natural), Whole, Integers, Rational, Irrational)
After yu cmplete each cncept give yurself a rating 1. 15 5 2 (5 3) 2. 2 4-8 (2 5) 3. Classify the fllwing Numbers (Cunting (natural), Whle, Integers, Ratinal, Irratinal) a. 7 b. 2 3 c. 2 4. Are negative
More informationLim f (x) e. Find the largest possible domain and its discontinuity points. Why is it discontinuous at those points (if any)?
THESE ARE SAMPLE QUESTIONS FOR EACH OF THE STUDENT LEARNING OUTCOMES (SLO) SET FOR THIS COURSE. SLO 1: Understand and use the cncept f the limit f a functin i. Use prperties f limits and ther techniques,
More informationResampling Methods. Chapter 5. Chapter 5 1 / 52
Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and
More informationEE247B/ME218: Introduction to MEMS Design Lecture 7m1: Lithography, Etching, & Doping CTN 2/6/18
EE247B/ME218 Intrductin t MEMS Design Lecture 7m1 Lithgraphy, Etching, & Dping Dping f Semicnductrs Semicnductr Dping Semicnductrs are nt intrinsically cnductive T make them cnductive, replace silicn atms
More informationCHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.
MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the
More information1 PreCalculus AP Unit G Rotational Trig (MCR) Name:
1 PreCalculus AP Unit G Rtatinal Trig (MCR) Name: Big idea In this unit yu will extend yur knwledge f SOH CAH TOA t wrk with btuse and reflex angles. This extensin will invlve the unit circle which will
More informationInference in the Multiple-Regression
Sectin 5 Mdel Inference in the Multiple-Regressin Kinds f hypthesis tests in a multiple regressin There are several distinct kinds f hypthesis tests we can run in a multiple regressin. Suppse that amng
More informationHomology groups of disks with holes
Hmlgy grups f disks with hles THEOREM. Let p 1,, p k } be a sequence f distinct pints in the interir unit disk D n where n 2, and suppse that fr all j the sets E j Int D n are clsed, pairwise disjint subdisks.
More informationEvaluating enterprise support: state of the art and future challenges. Dirk Czarnitzki KU Leuven, Belgium, and ZEW Mannheim, Germany
Evaluating enterprise supprt: state f the art and future challenges Dirk Czarnitzki KU Leuven, Belgium, and ZEW Mannheim, Germany Intrductin During the last decade, mircecnmetric ecnmetric cunterfactual
More informationPublic Key Cryptography. Tim van der Horst & Kent Seamons
Public Key Cryptgraphy Tim van der Hrst & Kent Seamns Last Updated: Oct 5, 2017 Asymmetric Encryptin Why Public Key Crypt is Cl Has a linear slutin t the key distributin prblem Symmetric crypt has an expnential
More informationWRITING THE REPORT. Organizing the report. Title Page. Table of Contents
WRITING THE REPORT Organizing the reprt Mst reprts shuld be rganized in the fllwing manner. Smetime there is a valid reasn t include extra chapters in within the bdy f the reprt. 1. Title page 2. Executive
More informationALE 21. Gibbs Free Energy. At what temperature does the spontaneity of a reaction change?
Name Chem 163 Sectin: Team Number: ALE 21. Gibbs Free Energy (Reference: 20.3 Silberberg 5 th editin) At what temperature des the spntaneity f a reactin change? The Mdel: The Definitin f Free Energy S
More informationCOMP 551 Applied Machine Learning Lecture 11: Support Vector Machines
COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse
More informationElements of Machine Intelligence - I
ECE-175A Elements f Machine Intelligence - I Ken Kreutz-Delgad Nun Vascncels ECE Department, UCSD Winter 2011 The curse The curse will cver basic, but imprtant, aspects f machine learning and pattern recgnitin
More informationLCAO APPROXIMATIONS OF ORGANIC Pi MO SYSTEMS The allyl system (cation, anion or radical).
Principles f Organic Chemistry lecture 5, page LCAO APPROIMATIONS OF ORGANIC Pi MO SYSTEMS The allyl system (catin, anin r radical).. Draw mlecule and set up determinant. 2 3 0 3 C C 2 = 0 C 2 3 0 = -
More informationAP Statistics Notes Unit Two: The Normal Distributions
AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).
More informationArtificial Neural Networks MLP, Backpropagation
Artificial Neural Netwrks MLP, Backprpagatin 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001 00100000
More informationDepartment of Economics, University of California, Davis Ecn 200C Micro Theory Professor Giacomo Bonanno. Insurance Markets
Department f Ecnmics, University f alifrnia, Davis Ecn 200 Micr Thery Prfessr Giacm Bnann Insurance Markets nsider an individual wh has an initial wealth f. ith sme prbability p he faces a lss f x (0
More informationCOMP9444 Neural Networks and Deep Learning 3. Backpropagation
COMP9444 Neural Netwrks and Deep Learning 3. Backprpagatin Tetbk, Sectins 4.3, 5.2, 6.5.2 COMP9444 17s2 Backprpagatin 1 Outline Supervised Learning Ockham s Razr (5.2) Multi-Layer Netwrks Gradient Descent
More informationTrigonometric Ratios Unit 5 Tentative TEST date
1 U n i t 5 11U Date: Name: Trignmetric Ratis Unit 5 Tentative TEST date Big idea/learning Gals In this unit yu will extend yur knwledge f SOH CAH TOA t wrk with btuse and reflex angles. This extensin
More informationFlipping Physics Lecture Notes: Simple Harmonic Motion Introduction via a Horizontal Mass-Spring System
Flipping Physics Lecture Ntes: Simple Harmnic Mtin Intrductin via a Hrizntal Mass-Spring System A Hrizntal Mass-Spring System is where a mass is attached t a spring, riented hrizntally, and then placed
More informationELT COMMUNICATION THEORY
ELT 41307 COMMUNICATION THEORY Matlab Exercise #2 Randm variables and randm prcesses 1 RANDOM VARIABLES 1.1 ROLLING A FAIR 6 FACED DICE (DISCRETE VALIABLE) Generate randm samples fr rlling a fair 6 faced
More informationCS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007
CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is
More informationMATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank
MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use
More informationAdmin. MDP Search Trees. Optimal Quantities. Reinforcement Learning
Admin Reinfrcement Learning Cntent adapted frm Berkeley CS188 MDP Search Trees Each MDP state prjects an expectimax-like search tree Optimal Quantities The value (utility) f a state s: V*(s) = expected
More informationA New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation
III-l III. A New Evaluatin Measure J. Jiner and L. Werner Abstract The prblems f evaluatin and the needed criteria f evaluatin measures in the SMART system f infrmatin retrieval are reviewed and discussed.
More informationEquilibrium of Stress
Equilibrium f Stress Cnsider tw perpendicular planes passing thrugh a pint p. The stress cmpnents acting n these planes are as shwn in ig. 3.4.1a. These stresses are usuall shwn tgether acting n a small
More informationLocalized Model Selection for Regression
Lcalized Mdel Selectin fr Regressin Yuhng Yang Schl f Statistics University f Minnesta Church Street S.E. Minneaplis, MN 5555 May 7, 007 Abstract Research n mdel/prcedure selectin has fcused n selecting
More informationMATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank
MATCHING TECHNIQUES Technical Track Sessin VI Céline Ferré The Wrld Bank When can we use matching? What if the assignment t the treatment is nt dne randmly r based n an eligibility index, but n the basis
More informationAdministrativia. Assignment 1 due thursday 9/23/2004 BEFORE midnight. Midterm exam 10/07/2003 in class. CS 460, Sessions 8-9 1
Administrativia Assignment 1 due thursday 9/23/2004 BEFORE midnight Midterm eam 10/07/2003 in class CS 460, Sessins 8-9 1 Last time: search strategies Uninfrmed: Use nly infrmatin available in the prblem
More informationInterference is when two (or more) sets of waves meet and combine to produce a new pattern.
Interference Interference is when tw (r mre) sets f waves meet and cmbine t prduce a new pattern. This pattern can vary depending n the riginal wave directin, wavelength, amplitude, etc. The tw mst extreme
More informationEDA Engineering Design & Analysis Ltd
EDA Engineering Design & Analysis Ltd THE FINITE ELEMENT METHOD A shrt tutrial giving an verview f the histry, thery and applicatin f the finite element methd. Intrductin Value f FEM Applicatins Elements
More informationActivity Guide Loops and Random Numbers
Unit 3 Lessn 7 Name(s) Perid Date Activity Guide Lps and Randm Numbers CS Cntent Lps are a relatively straightfrward idea in prgramming - yu want a certain chunk f cde t run repeatedly - but it takes a
More informationSIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A.
SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST Mark C. Ott Statistics Research Divisin, Bureau f the Census Washingtn, D.C. 20233, U.S.A. and Kenneth H. Pllck Department f Statistics, Nrth Carlina State
More informationPhysical Layer: Outline
18-: Intrductin t Telecmmunicatin Netwrks Lectures : Physical Layer Peter Steenkiste Spring 01 www.cs.cmu.edu/~prs/nets-ece Physical Layer: Outline Digital Representatin f Infrmatin Characterizatin f Cmmunicatin
More informationECE 2100 Circuit Analysis
ECE 2100 Circuit Analysis Lessn 25 Chapter 9 & App B: Passive circuit elements in the phasr representatin Daniel M. Litynski, Ph.D. http://hmepages.wmich.edu/~dlitynsk/ ECE 2100 Circuit Analysis Lessn
More informationOn Topological Structures and. Fuzzy Sets
L - ZHR UNIVERSIT - GZ DENSHIP OF GRDUTE STUDIES & SCIENTIFIC RESERCH On Tplgical Structures and Fuzzy Sets y Nashaat hmed Saleem Raab Supervised by Dr. Mhammed Jamal Iqelan Thesis Submitted in Partial
More informationEarly detection of mining truck failure by modelling its operation with neural networks classification algorithms
RU, Rand GOLOSINSKI, T.S. Early detectin f mining truck failure by mdelling its peratin with neural netwrks classificatin algrithms. Applicatin f Cmputers and Operatins Research ill the Minerals Industries,
More informationAdmissibility Conditions and Asymptotic Behavior of Strongly Regular Graphs
Admissibility Cnditins and Asympttic Behavir f Strngly Regular Graphs VASCO MOÇO MANO Department f Mathematics University f Prt Oprt PORTUGAL vascmcman@gmailcm LUÍS ANTÓNIO DE ALMEIDA VIEIRA Department
More informationLHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers
LHS Mathematics Department Hnrs Pre-alculus Final Eam nswers Part Shrt Prblems The table at the right gives the ppulatin f Massachusetts ver the past several decades Using an epnential mdel, predict the
More information22.54 Neutron Interactions and Applications (Spring 2004) Chapter 11 (3/11/04) Neutron Diffusion
.54 Neutrn Interactins and Applicatins (Spring 004) Chapter (3//04) Neutrn Diffusin References -- J. R. Lamarsh, Intrductin t Nuclear Reactr Thery (Addisn-Wesley, Reading, 966) T study neutrn diffusin
More informationMore Tutorial at
Answer each questin in the space prvided; use back f page if extra space is needed. Answer questins s the grader can READILY understand yur wrk; nly wrk n the exam sheet will be cnsidered. Write answers,
More informationA Few Basic Facts About Isothermal Mass Transfer in a Binary Mixture
Few asic Facts but Isthermal Mass Transfer in a inary Miture David Keffer Department f Chemical Engineering University f Tennessee first begun: pril 22, 2004 last updated: January 13, 2006 dkeffer@utk.edu
More informationSequential Allocation with Minimal Switching
In Cmputing Science and Statistics 28 (1996), pp. 567 572 Sequential Allcatin with Minimal Switching Quentin F. Stut 1 Janis Hardwick 1 EECS Dept., University f Michigan Statistics Dept., Purdue University
More informationAristotle I PHIL301 Prof. Oakes Winthrop University updated: 3/14/14 8:48 AM
Aristtle I PHIL301 Prf. Oakes Winthrp University updated: 3/14/14 8:48 AM The Categries - The Categries is ne f several imprtant wrks by Aristtle n metaphysics. His tpic here is the classificatin f beings
More information4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression
4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw
More informationAssessment Primer: Writing Instructional Objectives
Assessment Primer: Writing Instructinal Objectives (Based n Preparing Instructinal Objectives by Mager 1962 and Preparing Instructinal Objectives: A critical tl in the develpment f effective instructin
More information