LECTURE 17: Linear Discriminant Functions

Size: px
Start display at page:

Download "LECTURE 17: Linear Discriminant Functions"

Transcription

1 LECURE 7: Liear Discrimiat Fuctios Perceptro leari Miimum squared error (MSE) solutio Least-mea squares (LMS) rule Ho-Kashyap procedure Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity

2 Liear Discrimiat Fuctios () he objective of this lecture is to preset methods for leari liear discrimiat fuctios of the form ( x) = w x + w ( x) ( x) > < x ω x ω 2 where w is the weiht vector ad w is the threshold weiht or bias (ot to be cofused with that of the bias-variace dilemma) x 2 w x+w > x ( w x+w < d x x (2 w x Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 2

3 Liear Discrimiat Fuctios (2) Similar discrimiat fuctios were derived i Lecture 5 as a special case of the quadratic classifier I this lecture, the discrimiat fuctios will be derived i a o- parametric fashio, this is, o assumptios will be made about the uderlyi desities For coveiece, we will focus o the biary classificatio problem Extesio to the multicateory case ca be easily achieved by Usi ω i /ot ω i dichotomies Usi ω i /ω i dichotomies Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 3

4 Gradiet descet Gradiet descet is eeral method for fuctio miimizatio Recall that the miimum of a fuctio J(x) is defied by the zeros of the radiet Oly i very special cases this miimizatio fuctio has a closed form solutio I some other cases, a closed form solutio may exist, but is umerically illposed or impractical (e.., memory requiremets) Gradiet descet fids the miimum i a iterative fashio by movi i the directio of steepest descet where η is a leari rate [ ] J(x) = x* = armi J(x) x x. Start with a arbitrary solutio x() 2. Compute the radiet x J(x(k)) 3. Move i the directio of steepest descet: x( k + ) = x( k) η xj( x( k) ) 4. Go to (util coverece) x 2 2 Local miimum Iitial uess Global miimum x Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 4

5 Perceptro leari () Let s ow cosider the problem of leari a biary classificatio problem with a liear discrimiat fuctio As usual, assume we have a dataset X={x (,x (2, x (N } cotaii examples from the two classes For coveiece, we will absorb the itercept w by aumeti the feature vector x with a additioal costat dimesio: w x + w = [ w w ] = a y x Keep i mid that our objective is to fid a vector a such that > x ω ( x) = a y < x ω2 o simplify the derivatio, we will ormalize the traii set by replaci all examples from class ω 2 by their eative y [ y] y ω2 his allows us to iore class labels ad look for a weiht vector such that a y > y Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity From [Duda, Hart ad Stork, 2] 5

6 Perceptro leari (2) o fid this solutio we must first defie a objective fuctio J(a) A ood choice is what is kow as the Perceptro criterio fuctio where Y M is the set of examples misclassified by a. Note that J P (a) is o-eative sice a y< for all misclassified samples o fid the miimum of this fuctio, we use radiet descet he radiet is defied by Ad the radiet descet update rule becomes his is kow as the perceptro batch update rule. a J a P ( a) = ( a y) J P y Υ M ( a) = ( y) y Υ M ( k + ) = a( k) + η he weiht vector may also be updated i a o-lie fashio, this is, after the presetatio of each idividual example y Υ M y ( k ) Perceptro rule (i ( k + ) = a( k) ηy a + where y (i is a example that has bee misclassified by a(k) Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 6

7 Perceptro leari (3) If classes are liearly separable, the perceptro rule is uarateed to covere to a valid solutio Some versio of the perceptro rule use a variable leari rate η(k) I this case, coverece is uarateed oly uder certai coditios (for details refer to [Duda, Hart ad Stork, 2], pp ) However, if the two classes are ot liearly separable, the perceptro rule will ot covere Sice o weiht vector a ca correctly classify every sample i a o-separable dataset, the correctios i the perceptro rule will ever cease Oe ad-hoc solutio to this problem is to eforce coverece by usi variable leari rates η(k) that approach zero as k approaches ifiite Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 7

8 Miimum Squared Error solutio () he classical Miimum Squared Error (MSE) criterio provides a alterative to the perceptro rule he perceptro rule seeks a weiht vector a that satisfies the iequality a y (i > he perceptro rule oly cosiders misclassified samples, sice these are the oly oes that violate the above iequality Istead, the MSE criterio looks for a solutio to the equality a y (i =b (i, where b (i are some pre-specified taret values (e.., class labels) As a result, the MSE solutio uses ALL of the samples i the traii set: he system of equatios solved by MSE is where a is the weiht vector, each row i Y is a traii example, ad each row i b is the correspodi class label y y M M y ( (2 (N y y y ( (2 M M (N L L L y y y ( D (2 D M M (N D a a M a D b b = M M b Ya = b For cosistecy, we will cotiue assumi that examples from class ω 2 have bee replaced by their eative vector, althouh this is ot a requiremet for the MSE solutio ( (2 (N Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity From [Duda, Hart ad Stork, 2] 8

9 Miimum Squared Error solutio (2) A exact solutio to Ya=b ca sometimes be foud If the umber of (idepedet) equatios (N) is equal to the umber of ukows (D+), the exact solutio is defied by a = Y b I practice, however, Y will be siular so its iverse Y - does ot exist Y will commoly have more rows (examples) tha colums (ukow), which yields a over-determied system, for which a exact solutio caot be foud he solutio i this case is to fid a weiht vector that miimizes some fuctio of the error betwee the model (ay) ad the desired output (b) I particular, MSE seeks to Miimize the sum of the Squares of these Errors: N (i (i 2 2 ( a) ( a y b ) Ya -b J MSE = = i= which, as usual, ca be foud by setti its radiet to zero Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 9

10 he pseudo-iverse solutio he radiet of the objective fuctio is with zeros defied by N (i (i (i ( a) = 2( a y b ) y = 2Y ( Ya b) a JMSE = i= Ya = Y Notice that Y Y is ow a square matrix! If Y Y is osiular, the MSE solutio becomes Y b - ( Y Y) Y b = Y b a = Pseudo-iverse solutio where the matrix Y =(Y Y) - Y is kow as the pseudo-iverse of Y (Y Y=I) Note that, i eeral, YY I i eeral Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity

11 Ride-reressio solutio If the traii data is extremely correlated (colliearity problem), the matrix Y Y becomes ear siular As a result, the smaller eievalues (the oise) domiate the computatio of the iverse (Y Y) -, which results i umerical problems he colliearity problem ca be solved throuh reularizatio his is equivalet to addi a small multiple of the idetity matrix to the term Y Y, which results i tr ( ) ( Y Y) a = - ε Y Y + ε I Y b D Ride Reressio - where ε (<ε<) is a reularizatio parameter that cotrols the amout of shrikae to the idetity matrix. his is kow as the ride-reressio solutio If the features have siificatly differet variaces, the reularizatio term may be replaced by a diaoal matrix of the feature variaces Selectio of the reularizatio parameter For ε=, ride-reressio solutio is equivalet to the pseudo-iverse solutio For ε=, the ride-reressio solutio is a costat fuctio that predicts the averae classificatio rate across the etire dataset A appropriate value for ε is typically foud throuh cross-validatio Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity From [Gutierrez-Osua, 22]

12 Least-mea-squares solutio he objective fuctio J MSE (a)= Ya-b 2 ca also be foud usi a radiet descet procedure his avoids the problems that arise whe Y Y is siular I additio, it also avoids the eed for worki with lare matrices Looki at the expressio of the radiet, the obvious update rule is ( + ) = a( k) + η( k) Y ( b - Ya( k) ) a k It ca be show that if η(k)=η()/k, where η() is ay positive costat, this rule eerates a sequece of vectors that covere to a solutio to Y (Ya-b)= he storae requiremets of this alorithm ca be reduced by cosideri each sample sequetially ( (i ) y (i ( k ) = a( k) η( k) b - y (i a( k) a + + LMS rule his is kow as the Widrow-Hoff, least-mea-squares (LMS) or delta rule [Mitchell, 997] Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity From [Duda, Hart ad Stork, 2] 2

13 Numerical example Compute the perceptro ad MSE solutio for the followi dataset X = [ (,6), (7,2), (8,9), (9,9)] X2 = [ (2,), (2,2), (2,4), (7,)] x 2 Assume η= Y = 2 Assume a()=[.,.,.] Use a olie update rule 2 4 SOLUION 7 Normalize the dataset x Iterate throuh all the examples ad update a(k) o the oes that are misclassified. Y() [ 6]*[...] > o update 2. Y(2) [ 7 2]*[...] > o update Y(5) [- -2 -]*[...] < update a() = [...] + η*[- -2 -] = [ -. ] 5. Y(6) [- -2-2]*[ -. ] > o update Y() [ 6]*[ -. ] < update a(2) = [ -. ] +η*[ 6] = [..6] 8. Y(2) [ 7 2]*[..6] > o update 9.. I this example, the perceptro rule coveres after 75 iteratios to a=[ ] o covice yourself this is a solutio, compute Y*a (you will fid out that all of the terms are o-eative) Perceptro leai MSE he MSE solutio is foud i oe shot as a=(y Y) - Y b=[ ]; For the choice of tarets b = [ ] As you ca see i the fiure, the MSE solutio misclassifies oe of the samples 6 8 MSE Perceptro Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 3

14 Summary: Perceptro vs. MSE procedures Perceptro rule he perceptro rule always fids a solutio if the classes are liearly separable, but does ot covere if the classes are o-separable MSE criterio he MSE solutio has uarateed coverece, but it may ot fid a separati hyperplae if classes are liearly separable Notice that MSE tries to miimize the sum of the squares of the distaces of the traii data to the separati hyperplae, as opposed to fidi this hyperplae Perceptro x 2 LMS x Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 4

15 he Ho-Kashyap procedure () he mai limitatio of the MSE criterio is the lack of uaratees that a separati hyperplae will be foud i the liearly separable case All we ca say about the MSE rule is that it miimizes Ya-b 2 Whether MSE fids a separati hyperplae or ot depeds o how properly the taret outputs b (i are selected Now, if the two classes are liearly separable, there must exist vectors a* ad b* such that Ya*=b*> If b* were kow, oe could simply use the MSE solutio (a=y b) to compute the separati hyperplae However, sice b* is ukow, oe must the solve for BOH a ad b his idea ives rise to a alterative traii alorithm for liear discrimiat fuctios kow as the Ho-Kashyap procedure ) Fid the taret values b throuh radiet descet 2) Compute the weiht vector a from the MSE solutio 3) Repeat ) ad 2) util coverece Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity Here we also assume y [ y] y ω 2 ) 5

16 he Ho-Kashyap procedure (2) he radiet b J is defied by which suest a possible update rule for b Now, sice b is subject to the costrait b>, we are ot free to follow whichever directio the radiet may poit to he solutio is to start with a iitial solutio b>, ad refuse to reduce ay of its compoets his is accomplished by setti to zero all the positive compoets of the radiet b J b J MSE ( a,b) = 2( Ya b) J b = bj - bj 2 where deotes the absolute value of the compoets of the arumet vector ( ) [ ] Oce b is updated, the MSE solutio a=y b directly provides the zeros of a J he resulti iterative procedure is b a ( k + ) = b( k) η [ J- J] 2 ( k + ) = Y b( k + ) b b Ho-Kashyap procedure Itroductio to Patter Aalysis Ricardo Gutierrez-Osua exas A&M Uiversity 6

10/2/ , 5.9, Jacob Hays Amit Pillay James DeFelice

10/2/ , 5.9, Jacob Hays Amit Pillay James DeFelice 0//008 Liear Discrimiat Fuctios Jacob Hays Amit Pillay James DeFelice 5.8, 5.9, 5. Miimum Squared Error Previous methods oly worked o liear separable cases, by lookig at misclassified samples to correct

More information

Fast Power Flow Methods 1.0 Introduction

Fast Power Flow Methods 1.0 Introduction Fast ower Flow Methods. Itroductio What we have leared so far is the so-called full- ewto-raphso R power flow alorithm. The R alorithm is perhaps the most robust alorithm i the sese that it is most liely

More information

Pattern recognition systems Laboratory 10 Linear Classifiers and the Perceptron Algorithm

Pattern recognition systems Laboratory 10 Linear Classifiers and the Perceptron Algorithm Patter recogitio systems Laboratory 10 Liear Classifiers ad the Perceptro Algorithm 1. Objectives his laboratory sessio presets the perceptro learig algorithm for the liear classifier. We will apply gradiet

More information

Transient Response Analysis for Temperature Modulated Chemoresistors

Transient Response Analysis for Temperature Modulated Chemoresistors Trasiet Respose Aalysis for Temperature Modulated Chemoresistors R. Gutierrez-Osua 1,2, A. Gutierrez 1,2 ad N. Powar 2 1 Texas A&M Uiversity, Collee Statio, TX 2 Wriht State Uiversity, Dayto, OH Multi-frequecy

More information

Dimensionality Reduction Algorithms!! Lecture 15!

Dimensionality Reduction Algorithms!! Lecture 15! Dimesioality Reductio Alorithms!! Lecture 15! 1! Hih-dimesioal Spaces! Couter-ituitive properties! Volume of a d dimesioal sphere of radius r V d (r) = 2 d / 2 r d d# d & ) % 2( where Γ() is the amma fuctio.

More information

Pattern recognition systems Lab 10 Linear Classifiers and the Perceptron Algorithm

Pattern recognition systems Lab 10 Linear Classifiers and the Perceptron Algorithm Patter recogitio systems Lab 10 Liear Classifiers ad the Perceptro Algorithm 1. Objectives his lab sessio presets the perceptro learig algorithm for the liear classifier. We will apply gradiet descet ad

More information

Recurrence Relations

Recurrence Relations Recurrece Relatios Aalysis of recursive algorithms, such as: it factorial (it ) { if (==0) retur ; else retur ( * factorial(-)); } Let t be the umber of multiplicatios eeded to calculate factorial(). The

More information

The Method of Least Squares. To understand least squares fitting of data.

The Method of Least Squares. To understand least squares fitting of data. The Method of Least Squares KEY WORDS Curve fittig, least square GOAL To uderstad least squares fittig of data To uderstad the least squares solutio of icosistet systems of liear equatios 1 Motivatio Curve

More information

Introduction to Machine Learning DIS10

Introduction to Machine Learning DIS10 CS 189 Fall 017 Itroductio to Machie Learig DIS10 1 Fu with Lagrage Multipliers (a) Miimize the fuctio such that f (x,y) = x + y x + y = 3. Solutio: The Lagragia is: L(x,y,λ) = x + y + λ(x + y 3) Takig

More information

Support vector machine revisited

Support vector machine revisited 6.867 Machie learig, lecture 8 (Jaakkola) 1 Lecture topics: Support vector machie ad kerels Kerel optimizatio, selectio Support vector machie revisited Our task here is to first tur the support vector

More information

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise)

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise) Lecture 22: Review for Exam 2 Basic Model Assumptios (without Gaussia Noise) We model oe cotiuous respose variable Y, as a liear fuctio of p umerical predictors, plus oise: Y = β 0 + β X +... β p X p +

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Patter Recogitio Classificatio: No-Parametric Modelig Hamid R. Rabiee Jafar Muhammadi Sprig 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Ageda Parametric Modelig No-Parametric Modelig

More information

Perceptron. Inner-product scalar Perceptron. XOR problem. Gradient descent Stochastic Approximation to gradient descent 5/10/10

Perceptron. Inner-product scalar Perceptron. XOR problem. Gradient descent Stochastic Approximation to gradient descent 5/10/10 Perceptro Ier-product scalar Perceptro Perceptro learig rule XOR problem liear separable patters Gradiet descet Stochastic Approximatio to gradiet descet LMS Adalie 1 Ier-product et =< w, x >= w x cos(θ)

More information

1 Review of Probability & Statistics

1 Review of Probability & Statistics 1 Review of Probability & Statistics a. I a group of 000 people, it has bee reported that there are: 61 smokers 670 over 5 960 people who imbibe (drik alcohol) 86 smokers who imbibe 90 imbibers over 5

More information

10-701/ Machine Learning Mid-term Exam Solution

10-701/ Machine Learning Mid-term Exam Solution 0-70/5-78 Machie Learig Mid-term Exam Solutio Your Name: Your Adrew ID: True or False (Give oe setece explaatio) (20%). (F) For a cotiuous radom variable x ad its probability distributio fuctio p(x), it

More information

Markov Decision Processes

Markov Decision Processes Markov Decisio Processes Defiitios; Statioary policies; Value improvemet algorithm, Policy improvemet algorithm, ad liear programmig for discouted cost ad average cost criteria. Markov Decisio Processes

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

A PENALTY-CONJUGATE GRADIENT ALGORITHM FOR SOLVING CONSTRAINED OPTIMIZATION PROBLEMS

A PENALTY-CONJUGATE GRADIENT ALGORITHM FOR SOLVING CONSTRAINED OPTIMIZATION PROBLEMS IJRRAS 35 () April 8 wwwarpapresscom/volumes/vol35issue/ijrras_35 pdf A PENALY-CONJUGAE GRADIEN ALGORIHM FOR SOLVING CONSRAINED OPIMIZAION PROBLEMS Hoia Zha Zhuohua Lia & Weli Zhao School of Mathematics

More information

Outline. Linear regression. Regularization functions. Polynomial curve fitting. Stochastic gradient descent for regression. MLE for regression

Outline. Linear regression. Regularization functions. Polynomial curve fitting. Stochastic gradient descent for regression. MLE for regression REGRESSION 1 Outlie Liear regressio Regularizatio fuctios Polyomial curve fittig Stochastic gradiet descet for regressio MLE for regressio Step-wise forward regressio Regressio methods Statistical techiques

More information

Linear Support Vector Machines

Linear Support Vector Machines Liear Support Vector Machies David S. Roseberg The Support Vector Machie For a liear support vector machie (SVM), we use the hypothesis space of affie fuctios F = { f(x) = w T x + b w R d, b R } ad evaluate

More information

Jacob Hays Amit Pillay James DeFelice 4.1, 4.2, 4.3

Jacob Hays Amit Pillay James DeFelice 4.1, 4.2, 4.3 No-Parametric Techiques Jacob Hays Amit Pillay James DeFelice 4.1, 4.2, 4.3 Parametric vs. No-Parametric Parametric Based o Fuctios (e.g Normal Distributio) Uimodal Oly oe peak Ulikely real data cofies

More information

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3 MATH 337 Sequeces Dr. Neal, WKU Let X be a metric space with distace fuctio d. We shall defie the geeral cocept of sequece ad limit i a metric space, the apply the results i particular to some special

More information

Study the bias (due to the nite dimensional approximation) and variance of the estimators

Study the bias (due to the nite dimensional approximation) and variance of the estimators 2 Series Methods 2. Geeral Approach A model has parameters (; ) where is ite-dimesioal ad is oparametric. (Sometimes, there is o :) We will focus o regressio. The fuctio is approximated by a series a ite

More information

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT OCTOBER 7, 2016 LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT Geometry of LS We ca thik of y ad the colums of X as members of the -dimesioal Euclidea space R Oe ca

More information

Linear Classifiers III

Linear Classifiers III Uiversität Potsdam Istitut für Iformatik Lehrstuhl Maschielles Lere Liear Classifiers III Blaie Nelso, Tobias Scheffer Cotets Classificatio Problem Bayesia Classifier Decisio Liear Classifiers, MAP Models

More information

Pattern Classification, Ch4 (Part 1)

Pattern Classification, Ch4 (Part 1) Patter Classificatio All materials i these slides were take from Patter Classificatio (2d ed) by R O Duda, P E Hart ad D G Stork, Joh Wiley & Sos, 2000 with the permissio of the authors ad the publisher

More information

Introduction to Optimization Techniques. How to Solve Equations

Introduction to Optimization Techniques. How to Solve Equations Itroductio to Optimizatio Techiques How to Solve Equatios Iterative Methods of Optimizatio Iterative methods of optimizatio Solutio of the oliear equatios resultig form a optimizatio problem is usually

More information

Algebra of Least Squares

Algebra of Least Squares October 19, 2018 Algebra of Least Squares Geometry of Least Squares Recall that out data is like a table [Y X] where Y collects observatios o the depedet variable Y ad X collects observatios o the k-dimesioal

More information

REGRESSION (Physics 1210 Notes, Partial Modified Appendix A)

REGRESSION (Physics 1210 Notes, Partial Modified Appendix A) REGRESSION (Physics 0 Notes, Partial Modified Appedix A) HOW TO PERFORM A LINEAR REGRESSION Cosider the followig data poits ad their graph (Table I ad Figure ): X Y 0 3 5 3 7 4 9 5 Table : Example Data

More information

PC5215 Numerical Recipes with Applications - Review Problems

PC5215 Numerical Recipes with Applications - Review Problems PC55 Numerical Recipes with Applicatios - Review Problems Give the IEEE 754 sigle precisio bit patter (biary or he format) of the followig umbers: 0 0 05 00 0 00 Note that it has 8 bits for the epoet,

More information

Optimally Sparse SVMs

Optimally Sparse SVMs A. Proof of Lemma 3. We here prove a lower boud o the umber of support vectors to achieve geeralizatio bouds of the form which we cosider. Importatly, this result holds ot oly for liear classifiers, but

More information

LECTURE 7: Kernel Density Estimation

LECTURE 7: Kernel Density Estimation LECTURE 7: erel esity Estimatio o-parametric esity Estimatio Historams Parze Widows Smoot erels Product erel esity Estimatio Te aïve Bayes Classifier Itroductio to Patter Aalysis Ricardo Gutierrez-Osua

More information

DETERMINATION OF MECHANICAL PROPERTIES OF A NON- UNIFORM BEAM USING THE MEASUREMENT OF THE EXCITED LONGITUDINAL ELASTIC VIBRATIONS.

DETERMINATION OF MECHANICAL PROPERTIES OF A NON- UNIFORM BEAM USING THE MEASUREMENT OF THE EXCITED LONGITUDINAL ELASTIC VIBRATIONS. ICSV4 Cairs Australia 9- July 7 DTRMINATION OF MCHANICAL PROPRTIS OF A NON- UNIFORM BAM USING TH MASURMNT OF TH XCITD LONGITUDINAL LASTIC VIBRATIONS Pavel Aokhi ad Vladimir Gordo Departmet of the mathematics

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Principle Of Superposition

Principle Of Superposition ecture 5: PREIMINRY CONCEP O RUCUR NYI Priciple Of uperpositio Mathematically, the priciple of superpositio is stated as ( a ) G( a ) G( ) G a a or for a liear structural system, the respose at a give

More information

TEACHER CERTIFICATION STUDY GUIDE

TEACHER CERTIFICATION STUDY GUIDE COMPETENCY 1. ALGEBRA SKILL 1.1 1.1a. ALGEBRAIC STRUCTURES Kow why the real ad complex umbers are each a field, ad that particular rigs are ot fields (e.g., itegers, polyomial rigs, matrix rigs) Algebra

More information

Lesson 10: Limits and Continuity

Lesson 10: Limits and Continuity www.scimsacademy.com Lesso 10: Limits ad Cotiuity SCIMS Academy 1 Limit of a fuctio The cocept of limit of a fuctio is cetral to all other cocepts i calculus (like cotiuity, derivative, defiite itegrals

More information

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification INF 4300 90 Itroductio to classifictio Ae Solberg ae@ifiuioo Based o Chapter -6 i Duda ad Hart: atter Classificatio 90 INF 4300 Madator proect Mai task: classificatio You must implemet a classificatio

More information

Lecture #18

Lecture #18 18-1 Variatioal Method (See CTDL 1148-1155, [Variatioal Method] 252-263, 295-307[Desity Matrices]) Last time: Quasi-Degeeracy Diagoalize a part of ifiite H * sub-matrix : H (0) + H (1) * correctios for

More information

Linear Regression Demystified

Linear Regression Demystified Liear Regressio Demystified Liear regressio is a importat subject i statistics. I elemetary statistics courses, formulae related to liear regressio are ofte stated without derivatio. This ote iteds to

More information

6. Kalman filter implementation for linear algebraic equations. Karhunen-Loeve decomposition

6. Kalman filter implementation for linear algebraic equations. Karhunen-Loeve decomposition 6. Kalma filter implemetatio for liear algebraic equatios. Karhue-Loeve decompositio 6.1. Solvable liear algebraic systems. Probabilistic iterpretatio. Let A be a quadratic matrix (ot obligatory osigular.

More information

Tracking Performance of the MMax Conjugate Gradient Algorithm Bei Xie and Tamal Bose

Tracking Performance of the MMax Conjugate Gradient Algorithm Bei Xie and Tamal Bose racki Performace of the MMa Cojuate Gradiet Alorithm Bei Xie ad amal Bose Wireless@V Bradley Det. of Electrical ad Comuter Eieeri Viriia ech Outlie Motivatio Backroud Cojuate Gradiet CG Alorithm Partial

More information

Differentiable Convex Functions

Differentiable Convex Functions Differetiable Covex Fuctios The followig picture motivates Theorem 11. f ( x) f ( x) f '( x)( x x) ˆx x 1 Theorem 11 : Let f : R R be differetiable. The, f is covex o the covex set C R if, ad oly if for

More information

Matrix Representation of Data in Experiment

Matrix Representation of Data in Experiment Matrix Represetatio of Data i Experimet Cosider a very simple model for resposes y ij : y ij i ij, i 1,; j 1,,..., (ote that for simplicity we are assumig the two () groups are of equal sample size ) Y

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

Machine Learning Regression I Hamid R. Rabiee [Slides are based on Bishop Book] Spring

Machine Learning Regression I Hamid R. Rabiee [Slides are based on Bishop Book] Spring Machie Learig Regressio I Hamid R. Rabiee [Slides are based o Bishop Book] Sprig 015 http://ce.sharif.edu/courses/93-94//ce717-1 Liear Regressio Liear regressio: ivolves a respose variable ad a sigle predictor

More information

Correlation Regression

Correlation Regression Correlatio Regressio While correlatio methods measure the stregth of a liear relatioship betwee two variables, we might wish to go a little further: How much does oe variable chage for a give chage i aother

More information

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

Problem Set 4 Due Oct, 12

Problem Set 4 Due Oct, 12 EE226: Radom Processes i Systems Lecturer: Jea C. Walrad Problem Set 4 Due Oct, 12 Fall 06 GSI: Assae Gueye This problem set essetially reviews detectio theory ad hypothesis testig ad some basic otios

More information

Notes on iteration and Newton s method. Iteration

Notes on iteration and Newton s method. Iteration Notes o iteratio ad Newto s method Iteratio Iteratio meas doig somethig over ad over. I our cotet, a iteratio is a sequece of umbers, vectors, fuctios, etc. geerated by a iteratio rule of the type 1 f

More information

, then cv V. Differential Equations Elements of Lineaer Algebra Name: Consider the differential equation. and y2 cos( kx)

, then cv V. Differential Equations Elements of Lineaer Algebra Name: Consider the differential equation. and y2 cos( kx) Cosider the differetial equatio y '' k y 0 has particular solutios y1 si( kx) ad y cos( kx) I geeral, ay liear combiatio of y1 ad y, cy 1 1 cy where c1, c is also a solutio to the equatio above The reaso

More information

Problem Set 2 Solutions

Problem Set 2 Solutions CS271 Radomess & Computatio, Sprig 2018 Problem Set 2 Solutios Poit totals are i the margi; the maximum total umber of poits was 52. 1. Probabilistic method for domiatig sets 6pts Pick a radom subset S

More information

Topics Machine learning: lecture 2. Review: the learning problem. Hypotheses and estimation. Estimation criterion cont d. Estimation criterion

Topics Machine learning: lecture 2. Review: the learning problem. Hypotheses and estimation. Estimation criterion cont d. Estimation criterion .87 Machie learig: lecture Tommi S. Jaakkola MIT CSAIL tommi@csail.mit.edu Topics The learig problem hypothesis class, estimatio algorithm loss ad estimatio criterio samplig, empirical ad epected losses

More information

Definitions and Theorems. where x are the decision variables. c, b, and a are constant coefficients.

Definitions and Theorems. where x are the decision variables. c, b, and a are constant coefficients. Defiitios ad Theorems Remember the scalar form of the liear programmig problem, Miimize, Subject to, f(x) = c i x i a 1i x i = b 1 a mi x i = b m x i 0 i = 1,2,, where x are the decisio variables. c, b,

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Chapter 9 - CD companion 1. A Generic Implementation; The Common-Merge Amplifier. 1 τ is. ω ch. τ io

Chapter 9 - CD companion 1. A Generic Implementation; The Common-Merge Amplifier. 1 τ is. ω ch. τ io Chapter 9 - CD compaio CHAPTER NINE CD-9.2 CD-9.2. Stages With Voltage ad Curret Gai A Geeric Implemetatio; The Commo-Merge Amplifier The advaced method preseted i the text for approximatig cutoff frequecies

More information

1.010 Uncertainty in Engineering Fall 2008

1.010 Uncertainty in Engineering Fall 2008 MIT OpeCourseWare http://ocw.mit.edu.00 Ucertaity i Egieerig Fall 2008 For iformatio about citig these materials or our Terms of Use, visit: http://ocw.mit.edu.terms. .00 - Brief Notes # 9 Poit ad Iterval

More information

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015 ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],

More information

Chapter 1 Simple Linear Regression (part 6: matrix version)

Chapter 1 Simple Linear Regression (part 6: matrix version) Chapter Simple Liear Regressio (part 6: matrix versio) Overview Simple liear regressio model: respose variable Y, a sigle idepedet variable X Y β 0 + β X + ε Multiple liear regressio model: respose Y,

More information

Lecture 33: Bootstrap

Lecture 33: Bootstrap Lecture 33: ootstrap Motivatio To evaluate ad compare differet estimators, we eed cosistet estimators of variaces or asymptotic variaces of estimators. This is also importat for hypothesis testig ad cofidece

More information

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f. Lecture 5 Let us give oe more example of MLE. Example 3. The uiform distributio U[0, ] o the iterval [0, ] has p.d.f. { 1 f(x =, 0 x, 0, otherwise The likelihood fuctio ϕ( = f(x i = 1 I(X 1,..., X [0,

More information

Math 257: Finite difference methods

Math 257: Finite difference methods Math 257: Fiite differece methods 1 Fiite Differeces Remember the defiitio of a derivative f f(x + ) f(x) (x) = lim 0 Also recall Taylor s formula: (1) f(x + ) = f(x) + f (x) + 2 f (x) + 3 f (3) (x) +...

More information

State Space Representation

State Space Representation Optimal Cotrol, Guidace ad Estimatio Lecture 2 Overview of SS Approach ad Matrix heory Prof. Radhakat Padhi Dept. of Aerospace Egieerig Idia Istitute of Sciece - Bagalore State Space Represetatio Prof.

More information

Let l be an index for latent variables (l=1,2,3,4). Consider the latent variable z. vector of observed covariates (excluding a constant), α

Let l be an index for latent variables (l=1,2,3,4). Consider the latent variable z. vector of observed covariates (excluding a constant), α Olie Supplemet to MaaS i Car-Domiated Cities: Modeli the adoptio, frequecy, ad characteristics of ride-haili trips i Dallas, TX Patrícia S. Lavieri ad Chadra R. Bhat (correspodi author) A Overview of the

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

We will conclude the chapter with the study a few methods and techniques which are useful

We will conclude the chapter with the study a few methods and techniques which are useful Chapter : Coordiate geometry: I this chapter we will lear about the mai priciples of graphig i a dimesioal (D) Cartesia system of coordiates. We will focus o drawig lies ad the characteristics of the graphs

More information

CS284A: Representations and Algorithms in Molecular Biology

CS284A: Representations and Algorithms in Molecular Biology CS284A: Represetatios ad Algorithms i Molecular Biology Scribe Notes o Lectures 3 & 4: Motif Discovery via Eumeratio & Motif Represetatio Usig Positio Weight Matrix Joshua Gervi Based o presetatios by

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So,

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So, 0 2. OLS Part II The OLS residuals are orthogoal to the regressors. If the model icludes a itercept, the orthogoality of the residuals ad regressors gives rise to three results, which have limited practical

More information

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d Liear regressio Daiel Hsu (COMS 477) Maximum likelihood estimatio Oe of the simplest liear regressio models is the followig: (X, Y ),..., (X, Y ), (X, Y ) are iid radom pairs takig values i R d R, ad Y

More information

Introduction to Signals and Systems, Part V: Lecture Summary

Introduction to Signals and Systems, Part V: Lecture Summary EEL33: Discrete-Time Sigals ad Systems Itroductio to Sigals ad Systems, Part V: Lecture Summary Itroductio to Sigals ad Systems, Part V: Lecture Summary So far we have oly looked at examples of o-recursive

More information

Stochastic Matrices in a Finite Field

Stochastic Matrices in a Finite Field Stochastic Matrices i a Fiite Field Abstract: I this project we will explore the properties of stochastic matrices i both the real ad the fiite fields. We first explore what properties 2 2 stochastic matrices

More information

Math 25 Solutions to practice problems

Math 25 Solutions to practice problems Math 5: Advaced Calculus UC Davis, Sprig 0 Math 5 Solutios to practice problems Questio For = 0,,, 3,... ad 0 k defie umbers C k C k =! k!( k)! (for k = 0 ad k = we defie C 0 = C = ). by = ( )... ( k +

More information

IIT JAM Mathematical Statistics (MS) 2006 SECTION A

IIT JAM Mathematical Statistics (MS) 2006 SECTION A IIT JAM Mathematical Statistics (MS) 6 SECTION A. If a > for ad lim a / L >, the which of the followig series is ot coverget? (a) (b) (c) (d) (d) = = a = a = a a + / a lim a a / + = lim a / a / + = lim

More information

Chapter 7. Support Vector Machine

Chapter 7. Support Vector Machine Chapter 7 Support Vector Machie able of Cotet Margi ad support vectors SVM formulatio Slack variables ad hige loss SVM for multiple class SVM ith Kerels Relevace Vector Machie Support Vector Machie (SVM)

More information

Lecture 4. We also define the set of possible values for the random walk as the set of all x R d such that P(S n = x) > 0 for some n.

Lecture 4. We also define the set of possible values for the random walk as the set of all x R d such that P(S n = x) > 0 for some n. Radom Walks ad Browia Motio Tel Aviv Uiversity Sprig 20 Lecture date: Mar 2, 20 Lecture 4 Istructor: Ro Peled Scribe: Lira Rotem This lecture deals primarily with recurrece for geeral radom walks. We preset

More information

ECON 3150/4150, Spring term Lecture 3

ECON 3150/4150, Spring term Lecture 3 Itroductio Fidig the best fit by regressio Residuals ad R-sq Regressio ad causality Summary ad ext step ECON 3150/4150, Sprig term 2014. Lecture 3 Ragar Nymoe Uiversity of Oslo 21 Jauary 2014 1 / 30 Itroductio

More information

Real Variables II Homework Set #5

Real Variables II Homework Set #5 Real Variables II Homework Set #5 Name: Due Friday /0 by 4pm (at GOS-4) Istructios: () Attach this page to the frot of your homework assigmet you tur i (or write each problem before your solutio). () Please

More information

ENGI 9420 Engineering Analysis Assignment 3 Solutions

ENGI 9420 Engineering Analysis Assignment 3 Solutions ENGI 9 Egieerig Aalysis Assigmet Solutios Fall [Series solutio of ODEs, matri algebra; umerical methods; Chapters, ad ]. Fid a power series solutio about =, as far as the term i 7, to the ordiary differetial

More information

1 Inferential Methods for Correlation and Regression Analysis

1 Inferential Methods for Correlation and Regression Analysis 1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet

More information

Chapter Vectors

Chapter Vectors Chapter 4. Vectors fter readig this chapter you should be able to:. defie a vector. add ad subtract vectors. fid liear combiatios of vectors ad their relatioship to a set of equatios 4. explai what it

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss ECE 90 Lecture : Complexity Regularizatio ad the Squared Loss R. Nowak 5/7/009 I the previous lectures we made use of the Cheroff/Hoeffdig bouds for our aalysis of classifier errors. Hoeffdig s iequality

More information

1 Duality revisited. AM 221: Advanced Optimization Spring 2016

1 Duality revisited. AM 221: Advanced Optimization Spring 2016 AM 22: Advaced Optimizatio Sprig 206 Prof. Yaro Siger Sectio 7 Wedesday, Mar. 9th Duality revisited I this sectio, we will give a slightly differet perspective o duality. optimizatio program: f(x) x R

More information

Naïve Bayes. Naïve Bayes

Naïve Bayes. Naïve Bayes Statistical Data Miig ad Machie Learig Hilary Term 206 Dio Sejdiovic Departmet of Statistics Oxford Slides ad other materials available at: http://www.stats.ox.ac.uk/~sejdiov/sdmml : aother plug-i classifier

More information

Similarity Solutions to Unsteady Pseudoplastic. Flow Near a Moving Wall

Similarity Solutions to Unsteady Pseudoplastic. Flow Near a Moving Wall Iteratioal Mathematical Forum, Vol. 9, 04, o. 3, 465-475 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/0.988/imf.04.48 Similarity Solutios to Usteady Pseudoplastic Flow Near a Movig Wall W. Robi Egieerig

More information

CS537. Numerical Analysis and Computing

CS537. Numerical Analysis and Computing CS57 Numerical Aalysis ad Computig Lecture Locatig Roots o Equatios Proessor Ju Zhag Departmet o Computer Sciece Uiversity o Ketucky Leigto KY 456-6 Jauary 9 9 What is the Root May physical system ca be

More information

PAijpam.eu ON DERIVATION OF RATIONAL SOLUTIONS OF BABBAGE S FUNCTIONAL EQUATION

PAijpam.eu ON DERIVATION OF RATIONAL SOLUTIONS OF BABBAGE S FUNCTIONAL EQUATION Iteratioal Joural of Pure ad Applied Mathematics Volume 94 No. 204, 9-20 ISSN: 3-8080 (prited versio); ISSN: 34-3395 (o-lie versio) url: http://www.ijpam.eu doi: http://dx.doi.org/0.2732/ijpam.v94i.2 PAijpam.eu

More information

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences.

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences. Norwegia Uiversity of Sciece ad Techology Departmet of Mathematical Scieces Corrected 3 May ad 4 Jue Solutios TMA445 Statistics Saturday 6 May 9: 3: Problem Sow desity a The probability is.9.5 6x x dx

More information

(b) What is the probability that a particle reaches the upper boundary n before the lower boundary m?

(b) What is the probability that a particle reaches the upper boundary n before the lower boundary m? MATH 529 The Boudary Problem The drukard s walk (or boudary problem) is oe of the most famous problems i the theory of radom walks. Oe versio of the problem is described as follows: Suppose a particle

More information

6.3 Testing Series With Positive Terms

6.3 Testing Series With Positive Terms 6.3. TESTING SERIES WITH POSITIVE TERMS 307 6.3 Testig Series With Positive Terms 6.3. Review of what is kow up to ow I theory, testig a series a i for covergece amouts to fidig the i= sequece of partial

More information

Introduction to Artificial Intelligence CAP 4601 Summer 2013 Midterm Exam

Introduction to Artificial Intelligence CAP 4601 Summer 2013 Midterm Exam Itroductio to Artificial Itelligece CAP 601 Summer 013 Midterm Exam 1. Termiology (7 Poits). Give the followig task eviromets, eter their properties/characteristics. The properties/characteristics of the

More information

Empirical Process Theory and Oracle Inequalities

Empirical Process Theory and Oracle Inequalities Stat 928: Statistical Learig Theory Lecture: 10 Empirical Process Theory ad Oracle Iequalities Istructor: Sham Kakade 1 Risk vs Risk See Lecture 0 for a discussio o termiology. 2 The Uio Boud / Boferoi

More information

APPENDIX A SMO ALGORITHM

APPENDIX A SMO ALGORITHM AENDIX A SMO ALGORITHM Sequetial Miimal Optimizatio SMO) is a simple algorithm that ca quickly solve the SVM Q problem without ay extra matrix storage ad without usig time-cosumig umerical Q optimizatio

More information

11 Correlation and Regression

11 Correlation and Regression 11 Correlatio Regressio 11.1 Multivariate Data Ofte we look at data where several variables are recorded for the same idividuals or samplig uits. For example, at a coastal weather statio, we might record

More information

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator Slide Set 13 Liear Model with Edogeous Regressors ad the GMM estimator Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Friday

More information

Admin REGULARIZATION. Schedule. Midterm 9/29/16. Assignment 5. Midterm next week, due Friday (more on this in 1 min)

Admin REGULARIZATION. Schedule. Midterm 9/29/16. Assignment 5. Midterm next week, due Friday (more on this in 1 min) Admi Assigmet 5! Starter REGULARIZATION David Kauchak CS 158 Fall 2016 Schedule Midterm ext week, due Friday (more o this i 1 mi Assigmet 6 due Friday before fall break Midterm Dowload from course web

More information

1 Covariance Estimation

1 Covariance Estimation Eco 75 Lecture 5 Covariace Estimatio ad Optimal Weightig Matrices I this lecture, we cosider estimatio of the asymptotic covariace matrix B B of the extremum estimator b : Covariace Estimatio Lemma 4.

More information