6.867 Machine Learning
|
|
- Osborn Stanley
- 5 years ago
- Views:
Transcription
1 6.867 Mache Learg Problem set Due Frday, September 9, rectato Please address all questos ad commets about ths problem set to You do ot eed to use MATLAB for ths problem set though you ca certaly do so. We wll provde helpful hts alog the way but f you are ot famlar wth MATLAB ad wsh to use MATLAB ths problem set, please cosult ad the lks there. Part : Least-Squares Regresso Referece: Lectures 2 ad 3, chapters The goal of ths secto s to soldfy basc cocepts least squares regresso. Suppose we have some smple dataset, {(x, y ), =,..., }, where x ad y are real umbers. Our model of how y s related to x s gve by y = f(x; w) + e () f(x; w) = w φ(x) (2) where φ : R R d s a specfed fucto (see below) whch maps x to a d-dmesoal feature vector, φ(x) = (φ (x),..., φ d (x)) ; w s a d-dmesoal parameter vector w = (w,..., w d ) ; e s the predcto error, whch we do ot model explctly. We wll use w to deote the traspose of ay vector w, as s doe MATLAB. Note that our formulato above does ot explctly clude the offset parameter, or w 0, as was doe the lectures. We ca corporate the offset by defg φ (x) =. I the followg, we wsh to determe the least squares optmal parameters or ŵ. I other words, we mmze the followg squared predcto error: J(w) = (y f(x ; w)) 2 (3) By a smlar argumet as gve Lecture 2, the soluto to ths problem s ŵ = (X X) X y (4)
2 where X = (φ(x ),..., φ(x )) s a d matrx whose frst row s φ (x ),..., φ d (x ) ad the last row s gve by φ (x ),..., φ d (x ); The output vector y s defed as y = (y, y 2,..., y ). Note: We assume that the matrx (X X) s vertble so that the problem s well-posed,.e. there exsts a uque mmzer. Ths s true whe the feature vectors φ(x ),..., φ(x ) assocated wth the trag examples spa the d dmesoal feature space. Whe the feature vectors are log ad the umber of trag pots s small, ths s ot at all ecessarly the case. For example, t caot be the case wheever d >. Now, for ths estmate ŵ the resultg predcto errors ê = y f(x ; ŵ) should be ucorrelated wth the features: ê φ k (x ) = 0, k =,..., d (5) These codtos are obtaed by takg the dervatve of J(w) wth respect to each w, =,..., d, ad settg them to zero. Note that the predcto error eed ot be zero mea uless oe of the features s a costat,.e., f say φ (x) = for all x, so that ê φ (x ) = ê = 0 (6) The error s guarateed to be ucorrelated wth oly features actually cluded the predcto. You may woder why we are talkg about correlato the frst place. Let s explore ths a bt further, here ad the problems that follow. For jot samples (u, v ), =,...,, the sample covarace s defed as Σ u,v = (u ū)(v v) (7) where ū = (/) u s the sample mea of u ad smlarly for v. The samples are ucorrelated f the sample covarace s exactly zero, Σ u,v = 0. Covarace measures how well oe varable (or oe set of samples) s learly predctable from the other. Problem. (5pts) Assumg that the frst compoet of the feature vector s a costat,.e., φ (x) =, show that the jot samples (ê, φ k (x )), =,...,, are deed ucorrelated for all k =,..., d accordg to our defto above. 2. (5pts) Show that all lear fuctos of the bass fuctos,.e., fuctos of the form w φ(x) for some w R d, are also ucorrelated wth the predcto errors ê assocated wth the least squares optmal parameters ŵ. I other words, show that (ê, w φ(x )), =,...,, are ucorrelated for ay w. 2
3 3. (5pts) Yet aother way of uderstadg ths result s that f we try to ft a lear fucto (usg the same bass fuctos) to the predcto errors, we ca oly get zero. Let ŵ ad ê, =,...,, be defed as above. If we ow use ỹ = ê as the ew target outputs ad repeat the parameter estmato step usg these ew outputs ad the same set of bass fuctos, show that the resultg ew least squares parameters are deed detcally zero. 4. (5pts) Suppose we chage our feature represetato of examples by rescalg the bass fuctos,.e., use φ(x) = (a φ (x),..., a d φ d (x)) as the feature vector, where a, =,..., d are ay o-zero real umbers. Show that the uscaled soluto, the fucto ŵ φ(x), s stll optmal the sese that ŵ φ(x) = ˆ w φ(x), where ˆ w are the least squares optmal parameters for the scaled feature vectors. (Ht. use correlato). 5. (Optoal) Let s go through a small umercal example to get started wth MATLAB. We wll use the followg data (expressed MATLAB otato): x = [-2-0 2] ; y = [ ] ; (both are colum vectors). Let φ(x) = (, x, x 2 ). To fd the least squares parameters, say wh MATLAB, we ca costruct the X matrx smply as X = [oes(sze(x)),x,x.^2]; where the dot refers to a elemetwse operato. v(a) for ay vertble A. Matrx verse MATLAB s Fd the least squares optmal parameters wh ths case. Plot the sample pots ad the resultg fucto correspodg to the parameters. Verfy that the predcto error s deed ucorrelated wth the bass fuctos. Repeat the procedure for φ(x) = s(πx) (oly oe bass fucto). (ote: π MATLAB s smply a costat p). Does the result look reasoable? What should the aswer be? Problem 2 The predctos we make the regresso formulato eed ot be oe dmesoal. We ca just as easly make predctos that are vector valued. Cosder a smple example where the put x takes oly bary values x {0, } ad y s a two-dmesoal measuremet y R 2. Here, the model s y = f(x; W ) + e (8) f(x; W ) = W φ(x) (9) 3
4 where the feature vector s defed by φ(0) = (, 0) ad φ() = (0, ) ; W s a twoby-two matrx of model parameters; ad both the predcto f(x; W ) ad the predcto errors e = y f(x ; W ) are two-dmesoal vectors. We ow wsh to determe Ŵ that mmzes the squared error J(W ) = e 2 = e e (0). (0pts) Show that the least-squares estmate of W s Ŵ = (X X) X Y () where X = (φ(x )... φ(x )) ad Y = (y... y ). Ht. Show that the objectve decomposes so that each colum of W may be obtaed depedetly; you are essetally solvg two separate -dmesoal regresso problems. Next, cosder the data set: x y 0 (, ) 0 (, 2) 0 ( 2, ) (, ) (, 2) (2, ) 2. (5pts) Compute Ŵ. Plot the data pots y ad the colums of Ŵ = (ŷ 0 ŷ ) (ote the traspose). 3. (5pts) Verfy that ê φ(x ) = 0 (a 2x2 matrx ths case). What s the terpretato of the colums of Ŵ? Part 2: Probablstc Modelg ad Lkelhood Referece: Lecture 3, chapter 4 (up to eq 4.20) Frst a bt of backgroud. Suppose we have a probablty dstrbuto or desty p(x; θ), where x may be dscrete or cotuous depedg o the problem we are terested. θ specfes the parameters of ths dstrbuto such as the mea ad the varace of a oe dmesoal Gaussa. Dfferet settgs of the parameters mply dfferet dstrbutos over x. The avalable data, whe terpreted as samples x,..., x from oe such dstrbuto, should favor oe settg of the parameters over aother. We eed a formal crtero for gaugg how well ay potetal dstrbuto p( θ) explas or fts the data. Sce 4
5 p(x θ) s the probablty of reproducg ay observato x, t seems atural to try to maxmze ths probablty. Ths gves rse to the Maxmum Lkelhood estmato crtero for the parameters θ: ˆθ ML = argmax θ L(x,..., x ; θ) = argmax θ p(x θ) (2) = where we have assumed that each data pot x s draw depedetly from the same dstrbuto so that the lkelhood of the data s L(x,..., x ; θ) = = p(x ; θ). Lkelhood s vewed prmarly as a fucto of the parameters, a fucto that depeds o the data. The above expresso ca be qute complcated (depedg o the famly of dstrbutos we are cosderg), ad make maxmzato techcally challegg. However, ay mootocally creasg fucto of the lkelhood wll have the same maxma. Oe such fucto s log-lkelhood log L(x,..., x ; θ); takg the log turs the product to a sum, makg dervatves sgfcatly smpler. We wll maxmze the log-lkelhood stead of lkelhood. Problem 4 Let x {0, } deote the result of a co flp (x = 0 for tals, ad x = for heads ). The co s potetally based so that heads occurs wth probablty θ. Suppose also that someoe else observes the co flp ad reports to you heads or tals (deote ths report by y). But ths perso s urelable ad oly reports the result correctly wth probablty θ 2 (the correctess of the report s depedet of the co toss).. (5pts) Wrte dow the jot probablty dstrbuto P (x, y θ) for all x, y (a 2x2 matrx) as a fucto of the parameters θ = (θ, θ 2 ). Suppose we have access to the followg (jot) observatos of x ad y: x y (0pts) What are the maxmum-lkelhood (ML) values of θ ad θ 2? Provde the detals of your dervato as well as the aswer. Ht. You ca frst cofrm that P (x, y θ) = P (y x, θ 2 )P (x θ ), where the key observato s that the parameters ca be separated to the dfferet compoets. After all the dstrbuto of the co toss, govered by P (x θ ), s depedet of the accuracy of the report, cotaed P (y x, θ 2 ). Ths separato helps you to solate the estmato of each parameter the log-lkelhood crtero. 5
6 3. (0pts) Let ˆθ (x,..., x ) be the ML estmator of θ based o the observed data x,..., x, where the data s vewed as depedet samples from P (x θ ) for some fxed θ. We ca try to assess how well the estmator recovers the parameters θ. Oe useful measure s the bas of the estmator. Ths s defed as the expectato E [ ˆθ (X,..., X ) θ ], take wth respect to the true dstrbuto of X,..., X or P (X θ ). The bas measures whether the estmator systematcally devates from the true parameters θ that were used to geerate the data. A estmator s called ubased f ts bas s zero. Show that the ML estmator ˆθ s deed ubased ths sese. 4. (0pts) We have thus far used oly two parameters θ ad θ 2 to specfy the jot dstrbuto over (x, y). Ths was possble because of the assumpto that the accuracy of the report (whether y = x) s depedet of the co toss (what x s). It takes three parameters to specfy a ucostraed jot dstrbuto over (x, y). Whle there are four possble cofguratos of the varables, there are oly three parameters that ca be set depedetly (the fourth oe s determed due to ormalzato, x,y P (x, y) = ). We ca parameterze the jot dstrbuto symmetrcally terms of four umbers P (x, y) = θ x,y, that sum to oe x,y θ x,y =. Whe we estmate the maxmum lkelhood jot dstrbuto, we estmate the ML settg of the parameters ˆθ x,y. What s the maxmum lkelhood estmate of P (x, y) ths case? Whch model has the hgher log-lkelhood? 5. (Optoal) Show that the ML parameters ˆθ x,y are ubased estmates of θ x,y. 6. (Optoal) Suppose we are ot sure whch model s correct. Ca you exted the leave-oe-out cross-valdato procedure descrbed the lear regresso cotext to our settg here? Whch model would the resultg cross-valdato crtero choose ths case? Problem 5 Cosder a bvarate Gaussa dstrbuto x = (x, x 2 ) N(µ, Σ) wth probablty desty p(x; µ, Σ) = 2π Σ /2 exp{ 2 (x µ) Σ (x µ)} (3) where µ = E{x} s the two-dmesoal mea vector ad Σ = E{(x µ)(x µ) } s the two-by-two covarace matrx ( Σ s the determat of the covarace matrx). The Gaussa s fully specfed by the parameters (µ, Σ).. (0pts) Gve a collecto of depedet samples x, =,...,, we wsh to estmate the model parameters (µ, Σ). The maxmum-lkelhood estmates are chose so as to 6
7 maxmze the log-lkelhood J(µ, Σ) = log p(x,..., x ; µ, Σ) (4) = log p(x ; µ, Σ) (5) Show that the ML estmates based o data x,..., x are gve by the sample mea ad sample covarace: ˆµ = x (6) ˆΣ = (x ˆµ)(x ˆµ) (7) Hts. Start wth the mea estmate. Express the Gaussa dstrbuto terms of the verse covarace matrx A = Σ ad use the followg matrx dervatves d da (x µ) A(x µ) = (x µ)(x µ) d da log A = A (8) 2. (0pts) The b-varate Gaussa dstrbuto allows the two varables to be depedet o each other (the values of the varables co-vary). Ths depedece s fully descrbed by the covarace matrx. I lght of problem we suspect that ths depedece s captured by learly predctg oe from the other. Suppose we have access to x part of the samples from a Gaussa model (µ, Σ) ad wsh to use them to estmate x 2. Derve the least squares optmal estmate ˆx 2 (x ) that mmzes the expected squared error E{(x 2 ˆx 2 (x )) 2 } (the exectato s over samples (x, x 2 ) N(µ, Σ)). Ht. Use the fact that the best estmate s of the form ˆx 2 (x ) = E{x 2 x }, as dscussed the lecture. 7
Bayes (Naïve or not) Classifiers: Generative Approach
Logstc regresso Bayes (Naïve or ot) Classfers: Geeratve Approach What do we mea by Geeratve approach: Lear p(y), p(x y) ad the apply bayes rule to compute p(y x) for makg predctos Ths s essetally makg
More informationUNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS
UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS Postpoed exam: ECON430 Statstcs Date of exam: Jauary 0, 0 Tme for exam: 09:00 a.m. :00 oo The problem set covers 5 pages Resources allowed: All wrtte ad prted
More informationPoint Estimation: definition of estimators
Pot Estmato: defto of estmators Pot estmator: ay fucto W (X,..., X ) of a data sample. The exercse of pot estmato s to use partcular fuctos of the data order to estmate certa ukow populato parameters.
More informationMultivariate Transformation of Variables and Maximum Likelihood Estimation
Marquette Uversty Multvarate Trasformato of Varables ad Maxmum Lkelhood Estmato Dael B. Rowe, Ph.D. Assocate Professor Departmet of Mathematcs, Statstcs, ad Computer Scece Copyrght 03 by Marquette Uversty
More informationUNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS
UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS Exam: ECON430 Statstcs Date of exam: Frday, December 8, 07 Grades are gve: Jauary 4, 08 Tme for exam: 0900 am 00 oo The problem set covers 5 pages Resources allowed:
More informationTESTS BASED ON MAXIMUM LIKELIHOOD
ESE 5 Toy E. Smth. The Basc Example. TESTS BASED ON MAXIMUM LIKELIHOOD To llustrate the propertes of maxmum lkelhood estmates ad tests, we cosder the smplest possble case of estmatg the mea of the ormal
More informationMaximum Likelihood Estimation
Marquette Uverst Maxmum Lkelhood Estmato Dael B. Rowe, Ph.D. Professor Departmet of Mathematcs, Statstcs, ad Computer Scece Coprght 08 b Marquette Uverst Maxmum Lkelhood Estmato We have bee sag that ~
More informationEconometric Methods. Review of Estimation
Ecoometrc Methods Revew of Estmato Estmatg the populato mea Radom samplg Pot ad terval estmators Lear estmators Ubased estmators Lear Ubased Estmators (LUEs) Effcecy (mmum varace) ad Best Lear Ubased Estmators
More informationLecture 3 Probability review (cont d)
STATS 00: Itroducto to Statstcal Iferece Autum 06 Lecture 3 Probablty revew (cot d) 3. Jot dstrbutos If radom varables X,..., X k are depedet, the ther dstrbuto may be specfed by specfyg the dvdual dstrbuto
More informationSTATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ " 1
STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS Recall Assumpto E(Y x) η 0 + η x (lear codtoal mea fucto) Data (x, y ), (x 2, y 2 ),, (x, y ) Least squares estmator ˆ E (Y x) ˆ " 0 + ˆ " x, where ˆ
More informationChapter 14 Logistic Regression Models
Chapter 4 Logstc Regresso Models I the lear regresso model X β + ε, there are two types of varables explaatory varables X, X,, X k ad study varable y These varables ca be measured o a cotuous scale as
More informationDimensionality Reduction and Learning
CMSC 35900 (Sprg 009) Large Scale Learg Lecture: 3 Dmesoalty Reducto ad Learg Istructors: Sham Kakade ad Greg Shakharovch L Supervsed Methods ad Dmesoalty Reducto The theme of these two lectures s that
More informationQualifying Exam Statistical Theory Problem Solutions August 2005
Qualfyg Exam Statstcal Theory Problem Solutos August 5. Let X, X,..., X be d uform U(,),
More informationSTK4011 and STK9011 Autumn 2016
STK4 ad STK9 Autum 6 Pot estmato Covers (most of the followg materal from chapter 7: Secto 7.: pages 3-3 Secto 7..: pages 3-33 Secto 7..: pages 35-3 Secto 7..3: pages 34-35 Secto 7.3.: pages 33-33 Secto
More informationX X X E[ ] E X E X. is the ()m n where the ( i,)th. j element is the mean of the ( i,)th., then
Secto 5 Vectors of Radom Varables Whe workg wth several radom varables,,..., to arrage them vector form x, t s ofte coveet We ca the make use of matrx algebra to help us orgaze ad mapulate large umbers
More informationIntroduction to local (nonparametric) density estimation. methods
Itroducto to local (oparametrc) desty estmato methods A slecture by Yu Lu for ECE 66 Sprg 014 1. Itroducto Ths slecture troduces two local desty estmato methods whch are Parze desty estmato ad k-earest
More informationLecture 02: Bounding tail distributions of a random variable
CSCI-B609: A Theorst s Toolkt, Fall 206 Aug 25 Lecture 02: Boudg tal dstrbutos of a radom varable Lecturer: Yua Zhou Scrbe: Yua Xe & Yua Zhou Let us cosder the ubased co flps aga. I.e. let the outcome
More informationSummary of the lecture in Biostatistics
Summary of the lecture Bostatstcs Probablty Desty Fucto For a cotuos radom varable, a probablty desty fucto s a fucto such that: 0 dx a b) b a dx A probablty desty fucto provdes a smple descrpto of the
More informationChapter 5 Properties of a Random Sample
Lecture 6 o BST 63: Statstcal Theory I Ku Zhag, /0/008 Revew for the prevous lecture Cocepts: t-dstrbuto, F-dstrbuto Theorems: Dstrbutos of sample mea ad sample varace, relatoshp betwee sample mea ad sample
More informationOverview. Basic concepts of Bayesian learning. Most probable model given data Coin tosses Linear regression Logistic regression
Overvew Basc cocepts of Bayesa learg Most probable model gve data Co tosses Lear regresso Logstc regresso Bayesa predctos Co tosses Lear regresso 30 Recap: regresso problems Iput to learg problem: trag
More informationObjectives of Multiple Regression
Obectves of Multple Regresso Establsh the lear equato that best predcts values of a depedet varable Y usg more tha oe eplaator varable from a large set of potetal predctors {,,... k }. Fd that subset of
More informationENGI 4421 Joint Probability Distributions Page Joint Probability Distributions [Navidi sections 2.5 and 2.6; Devore sections
ENGI 441 Jot Probablty Dstrbutos Page 7-01 Jot Probablty Dstrbutos [Navd sectos.5 ad.6; Devore sectos 5.1-5.] The jot probablty mass fucto of two dscrete radom quattes, s, P ad p x y x y The margal probablty
More informationLinear Regression Linear Regression with Shrinkage. Some slides are due to Tommi Jaakkola, MIT AI Lab
Lear Regresso Lear Regresso th Shrkage Some sldes are due to Tomm Jaakkola, MIT AI Lab Itroducto The goal of regresso s to make quattatve real valued predctos o the bass of a vector of features or attrbutes.
More informationTHE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE
THE ROYAL STATISTICAL SOCIETY 00 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER I STATISTICAL THEORY The Socety provdes these solutos to assst caddates preparg for the examatos future years ad for the
More informationECONOMETRIC THEORY. MODULE VIII Lecture - 26 Heteroskedasticity
ECONOMETRIC THEORY MODULE VIII Lecture - 6 Heteroskedastcty Dr. Shalabh Departmet of Mathematcs ad Statstcs Ida Isttute of Techology Kapur . Breusch Paga test Ths test ca be appled whe the replcated data
More informationLecture Note to Rice Chapter 8
ECON 430 HG revsed Nov 06 Lecture Note to Rce Chapter 8 Radom matrces Let Y, =,,, m, =,,, be radom varables (r.v. s). The matrx Y Y Y Y Y Y Y Y Y Y = m m m s called a radom matrx ( wth a ot m-dmesoal dstrbuto,
More informationENGI 3423 Simple Linear Regression Page 12-01
ENGI 343 mple Lear Regresso Page - mple Lear Regresso ometmes a expermet s set up where the expermeter has cotrol over the values of oe or more varables X ad measures the resultg values of aother varable
More informationLecture 3. Sampling, sampling distributions, and parameter estimation
Lecture 3 Samplg, samplg dstrbutos, ad parameter estmato Samplg Defto Populato s defed as the collecto of all the possble observatos of terest. The collecto of observatos we take from the populato s called
More informationOrdinary Least Squares Regression. Simple Regression. Algebra and Assumptions.
Ordary Least Squares egresso. Smple egresso. Algebra ad Assumptos. I ths part of the course we are gog to study a techque for aalysg the lear relatoshp betwee two varables Y ad X. We have pars of observatos
More informationFeature Selection: Part 2. 1 Greedy Algorithms (continued from the last lecture)
CSE 546: Mache Learg Lecture 6 Feature Selecto: Part 2 Istructor: Sham Kakade Greedy Algorthms (cotued from the last lecture) There are varety of greedy algorthms ad umerous amg covetos for these algorthms.
More information{ }{ ( )} (, ) = ( ) ( ) ( ) Chapter 14 Exercises in Sampling Theory. Exercise 1 (Simple random sampling): Solution:
Chapter 4 Exercses Samplg Theory Exercse (Smple radom samplg: Let there be two correlated radom varables X ad A sample of sze s draw from a populato by smple radom samplg wthout replacemet The observed
More informationLINEAR REGRESSION ANALYSIS
LINEAR REGRESSION ANALYSIS MODULE V Lecture - Correctg Model Iadequaces Through Trasformato ad Weghtg Dr. Shalabh Departmet of Mathematcs ad Statstcs Ida Isttute of Techology Kapur Aalytcal methods for
More informationMultiple Linear Regression Analysis
LINEA EGESSION ANALYSIS MODULE III Lecture - 4 Multple Lear egresso Aalyss Dr. Shalabh Departmet of Mathematcs ad Statstcs Ida Isttute of Techology Kapur Cofdece terval estmato The cofdece tervals multple
More informationLecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model
Lecture 7. Cofdece Itervals ad Hypothess Tests the Smple CLR Model I lecture 6 we troduced the Classcal Lear Regresso (CLR) model that s the radom expermet of whch the data Y,,, K, are the outcomes. The
More informationbest estimate (mean) for X uncertainty or error in the measurement (systematic, random or statistical) best
Error Aalyss Preamble Wheever a measuremet s made, the result followg from that measuremet s always subject to ucertaty The ucertaty ca be reduced by makg several measuremets of the same quatty or by mprovg
More informationSpecial Instructions / Useful Data
JAM 6 Set of all real umbers P A..d. B, p Posso Specal Istructos / Useful Data x,, :,,, x x Probablty of a evet A Idepedetly ad detcally dstrbuted Bomal dstrbuto wth parameters ad p Posso dstrbuto wth
More information1 Solution to Problem 6.40
1 Soluto to Problem 6.40 (a We wll wrte T τ (X 1,...,X where the X s are..d. wth PDF f(x µ, σ 1 ( x µ σ g, σ where the locato parameter µ s ay real umber ad the scale parameter σ s > 0. Lettg Z X µ σ we
More informationENGI 4421 Propagation of Error Page 8-01
ENGI 441 Propagato of Error Page 8-01 Propagato of Error [Navd Chapter 3; ot Devore] Ay realstc measuremet procedure cotas error. Ay calculatos based o that measuremet wll therefore also cota a error.
More informationDiscrete Mathematics and Probability Theory Fall 2016 Seshia and Walrand DIS 10b
CS 70 Dscrete Mathematcs ad Probablty Theory Fall 206 Sesha ad Walrad DIS 0b. Wll I Get My Package? Seaky delvery guy of some compay s out delverg packages to customers. Not oly does he had a radom package
More informationCHAPTER VI Statistical Analysis of Experimental Data
Chapter VI Statstcal Aalyss of Expermetal Data CHAPTER VI Statstcal Aalyss of Expermetal Data Measuremets do ot lead to a uque value. Ths s a result of the multtude of errors (maly radom errors) that ca
More informationTHE ROYAL STATISTICAL SOCIETY 2016 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5
THE ROYAL STATISTICAL SOCIETY 06 EAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5 The Socety s provdg these solutos to assst cadtes preparg for the examatos 07. The solutos are teded as learg ads ad should
More informationLecture 8: Linear Regression
Lecture 8: Lear egresso May 4, GENOME 56, Sprg Goals Develop basc cocepts of lear regresso from a probablstc framework Estmatg parameters ad hypothess testg wth lear models Lear regresso Su I Lee, CSE
More informationIntroduction to Matrices and Matrix Approach to Simple Linear Regression
Itroducto to Matrces ad Matrx Approach to Smple Lear Regresso Matrces Defto: A matrx s a rectagular array of umbers or symbolc elemets I may applcatos, the rows of a matrx wll represet dvduals cases (people,
More informationCHAPTER 4 RADICAL EXPRESSIONS
6 CHAPTER RADICAL EXPRESSIONS. The th Root of a Real Number A real umber a s called the th root of a real umber b f Thus, for example: s a square root of sce. s also a square root of sce ( ). s a cube
More informationAssignment 5/MATH 247/Winter Due: Friday, February 19 in class (!) (answers will be posted right after class)
Assgmet 5/MATH 7/Wter 00 Due: Frday, February 9 class (!) (aswers wll be posted rght after class) As usual, there are peces of text, before the questos [], [], themselves. Recall: For the quadratc form
More informationChapter 4 Multiple Random Variables
Revew for the prevous lecture: Theorems ad Examples: How to obta the pmf (pdf) of U = g (, Y) ad V = g (, Y) Chapter 4 Multple Radom Varables Chapter 44 Herarchcal Models ad Mxture Dstrbutos Examples:
More informationGenerative classification models
CS 75 Mache Learg Lecture Geeratve classfcato models Mlos Hauskrecht mlos@cs.ptt.edu 539 Seott Square Data: D { d, d,.., d} d, Classfcato represets a dscrete class value Goal: lear f : X Y Bar classfcato
More informationClass 13,14 June 17, 19, 2015
Class 3,4 Jue 7, 9, 05 Pla for Class3,4:. Samplg dstrbuto of sample mea. The Cetral Lmt Theorem (CLT). Cofdece terval for ukow mea.. Samplg Dstrbuto for Sample mea. Methods used are based o CLT ( Cetral
More informationFunctions of Random Variables
Fuctos of Radom Varables Chapter Fve Fuctos of Radom Varables 5. Itroducto A geeral egeerg aalyss model s show Fg. 5.. The model output (respose) cotas the performaces of a system or product, such as weght,
More informationρ < 1 be five real numbers. The
Lecture o BST 63: Statstcal Theory I Ku Zhag, /0/006 Revew for the prevous lecture Deftos: covarace, correlato Examples: How to calculate covarace ad correlato Theorems: propertes of correlato ad covarace
More informationA tighter lower bound on the circuit size of the hardest Boolean functions
Electroc Colloquum o Computatoal Complexty, Report No. 86 2011) A tghter lower boud o the crcut sze of the hardest Boolea fuctos Masak Yamamoto Abstract I [IPL2005], Fradse ad Mlterse mproved bouds o the
More informationMOLECULAR VIBRATIONS
MOLECULAR VIBRATIONS Here we wsh to vestgate molecular vbratos ad draw a smlarty betwee the theory of molecular vbratos ad Hückel theory. 1. Smple Harmoc Oscllator Recall that the eergy of a oe-dmesoal
More informationUnimodality Tests for Global Optimization of Single Variable Functions Using Statistical Methods
Malaysa Umodalty Joural Tests of Mathematcal for Global Optmzato Sceces (): of 05 Sgle - 5 Varable (007) Fuctos Usg Statstcal Methods Umodalty Tests for Global Optmzato of Sgle Varable Fuctos Usg Statstcal
More information1 Mixed Quantum State. 2 Density Matrix. CS Density Matrices, von Neumann Entropy 3/7/07 Spring 2007 Lecture 13. ψ = α x x. ρ = p i ψ i ψ i.
CS 94- Desty Matrces, vo Neuma Etropy 3/7/07 Sprg 007 Lecture 3 I ths lecture, we wll dscuss the bascs of quatum formato theory I partcular, we wll dscuss mxed quatum states, desty matrces, vo Neuma etropy
More informationECON 482 / WH Hong The Simple Regression Model 1. Definition of the Simple Regression Model
ECON 48 / WH Hog The Smple Regresso Model. Defto of the Smple Regresso Model Smple Regresso Model Expla varable y terms of varable x y = β + β x+ u y : depedet varable, explaed varable, respose varable,
More informationThe Mathematical Appendix
The Mathematcal Appedx Defto A: If ( Λ, Ω, where ( λ λ λ whch the probablty dstrbutos,,..., Defto A. uppose that ( Λ,,..., s a expermet type, the σ-algebra o λ λ λ are defed s deoted by ( (,,...,, σ Ω.
More informationUnsupervised Learning and Other Neural Networks
CSE 53 Soft Computg NOT PART OF THE FINAL Usupervsed Learg ad Other Neural Networs Itroducto Mture Destes ad Idetfablty ML Estmates Applcato to Normal Mtures Other Neural Networs Itroducto Prevously, all
More informationTraining Sample Model: Given n observations, [[( Yi, x i the sample model can be expressed as (1) where, zero and variance σ
Stat 74 Estmato for Geeral Lear Model Prof. Goel Broad Outle Geeral Lear Model (GLM): Trag Samle Model: Gve observatos, [[( Y, x ), x = ( x,, xr )], =,,, the samle model ca be exressed as Y = µ ( x, x,,
More informationå 1 13 Practice Final Examination Solutions - = CS109 Dec 5, 2018
Chrs Pech Fal Practce CS09 Dec 5, 08 Practce Fal Examato Solutos. Aswer: 4/5 8/7. There are multle ways to obta ths aswer; here are two: The frst commo method s to sum over all ossbltes for the rak of
More information( ) = ( ) ( ) Chapter 13 Asymptotic Theory and Stochastic Regressors. Stochastic regressors model
Chapter 3 Asmptotc Theor ad Stochastc Regressors The ature of eplaator varable s assumed to be o-stochastc or fed repeated samples a regresso aalss Such a assumpto s approprate for those epermets whch
More informationChapter 4 (Part 1): Non-Parametric Classification (Sections ) Pattern Classification 4.3) Announcements
Aoucemets No-Parametrc Desty Estmato Techques HW assged Most of ths lecture was o the blacboard. These sldes cover the same materal as preseted DHS Bometrcs CSE 90-a Lecture 7 CSE90a Fall 06 CSE90a Fall
More informationBayesian Classification. CS690L Data Mining: Classification(2) Bayesian Theorem: Basics. Bayesian Theorem. Training dataset. Naïve Bayes Classifier
Baa Classfcato CS6L Data Mg: Classfcato() Referece: J. Ha ad M. Kamber, Data Mg: Cocepts ad Techques robablstc learg: Calculate explct probabltes for hypothess, amog the most practcal approaches to certa
More informationThe equation is sometimes presented in form Y = a + b x. This is reasonable, but it s not the notation we use.
INTRODUCTORY NOTE ON LINEAR REGREION We have data of the form (x y ) (x y ) (x y ) These wll most ofte be preseted to us as two colum of a spreadsheet As the topc develops we wll see both upper case ad
More informationRecall MLR 5 Homskedasticity error u has the same variance given any values of the explanatory variables Var(u x1,...,xk) = 2 or E(UU ) = 2 I
Chapter 8 Heterosedastcty Recall MLR 5 Homsedastcty error u has the same varace gve ay values of the eplaatory varables Varu,..., = or EUU = I Suppose other GM assumptos hold but have heterosedastcty.
More information18.413: Error Correcting Codes Lab March 2, Lecture 8
18.413: Error Correctg Codes Lab March 2, 2004 Lecturer: Dael A. Spelma Lecture 8 8.1 Vector Spaces A set C {0, 1} s a vector space f for x all C ad y C, x + y C, where we take addto to be compoet wse
More informationModule 7. Lecture 7: Statistical parameter estimation
Lecture 7: Statstcal parameter estmato Parameter Estmato Methods of Parameter Estmato 1) Method of Matchg Pots ) Method of Momets 3) Mamum Lkelhood method Populato Parameter Sample Parameter Ubased estmato
More informationECON 5360 Class Notes GMM
ECON 560 Class Notes GMM Geeralzed Method of Momets (GMM) I beg by outlg the classcal method of momets techque (Fsher, 95) ad the proceed to geeralzed method of momets (Hase, 98).. radtoal Method of Momets
More informationInvestigating Cellular Automata
Researcher: Taylor Dupuy Advsor: Aaro Wootto Semester: Fall 4 Ivestgatg Cellular Automata A Overvew of Cellular Automata: Cellular Automata are smple computer programs that geerate rows of black ad whte
More informationLecture Notes 2. The ability to manipulate matrices is critical in economics.
Lecture Notes. Revew of Matrces he ablt to mapulate matrces s crtcal ecoomcs.. Matr a rectagular arra of umbers, parameters, or varables placed rows ad colums. Matrces are assocated wth lear equatos. lemets
More informationLecture 2 - What are component and system reliability and how it can be improved?
Lecture 2 - What are compoet ad system relablty ad how t ca be mproved? Relablty s a measure of the qualty of the product over the log ru. The cocept of relablty s a exteded tme perod over whch the expected
More informationGeneralized Linear Regression with Regularization
Geeralze Lear Regresso wth Regularzato Zoya Bylsk March 3, 05 BASIC REGRESSION PROBLEM Note: I the followg otes I wll make explct what s a vector a what s a scalar usg vec t or otato, to avo cofuso betwee
More informationAn Introduction to. Support Vector Machine
A Itroducto to Support Vector Mache Support Vector Mache (SVM) A classfer derved from statstcal learg theory by Vapk, et al. 99 SVM became famous whe, usg mages as put, t gave accuracy comparable to eural-etwork
More information2SLS Estimates ECON In this case, begin with the assumption that E[ i
SLS Estmates ECON 3033 Bll Evas Fall 05 Two-Stage Least Squares (SLS Cosder a stadard lear bvarate regresso model y 0 x. I ths case, beg wth the assumto that E[ x] 0 whch meas that OLS estmates of wll
More information(This summarizes what you basically need to know about joint distributions in this course.)
HG Ot. ECON 430 H Extra exerses for o-semar week 4 (Solutos wll be put o the et at the ed of the week) Itroduto: Revew of multdmesoal dstrbutos (Ths summarzes what you basally eed to kow about jot dstrbutos
More informationSTA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #1
STA 08 Appled Lear Models: Regresso Aalyss Sprg 0 Soluto for Homework #. Let Y the dollar cost per year, X the umber of vsts per year. The the mathematcal relato betwee X ad Y s: Y 300 + X. Ths s a fuctoal
More informationDepartment of Agricultural Economics. PhD Qualifier Examination. August 2011
Departmet of Agrcultural Ecoomcs PhD Qualfer Examato August 0 Istructos: The exam cossts of sx questos You must aswer all questos If you eed a assumpto to complete a questo, state the assumpto clearly
More informationESS Line Fitting
ESS 5 014 17. Le Fttg A very commo problem data aalyss s lookg for relatoshpetwee dfferet parameters ad fttg les or surfaces to data. The smplest example s fttg a straght le ad we wll dscuss that here
More informationChapter 3 Sampling For Proportions and Percentages
Chapter 3 Samplg For Proportos ad Percetages I may stuatos, the characterstc uder study o whch the observatos are collected are qualtatve ature For example, the resposes of customers may marketg surveys
More informationMidterm Exam 1, section 2 (Solution) Thursday, February hour, 15 minutes
coometrcs, CON Sa Fracsco State Uverst Mchael Bar Sprg 5 Mdterm xam, secto Soluto Thursda, Februar 6 hour, 5 mutes Name: Istructos. Ths s closed book, closed otes exam.. No calculators of a kd are allowed..
More information1. The weight of six Golden Retrievers is 66, 61, 70, 67, 92 and 66 pounds. The weight of six Labrador Retrievers is 54, 60, 72, 78, 84 and 67.
Ecoomcs 3 Itroducto to Ecoometrcs Sprg 004 Professor Dobk Name Studet ID Frst Mdterm Exam You must aswer all the questos. The exam s closed book ad closed otes. You may use your calculators but please
More information12.2 Estimating Model parameters Assumptions: ox and y are related according to the simple linear regression model
1. Estmatg Model parameters Assumptos: ox ad y are related accordg to the smple lear regresso model (The lear regresso model s the model that says that x ad y are related a lear fasho, but the observed
More informationLecture 9: Tolerant Testing
Lecture 9: Tolerat Testg Dael Kae Scrbe: Sakeerth Rao Aprl 4, 07 Abstract I ths lecture we prove a quas lear lower boud o the umber of samples eeded to do tolerat testg for L dstace. Tolerat Testg We have
More informationLecture Notes Types of economic variables
Lecture Notes 3 1. Types of ecoomc varables () Cotuous varable takes o a cotuum the sample space, such as all pots o a le or all real umbers Example: GDP, Polluto cocetrato, etc. () Dscrete varables fte
More informationChapter 9 Jordan Block Matrices
Chapter 9 Jorda Block atrces I ths chapter we wll solve the followg problem. Gve a lear operator T fd a bass R of F such that the matrx R (T) s as smple as possble. f course smple s a matter of taste.
More informationX ε ) = 0, or equivalently, lim
Revew for the prevous lecture Cocepts: order statstcs Theorems: Dstrbutos of order statstcs Examples: How to get the dstrbuto of order statstcs Chapter 5 Propertes of a Radom Sample Secto 55 Covergece
More informationLecture Notes to Rice Chapter 5
ECON 430 Revsed Sept. 06 Lecture Notes to Rce Chapter 5 By H. Goldste. Chapter 5 gves a troducto to probablstc approxmato methods, but s suffcet for the eeds of a adequate study of ecoometrcs. The commo
More informationPoint Estimation: definition of estimators
Pot Estmato: defto of estmators Pot estmator: ay fucto W (X,..., X ) of a data sample. The exercse of pot estmato s to use partcular fuctos of the data order to estmate certa ukow populato parameters.
More informationRademacher Complexity. Examples
Algorthmc Foudatos of Learg Lecture 3 Rademacher Complexty. Examples Lecturer: Patrck Rebesch Verso: October 16th 018 3.1 Itroducto I the last lecture we troduced the oto of Rademacher complexty ad showed
More information4. Standard Regression Model and Spatial Dependence Tests
4. Stadard Regresso Model ad Spatal Depedece Tests Stadard regresso aalss fals the presece of spatal effects. I case of spatal depedeces ad/or spatal heterogeet a stadard regresso model wll be msspecfed.
More informationTHE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA
THE ROYAL STATISTICAL SOCIETY 3 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA PAPER I STATISTICAL THEORY & METHODS The Socety provdes these solutos to assst caddates preparg for the examatos future years ad
More informationLogistic regression (continued)
STAT562 page 138 Logstc regresso (cotued) Suppose we ow cosder more complex models to descrbe the relatoshp betwee a categorcal respose varable (Y) that takes o two (2) possble outcomes ad a set of p explaatory
More information22 Nonparametric Methods.
22 oparametrc Methods. I parametrc models oe assumes apror that the dstrbutos have a specfc form wth oe or more ukow parameters ad oe tres to fd the best or atleast reasoably effcet procedures that aswer
More informationSome Different Perspectives on Linear Least Squares
Soe Dfferet Perspectves o Lear Least Squares A stadard proble statstcs s to easure a respose or depedet varable, y, at fed values of oe or ore depedet varables. Soetes there ests a deterstc odel y f (,,
More informationFridayʼs lecture" Problem solutions" Joint densities" 1."E(X) xf (x) dx (x,y) dy X,Y Marginal distributions" The distribution of a ratio" Problems"
Frdayʼs lecture" Jot destes" Margal dstrbutos" The dstrbuto of a rato" Problems" Problem solutos" 1." E(X) = xf X (x)dx = x f X,Y (x,y)dy dx 2. E(X) " = kp X (k) = p X (1) + 2p X (2) +... = xf X,Y (x,y)dxdy
More informationCS 2750 Machine Learning. Lecture 8. Linear regression. CS 2750 Machine Learning. Linear regression. is a linear combination of input components x
CS 75 Mache Learg Lecture 8 Lear regresso Mlos Hauskrecht mlos@cs.ptt.edu 539 Seott Square CS 75 Mache Learg Lear regresso Fucto f : X Y s a lear combato of put compoets f + + + K d d K k - parameters
More information16 Homework lecture 16
Quees College, CUNY, Departmet of Computer Scece Numercal Methods CSCI 361 / 761 Fall 2018 Istructor: Dr. Sateesh Mae c Sateesh R. Mae 2018 16 Homework lecture 16 Please emal your soluto, as a fle attachmet,
More informationSupervised learning: Linear regression Logistic regression
CS 57 Itroducto to AI Lecture 4 Supervsed learg: Lear regresso Logstc regresso Mlos Hauskrecht mlos@cs.ptt.edu 539 Seott Square CS 57 Itro to AI Data: D { D D.. D D Supervsed learg d a set of eamples s
More informationLecture 2: Linear Least Squares Regression
Lecture : Lear Least Squares Regresso Dave Armstrog UW Mlwaukee February 8, 016 Is the Relatoshp Lear? lbrary(car) data(davs) d 150) Davs$weght[d]
More informationThe Occupancy and Coupon Collector problems
Chapter 4 The Occupacy ad Coupo Collector problems By Sarel Har-Peled, Jauary 9, 08 4 Prelmares [ Defto 4 Varace ad Stadard Devato For a radom varable X, let V E [ X [ µ X deote the varace of X, where
More informationMultiple Regression. More than 2 variables! Grade on Final. Multiple Regression 11/21/2012. Exam 2 Grades. Exam 2 Re-grades
STAT 101 Dr. Kar Lock Morga 11/20/12 Exam 2 Grades Multple Regresso SECTIONS 9.2, 10.1, 10.2 Multple explaatory varables (10.1) Parttog varablty R 2, ANOVA (9.2) Codtos resdual plot (10.2) Trasformatos
More information