7: Estimation: The Least-Squares Method

Size: px
Start display at page:

Download "7: Estimation: The Least-Squares Method"

Transcription

1 7: Estmaton: The Least-Squares Method Generaltes. Outlne of the method Suppose that a data sample conssts of a N pars of ponts {x, y }; the x s are known exactly, whle the y s have been measured wth the resoluton σ, wth ndependent errors. Let s frst assumes that the errors are gaussan,and that y s gven by a theoretcal functon y = f(x; a,, a p ). Thus the ndvdual probablty for the pont s and the logarthm of the lkelhood s: P (y ; a) = (y f(x,a) σ π e σ log L = ( y f(x, a) ) σ log( πσ ) The Maxmum lkelhood method conssts of mnmzng the quantty log L, whch corresponds to wth respect to each of the p parameters a k : [ y f(x, a) σ ] χ k, [ y f(x, a) ] = 0 a k σ The use of the χ notaton, n the above defnton, wll be justfed further: t wll be shown that for the case of errors ndeed dstrbuted accordng to a normal law, χ s dstrbuted accordng to the χ dstrbuton.. Least Squares Method when errors are not Gaussan In the general case where no assumpton can be made on the dstrbuton of the errors the method s stll the same,.e t conssts of mnmzng the quantty X = [ y f(x, a) σ ] The notaton X nstead of χ remnds that the sum s not dstrbuted accordng to the χ dstrbuton. Qute frequently, when the observatons are known to be of dfferent accuracy, one can not clam a precse knowledge of the errors. Instead one must estmate the error from each ndvdual measurement. Suppose for example that the measurement y s the number of events n a class. One could use the approxmaton σ = y, equvalent to consderng the true f n a Posson dstrbuton of mean and varance f. In such a case, nstead of the rgorous formula X = [y f(x, a)] f

2 one could use the smplfed LS estmator whch conssts of mnmzng the quantty X [y f(x, a)] y One should also note that the LS estmaton method makes no requrement about the dstrbutonal propertes of the observables. In ths sense, t s dstrbuton-free..3 Least Squares Method when measurements are correlated If the observatons are correlated, wth errors and covarance terms gven n the covarance matrx V = V (y), the Least Squares prncple for estmatng the parameters s formulated as N N X = (y f ) Vj (y j f j ) = mnmum = j= The most classc example: the straght lne ft. Estmators of the slope and the y-ntercept Assume that you want to ft your data ponts to the functon y = ax + b. The χ and ts partal dervatves wth respect to a and b can be wrtten: χ = χ b χ a = = (y ax b) σ σ σ (y ax b) x (y ax b) One has to solve the system of equatons: 0 = y σ â x σ b σ 0 = y x σ â x σ b x σ Introducng the X notaton: x /σ x w x = /σ = w ỹ = y /σ y w /σ = w x x = /σ x /σ = w w x y /σ x y w xy = /σ = w

3 Ths can be rewrtten as: 0 = ỹ â x b 0 = xy â x b x â = xy xỹ x x b = ỹ â x b = ỹ x x xy x x Note that n the smple case where all σ s are equal to a common uncertanty σ, the X notaton reduces to x = x, ỹ = y, ỹ = y, xy = xy,. Parameters covarance n the case of the straght lne ft Snce the least squares s nothng by a specal case of the Maxmum Lkelhood method, one can use the formalsm used n the prevous chapter. Recall that the nverse covarance matrx of the parameters a s: cov (a, a j ) = < log L log L >= < log L > a a j a a j Therefore one can wrte, usng the notaton w = /σ : Thus, the covarance matrx V ab s log L = w (y ax b) log L a = w x (y ax b) log L b = w (y ax b) log L a = w x log L b = w log L a b = w x V ab = V ab = ( ) x x w x ( x x x ) w x x Ths can be rewrtten as: σ (â) = w x x 3

4 σ ( b) = cov(â, b) = ρâ b And n the case of all σ s equal to a common σ: w w = x x σ (â) = σ N σ ( b) = σ N cov(â, b) = σ N ρâ b = x x.3 χ n the case of the straght lne ft x x x x x x x x x x x x x x In the case of a straght lne ft, the actual χ can be evaluated even wthout actually computng â and b. The proof s lengthy, but nstructve: χ = w (y âx b) = w y + â w x + b w â w y x b w y + â b w x χ w = ỹ + â x + b â xỹ bỹ + â b x Notng that b = ỹ â x and replacng â by ts value â = xy xỹ x x χ w = ỹ + â x + ỹ + â x â xỹ â xy ỹ + â xỹ + â xỹ â x = ỹ ỹ + â ( x x ) â( xỹ xy) = ỹ ỹ + = ỹ ỹ = V (y) ( ρ xy) ( xy xỹ) x x ( xy xỹ) x x ( xy xỹ) x x In the case where all σ s are equal to a common σ, ths becomes: χ = N V (y) σ ( ρ xy) These formulae are sometmes useful, when one s nterested n a goodness of ft test (see next chapter) wthout havng to actually compute the estmators a and b. 4

5 .4 Straght lne ft wth uncertantes on both x and y Let s start wth the smple case where both x s and y s have equal uncertantes. If we apply the maxmum lkelhood method, the probablty that a true pont A = (x A, y A ), lyng on the theoretcal lne y = ax + b, s measured as B = (x B, y B ) s just: whch s just: P (A B) = πσ e (x A x B ) /σ e (y A y B ) /σ P (A B) = πσ e u /σ e h /σ To fnd the total probablty that B could come from any pont A, one has to ntegrate P (A B) over all u. The ntegral of e u /σ just gves a constant, σ π, so the probablty of a pont occurrng at B s proportonal to e h /σ, where h s the dstance of the pont B to the lne y = ax + b. In formng the χ to be mnmzed one sums the squared dstances from each measured pont to the pont on the lne from whch t s most lkely to have been produced. The fact that t may actually come from another pont on the lne need not be explctly taken nto account. Specfcally, the dstance between a pont (x, y ) and the lne y = ax + b s h = y ax b + a Thus the χ can be wrtten χ = (y ax b) + a Dfferentatng wth respect to a and b gves, at the cost of a panful calculaton: y = âx + b â = A ± A + A = V (y) V (x) cov(x, y) Out of the solutons, one has to choose the â whose sgn s that of cov(x, y). It must be stressed that ths method works because we assumed that the a pror dstrbuton of ponts s flat and nfnte along a straght lne. In the case that x and y accuraces and not the same and/or the x and y values at a pont are correlated, a change of scale and/or a rotaton can brng us back to the prevous case. The concluson perssts: one needs only to consder the pont on the lne correspondng to the hghest probablty of producng the data pont. 3 The χ dstrbuton 3. Mult-Normal dstrbuton wth, and 3 varables Let s address the followng queston: gven, or 3 ndependent normally dstrbuted varables x, y and z, one can form the quanttes r = x, x + y, or x + y + z. What wll be the dstrbutons of r and r? 5

6 In one dmenson: The standard dstrbuton s f(x) = e x π Under the change of varable r r = h(r), g(r ) = f(r) h (r) = f(r) r g(r ) = r e r = π π (r ) e r = Γ( ) (r ) e Where the frst factor comes from the fact that the equaton y = r has two solutons n r = x, and usng the well known(!) expresson of the Γ functon: Γ( ) = π In two dmensons: We are now consderng the dstrbuton of r = x + y on a crcle n the xy plane, where x and y are normally dstrbuted. On the crcle the dstrbuton s constant and ntegratng over π gves: f(r) = πr e x π r r g(r ) = r re r = π e y = r e r e r = Γ() (r ) 0 e r r In three dmensons: of a sphere and get: Now we consder r = x + y + y ; we now ntegrate over the surface f(r) = 4πr π π e r = r e r π g(r ) = f(r) r = r e r = ) r π 3 Γ( 3 e )(r usng the the well known(!) relaton: Γ( 3 ) = Γ( ) = π/ 3. Defnton of the χ dstrbuton By defnton, the χ dstrbuton wth n degrees of freedom s the functon of the varable χ defned as the sum of the squares of n ndependent varables normally dstrbuted: Let s proof by nducton that: n=, n=, n=3 : see above n f(χ ; n) = N (x; 0, ) = f(χ ; n) = n Γ( n ) (χ ) n e χ 6

7 from n to n+ : Assume that the formula s true for f(χ ; n); for n + the dstrbuton becomes (f g)(x) = Q.E.D = = x t=0 n Γ( n e )tn n Γ( n ) n x n e x n+ Γ( n+ )x t n+ e x x t e dt For ths, we ve used the well-known(!) formula whch derves from ntegraton by parts : Γ(α + ) = = t=0 t=0 = αγ(α) e t t (α+) dt e t αt α dt Note that one can fnd ths result ntutvely: the probablty densty functon s P (x, x,...x n ) = P (x ) P (x )... P (x n ) P (x, x,...x n ) = A e x / e x /... e x n/ P (x, x,...x n ) = A e χ / To get the densty n χ, one has to ntegrate over the hyper-sphere of dmenson χ N and snce one s nterested n a p.d.f of χ, a change of varable wll eat another power of χ. The overall dstrbuton s then B (χ ) n e χ where B s found by the normalzaton condton. Ths s llustrated n the code ch.r 3.3 Basc propertes of the χ dstrbuton The χ dstrbuton s a decreasng functon of χ for n = and n = ; for n >, t starts at zero, reaches a maxmum, then decreases. Its characterstc functon s Performng a change of varable Φ(t) =< e tx >= n Γ( n ) 0 x n e ( x tx) dx y = x( t), dx = dy one gets, usng the defnton Γ(α) = 0 x α e x dx: Φ(t) = n Γ( n ) 0 t y n e y ( t) n ( t)dy = n ( t) n 7 = ( t) n

8 We fnd agan that the χ dstrbuton s nductve (though t was obvous from the defnton): χ (n) χ (n ) = χ (n + n ) From the characterstc functon one can mmedately get ts mean and varance: µ = Φ(t) t=0 = t σ = Φ(t) t=0 µ = t n ( t) n + t=0 = n n(n + ) ( t) n + t=0 n = n + n n = n The result concernng the mean s straghtforward but the one of the varance s more subtle: why a factor? The answer can be found by consderng the varance of a χ wth degree of freedom: t s just the varance of the second central moment of the normal dstrbuton. It s easy to show that the varance of the r th moment of any dstrbuton s: V (x r ) =< (x ) r > < x r > = µ r µ r For the second moment of a normal dstrbuton, one has µ 4 = 3σ 4 and ths becomes, notng that all moments also are central moments: Hence the factor. V (x ) = µ 4 µ = 3σ 4 σ 4 = σ The χ dstrbuton as a specal case of the Γ dstrbuton As we saw, the χ dstrbuton s wrtten: p(χ ; n) = n Γ( n ) (χ ) n e χ In ths form s a the dstrbuton of χ and not χ. Just posng z=χ, t becomes: p(z; n) = n Γ( n ) z n e z/ = Γ(z; α = n, β = ) where Γ s the Gamma dstrbuton we already studed. It follows that Its mean s < χ >= αβ = n Its varance s v(χ ) = αβ = n Its characterstc functon s Φ(t) = ( t) n/ 8

9 3.5 Generatng a varable accordng to a χ dstrbuton A very nce result s that, consderng n ndependent varables (x, x,..., x n ), unformly dstrbuted between 0 and, then u = log(x x...x n ) = χ (n) Ths can be easly shown for n = : y = log x, g(y) = f(x) h (y) = x = e y = χ (y; ) and then by nducton from n to n + ; the algorthm s thus two fold, dependng on whether the number of degrees of freedom one consders s odd or even: n even : generate n/ ndependent varables and apply the above formula n odd : go back to the prevous case and add to the result z, where z s a pseudo random varable dstrbuted accordng to the normal law. For large values of n (typcally n > 0) one can use the fact that χ s very close to a Gaussan dstrbuton of mean n and unt varance. See a graphcal llustraton: chsquare.m 3.6 A fantastc property lnkng the Posson and χ dstrbutons Let s consder a Posson dstrbuton of the nteger varable r and mean µ; a very useful result lnks the probablty that r n (or equvalently r > n) and the probablty that a χ varable s greater than some cut: Proof: P µ (r n >= P [χ (x; n + ) > µ] P [χ (x; n + ) > µ] = x=µ x n e x n! n+ dx = t=µ t n e t dt n! After the change of varable t = x/, dx = dt; now ntegratng by parts and notng that one gets: t=µ t n e t dt = n! x n e x dx = [e x x n ] + n x n e x dx t n e t dt + e µ µ n (n )! t=µ n! = r=n r=0 e µ µ r whch s exactly the sum of the r + ndvdual probabltes of the Posson dstrbuton for r = 0,,...n. Ths property s very mportant from a computatonal pont of vew snce the cumulatve χ probablty s avalable n most computer languages scentfc lbrares, as depcted by the followng Exercse: You are runnng an hgh energy partcle physcs experment, n search for the elusve Hggs boson; you expect b = 3.86 fake background events and you observe n = 9 of them. What s the probablty that your observaton s just a statstcal fluctuaton? (answer: ) 9 r!

10 4 Dstrbuton of the χ of a least squares ft The key (and well known!) result s that the dstrbuton of the χ of a least squares ft, n the case of Gaussan errors, s ndeed a χ dstrbuton of N p degrees of freedom: f(χ N p ; N p) = N p Γ( N p ) (χ ) e χ where N s the number of ponts and p the number of parameters evaluated by the ft. In the case of a straght lne ft, p =. The χ s a sum of N terms u = (y f(x )) σ each of them beng the square of a normal dstrbuton. The ponts of constant probablty le on a N dmensonal hyper-sphere of radus χ. The probablty that a result occurs between χ and χ + δχ thus depends on e χ /, multpled by the volume of the regon concerned, so After change of varable ths becomes P (χ) = αχ N e χ / P (χ ) = β χ N e χ / In secton. dealng wth the case of the ft to a straght lne, we saw that the u are constraned by equatons: 0 = u 0 = u x Thus the ponts u le on an (N )-dmensonal subspace of the orgnal, reducng the dmenson of a regon of equprobablty by two powers of χ. Thus the probablty densty functon s proportonal to P (χ ) = γ χ N 4 e χ / whch s a χ dstrbuton for N degrees of freedom. In case the ft nvolves p parameters nstead of, the above analyss stll holds, one degree of freedom beng lost for each extra parameter. Note that n the case p = N, the ft s not longer a ft but a system of N equatons wth N unknowns. Clearly, f the number of parameters s greater than the number of ponts, the problem s not solvable. 5 Lnear least squares fts A least squares ft s called lnear f the theoretcal functon f(x; a a p ) lnear wth respect to the parameters a ; for example, estmatng a, a, a 3, a 4 n s ndeed a lnear problem, whle estmatng a n y = a + a sn x + a 3 sn(x ) + a 4 x y = cos(ax) s obvously not. The mportant result we want to show s that the LS method provdes an analytc soluton for the parameters a whch can be expressed n a smple closed form. Moreover, the LS estmator n the lnear case s unque, unbased, and s a lnear functon of the observatons y. Fnally, among all estmators whch can be expressed as a lnear functon of the observatons, t s the most effcent. 0

11 5. Normal equatons for least squares fts If many (> ) parameters are to be estmated n a lnear ft, matrx notaton produces elegant results. Let s call a = a,, a p the vector of the p parameters, y = y,, y N the vector of the N measured quanttes and smlarly for x: The ft wll be called lnear f the theoretcal functon f can be expressed as a lnear combnaton of any functons of x; p f(x; a) = c r (x)a r In the general case the X can be expressed as a sum of exponents of the multnormal functons X = [y f(x ; a)]vj [y j f(x j ; a)] j whch can be expressed n matrx notaton: r= X = ( y f) T V (y f) Frst assumng that the y s are ndependent, each of them wth an uncertanty σ, we can wrte N X = [ y p r= a rc r (x ) ] σ Dfferentatng ths wth respect to a r gves the so-called normal equatons: c r (x ) y s âsc s (x ) σ = 0 whch can be expressed n the matrx notaton, ntroducng the matrx C r = c r (x ) f = Ca C T V C â = C T V y â = ( C T V C) C T V y In the above, the dmensons of the varous matrces are: y : N a : p C : N p V : N N C T V : p N C T V y : p C T V C : p p In the case the y s are not ndependent,.e. the matrx V s not dagonal, ths result stll holds: one has to dfferentate the quantty X = ( y Ca) T V (y Ca) wth respect to each of the a k : 0 = â[( y Ca) T V (y Câ)] 0 = C T V (y Câ) C T V y = C T V Câ

12 5. Propertes of Lnear Least Squares Fts The frst rather obvous property s that the Lnear Least Square method provdes a unque soluton, whch s lnear wth respect to the observables y ; secondly, ths estmator s unbased because t s lnear: < â > = < ( C T V C) C T V y > = ( C T V C) C T V < y > = ( C T V C) C T V Ca = a What s the covarance matrx of â? The answer s smple: we saw that f z = Mx, then V (z) = MV (x) M T. In our case M = (C T V C) C T V V (â) = (C T V C) C T V V V C( C T V C) = ( C T V C) 5.3 The Gauss-Markov theorem An nterestng optmum property of the lnear Least Squares estmator s contaned n the Gauss- Markov theorem : among all unbased estmators whch are lnear functons of the observatons, the LS estmators have the smallest varance. To prove t, consder any estmator t of the parameters a, lnear n the observatons y, t = Sy. Its expectaton t s: < t >= S < y >= SCa where a s the orgnal vector of parameters. For t to be unbased, we need < t >= a, a Hence we must have SC =. Now we want to fnd out when the covarance matrx V ( t) for the new estmator has mnmal dagonal term; one has V ( t) = SV (y)s T. For fndng the mnmum we consder the followng dentty, notng V (y) as V, and usng the facts that SC =, V = V T, and (C T V C) T = (C T V C): qtty = [(C T V C) C T V ]V [(C T V C) C T V ] T +[S (C T V C) C T V ]V [S (C T V C) C T V ] T = [(C T V C) C T V ]V [V C(C T V C) ] + SV S T SV V C(C T V C) +[(C T V C) C T V ]V [V C(C T V C) ] (C T V C) C T V V S T = (C T V C) + SV S T (C T V C) + (C T V C) (C T V C) qtty = SV S T = V ( t) Here each of the two terms on the rght hand sde n the defnton of qtty s of the quadratc form UV U T, whch mples non-negatve dagonal elements. Only the second term s a functon of S, and the sum of the two terms wll have strctly mnmum dagonal elements when the second term has vanshng elements on the dagonal. Ths occurs when S = (C T V C) C T V Therefore the mnmum varance unbased estmator for a s where â LS s the Least-Squares estmator for a. t = (C T V C) C T V y = â LS

13 5.4 Lnear Least squares fts wth constrants In frequent stuatons, one s faced wth the problem of estmatng parameters for whch one or more constrants exst. Two approaches can handle such a problem. The Elmnaton method. Lagrange multplers 5.4. Elmnaton method Ths method just conssts, (whenever possble!) n elmnatng some of the parameters by algebrac manpulaton. A very smple example wll llustrate the method: the three angles of a trangle θ, θ and θ 3 have been measured ndependently wth the same uncertanty σ. What s the constraned ( θ = π) LS estmators of the angles? For smplcty, let s call the measurements y, y and y 3 and assume that the measurements are performed n degrees. One wants to solve: X (θ, θ, θ 3 ) = 3 θ = 80 = 3 ( y θ ) = mnmum σ = In ths case, the algebrac manpulaton to elmnate, say, θ 3, s not very dffcult: X = ( y θ σ ) + ( y θ σ ) + [ y 3 (80 θ θ ) ] σ 0 = σ X θ = (y θ ) + [y 3 (80 θ θ )] 0 = σ X θ = (y θ ) + [y 3 (80 θ θ )] θ = 3 (80 + y y y 3 ) θ = 3 (80 y + y y 3 ) θ 3 = 3 (80 y y + y 3 ) An mproved measurement was obtaned by subtractng to each of the measurements y one thrd of the excess y + y + y Ths mprovement s drectly vsble n the varance of the estmators: V ( θ ) = V ( θ ) = V ( θ 3 ) = σ = 9 3 σ Ths result can be ntutvely explaned n terms of nformaton: one could have estmated θ, θ and θ 3 wth two measurements only, say y and y ; n ths case, y 3 would have been computed as y 3 = 80 y y. Addng a thrd measurement ncreases the nformaton by 50%, and we saw, n the secton devoted to the Cramer Rao bound (4.3.), that the mnmum varance s nversely proportonal to the amount of nformaton. 3

14 5.4. Lagrange Multplers Let s assume that we agan want to estmate p parameters a, a a n, lnked together by m constrants: D k (a) = 0, k =, m The basc trck of the method s to replace the p-component vector of the parameters a by a p+mcomponent vector whose frst p components are the usual a and the remanng ones λ k, k =, m, such as the new X s now: X = ( y Ca) T V (y Ca) + λ T D = mnmum If we here equate to zero the dervatves of X wth respect to a, a a p and λ k, = m, we get the normal equatons: âx = (C T V y C T V Câ) + ( âd) T λ = 0 λ X = D = 0 We are now left wth a set of p + m equatons. The second ones are the constrants, whle the frst ones are the analog of the normal equatons for the unconstraned case, now modfed by the λ term due to the constrants. In the case the constrants themselves are lnear, they can be expressed as Ba = b and the set of the modfed p + k normal equatons becomes 0 = (C T V y C T V Câ) + B T λ 0 = Bâ b whch s a set of lnear equatons. After some heavy algebra, one can show that the varance matrx of the estmator â s: V (â) = (C T V C) [B(C T V C) ] T [B T V B] [B(C T V C) ] Ths expresson can not be smplfed any further because nether C or B are square matrces! Note that the matrx (B T V B) has non negatve dagonal terms, and that the same s true for the second term of the above equaton. As a result, the constrant equatons wll always lead to a reducton of the parameter errors (.e. the dagonal terms) compared to the unconstraned case. Back to the prevous exercse, runnng the machnery yelds: χ = 3 = ( y θ σ 0 = χ θ = σ (y θ ) + λ 0 = χ θ = σ (y θ ) + λ 0 = χ θ 3 = σ (y 3 θ 3 ) + λ 4 ) + λ( θ 80)

15 0 = χ = ( θ 80) λ 0 = ( y 80) 3 λσ λ = 3σ ( y 80) θ = y λσ = y 3 ( y 80) 6 Non lnear least squares fts If the functon f(x; a) s not lnear n the a r, you generally have to use an teratve technque to solve the equatons. Gven a frst guess a 0, a Taylor expanson around a 0 wll yeld a new set of values a such as X a r a=a 0 and further teratons wll lead to the soluton. In such teratve methods, havng a good frst guess s wnnng 90% of the battle. The best advce s to fnd a relable software lbrary rather than beng tempted to wrte one s own package. 5

1 Matrix representations of canonical matrices

1 Matrix representations of canonical matrices 1 Matrx representatons of canoncal matrces 2-d rotaton around the orgn: ( ) cos θ sn θ R 0 = sn θ cos θ 3-d rotaton around the x-axs: R x = 1 0 0 0 cos θ sn θ 0 sn θ cos θ 3-d rotaton around the y-axs:

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede Fall 0 Analyss of Expermental easurements B. Esensten/rev. S. Errede We now reformulate the lnear Least Squares ethod n more general terms, sutable for (eventually extendng to the non-lnear case, and also

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Formulas for the Determinant

Formulas for the Determinant page 224 224 CHAPTER 3 Determnants e t te t e 2t 38 A = e t 2te t e 2t e t te t 2e 2t 39 If 123 A = 345, 456 compute the matrx product A adj(a) What can you conclude about det(a)? For Problems 40 43, use

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

SIO 224. m(r) =(ρ(r),k s (r),µ(r)) SIO 224 1. A bref look at resoluton analyss Here s some background for the Masters and Gubbns resoluton paper. Global Earth models are usually found teratvely by assumng a startng model and fndng small

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Lecture 3 Stat102, Spring 2007

Lecture 3 Stat102, Spring 2007 Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

APPENDIX A Some Linear Algebra

APPENDIX A Some Linear Algebra APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,

More information

Integrals and Invariants of Euler-Lagrange Equations

Integrals and Invariants of Euler-Lagrange Equations Lecture 16 Integrals and Invarants of Euler-Lagrange Equatons ME 256 at the Indan Insttute of Scence, Bengaluru Varatonal Methods and Structural Optmzaton G. K. Ananthasuresh Professor, Mechancal Engneerng,

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Difference Equations

Difference Equations Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Lecture 3. Ax x i a i. i i

Lecture 3. Ax x i a i. i i 18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Solutions Homework 4 March 5, 2018

Solutions Homework 4 March 5, 2018 1 Solutons Homework 4 March 5, 018 Soluton to Exercse 5.1.8: Let a IR be a translaton and c > 0 be a re-scalng. ˆb1 (cx + a) cx n + a (cx 1 + a) c x n x 1 cˆb 1 (x), whch shows ˆb 1 s locaton nvarant and

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

3.1 ML and Empirical Distribution

3.1 ML and Empirical Distribution 67577 Intro. to Machne Learnng Fall semester, 2008/9 Lecture 3: Maxmum Lkelhood/ Maxmum Entropy Dualty Lecturer: Amnon Shashua Scrbe: Amnon Shashua 1 In the prevous lecture we defned the prncple of Maxmum

More information

SELECTED PROOFS. DeMorgan s formulas: The first one is clear from Venn diagram, or the following truth table:

SELECTED PROOFS. DeMorgan s formulas: The first one is clear from Venn diagram, or the following truth table: SELECTED PROOFS DeMorgan s formulas: The frst one s clear from Venn dagram, or the followng truth table: A B A B A B Ā B Ā B T T T F F F F T F T F F T F F T T F T F F F F F T T T T The second one can be

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Solutions to exam in SF1811 Optimization, Jan 14, 2015 Solutons to exam n SF8 Optmzaton, Jan 4, 25 3 3 O------O -4 \ / \ / The network: \/ where all lnks go from left to rght. /\ / \ / \ 6 O------O -5 2 4.(a) Let x = ( x 3, x 4, x 23, x 24 ) T, where the varable

More information

Section 8.3 Polar Form of Complex Numbers

Section 8.3 Polar Form of Complex Numbers 80 Chapter 8 Secton 8 Polar Form of Complex Numbers From prevous classes, you may have encountered magnary numbers the square roots of negatve numbers and, more generally, complex numbers whch are the

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011 Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore 8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø

More information

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2) 1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

More metrics on cartesian products

More metrics on cartesian products More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

Radar Trackers. Study Guide. All chapters, problems, examples and page numbers refer to Applied Optimal Estimation, A. Gelb, Ed.

Radar Trackers. Study Guide. All chapters, problems, examples and page numbers refer to Applied Optimal Estimation, A. Gelb, Ed. Radar rackers Study Gude All chapters, problems, examples and page numbers refer to Appled Optmal Estmaton, A. Gelb, Ed. Chapter Example.0- Problem Statement wo sensors Each has a sngle nose measurement

More information

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University PHYS 45 Sprng semester 7 Lecture : Dealng wth Expermental Uncertantes Ron Refenberger Brck anotechnology Center Purdue Unversty Lecture Introductory Comments Expermental errors (really expermental uncertantes)

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

Linear Regression Analysis: Terminology and Notation

Linear Regression Analysis: Terminology and Notation ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented

More information

Assortment Optimization under MNL

Assortment Optimization under MNL Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.

More information

4DVAR, according to the name, is a four-dimensional variational method.

4DVAR, according to the name, is a four-dimensional variational method. 4D-Varatonal Data Assmlaton (4D-Var) 4DVAR, accordng to the name, s a four-dmensonal varatonal method. 4D-Var s actually a drect generalzaton of 3D-Var to handle observatons that are dstrbuted n tme. The

More information

4.3 Poisson Regression

4.3 Poisson Regression of teratvely reweghted least squares regressons (the IRLS algorthm). We do wthout gvng further detals, but nstead focus on the practcal applcaton. > glm(survval~log(weght)+age, famly="bnomal", data=baby)

More information

ELASTIC WAVE PROPAGATION IN A CONTINUOUS MEDIUM

ELASTIC WAVE PROPAGATION IN A CONTINUOUS MEDIUM ELASTIC WAVE PROPAGATION IN A CONTINUOUS MEDIUM An elastc wave s a deformaton of the body that travels throughout the body n all drectons. We can examne the deformaton over a perod of tme by fxng our look

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

MMA and GCMMA two methods for nonlinear optimization

MMA and GCMMA two methods for nonlinear optimization MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons

More information

Module 2. Random Processes. Version 2 ECE IIT, Kharagpur

Module 2. Random Processes. Version 2 ECE IIT, Kharagpur Module Random Processes Lesson 6 Functons of Random Varables After readng ths lesson, ou wll learn about cdf of functon of a random varable. Formula for determnng the pdf of a random varable. Let, X be

More information

Singular Value Decomposition: Theory and Applications

Singular Value Decomposition: Theory and Applications Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real

More information

Canonical transformations

Canonical transformations Canoncal transformatons November 23, 2014 Recall that we have defned a symplectc transformaton to be any lnear transformaton M A B leavng the symplectc form nvarant, Ω AB M A CM B DΩ CD Coordnate transformatons,

More information

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

Chapter 6. Supplemental Text Material

Chapter 6. Supplemental Text Material Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.

More information

Economics 130. Lecture 4 Simple Linear Regression Continued

Economics 130. Lecture 4 Simple Linear Regression Continued Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do

More information

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0 MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

MA 323 Geometric Modelling Course Notes: Day 13 Bezier Curves & Bernstein Polynomials

MA 323 Geometric Modelling Course Notes: Day 13 Bezier Curves & Bernstein Polynomials MA 323 Geometrc Modellng Course Notes: Day 13 Bezer Curves & Bernsten Polynomals Davd L. Fnn Over the past few days, we have looked at de Casteljau s algorthm for generatng a polynomal curve, and we have

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980 MT07: Multvarate Statstcal Methods Mke Tso: emal mke.tso@manchester.ac.uk Webpage for notes: http://www.maths.manchester.ac.uk/~mkt/new_teachng.htm. Introducton to multvarate data. Books Chat eld, C. and

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,

More information

PHYS 705: Classical Mechanics. Calculus of Variations II

PHYS 705: Classical Mechanics. Calculus of Variations II 1 PHYS 705: Classcal Mechancs Calculus of Varatons II 2 Calculus of Varatons: Generalzaton (no constrant yet) Suppose now that F depends on several dependent varables : We need to fnd such that has a statonary

More information

Important Instructions to the Examiners:

Important Instructions to the Examiners: Summer 0 Examnaton Subject & Code: asc Maths (70) Model Answer Page No: / Important Instructons to the Examners: ) The Answers should be examned by key words and not as word-to-word as gven n the model

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

Laboratory 3: Method of Least Squares

Laboratory 3: Method of Least Squares Laboratory 3: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly they are correlated wth

More information

Bernoulli Numbers and Polynomials

Bernoulli Numbers and Polynomials Bernoull Numbers and Polynomals T. Muthukumar tmk@tk.ac.n 17 Jun 2014 The sum of frst n natural numbers 1, 2, 3,..., n s n n(n + 1 S 1 (n := m = = n2 2 2 + n 2. Ths formula can be derved by notng that

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models

More information

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016 U.C. Berkeley CS94: Spectral Methods and Expanders Handout 8 Luca Trevsan February 7, 06 Lecture 8: Spectral Algorthms Wrap-up In whch we talk about even more generalzatons of Cheeger s nequaltes, and

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

β0 + β1xi. You are interested in estimating the unknown parameters β

β0 + β1xi. You are interested in estimating the unknown parameters β Ordnary Least Squares (OLS): Smple Lnear Regresson (SLR) Analytcs The SLR Setup Sample Statstcs Ordnary Least Squares (OLS): FOCs and SOCs Back to OLS and Sample Statstcs Predctons (and Resduals) wth OLS

More information

= = = (a) Use the MATLAB command rref to solve the system. (b) Let A be the coefficient matrix and B be the right-hand side of the system.

= = = (a) Use the MATLAB command rref to solve the system. (b) Let A be the coefficient matrix and B be the right-hand side of the system. Chapter Matlab Exercses Chapter Matlab Exercses. Consder the lnear system of Example n Secton.. x x x y z y y z (a) Use the MATLAB command rref to solve the system. (b) Let A be the coeffcent matrx and

More information

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1 C/CS/Phy9 Problem Set 3 Solutons Out: Oct, 8 Suppose you have two qubts n some arbtrary entangled state ψ You apply the teleportaton protocol to each of the qubts separately What s the resultng state obtaned

More information

Physics 5153 Classical Mechanics. D Alembert s Principle and The Lagrangian-1

Physics 5153 Classical Mechanics. D Alembert s Principle and The Lagrangian-1 P. Guterrez Physcs 5153 Classcal Mechancs D Alembert s Prncple and The Lagrangan 1 Introducton The prncple of vrtual work provdes a method of solvng problems of statc equlbrum wthout havng to consder the

More information

2.3 Nilpotent endomorphisms

2.3 Nilpotent endomorphisms s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms

More information

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14 APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce

More information

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of Chapter 7 Generalzed and Weghted Least Squares Estmaton The usual lnear regresson model assumes that all the random error components are dentcally and ndependently dstrbuted wth constant varance. When

More information

Feb 14: Spatial analysis of data fields

Feb 14: Spatial analysis of data fields Feb 4: Spatal analyss of data felds Mappng rregularly sampled data onto a regular grd Many analyss technques for geophyscal data requre the data be located at regular ntervals n space and/or tme. hs s

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

Moments of Inertia. and reminds us of the analogous equation for linear momentum p= mv, which is of the form. The kinetic energy of the body is.

Moments of Inertia. and reminds us of the analogous equation for linear momentum p= mv, which is of the form. The kinetic energy of the body is. Moments of Inerta Suppose a body s movng on a crcular path wth constant speed Let s consder two quanttes: the body s angular momentum L about the center of the crcle, and ts knetc energy T How are these

More information

Modelli Clamfim Equazione del Calore Lezione ottobre 2014

Modelli Clamfim Equazione del Calore Lezione ottobre 2014 CLAMFIM Bologna Modell 1 @ Clamfm Equazone del Calore Lezone 17 15 ottobre 2014 professor Danele Rtell danele.rtell@unbo.t 1/24? Convoluton The convoluton of two functons g(t) and f(t) s the functon (g

More information

Probability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n!

Probability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n! 8333: Statstcal Mechancs I Problem Set # 3 Solutons Fall 3 Characterstc Functons: Probablty Theory The characterstc functon s defned by fk ep k = ep kpd The nth coeffcent of the Taylor seres of fk epanded

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13 CME 30: NUMERICAL LINEAR ALGEBRA FALL 005/06 LECTURE 13 GENE H GOLUB 1 Iteratve Methods Very large problems (naturally sparse, from applcatons): teratve methods Structured matrces (even sometmes dense,

More information

Notes on Frequency Estimation in Data Streams

Notes on Frequency Estimation in Data Streams Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to

More information

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence

More information

A New Refinement of Jacobi Method for Solution of Linear System Equations AX=b

A New Refinement of Jacobi Method for Solution of Linear System Equations AX=b Int J Contemp Math Scences, Vol 3, 28, no 17, 819-827 A New Refnement of Jacob Method for Soluton of Lnear System Equatons AX=b F Naem Dafchah Department of Mathematcs, Faculty of Scences Unversty of Gulan,

More information

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2 Salmon: Lectures on partal dfferental equatons 5. Classfcaton of second-order equatons There are general methods for classfyng hgher-order partal dfferental equatons. One s very general (applyng even to

More information

Affine transformations and convexity

Affine transformations and convexity Affne transformatons and convexty The purpose of ths document s to prove some basc propertes of affne transformatons nvolvng convex sets. Here are a few onlne references for background nformaton: http://math.ucr.edu/

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017 U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that

More information

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng

More information

Vector Norms. Chapter 7 Iterative Techniques in Matrix Algebra. Cauchy-Bunyakovsky-Schwarz Inequality for Sums. Distances. Convergence.

Vector Norms. Chapter 7 Iterative Techniques in Matrix Algebra. Cauchy-Bunyakovsky-Schwarz Inequality for Sums. Distances. Convergence. Vector Norms Chapter 7 Iteratve Technques n Matrx Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematcs Unversty of Calforna, Berkeley Math 128B Numercal Analyss Defnton A vector norm

More information

Effects of Ignoring Correlations When Computing Sample Chi-Square. John W. Fowler February 26, 2012

Effects of Ignoring Correlations When Computing Sample Chi-Square. John W. Fowler February 26, 2012 Effects of Ignorng Correlatons When Computng Sample Ch-Square John W. Fowler February 6, 0 It can happen that ch-square must be computed for a sample whose elements are correlated to an unknown extent.

More information

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities Supplementary materal: Margn based PU Learnng We gve the complete proofs of Theorem and n Secton We frst ntroduce the well-known concentraton nequalty, so the covarance estmator can be bounded Then we

More information