7: Estimation: The Least-Squares Method
|
|
- Jacob Henry
- 5 years ago
- Views:
Transcription
1 7: Estmaton: The Least-Squares Method Generaltes. Outlne of the method Suppose that a data sample conssts of a N pars of ponts {x, y }; the x s are known exactly, whle the y s have been measured wth the resoluton σ, wth ndependent errors. Let s frst assumes that the errors are gaussan,and that y s gven by a theoretcal functon y = f(x; a,, a p ). Thus the ndvdual probablty for the pont s and the logarthm of the lkelhood s: P (y ; a) = (y f(x,a) σ π e σ log L = ( y f(x, a) ) σ log( πσ ) The Maxmum lkelhood method conssts of mnmzng the quantty log L, whch corresponds to wth respect to each of the p parameters a k : [ y f(x, a) σ ] χ k, [ y f(x, a) ] = 0 a k σ The use of the χ notaton, n the above defnton, wll be justfed further: t wll be shown that for the case of errors ndeed dstrbuted accordng to a normal law, χ s dstrbuted accordng to the χ dstrbuton.. Least Squares Method when errors are not Gaussan In the general case where no assumpton can be made on the dstrbuton of the errors the method s stll the same,.e t conssts of mnmzng the quantty X = [ y f(x, a) σ ] The notaton X nstead of χ remnds that the sum s not dstrbuted accordng to the χ dstrbuton. Qute frequently, when the observatons are known to be of dfferent accuracy, one can not clam a precse knowledge of the errors. Instead one must estmate the error from each ndvdual measurement. Suppose for example that the measurement y s the number of events n a class. One could use the approxmaton σ = y, equvalent to consderng the true f n a Posson dstrbuton of mean and varance f. In such a case, nstead of the rgorous formula X = [y f(x, a)] f
2 one could use the smplfed LS estmator whch conssts of mnmzng the quantty X [y f(x, a)] y One should also note that the LS estmaton method makes no requrement about the dstrbutonal propertes of the observables. In ths sense, t s dstrbuton-free..3 Least Squares Method when measurements are correlated If the observatons are correlated, wth errors and covarance terms gven n the covarance matrx V = V (y), the Least Squares prncple for estmatng the parameters s formulated as N N X = (y f ) Vj (y j f j ) = mnmum = j= The most classc example: the straght lne ft. Estmators of the slope and the y-ntercept Assume that you want to ft your data ponts to the functon y = ax + b. The χ and ts partal dervatves wth respect to a and b can be wrtten: χ = χ b χ a = = (y ax b) σ σ σ (y ax b) x (y ax b) One has to solve the system of equatons: 0 = y σ â x σ b σ 0 = y x σ â x σ b x σ Introducng the X notaton: x /σ x w x = /σ = w ỹ = y /σ y w /σ = w x x = /σ x /σ = w w x y /σ x y w xy = /σ = w
3 Ths can be rewrtten as: 0 = ỹ â x b 0 = xy â x b x â = xy xỹ x x b = ỹ â x b = ỹ x x xy x x Note that n the smple case where all σ s are equal to a common uncertanty σ, the X notaton reduces to x = x, ỹ = y, ỹ = y, xy = xy,. Parameters covarance n the case of the straght lne ft Snce the least squares s nothng by a specal case of the Maxmum Lkelhood method, one can use the formalsm used n the prevous chapter. Recall that the nverse covarance matrx of the parameters a s: cov (a, a j ) = < log L log L >= < log L > a a j a a j Therefore one can wrte, usng the notaton w = /σ : Thus, the covarance matrx V ab s log L = w (y ax b) log L a = w x (y ax b) log L b = w (y ax b) log L a = w x log L b = w log L a b = w x V ab = V ab = ( ) x x w x ( x x x ) w x x Ths can be rewrtten as: σ (â) = w x x 3
4 σ ( b) = cov(â, b) = ρâ b And n the case of all σ s equal to a common σ: w w = x x σ (â) = σ N σ ( b) = σ N cov(â, b) = σ N ρâ b = x x.3 χ n the case of the straght lne ft x x x x x x x x x x x x x x In the case of a straght lne ft, the actual χ can be evaluated even wthout actually computng â and b. The proof s lengthy, but nstructve: χ = w (y âx b) = w y + â w x + b w â w y x b w y + â b w x χ w = ỹ + â x + b â xỹ bỹ + â b x Notng that b = ỹ â x and replacng â by ts value â = xy xỹ x x χ w = ỹ + â x + ỹ + â x â xỹ â xy ỹ + â xỹ + â xỹ â x = ỹ ỹ + â ( x x ) â( xỹ xy) = ỹ ỹ + = ỹ ỹ = V (y) ( ρ xy) ( xy xỹ) x x ( xy xỹ) x x ( xy xỹ) x x In the case where all σ s are equal to a common σ, ths becomes: χ = N V (y) σ ( ρ xy) These formulae are sometmes useful, when one s nterested n a goodness of ft test (see next chapter) wthout havng to actually compute the estmators a and b. 4
5 .4 Straght lne ft wth uncertantes on both x and y Let s start wth the smple case where both x s and y s have equal uncertantes. If we apply the maxmum lkelhood method, the probablty that a true pont A = (x A, y A ), lyng on the theoretcal lne y = ax + b, s measured as B = (x B, y B ) s just: whch s just: P (A B) = πσ e (x A x B ) /σ e (y A y B ) /σ P (A B) = πσ e u /σ e h /σ To fnd the total probablty that B could come from any pont A, one has to ntegrate P (A B) over all u. The ntegral of e u /σ just gves a constant, σ π, so the probablty of a pont occurrng at B s proportonal to e h /σ, where h s the dstance of the pont B to the lne y = ax + b. In formng the χ to be mnmzed one sums the squared dstances from each measured pont to the pont on the lne from whch t s most lkely to have been produced. The fact that t may actually come from another pont on the lne need not be explctly taken nto account. Specfcally, the dstance between a pont (x, y ) and the lne y = ax + b s h = y ax b + a Thus the χ can be wrtten χ = (y ax b) + a Dfferentatng wth respect to a and b gves, at the cost of a panful calculaton: y = âx + b â = A ± A + A = V (y) V (x) cov(x, y) Out of the solutons, one has to choose the â whose sgn s that of cov(x, y). It must be stressed that ths method works because we assumed that the a pror dstrbuton of ponts s flat and nfnte along a straght lne. In the case that x and y accuraces and not the same and/or the x and y values at a pont are correlated, a change of scale and/or a rotaton can brng us back to the prevous case. The concluson perssts: one needs only to consder the pont on the lne correspondng to the hghest probablty of producng the data pont. 3 The χ dstrbuton 3. Mult-Normal dstrbuton wth, and 3 varables Let s address the followng queston: gven, or 3 ndependent normally dstrbuted varables x, y and z, one can form the quanttes r = x, x + y, or x + y + z. What wll be the dstrbutons of r and r? 5
6 In one dmenson: The standard dstrbuton s f(x) = e x π Under the change of varable r r = h(r), g(r ) = f(r) h (r) = f(r) r g(r ) = r e r = π π (r ) e r = Γ( ) (r ) e Where the frst factor comes from the fact that the equaton y = r has two solutons n r = x, and usng the well known(!) expresson of the Γ functon: Γ( ) = π In two dmensons: We are now consderng the dstrbuton of r = x + y on a crcle n the xy plane, where x and y are normally dstrbuted. On the crcle the dstrbuton s constant and ntegratng over π gves: f(r) = πr e x π r r g(r ) = r re r = π e y = r e r e r = Γ() (r ) 0 e r r In three dmensons: of a sphere and get: Now we consder r = x + y + y ; we now ntegrate over the surface f(r) = 4πr π π e r = r e r π g(r ) = f(r) r = r e r = ) r π 3 Γ( 3 e )(r usng the the well known(!) relaton: Γ( 3 ) = Γ( ) = π/ 3. Defnton of the χ dstrbuton By defnton, the χ dstrbuton wth n degrees of freedom s the functon of the varable χ defned as the sum of the squares of n ndependent varables normally dstrbuted: Let s proof by nducton that: n=, n=, n=3 : see above n f(χ ; n) = N (x; 0, ) = f(χ ; n) = n Γ( n ) (χ ) n e χ 6
7 from n to n+ : Assume that the formula s true for f(χ ; n); for n + the dstrbuton becomes (f g)(x) = Q.E.D = = x t=0 n Γ( n e )tn n Γ( n ) n x n e x n+ Γ( n+ )x t n+ e x x t e dt For ths, we ve used the well-known(!) formula whch derves from ntegraton by parts : Γ(α + ) = = t=0 t=0 = αγ(α) e t t (α+) dt e t αt α dt Note that one can fnd ths result ntutvely: the probablty densty functon s P (x, x,...x n ) = P (x ) P (x )... P (x n ) P (x, x,...x n ) = A e x / e x /... e x n/ P (x, x,...x n ) = A e χ / To get the densty n χ, one has to ntegrate over the hyper-sphere of dmenson χ N and snce one s nterested n a p.d.f of χ, a change of varable wll eat another power of χ. The overall dstrbuton s then B (χ ) n e χ where B s found by the normalzaton condton. Ths s llustrated n the code ch.r 3.3 Basc propertes of the χ dstrbuton The χ dstrbuton s a decreasng functon of χ for n = and n = ; for n >, t starts at zero, reaches a maxmum, then decreases. Its characterstc functon s Performng a change of varable Φ(t) =< e tx >= n Γ( n ) 0 x n e ( x tx) dx y = x( t), dx = dy one gets, usng the defnton Γ(α) = 0 x α e x dx: Φ(t) = n Γ( n ) 0 t y n e y ( t) n ( t)dy = n ( t) n 7 = ( t) n
8 We fnd agan that the χ dstrbuton s nductve (though t was obvous from the defnton): χ (n) χ (n ) = χ (n + n ) From the characterstc functon one can mmedately get ts mean and varance: µ = Φ(t) t=0 = t σ = Φ(t) t=0 µ = t n ( t) n + t=0 = n n(n + ) ( t) n + t=0 n = n + n n = n The result concernng the mean s straghtforward but the one of the varance s more subtle: why a factor? The answer can be found by consderng the varance of a χ wth degree of freedom: t s just the varance of the second central moment of the normal dstrbuton. It s easy to show that the varance of the r th moment of any dstrbuton s: V (x r ) =< (x ) r > < x r > = µ r µ r For the second moment of a normal dstrbuton, one has µ 4 = 3σ 4 and ths becomes, notng that all moments also are central moments: Hence the factor. V (x ) = µ 4 µ = 3σ 4 σ 4 = σ The χ dstrbuton as a specal case of the Γ dstrbuton As we saw, the χ dstrbuton s wrtten: p(χ ; n) = n Γ( n ) (χ ) n e χ In ths form s a the dstrbuton of χ and not χ. Just posng z=χ, t becomes: p(z; n) = n Γ( n ) z n e z/ = Γ(z; α = n, β = ) where Γ s the Gamma dstrbuton we already studed. It follows that Its mean s < χ >= αβ = n Its varance s v(χ ) = αβ = n Its characterstc functon s Φ(t) = ( t) n/ 8
9 3.5 Generatng a varable accordng to a χ dstrbuton A very nce result s that, consderng n ndependent varables (x, x,..., x n ), unformly dstrbuted between 0 and, then u = log(x x...x n ) = χ (n) Ths can be easly shown for n = : y = log x, g(y) = f(x) h (y) = x = e y = χ (y; ) and then by nducton from n to n + ; the algorthm s thus two fold, dependng on whether the number of degrees of freedom one consders s odd or even: n even : generate n/ ndependent varables and apply the above formula n odd : go back to the prevous case and add to the result z, where z s a pseudo random varable dstrbuted accordng to the normal law. For large values of n (typcally n > 0) one can use the fact that χ s very close to a Gaussan dstrbuton of mean n and unt varance. See a graphcal llustraton: chsquare.m 3.6 A fantastc property lnkng the Posson and χ dstrbutons Let s consder a Posson dstrbuton of the nteger varable r and mean µ; a very useful result lnks the probablty that r n (or equvalently r > n) and the probablty that a χ varable s greater than some cut: Proof: P µ (r n >= P [χ (x; n + ) > µ] P [χ (x; n + ) > µ] = x=µ x n e x n! n+ dx = t=µ t n e t dt n! After the change of varable t = x/, dx = dt; now ntegratng by parts and notng that one gets: t=µ t n e t dt = n! x n e x dx = [e x x n ] + n x n e x dx t n e t dt + e µ µ n (n )! t=µ n! = r=n r=0 e µ µ r whch s exactly the sum of the r + ndvdual probabltes of the Posson dstrbuton for r = 0,,...n. Ths property s very mportant from a computatonal pont of vew snce the cumulatve χ probablty s avalable n most computer languages scentfc lbrares, as depcted by the followng Exercse: You are runnng an hgh energy partcle physcs experment, n search for the elusve Hggs boson; you expect b = 3.86 fake background events and you observe n = 9 of them. What s the probablty that your observaton s just a statstcal fluctuaton? (answer: ) 9 r!
10 4 Dstrbuton of the χ of a least squares ft The key (and well known!) result s that the dstrbuton of the χ of a least squares ft, n the case of Gaussan errors, s ndeed a χ dstrbuton of N p degrees of freedom: f(χ N p ; N p) = N p Γ( N p ) (χ ) e χ where N s the number of ponts and p the number of parameters evaluated by the ft. In the case of a straght lne ft, p =. The χ s a sum of N terms u = (y f(x )) σ each of them beng the square of a normal dstrbuton. The ponts of constant probablty le on a N dmensonal hyper-sphere of radus χ. The probablty that a result occurs between χ and χ + δχ thus depends on e χ /, multpled by the volume of the regon concerned, so After change of varable ths becomes P (χ) = αχ N e χ / P (χ ) = β χ N e χ / In secton. dealng wth the case of the ft to a straght lne, we saw that the u are constraned by equatons: 0 = u 0 = u x Thus the ponts u le on an (N )-dmensonal subspace of the orgnal, reducng the dmenson of a regon of equprobablty by two powers of χ. Thus the probablty densty functon s proportonal to P (χ ) = γ χ N 4 e χ / whch s a χ dstrbuton for N degrees of freedom. In case the ft nvolves p parameters nstead of, the above analyss stll holds, one degree of freedom beng lost for each extra parameter. Note that n the case p = N, the ft s not longer a ft but a system of N equatons wth N unknowns. Clearly, f the number of parameters s greater than the number of ponts, the problem s not solvable. 5 Lnear least squares fts A least squares ft s called lnear f the theoretcal functon f(x; a a p ) lnear wth respect to the parameters a ; for example, estmatng a, a, a 3, a 4 n s ndeed a lnear problem, whle estmatng a n y = a + a sn x + a 3 sn(x ) + a 4 x y = cos(ax) s obvously not. The mportant result we want to show s that the LS method provdes an analytc soluton for the parameters a whch can be expressed n a smple closed form. Moreover, the LS estmator n the lnear case s unque, unbased, and s a lnear functon of the observatons y. Fnally, among all estmators whch can be expressed as a lnear functon of the observatons, t s the most effcent. 0
11 5. Normal equatons for least squares fts If many (> ) parameters are to be estmated n a lnear ft, matrx notaton produces elegant results. Let s call a = a,, a p the vector of the p parameters, y = y,, y N the vector of the N measured quanttes and smlarly for x: The ft wll be called lnear f the theoretcal functon f can be expressed as a lnear combnaton of any functons of x; p f(x; a) = c r (x)a r In the general case the X can be expressed as a sum of exponents of the multnormal functons X = [y f(x ; a)]vj [y j f(x j ; a)] j whch can be expressed n matrx notaton: r= X = ( y f) T V (y f) Frst assumng that the y s are ndependent, each of them wth an uncertanty σ, we can wrte N X = [ y p r= a rc r (x ) ] σ Dfferentatng ths wth respect to a r gves the so-called normal equatons: c r (x ) y s âsc s (x ) σ = 0 whch can be expressed n the matrx notaton, ntroducng the matrx C r = c r (x ) f = Ca C T V C â = C T V y â = ( C T V C) C T V y In the above, the dmensons of the varous matrces are: y : N a : p C : N p V : N N C T V : p N C T V y : p C T V C : p p In the case the y s are not ndependent,.e. the matrx V s not dagonal, ths result stll holds: one has to dfferentate the quantty X = ( y Ca) T V (y Ca) wth respect to each of the a k : 0 = â[( y Ca) T V (y Câ)] 0 = C T V (y Câ) C T V y = C T V Câ
12 5. Propertes of Lnear Least Squares Fts The frst rather obvous property s that the Lnear Least Square method provdes a unque soluton, whch s lnear wth respect to the observables y ; secondly, ths estmator s unbased because t s lnear: < â > = < ( C T V C) C T V y > = ( C T V C) C T V < y > = ( C T V C) C T V Ca = a What s the covarance matrx of â? The answer s smple: we saw that f z = Mx, then V (z) = MV (x) M T. In our case M = (C T V C) C T V V (â) = (C T V C) C T V V V C( C T V C) = ( C T V C) 5.3 The Gauss-Markov theorem An nterestng optmum property of the lnear Least Squares estmator s contaned n the Gauss- Markov theorem : among all unbased estmators whch are lnear functons of the observatons, the LS estmators have the smallest varance. To prove t, consder any estmator t of the parameters a, lnear n the observatons y, t = Sy. Its expectaton t s: < t >= S < y >= SCa where a s the orgnal vector of parameters. For t to be unbased, we need < t >= a, a Hence we must have SC =. Now we want to fnd out when the covarance matrx V ( t) for the new estmator has mnmal dagonal term; one has V ( t) = SV (y)s T. For fndng the mnmum we consder the followng dentty, notng V (y) as V, and usng the facts that SC =, V = V T, and (C T V C) T = (C T V C): qtty = [(C T V C) C T V ]V [(C T V C) C T V ] T +[S (C T V C) C T V ]V [S (C T V C) C T V ] T = [(C T V C) C T V ]V [V C(C T V C) ] + SV S T SV V C(C T V C) +[(C T V C) C T V ]V [V C(C T V C) ] (C T V C) C T V V S T = (C T V C) + SV S T (C T V C) + (C T V C) (C T V C) qtty = SV S T = V ( t) Here each of the two terms on the rght hand sde n the defnton of qtty s of the quadratc form UV U T, whch mples non-negatve dagonal elements. Only the second term s a functon of S, and the sum of the two terms wll have strctly mnmum dagonal elements when the second term has vanshng elements on the dagonal. Ths occurs when S = (C T V C) C T V Therefore the mnmum varance unbased estmator for a s where â LS s the Least-Squares estmator for a. t = (C T V C) C T V y = â LS
13 5.4 Lnear Least squares fts wth constrants In frequent stuatons, one s faced wth the problem of estmatng parameters for whch one or more constrants exst. Two approaches can handle such a problem. The Elmnaton method. Lagrange multplers 5.4. Elmnaton method Ths method just conssts, (whenever possble!) n elmnatng some of the parameters by algebrac manpulaton. A very smple example wll llustrate the method: the three angles of a trangle θ, θ and θ 3 have been measured ndependently wth the same uncertanty σ. What s the constraned ( θ = π) LS estmators of the angles? For smplcty, let s call the measurements y, y and y 3 and assume that the measurements are performed n degrees. One wants to solve: X (θ, θ, θ 3 ) = 3 θ = 80 = 3 ( y θ ) = mnmum σ = In ths case, the algebrac manpulaton to elmnate, say, θ 3, s not very dffcult: X = ( y θ σ ) + ( y θ σ ) + [ y 3 (80 θ θ ) ] σ 0 = σ X θ = (y θ ) + [y 3 (80 θ θ )] 0 = σ X θ = (y θ ) + [y 3 (80 θ θ )] θ = 3 (80 + y y y 3 ) θ = 3 (80 y + y y 3 ) θ 3 = 3 (80 y y + y 3 ) An mproved measurement was obtaned by subtractng to each of the measurements y one thrd of the excess y + y + y Ths mprovement s drectly vsble n the varance of the estmators: V ( θ ) = V ( θ ) = V ( θ 3 ) = σ = 9 3 σ Ths result can be ntutvely explaned n terms of nformaton: one could have estmated θ, θ and θ 3 wth two measurements only, say y and y ; n ths case, y 3 would have been computed as y 3 = 80 y y. Addng a thrd measurement ncreases the nformaton by 50%, and we saw, n the secton devoted to the Cramer Rao bound (4.3.), that the mnmum varance s nversely proportonal to the amount of nformaton. 3
14 5.4. Lagrange Multplers Let s assume that we agan want to estmate p parameters a, a a n, lnked together by m constrants: D k (a) = 0, k =, m The basc trck of the method s to replace the p-component vector of the parameters a by a p+mcomponent vector whose frst p components are the usual a and the remanng ones λ k, k =, m, such as the new X s now: X = ( y Ca) T V (y Ca) + λ T D = mnmum If we here equate to zero the dervatves of X wth respect to a, a a p and λ k, = m, we get the normal equatons: âx = (C T V y C T V Câ) + ( âd) T λ = 0 λ X = D = 0 We are now left wth a set of p + m equatons. The second ones are the constrants, whle the frst ones are the analog of the normal equatons for the unconstraned case, now modfed by the λ term due to the constrants. In the case the constrants themselves are lnear, they can be expressed as Ba = b and the set of the modfed p + k normal equatons becomes 0 = (C T V y C T V Câ) + B T λ 0 = Bâ b whch s a set of lnear equatons. After some heavy algebra, one can show that the varance matrx of the estmator â s: V (â) = (C T V C) [B(C T V C) ] T [B T V B] [B(C T V C) ] Ths expresson can not be smplfed any further because nether C or B are square matrces! Note that the matrx (B T V B) has non negatve dagonal terms, and that the same s true for the second term of the above equaton. As a result, the constrant equatons wll always lead to a reducton of the parameter errors (.e. the dagonal terms) compared to the unconstraned case. Back to the prevous exercse, runnng the machnery yelds: χ = 3 = ( y θ σ 0 = χ θ = σ (y θ ) + λ 0 = χ θ = σ (y θ ) + λ 0 = χ θ 3 = σ (y 3 θ 3 ) + λ 4 ) + λ( θ 80)
15 0 = χ = ( θ 80) λ 0 = ( y 80) 3 λσ λ = 3σ ( y 80) θ = y λσ = y 3 ( y 80) 6 Non lnear least squares fts If the functon f(x; a) s not lnear n the a r, you generally have to use an teratve technque to solve the equatons. Gven a frst guess a 0, a Taylor expanson around a 0 wll yeld a new set of values a such as X a r a=a 0 and further teratons wll lead to the soluton. In such teratve methods, havng a good frst guess s wnnng 90% of the battle. The best advce s to fnd a relable software lbrary rather than beng tempted to wrte one s own package. 5
1 Matrix representations of canonical matrices
1 Matrx representatons of canoncal matrces 2-d rotaton around the orgn: ( ) cos θ sn θ R 0 = sn θ cos θ 3-d rotaton around the x-axs: R x = 1 0 0 0 cos θ sn θ 0 sn θ cos θ 3-d rotaton around the y-axs:
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationFall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede
Fall 0 Analyss of Expermental easurements B. Esensten/rev. S. Errede We now reformulate the lnear Least Squares ethod n more general terms, sutable for (eventually extendng to the non-lnear case, and also
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationFormulas for the Determinant
page 224 224 CHAPTER 3 Determnants e t te t e 2t 38 A = e t 2te t e 2t e t te t 2e 2t 39 If 123 A = 345, 456 compute the matrx product A adj(a) What can you conclude about det(a)? For Problems 40 43, use
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationFirst Year Examination Department of Statistics, University of Florida
Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve
More informationSIO 224. m(r) =(ρ(r),k s (r),µ(r))
SIO 224 1. A bref look at resoluton analyss Here s some background for the Masters and Gubbns resoluton paper. Global Earth models are usually found teratvely by assumng a startng model and fndng small
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationChapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems
Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons
More informationLecture 3 Stat102, Spring 2007
Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationAPPENDIX A Some Linear Algebra
APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,
More informationIntegrals and Invariants of Euler-Lagrange Equations
Lecture 16 Integrals and Invarants of Euler-Lagrange Equatons ME 256 at the Indan Insttute of Scence, Bengaluru Varatonal Methods and Structural Optmzaton G. K. Ananthasuresh Professor, Mechancal Engneerng,
More informationLecture 12: Discrete Laplacian
Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly
More informationDifference Equations
Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationLecture 3. Ax x i a i. i i
18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationSolutions Homework 4 March 5, 2018
1 Solutons Homework 4 March 5, 018 Soluton to Exercse 5.1.8: Let a IR be a translaton and c > 0 be a re-scalng. ˆb1 (cx + a) cx n + a (cx 1 + a) c x n x 1 cˆb 1 (x), whch shows ˆb 1 s locaton nvarant and
More informationChapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the
More informationSTAT 3008 Applied Regression Analysis
STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,
More information3.1 ML and Empirical Distribution
67577 Intro. to Machne Learnng Fall semester, 2008/9 Lecture 3: Maxmum Lkelhood/ Maxmum Entropy Dualty Lecturer: Amnon Shashua Scrbe: Amnon Shashua 1 In the prevous lecture we defned the prncple of Maxmum
More informationSELECTED PROOFS. DeMorgan s formulas: The first one is clear from Venn diagram, or the following truth table:
SELECTED PROOFS DeMorgan s formulas: The frst one s clear from Venn dagram, or the followng truth table: A B A B A B Ā B Ā B T T T F F F F T F T F F T F F T T F T F F F F F T T T T The second one can be
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationSolutions to exam in SF1811 Optimization, Jan 14, 2015
Solutons to exam n SF8 Optmzaton, Jan 4, 25 3 3 O------O -4 \ / \ / The network: \/ where all lnks go from left to rght. /\ / \ / \ 6 O------O -5 2 4.(a) Let x = ( x 3, x 4, x 23, x 24 ) T, where the varable
More informationSection 8.3 Polar Form of Complex Numbers
80 Chapter 8 Secton 8 Polar Form of Complex Numbers From prevous classes, you may have encountered magnary numbers the square roots of negatve numbers and, more generally, complex numbers whch are the
More informationThe Geometry of Logit and Probit
The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationStanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011
Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected
More informationThe Expectation-Maximization Algorithm
The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.
More information4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA
4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More information8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore
8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø
More informationMATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)
1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons
More informationConjugacy and the Exponential Family
CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the
More informationMore metrics on cartesian products
More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of
More informationx i1 =1 for all i (the constant ).
Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by
More informationRadar Trackers. Study Guide. All chapters, problems, examples and page numbers refer to Applied Optimal Estimation, A. Gelb, Ed.
Radar rackers Study Gude All chapters, problems, examples and page numbers refer to Appled Optmal Estmaton, A. Gelb, Ed. Chapter Example.0- Problem Statement wo sensors Each has a sngle nose measurement
More informationPHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University
PHYS 45 Sprng semester 7 Lecture : Dealng wth Expermental Uncertantes Ron Refenberger Brck anotechnology Center Purdue Unversty Lecture Introductory Comments Expermental errors (really expermental uncertantes)
More informatione i is a random error
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown
More informationLinear Regression Analysis: Terminology and Notation
ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented
More informationAssortment Optimization under MNL
Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.
More information4DVAR, according to the name, is a four-dimensional variational method.
4D-Varatonal Data Assmlaton (4D-Var) 4DVAR, accordng to the name, s a four-dmensonal varatonal method. 4D-Var s actually a drect generalzaton of 3D-Var to handle observatons that are dstrbuted n tme. The
More information4.3 Poisson Regression
of teratvely reweghted least squares regressons (the IRLS algorthm). We do wthout gvng further detals, but nstead focus on the practcal applcaton. > glm(survval~log(weght)+age, famly="bnomal", data=baby)
More informationELASTIC WAVE PROPAGATION IN A CONTINUOUS MEDIUM
ELASTIC WAVE PROPAGATION IN A CONTINUOUS MEDIUM An elastc wave s a deformaton of the body that travels throughout the body n all drectons. We can examne the deformaton over a perod of tme by fxng our look
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationMMA and GCMMA two methods for nonlinear optimization
MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons
More informationModule 2. Random Processes. Version 2 ECE IIT, Kharagpur
Module Random Processes Lesson 6 Functons of Random Varables After readng ths lesson, ou wll learn about cdf of functon of a random varable. Formula for determnng the pdf of a random varable. Let, X be
More informationSingular Value Decomposition: Theory and Applications
Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real
More informationCanonical transformations
Canoncal transformatons November 23, 2014 Recall that we have defned a symplectc transformaton to be any lnear transformaton M A B leavng the symplectc form nvarant, Ω AB M A CM B DΩ CD Coordnate transformatons,
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More informationComputation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models
Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,
More informationChapter 6. Supplemental Text Material
Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.
More informationEconomics 130. Lecture 4 Simple Linear Regression Continued
Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do
More informationn α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0
MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More informationMA 323 Geometric Modelling Course Notes: Day 13 Bezier Curves & Bernstein Polynomials
MA 323 Geometrc Modellng Course Notes: Day 13 Bezer Curves & Bernsten Polynomals Davd L. Fnn Over the past few days, we have looked at de Casteljau s algorthm for generatng a polynomal curve, and we have
More informationNUMERICAL DIFFERENTIATION
NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the
More informationChat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980
MT07: Multvarate Statstcal Methods Mke Tso: emal mke.tso@manchester.ac.uk Webpage for notes: http://www.maths.manchester.ac.uk/~mkt/new_teachng.htm. Introducton to multvarate data. Books Chat eld, C. and
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationInner Product. Euclidean Space. Orthonormal Basis. Orthogonal
Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,
More informationPHYS 705: Classical Mechanics. Calculus of Variations II
1 PHYS 705: Classcal Mechancs Calculus of Varatons II 2 Calculus of Varatons: Generalzaton (no constrant yet) Suppose now that F depends on several dependent varables : We need to fnd such that has a statonary
More informationImportant Instructions to the Examiners:
Summer 0 Examnaton Subject & Code: asc Maths (70) Model Answer Page No: / Important Instructons to the Examners: ) The Answers should be examned by key words and not as word-to-word as gven n the model
More informationLecture 3: Probability Distributions
Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the
More informationLaboratory 3: Method of Least Squares
Laboratory 3: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly they are correlated wth
More informationBernoulli Numbers and Polynomials
Bernoull Numbers and Polynomals T. Muthukumar tmk@tk.ac.n 17 Jun 2014 The sum of frst n natural numbers 1, 2, 3,..., n s n n(n + 1 S 1 (n := m = = n2 2 2 + n 2. Ths formula can be derved by notng that
More informationHidden Markov Models
Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,
More informationMaximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models
ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models
More informationU.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016
U.C. Berkeley CS94: Spectral Methods and Expanders Handout 8 Luca Trevsan February 7, 06 Lecture 8: Spectral Algorthms Wrap-up In whch we talk about even more generalzatons of Cheeger s nequaltes, and
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationNumerical Heat and Mass Transfer
Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and
More informationβ0 + β1xi. You are interested in estimating the unknown parameters β
Ordnary Least Squares (OLS): Smple Lnear Regresson (SLR) Analytcs The SLR Setup Sample Statstcs Ordnary Least Squares (OLS): FOCs and SOCs Back to OLS and Sample Statstcs Predctons (and Resduals) wth OLS
More information= = = (a) Use the MATLAB command rref to solve the system. (b) Let A be the coefficient matrix and B be the right-hand side of the system.
Chapter Matlab Exercses Chapter Matlab Exercses. Consder the lnear system of Example n Secton.. x x x y z y y z (a) Use the MATLAB command rref to solve the system. (b) Let A be the coeffcent matrx and
More informationC/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1
C/CS/Phy9 Problem Set 3 Solutons Out: Oct, 8 Suppose you have two qubts n some arbtrary entangled state ψ You apply the teleportaton protocol to each of the qubts separately What s the resultng state obtaned
More informationPhysics 5153 Classical Mechanics. D Alembert s Principle and The Lagrangian-1
P. Guterrez Physcs 5153 Classcal Mechancs D Alembert s Prncple and The Lagrangan 1 Introducton The prncple of vrtual work provdes a method of solvng problems of statc equlbrum wthout havng to consder the
More information2.3 Nilpotent endomorphisms
s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms
More informationAPPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14
APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce
More informationChapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of
Chapter 7 Generalzed and Weghted Least Squares Estmaton The usual lnear regresson model assumes that all the random error components are dentcally and ndependently dstrbuted wth constant varance. When
More informationFeb 14: Spatial analysis of data fields
Feb 4: Spatal analyss of data felds Mappng rregularly sampled data onto a regular grd Many analyss technques for geophyscal data requre the data be located at regular ntervals n space and/or tme. hs s
More informationParametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010
Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton
More informationProblem Set 9 Solutions
Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem
More informationMoments of Inertia. and reminds us of the analogous equation for linear momentum p= mv, which is of the form. The kinetic energy of the body is.
Moments of Inerta Suppose a body s movng on a crcular path wth constant speed Let s consder two quanttes: the body s angular momentum L about the center of the crcle, and ts knetc energy T How are these
More informationModelli Clamfim Equazione del Calore Lezione ottobre 2014
CLAMFIM Bologna Modell 1 @ Clamfm Equazone del Calore Lezone 17 15 ottobre 2014 professor Danele Rtell danele.rtell@unbo.t 1/24? Convoluton The convoluton of two functons g(t) and f(t) s the functon (g
More informationProbability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n!
8333: Statstcal Mechancs I Problem Set # 3 Solutons Fall 3 Characterstc Functons: Probablty Theory The characterstc functon s defned by fk ep k = ep kpd The nth coeffcent of the Taylor seres of fk epanded
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13
CME 30: NUMERICAL LINEAR ALGEBRA FALL 005/06 LECTURE 13 GENE H GOLUB 1 Iteratve Methods Very large problems (naturally sparse, from applcatons): teratve methods Structured matrces (even sometmes dense,
More informationNotes on Frequency Estimation in Data Streams
Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to
More informationLOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin
Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence
More informationA New Refinement of Jacobi Method for Solution of Linear System Equations AX=b
Int J Contemp Math Scences, Vol 3, 28, no 17, 819-827 A New Refnement of Jacob Method for Soluton of Lnear System Equatons AX=b F Naem Dafchah Department of Mathematcs, Faculty of Scences Unversty of Gulan,
More informationSalmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2
Salmon: Lectures on partal dfferental equatons 5. Classfcaton of second-order equatons There are general methods for classfyng hgher-order partal dfferental equatons. One s very general (applyng even to
More informationAffine transformations and convexity
Affne transformatons and convexty The purpose of ths document s to prove some basc propertes of affne transformatons nvolvng convex sets. Here are a few onlne references for background nformaton: http://math.ucr.edu/
More informationP R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /
Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that
More informationTransfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system
Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng
More informationVector Norms. Chapter 7 Iterative Techniques in Matrix Algebra. Cauchy-Bunyakovsky-Schwarz Inequality for Sums. Distances. Convergence.
Vector Norms Chapter 7 Iteratve Technques n Matrx Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematcs Unversty of Calforna, Berkeley Math 128B Numercal Analyss Defnton A vector norm
More informationEffects of Ignoring Correlations When Computing Sample Chi-Square. John W. Fowler February 26, 2012
Effects of Ignorng Correlatons When Computng Sample Ch-Square John W. Fowler February 6, 0 It can happen that ch-square must be computed for a sample whose elements are correlated to an unknown extent.
More informationSupplementary material: Margin based PU Learning. Matrix Concentration Inequalities
Supplementary materal: Margn based PU Learnng We gve the complete proofs of Theorem and n Secton We frst ntroduce the well-known concentraton nequalty, so the covarance estmator can be bounded Then we
More information