Monotonic convergence of a general algorithm for computing optimal designs
|
|
- Pamela Turner
- 6 years ago
- Views:
Transcription
1 Monotonc convergence of a general algorthm for computng optmal desgns Yamng Yu Department of Statstcs Unversty of Calforna Irvne, CA 92697, USA yamngy@uc.edu Abstract Monotonc convergence s establshed for a general class of multplcatve algorthms ntroduced by Slvey et al. 1978) for computng optmal desgns. A conjecture of Ttterngton 1978) s confrmed as a consequence. Optmal desgns for logstc regresson are used as an llustraton. Keywords: A-optmalty; auxlary varables; c-optmalty; D-optmalty; expermental desgn; generalzed lnear models; multplcatve algorthm. 1 A general class of algorthms Optmal expermental desgn approxmate theory) s a well-developed area and we refer to Kefer 1974), Slvey 1980), Pázman 1986), and Pukelshem 1993) for general ntroducton and basc results. We consder computatonal aspects of optmal desgns, focusng on a fnte desgn space X = {x 1,..., x n }. Suppose the probablty densty or mass functon of the response s specfed as py x, θ), where θ s the m 1 parameter of nterest. Let A denote the m m expected Fsher nformaton matrx from a unt assgned to x the expectaton s wth respect to py x, θ)): A = E[sθ; y, x )s θ; y, x )], sθ; y, x ) = log py x, θ). θ The moment matrx, as a functon of the desgn measure w = w 1,..., w n ), s defned as Mw) = w A, whch s proportonal to the Fsher nformaton for θ when the number of unts assgned to x s proportonal to w. Here w Ω, and Ω denotes the closure of Ω = {w : w > 0, n =1 w = 1}. Throughout we assume that A are well-defned and nonnegatve defnte for fxed θ). The set =1 Ω + {w Ω : Mw) > 0 postve defnte)} s assumed nonempty. Our approach may concevably extend to the case where Mw) s allowed sngular, by usng generalzed nverses, although we do not pursue ths here. Gven an optmalty crteron φ, defned on postve defnte matrces, the goal s to maxmze φmw)) wth respect to w Ω +. Typcal optmalty crtera nclude 1
2 ) the D-crteron φ 0 M) = log detm), ) the A-crteron φ 1 M) = trm 1 ), ) more generally, the pth mean crteron φ p M) = trm p ), p < 0, and v) the c-crteron φ c M) = c M 1 c, where c s a nonzero constant vector. Our notaton s mostly for convenence and may not correspond exactly to those of well-known texts.) Often only a lnear combnaton K θ, e.g., a subvector of θ, s of nterest. Assumng nvertblty, the Fsher nformaton for K θ s naturally defned as K M 1 K) 1 Pukelshem, 1993). We may therefore consder the D- and A-crtera for K θ defned respectvely as φ 0,K M) = log detk M 1 K); φ 1,K M) = trk M 1 K). 1) The c-crteron s a specal case of φ 1,K M). Motvatons for such optmalty crtera are wellknown. For example, n a lnear problem, the A-crteron seeks to mnmze the sum of varances of the best lnear unbased estmators BLUEs) for all coordnates of θ, whle the c-crteron seeks to mnmze the varance of the BLUE for c θ. Smlar nterpretatons wth asymptotc arguments) apply to nonlnear problems. In general Mw) also depends on the unknown parameter θ, whch complcates the defnton of an optmalty crteron. A smple soluton s to maxmze φmw)) wth θ fxed at a pror guess θ ; ths leads to local optmalty Chernoff 1953). Local optmalty may be crtczed for gnorng uncertanty n θ. However, n a stuaton where real pror nformaton s avalable, or where the dependence of M on θ s weak, t s nevertheless a vable approach, and has been adopted routnely see, for example, L and Majumdar 2008). Henceforth we assume a fxed θ and suppress the dependence of M on θ. Possble extensons are mentoned n Secton 5. Optmal desgns do not usually come n closed form. As early as Wynn 1972), Fedorov 1972) and Atwood 1973), and as late as Torsney 2007), Harman and Pronzato 2007), and Dette 2008), varous procedures have been studed for numercal computaton. We shall focus on the followng multplcatve algorthm Ttterngton 1976, 1978; Slvey 1978), whch s specfed through a power parameter λ 0, 1]. Algorthm I Set λ 0, 1] and w 0) Ω. For t = 0, 1,..., compute w t) ) where w t+1) = w t) d λ n j=1 wt) j d λ j w t) ), = 1,..., n, 2) d w) = trφ Mw))A ), φ M) φm) M. 3) Iterate untl convergence. We defer the dscusson on convergence crtera untl Secton 4. For a heurstc explanaton, observe that 2) s equvalent to w t+1) w t) φmw)) λ w, = 1,..., n. 4) w=w t)) The value of φmw))/ w ndcates the amount of gan n nformaton, as measured by φ, by a slght ncrease n w, the weght on the th desgn pont. So 4) can be seen as adjustng w so that relatvely more weght s placed on desgn ponts whose ncreased weght may result n a larger gan n φ. 2
3 Algorthm I s remarkable n ts generalty. For example, lttle restrcton s placed on the underlyng model py x, θ). Part of the reason, of course, s that we focus on Fsher nformaton and local optmalty, whch essentally reduces the problem to a lnear one. There exsts a large lterature on Algorthm I and ts relatves; see, for example, Ttterngton 1976, 1978), Slvey et al. 1978), Pázman 1986), Fellman 1989), Pukelshem and Torsney 1991), Torsney and Mandal 2006), Harman and Pronzato 2007), and Dette et al. 2008). One feature that has attracted much attenton s that Algorthm I appears to be monotonc,.e., φmw t) )) ncreases n t, at least n some specal cases. For example, when φ = φ 0 for D- optmalty) and λ = 1, Ttterngton 1976) and Pázman 1986) have shown monotoncty usng clever probablstc and analytc nequaltes; see also Dette 2008) and Harman and Trnovská 2009). Algorthm I s also known to be monotonc for φ = φ 1,K as n 1), assumng λ = 1/2 and A are rank-one Fellman 1974; Torsney 1983). Monotoncty s mportant because convergence then holds under mld assumptons see Secton 4). Results n these specal cases suggest a monotonc convergence theory for a broad class of φ, whch s also supported by numercal evdence presented n some of the references above. 2 Man result We am to state general condtons on φ that ensure that Algorthm I converges monotoncally. As a consequence certan known theoretcal results are unfed and generalzed, and one partcular conjecture Ttterngton 1978) s confrmed. Defne ψm) φm 1 ), M > 0. The functons φ and ψ are assumed to be dfferentable on postve defnte matrces. Our condtons are convenently stated n terms of ψ. As usual, for two symmetrc matrces, M 1 <)M 2 means M 2 M 1 s nonnegatve postve) defnte. ψm) s ncreasng: 0 < M 1 M 2 = ψm 1 ) ψm 2 ), 5) or, equvalently, ψ M) s nonnegatve defnte for postve defnte M. ψm) s concave: αψm 1 ) + 1 α)ψm 2 ) ψαm α)m 2 ), 6) for α [0, 1], M 1, M 2 > 0. Equvalently, ψm 2 ) ψm 1 ) + trψ M 1 )M 2 M 1 )), M 1, M 2 > 0. 7) Condton 5) s usually satsfed by any reasonable nformaton crteron Pukelshem 1993). Also note that, f 5) fals, then φmw))/ w on the rght hand sde of 4) s not even guaranteed to be nonnegatve. The real restrcton s the concavty condton 6). For example, 6) s not satsfed by ψ p M) = φ p M 1 ) the pth mean crteron) when p < 1. It s usually assumed that φm), rather than ψm), s concave.) Nevertheless, 6) s satsfed by a wde range of crtera, ncludng the commonly used D-, A- or c-crtera see Cases ) and ) n the llustraton of the man result below). Our man result s as follows. 3
4 Theorem 1 General monotoncty). Assume 5) and 6). Assume that n teraton 2), Mw t) ) > 0, φ Mw t) )) 0, and Mw t+1) ) > 0. Then φmw t+1) )) φmw t) )). In other words, under mld condtons that ensure that 2) s well-defned specfcally, the denomnator n 2) s nonzero), 5) and 6) mply that 2) monotoncally ncreases the crteron φ. Let us llustrate Theorem 1 wth some examples. For smplcty, n ) v) we dsplay formulae for λ = 1 only, although the dscusson apples to all λ 0, 1]. ) Take { log det M, p = 0; φ p M) = trm p ), p [ 1, 0). Then ψ p M) φ p M 1 ) satsfes 5) and 6) see Appendx A). By Theorem 1, Algorthm I s monotonc for φ = φ p, p [ 1, 0]. Ths generalzes the prevously known cases p = 0 and p = 1 wth partcular values of λ). The teraton 2) reads w t+1) = w t) trm p 1 w t) )A ) trm p w t), = 1,..., n. )) ) More generally, gven a full rank m r matrx K r m), consder { ψ p,k M 1 log detk ) φ p,k M) = M 1 K), p = 0; trk M 1 K) p ), p [ 1, 0). Then ψ p,k M) satsfes 5) and 6) the proof s the same as n Case )). By Theorem 1, Algorthm I s monotonc for φ = φ p,k, p [ 1, 0]. The teraton 2) reads w t+1) = w t) trm 1 KK M 1 K) p 1 K M 1 A ) tr K M 1 K) p ). 8) M=Mw t) ) ) In partcular, takng r = 1, K = c an m 1 vector) and p = 1 n Case ), we obtan that Algorthm I s monotonc for the c-crteron φ c. The teraton 8) reduces to compare wth Fellman 1974) w t+1) = w t) c M 1 w t) )A M 1 w t) )c c M 1 w t), = 1,..., n. )c v) Consder another example of Case ), wth p = 0, r = m 1 and K = 0 r, I r ). Henceforth 0 r denotes the r 1 vector of zeros, and I r denotes the r r dentty matrx. Assume A = x x, x = 1, z ) and z s m 1) 1. Ths corresponds to a D-optmal desgn problem for the lnear model wth ntercept y x, θ) Nx θ, σ 2 ), x = 1, z ), where the parameter s θ = θ 0, θ 1,..., θ m 1 ) but nterest centers on θ 1,..., θ m 1 ), not θ 0. Nevertheless, as far as the desgn measure w s concerned, the optmalty crteron, φ 0,K M), concdes wth φ 0 M),.e., log detk M 1 w)k) = log det Mw). 4
5 After some algebra, 8) reduces to where w t+1) z = = w t) =1 z z) M 1 c w t) z ; M c w t) ) = w t) )z z), = 1,..., n, 9) m 1 =1 w t) z z)z z). Thus 9) satsfes det Mw t+1) ) det Mw t) ). Monotoncty of 9) has been conjectured snce Ttterngton 1978), and consderable numercal evdence has accumulated over the years. Recently, extendng the arguments of Pázman 1986), Dette et al. 2008) have obtaned results that come very close to resolvng Ttterngton s conjecture. Nevertheless, we have been unable to extend ther arguments any further. Instead we prove the more general Theorem 1 usng a dfferent approach, and settle ths conjecture as a consequence. The proof of Theorem 1 s acheved by usng a method of auxlary varables. When a functon fw) e.g., det Mw)) to be mnmzed s complcated, we ntroduce a new varable Q and a functon gw, Q) such that mn Q gw, Q) = fw) for all w, thus transformng the problem nto mnmzng gw, Q) over w and Q jontly. Then we may use an teratve condtonal mnmzaton strategy on gw, Q). Ths s nspred by the EM algorthm Dempster et al. 1977; Meng and van Dyk 1997); n partcular, see Csszár and Tusnady s [6] nterpretaton see [30] for a related nterpretaton of the data augmentaton algorthm). In Secton 3 we analyze Algorthm I usng ths strategy. Although attenton s pad to the mathematcs, our focus s on ntutvely appealng nterpretatons, whch may lead to further extensons of Algorthm I wth the same desrable monotoncty propertes. If the algorthm s monotonc, then convergence can be establshed under mld condtons Secton 4). Secton 5 contans an llustraton wth locally optmal desgns for generalzed lnear models. 3 Explanng the monotoncty A key observaton s that the problem of maxmzng φmw)), or, equvalently, mnmzng ψm 1 w)) can be formulated as a jont mnmzaton over both the desgn and the estmator. Specfcally, let us compare the orgnal Problem P1 wth ts companon P2. Throughout A 1/2 denotes the symmetrc nonnegatve defnte SNND) square root of an SNND matrx A. Problem P1: Mnmze φmw)) ψ n =1 w A ) 1 ) over w Ω. Problem P2: Mnmze gw, Q) ψ Q w Q ) 10) over w Ω and Q an m mn) matrx), subject to QG = I m, where w Dagw 1 1,..., w 1 n ) I m ; G A 1/2 1,..., A 1/2 n ). Though not mmedately obvous, P1 and P2 are equvalent, and ths may be explaned n statstcal terms as follows. In 10), Q w Q s smply the varance matrx of a lnear unbased estmator, QY, of the m 1 parameter θ n the model Y = Gθ + ɛ, ɛ N0, w ), 5
6 where Y s the mn) 1 vector of observatons. The constrant QG = I m ensures unbasedness. Note that G s full-rank snce Mw) s nonsngular by assumpton.) Of course, the weghted least squares WLS) estmator s the best lnear unbased estmator, havng the smallest varance matrx n the sense of postve defnte orderng) and, by 5), the smallest ψ for that matrx. It follows that, for fxed w, gw, Q) s mnmzed by choosng QY as the WLS estmator: However, from 10) and 11) we get gw, ˆQ W LS ) = nf gw, Q), 11) QG=I m ) ˆQ W LS = M 1 w) w 1 A 1/2 1,..., w n An 1/2. gw, ˆQ W LS ) = ψm 1 w)). 12) That s, P2 reduces to P1 upon mnmzng over Q. Snce P2 s not mmedately solvable, t s natural to consder the subproblems: ) mnmzng gw, Q) over Q for fxed w, and ) mnmzng gw, Q) over w for fxed Q. Part ) s agan formulated as a jont mnmzaton problem. For a fxed m mn) matrx Q such that QG = I m, let us consder Problems P3 and P4. Problem P3: Mnmze gw, Q) as n 10) over w Ω. Problem P4: Mnmze the functon )) hσ, w, Q) = ψσ) + tr ψ Σ) Q w Q Σ 13) over w Ω and the m m postve-defnte matrx Σ. To relate P4 to P3, note that for fxed w and Q, the concavty assumpton 7) mples that hσ, w, Q) ψ Q w Q ) 14) wth equalty when Σ = Q w Q,.e., Problem P4 reduces to P3 upon mnmzng over Σ. Snce P4 s not mmedately solvable, t s natural to consder the subproblems: ) mnmzng hσ, w, Q) over Σ for fxed w and Q, and ) mnmzng hσ, w, Q) over w for fxed Σ and Q. Part ), whch amounts to mnmzng tr ψ Σ)Q w Q ) = tr Q ψ Σ)Q w ), admts a closed-form soluton: f we wrte Q = Q 1,..., Q n ) where each Q s m m, then w 2 should be proportonal to trq ψ Σ)Q ). But algorthm I may not perform an exact mnmzaton here; see 15). Based on the above dscusson, we can express Algorthm I as an teratve condtonal mnmzaton algorthm nvolvng w, Q and Σ. At teraton t, defne Q t) = Q t) 1,..., Qt) n ); Q t) = w t) M 1 w t) )A 1/2, = 1,..., n; Σ t) = Q t) w t)q t) = M 1 w t) ). 6
7 Then we have ψm 1 w t) )) = gw t), Q t) ) by 12)) = hσ t), w t), Q t) ) by 13)) hσ t), w t+1), Q t) ) see below) 15) gw t+1), Q t) ) by 14), 10)) 16) ψm 1 w t+1) )) by 11), 12)). 17) The choce of w t+1) leads to 15) as follows. After smple algebra, the teraton 2) becomes where w t+1) = r λ w1 2λ n j=1 rλ j w1 2λ j w w t), r tr Snce 0 < λ 1, Jensen s nequalty yelds Q t), = 1,..., n, ) ψ Σ t) )Q t). r w =1 ) 1 λ =1 ) 1 λ r w w 2 ; r w =1 ) λ =1 ) λ r w w 2. That s, Hence =1 r w =1 tr ψ Σ t) )Q t) w t)q t) ) = r 1 λ w 2λ 1 =1 =1 r w t) r w t+1) ) =1 r λ w 1 2λ ) = tr ψ Σ t) )Q t) w t+1)q t) ),. whch produces 15). Choosng λ = 1/2,.e., w t+1) r, leads to exact mnmzaton n 15); choosng λ = 1 yelds equalty n 15). But any choce of w t+1) that decreases hσ t), w, Q t) ) at 15) would have resulted n the desred nequalty ψm 1 w t) )) ψm 1 w t+1) )). We may allow λ to change from teraton to teraton, and monotoncty stll holds, as long as λ 0, 1]. See Slvey et al. 1978) and Fellman 1989) for nvestgatons concernng the choce of λ. Also note that we assume w t), w t+1) > 0 for all. Ths s not essental, however, because ) the possblty of w t) = 0 can be handled by restrctng our analyss to all desgn ponts such that w t) > 0, and ) the possblty of w t+1) = 0 can be handled by a standard lmtng argument. Monotoncty holds as long as Mw t) ) and Mw t+1) ) are both postve defnte, as noted n the statement of Theorem 1. 7
8 4 Global convergence Let us revew characterzatons of optmal desgns, commonly referred to as general equvalence theorems; see Kefer and Wolfowtz 1960) and Whttle 1973). Theorem 2. Assume the crteron φ s dfferentable and concave. Defne dw) w d w) = trφ Mw))Mw)), =1 where, as n 3), d w) trφ Mw))A ). Then for every w Ω +, the followng are equvalent: a) w maxmzes φmw)) on Ω + ; b) d w ) dw ) for all ; c) for all, f w 0 then d w ) = dw ), and f w = 0 then d w ) dw ). Theorem 2 suggests the followng convergence crteron for Algorthm I: stop when w t)) 1 + δ) d max 1 n d w t)), 18) for some small δ > 0. Theorem 2 also plays an mportant role n the proof of our man convergence result Theorem 3). Of course, the drvng force behnd Theorem 3 s monotoncty Theorem 1). Theorem 3 Global convergence). Denote the mappng 2) by T. a) Assume φ Mw)) 0; φ Mw))A 0, w Ω +, = 1,..., n. b) Assume 2) s strctly monotonc,.e., w Ω +, T w w = φmt w)) > φmw)). 19) c) Assume φ s strctly concave and φ s contnuous on postve defnte matrces. d) Assume that, f M a postve defnte matrx) tends to M such that φm) ncreases monotoncally, then M s nonsngular. Let w t) be generated by 2) wth w 0) > 0 for all. Then ) all lmt ponts of w t) are global maxma of φmw)), and ) as t, φmw t) )) ncreases monotoncally to sup w Ω+ φmw)). The proof of Theorem 3 s somewhat subtle. Standard arguments see Lemma 1 n Appendx B) show that all lmt ponts of w t) are fxed ponts of the mappng T. Ths alone does not mply convergence to a global maxmum, however, because there often exst sub-optmal fxed ponts on the boundary of Ω. Global maxma occur routnely on the boundary also.) Our goal s therefore to rule out possble convergence to such sub-optmal ponts; Appendx B presents the detals. We shall comment on Condtons a) d). Condton a) ensures that startng wth w 0) Ω +, all teratons are well-defned. Moreover, f w 0) > 0 for all, then w t) > 0 for all t and. Ths hghlghts the followng basc dea. In order to converge to a global maxmum w, the startng value w 0) must assgn postve weght to every support pont of w. Such a requrement s not necessary for monotoncty. On the 8
9 other hand, assgnng weght to non-supportng ponts of w tends to slow the algorthm down. Hence methods that quckly elmnate non-optmal support ponts are valuable Harman and Pronzato, 2007). Condton b) smply says that unless w s a fxed pont, the mappng T should produce a better soluton. Let us assume 5), 7) and Condton a), so that Theorem 1 apples. Then, by checkng the equalty condton n 15), t s easy to see that Condton b) s satsfed f 0 < λ < 1. The argument leadng to 19) techncally assumes that all coordnates of w are nonzero, but we can apply t to the approprate subvector of w.) If λ = 1, then 15) reduces to an equalty. However, by checkng the equalty condtons n 16) and 17), we can show that Condton b) s satsfed f ψ s strctly ncreasng and strctly concave: M 2 M 1 > 0, M 1 M 2 = ψm 1 ) < ψm 2 ); 20) M 1, M 2 > 0, M 1 M 2 = ψm 2 ) < ψm 1 ) + trψ M 1 )M 2 M 1 )). 21) Condtons c) and d) are techncal requrements that concern φ alone. Condton c) ensures unqueness of the optmal moment matrx, whch smplfes the analyss. Condton d) ensures that postve defnteness of Mw) s mantaned n the lmt. Condtons c) and d) are satsfed by φ = φ p wth p 0, for example. Let us menton a typcal example of Theorem 3. Corollary 1. Assume A 0, w 0) > 0, = 1,..., n, and Mw 0) ) > 0. Then the concluson of Theorem 3 holds for Algorthm I wth φ = φ 0. Proof. Condtons a), c) and d) are readly verfed. Condton b) s satsfed by 20) and 21). The clam follows from Theorem 3. A more specalzed example concerns 9). Proposton 1. Assume that X = x 1,..., x n ) as n 9) has rank m, and that w 0) all. Then the concluson of Theorem 3, wth φ = φ 0, holds for the teraton 9). > 0 for The teraton 9) techncally does not satsfy the assumptons of Theorem 3. For example, we do not have w t+1) > 0 even f w t) > 0 for all. However, nspecton of 9) shows that w t+1) s set to zero only when z = z, n whch case t can be shown that an optmal desgn need not nclude x as a support pont,.e., x s safely elmnated. Proposton 1 can be establshed by followng the proof of Theorem 3 step by step, rather than appealng to Theorem 3 drectly. We omt the detals. 5 Remarks and llustratons The lterature abounds wth numercal examples of Algorthm I and ts relatves. There are several reasons for such wde nterest. Smlar to the EM algorthm, Algorthm I s smple, easy to mplement, and monotoncally convergent for a large class of optmalty crtera although ths was not proved n the present generalty). Algorthm I s known to be slow sometmes. But t serves as a foundaton upon whch more effectve varants can be bult see, e.g., Harman and Pronzato 2007, and Dette et al. 2008). Whle solvng the conjectured monotoncty of 9) holds mathematcal nterest, our man contrbuton s a way of nterpretng such algorthms as 9
10 optmzaton on augmented spaces. Ths opens up new possbltes n constructng algorthms wth the same desrable monotonc convergence propertes. As a numercal example, consder the logstc regresson model py x, θ) = expyx θ) 1 + expx, y = 0, 1. θ) The expected Fsher nformaton for θ from a unt assgned to x s A = x expx θ) 1 + expx θ))2 x. We compute locally optmal desgns wth pror guess θ = 1, 1) m = 2), and desgn spaces X 1 = {x = 1, /20), = 1,..., 20}; X 2 = {x = 1, /10), = 1,..., 30}. The desgn crtera consdered are φ 0 for D-optmalty) and φ 2. We use Algorthm I wth λ = 1, startng wth equally weghted desgns. For φ 0, Corollary 1 guarantees monotonc convergence. Ths s llustrated by Fgure 1, the frst row, where φ 0 = log det Mw) s plotted aganst teraton t. Usng the convergence crteron 18) wth δ = , the number of teratons untl convergence s 93 for X 1 and 2121 for X 2. The actual locally D-optmal desgns are w 1 = w 20 = 0.5 for X 1 and w 1 = w 23 = 0.5 for X 2, as can be verfed usng the general equvalence theorem. Ths smple example serves to llustrate both the monotoncty of Algorthm I when Theorem 1 apples) and ts potental slow convergence. For φ 2, although Algorthm I can be mplemented just as easly, Theorem 1 does not apply, because the concavty condton 7) no longer holds. Indeed, Algorthm I wth λ = 1) s not monotonc, as s evdent from Fgure 1, the second row, where φ 2 = trm 2 w)) s plotted aganst teraton t. Ths shows the potental danger of usng Algorthm I when monotoncty s not guaranteed. It seems worthwhle to nvestgate modfcatons that may agan lead to monotonc convergence n such stuatons. We have focused on local optmalty crtera. An alternatve, Bayesan optmalty Chaloner and Larntz, 1989; Chaloner and Verdnell, 1995), seeks to maxmze the expected value of φmθ; w)) over a pror dstrbuton πθ). The notaton Mθ; w) emphaszes the dependence of the moment matrx on the parameter θ. It would be worthwhle to extend our strategy n Secton 3 to Bayesan optmalty, and we plan to report both theoretcal and emprcal evaluatons of such extensons n future work. Acknowledgement The author would lke to thank Professors Donald Rubn, Xao-L Meng, and Davd van Dyk for ntroducng hm to the feld of statstcal computng. Appendx A: Detals of Case ) We only verfy 6), snce 5) s easy. For ψ p M) = trm p ), p [ 1, 0), 6) holds because the functon fx) = x p, x [0, ), p [ 1, 0), s operator concave Bhata 1997, Chapter V). For 10
11 desgn space 1 desgn space 2 log det M log det M teraton teraton ph crteron p= 2 1.2e e e+00 desgn space ph crteron p= desgn space teraton teraton Fgure 1: Values of φ 0 = log det M row 1) and φ 2 = trm 2 ) row 2) for Algorthm I wth desgn spaces X 1 left) and X 2 rght). 11
12 ψ 0 M) = log det M, a smple proof of 6) s to choose an m m nvertble matrx U such that U M 1 U = Daga 1,..., a m ); U M 2 U = Dagb 1,..., b m ). Then 6) reduces to m m m α log a + 1 α) log b logαa + 1 α)b ), whch holds by Jensen s nequalty. =1 =1 =1 Appendx B: Proof of Theorem 3 Lemma 1. Any lmt pont w of w t) s a fxed pont of T,.e., T w = w. Proof. Let w t j ) be a subsequence convergng to w. By Theorem 1 and Condton d), Mw ) s postve defnte,.e., w Ω +. Moreover, snce both T and the functon φm )) are contnuous on Ω +, we have φmw )) = lm j φmw t j ) )) = lm j φmw t j +1) )) = φmt w )), where the two lmts are equal by monotoncty. From Condton b) we deduce that T w = w. Lemma 2. Suppose w s a lmt pont of w t), and defne S + = {w Ω + : w = 0 = w = 0},.e., S + collects all w that are absolutely contnuous wth respect to w and satsfy Mw) > 0. Then we have φmw )) = sup w S + φmw)); 22) ŵ S +, φmŵ)) = sup w S + φmw)) = Mŵ) = Mw ). 23) Proof. By Lemma 1, w s a fxed pont of T. That s, w 0 = trφ Mw ))A ) = trφ Mw ))Mw )). By Theorem 2, w maxmzes φmw)) on S +, and 22) s proved. The mplcaton 23) holds because we assume φ ) s strctly concave Condton c)). Lemma 3. The sequence Mw t) ) has fntely many lmt ponts. Proof. Snce M ) s contnuous, any lmt pont of Mw t) ) s of the form Mw ) for some lmt pont w of w t). By Lemma 2, Mw ) s the unque maxmzer of φm) among all M > 0 that can be wrtten as M = Mw) wth w w. Dependng on whch coordnates of w are zero, there are fewer than 2 n such degenerate maxmzers. Lemma 4. The lmt M lm t Mw t) ) exsts. 12
13 Proof. Assume Mw t) ) has L < lmt ponts, and let B, = 1,..., L, be non-ntersectng balls neghborhoods) centered on these. We know L 1 because Mw t) ) s bounded; the choce of a metrc s mmateral. Agan by boundedness, for large enough t, each Mw t) ) belongs to exactly one of B. Assume L 2,.e., Mw t) ) does not converge. Then there exsts a subsequence such that Mw t j ) ), Mw t j +1) ) always belong to dfferent B. By passng through a sub-subsequence f necessary, we may assume w t j ) w ; by Lemma 1, w t j +1) w. It follows that Mw t j ) ) Mw ) and Mw t j +1) ) Mw ), whch contradcts the assumpton of dstnct neghborhoods. Lemma 5. The lmt M as defned n Lemma 4 satsfes φm ) = sup w Ω + φmw)). Proof. Let us check the condtons of the general equvalence theorem,.e., trφ M )A ) trφ M )M ), = 1,..., n. Suppose ths fals for = 1, say, then by Lemma 4 there exsts δ > 0 such that for suffcently large t, we have ) ) tr φ Mw t) ))A 1 > 1 + δ)tr φ Mw t) ))Mw t) ). It follows from the defnton of T that w t+1) 1 w t) 1 > 1 + δ) λ n =1 w d ) λ n =1 w d λ, 24) where d = tr φ Mw t) ) ))A. However, the rght hand sde of 24) s at least 1 + δ) λ due to Jensen s nequalty. That s, w t+1) 1 > 1 + δ) λ w t) 1 for all t large enough, whch contradcts the obvous constrant 0 < w t) 1 1. Theorem 3 then follows from Lemma 5. References [1] C.L. Atwood, Sequences convergng to D-optmal desgns of experments, Ann. Statst ) pp [2] R. Bhata, Matrx Analyss 1997) Sprnger. [3] K. Chaloner and K. Larntz, Optmal Bayesan desgn appled to logstc regresson experments, J. Statst. Plann. Inference ) pp [4] K. Chaloner and I. Verdnell, Bayesan expermental desgn: A revew, Statst. Sc ) pp [5] H. Chernoff, Locally optmal desgn for estmatng parameters, Ann. Math. Statst ) pp
14 [6] I. Csszár and G. Tusnady, Informaton geometry and alternatng mnmzaton procedures, Statstcs & Decsons, Supplement Issue ) pp [7] A.P. Dempster, N.M. Lard and D.B. Rubn, Maxmum lkelhood from ncomplete data va the EM algorthm wth dscusson), J. Roy. Statst. Soc. B ) pp [8] H. Dette, A. Pepelyshev and A. Zhgljavsky, Improvng updatng rules n multplcatve algorthms for computng D-optmal desgns, Computatonal Statstcs & Data Analyss ) pp [9] V.V. Fedorov, Theory of Optmal Experments 1972) Academc Press, New York. [10] J. Fellman, On the allocaton of lnear observatons Thess), Comment. Phys.-Math ) pp [11] J. Fellman, An emprcal study of a class of teratve searches for optmal desgns, J. Statst. Plannng Infer ) pp [12] R. Harman and L. Pronzato, Improvements on removng nonoptmal support ponts n D-optmum desgn algorthms, Statst. Probab. Lett ) pp [13] R. Harman and M. Trnovská, Approxmate D-optmal desgns of experments on the convex hull of a fnte set of nformaton matrces, Accepted, Mathematca Slovaca. [14] J. Kefer, General equvalence theory for optmum desgns approxmate theory), Ann. Statst ) [15] J. Kefer and J. Wolfowtz, The equvalence of two extremum problems, Canad. J. Math ) pp [16] G. L and D. Majumdar, D-optmal desgns for logstc models wth three and four parameters, Journal of Statstcal Plannng and Inference ) pp [17] X.-L. Meng and D. van Dyk, The EM algorthm an old folk-song sung to a fast new tune wth dscusson), J. Roy. Statst. Soc. B ) pp [18] A. Pázman, Foundatons of Optmum Expermental Desgn, Redel, Dordrecht 1986). [19] F. Pukelshem, Optmal Desgn of Experments, John Wley & Sons Inc, New York 1993). [20] F. Pukelshem and B. Torsney, Optmal weghts for expermental desgns on lnearly ndependent support ponts, Ann. Statst ) pp [21] S.D. Slvey, Optmal Desgn, Chapman & Hall, London 1980). [22] S.D. Slvey, D.M. Ttterngton and B. Torsney, An algorthm for optmal desgns on a fnte desgn space, Commun. Stat. Theory Methods ) pp [23] D.M. Ttterngton, Algorthms for computng D-optmal desgn on fnte desgn spaces. In Proc. of the 1976 Conf. on Informaton Scence and Systems, John Hopkns Unversty, ) pp
15 [24] D.M. Ttterngton, Estmaton of correlaton coeffcents by ellpsodal trmmng, Appl. Stat ) pp [25] B. Torsney, A moment nequalty and monotoncty of an algorthm. In: Kortanek, K.O. and Facco, A.V. Eds.), Proceedngs of the Internatonal Symposum on Sem-Infnte Programmng and Appl., Lecture Notes n Economcs and Mathematcal Systems, 215. Unversty of Texas at Austn 1983) pp [26] B. Torsney, W-teratons and rpples therefrom, In: Pronzato, L., Zhgljavsky, A. Eds.), Optmal Desgn and Related Areas n Optmzaton and Statstcs, Sprnger-Verlag, N.Y 2007) pp [27] B. Torsney and S. Mandal, Two classes of multplcatve algorthms for constructng optmzng dstrbutons, Computatonal Statstcs & Data Analyss ) pp [28] P. Whttle, Some general ponts n the theory of optmal expermental desgn. J. R. Statst. Soc. B ) [29] H.P. Wynn, Results n the theory and constructon of D-optmum expermental desgns, J. Roy. Statst. Soc. Ser. B ) pp [30] Y. Yu, A bt of nformaton theory, and the data augmentaton algorthm converges, IEEE Trans. Inform. Theory ) pp
Approximate D-optimal designs of experiments on the convex hull of a finite set of information matrices
Approxmate D-optmal desgns of experments on the convex hull of a fnte set of nformaton matrces Radoslav Harman, Mára Trnovská Department of Appled Mathematcs and Statstcs Faculty of Mathematcs, Physcs
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationAPPENDIX A Some Linear Algebra
APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,
More informationAssortment Optimization under MNL
Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.
More informationMore metrics on cartesian products
More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationAffine transformations and convexity
Affne transformatons and convexty The purpose of ths document s to prove some basc propertes of affne transformatons nvolvng convex sets. Here are a few onlne references for background nformaton: http://math.ucr.edu/
More informationCOS 521: Advanced Algorithms Game Theory and Linear Programming
COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton
More informationInner Product. Euclidean Space. Orthonormal Basis. Orthogonal
Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationYong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )
Kangweon-Kyungk Math. Jour. 4 1996), No. 1, pp. 7 16 AN ITERATIVE ROW-ACTION METHOD FOR MULTICOMMODITY TRANSPORTATION PROBLEMS Yong Joon Ryang Abstract. The optmzaton problems wth quadratc constrants often
More informationSTAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 16
STAT 39: MATHEMATICAL COMPUTATIONS I FALL 218 LECTURE 16 1 why teratve methods f we have a lnear system Ax = b where A s very, very large but s ether sparse or structured (eg, banded, Toepltz, banded plus
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that
More informationPerfect Competition and the Nash Bargaining Solution
Perfect Competton and the Nash Barganng Soluton Renhard John Department of Economcs Unversty of Bonn Adenauerallee 24-42 53113 Bonn, Germany emal: rohn@un-bonn.de May 2005 Abstract For a lnear exchange
More informationMMA and GCMMA two methods for nonlinear optimization
MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons
More informationGeneral viscosity iterative method for a sequence of quasi-nonexpansive mappings
Avalable onlne at www.tjnsa.com J. Nonlnear Sc. Appl. 9 (2016), 5672 5682 Research Artcle General vscosty teratve method for a sequence of quas-nonexpansve mappngs Cuje Zhang, Ynan Wang College of Scence,
More informationDeriving the X-Z Identity from Auxiliary Space Method
Dervng the X-Z Identty from Auxlary Space Method Long Chen Department of Mathematcs, Unversty of Calforna at Irvne, Irvne, CA 92697 chenlong@math.uc.edu 1 Iteratve Methods In ths paper we dscuss teratve
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More informationModule 9. Lecture 6. Duality in Assignment Problems
Module 9 1 Lecture 6 Dualty n Assgnment Problems In ths lecture we attempt to answer few other mportant questons posed n earler lecture for (AP) and see how some of them can be explaned through the concept
More informationEstimation: Part 2. Chapter GREG estimation
Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the
More informationComputing MLE Bias Empirically
Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationLecture 4: Universal Hash Functions/Streaming Cont d
CSE 5: Desgn and Analyss of Algorthms I Sprng 06 Lecture 4: Unversal Hash Functons/Streamng Cont d Lecturer: Shayan Oves Gharan Aprl 6th Scrbe: Jacob Schreber Dsclamer: These notes have not been subjected
More informationThe Minimum Universal Cost Flow in an Infeasible Flow Network
Journal of Scences, Islamc Republc of Iran 17(2): 175-180 (2006) Unversty of Tehran, ISSN 1016-1104 http://jscencesutacr The Mnmum Unversal Cost Flow n an Infeasble Flow Network H Saleh Fathabad * M Bagheran
More informationA New Refinement of Jacobi Method for Solution of Linear System Equations AX=b
Int J Contemp Math Scences, Vol 3, 28, no 17, 819-827 A New Refnement of Jacob Method for Soluton of Lnear System Equatons AX=b F Naem Dafchah Department of Mathematcs, Faculty of Scences Unversty of Gulan,
More informationThe Order Relation and Trace Inequalities for. Hermitian Operators
Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence
More informationISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013
ISSN: 2277-375 Constructon of Trend Free Run Orders for Orthogonal rrays Usng Codes bstract: Sometmes when the expermental runs are carred out n a tme order sequence, the response can depend on the run
More informationThe Geometry of Logit and Probit
The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.
More informationGames of Threats. Elon Kohlberg Abraham Neyman. Working Paper
Games of Threats Elon Kohlberg Abraham Neyman Workng Paper 18-023 Games of Threats Elon Kohlberg Harvard Busness School Abraham Neyman The Hebrew Unversty of Jerusalem Workng Paper 18-023 Copyrght 2017
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More informationDifference Equations
Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1
More informationFeature Selection: Part 1
CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?
More informationThe Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction
ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also
More informationOn an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1
On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationChapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems
Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons
More informationLinear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.
Lnear, affne, and convex sets and hulls In the sequel, unless otherwse specfed, X wll denote a real vector space. Lnes and segments. Gven two ponts x, y X, we defne xy = {x + t(y x) : t R} = {(1 t)x +
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationSolutions to exam in SF1811 Optimization, Jan 14, 2015
Solutons to exam n SF8 Optmzaton, Jan 4, 25 3 3 O------O -4 \ / \ / The network: \/ where all lnks go from left to rght. /\ / \ / \ 6 O------O -5 2 4.(a) Let x = ( x 3, x 4, x 23, x 24 ) T, where the varable
More informationSolutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.
Solutons HW #2 Dual of general LP. Fnd the dual functon of the LP mnmze subject to c T x Gx h Ax = b. Gve the dual problem, and make the mplct equalty constrants explct. Soluton. 1. The Lagrangan s L(x,
More informationA note on almost sure behavior of randomly weighted sums of φ-mixing random variables with φ-mixing weights
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 7, Number 2, December 203 Avalable onlne at http://acutm.math.ut.ee A note on almost sure behavor of randomly weghted sums of φ-mxng
More informationPower law and dimension of the maximum value for belief distribution with the max Deng entropy
Power law and dmenson of the maxmum value for belef dstrbuton wth the max Deng entropy Bngy Kang a, a College of Informaton Engneerng, Northwest A&F Unversty, Yanglng, Shaanx, 712100, Chna. Abstract Deng
More informationMATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)
1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons
More informationSupplement: Proofs and Technical Details for The Solution Path of the Generalized Lasso
Supplement: Proofs and Techncal Detals for The Soluton Path of the Generalzed Lasso Ryan J. Tbshran Jonathan Taylor In ths document we gve supplementary detals to the paper The Soluton Path of the Generalzed
More informationThe Second Anti-Mathima on Game Theory
The Second Ant-Mathma on Game Theory Ath. Kehagas December 1 2006 1 Introducton In ths note we wll examne the noton of game equlbrum for three types of games 1. 2-player 2-acton zero-sum games 2. 2-player
More informationTime-Varying Systems and Computations Lecture 6
Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy
More informationSome modelling aspects for the Matlab implementation of MMA
Some modellng aspects for the Matlab mplementaton of MMA Krster Svanberg krlle@math.kth.se Optmzaton and Systems Theory Department of Mathematcs KTH, SE 10044 Stockholm September 2004 1. Consdered optmzaton
More information/ n ) are compared. The logic is: if the two
STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence
More informationExplaining the Stein Paradox
Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten
More informationApproximate Smallest Enclosing Balls
Chapter 5 Approxmate Smallest Enclosng Balls 5. Boundng Volumes A boundng volume for a set S R d s a superset of S wth a smple shape, for example a box, a ball, or an ellpsod. Fgure 5.: Boundng boxes Q(P
More informationPrimer on High-Order Moment Estimators
Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc
More informationLecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.
prnceton u. sp 02 cos 598B: algorthms and complexty Lecture 20: Lft and Project, SDP Dualty Lecturer: Sanjeev Arora Scrbe:Yury Makarychev Today we wll study the Lft and Project method. Then we wll prove
More informationAppendix B. Criterion of Riemann-Stieltjes Integrability
Appendx B. Crteron of Remann-Steltes Integrablty Ths note s complementary to [R, Ch. 6] and [T, Sec. 3.5]. The man result of ths note s Theorem B.3, whch provdes the necessary and suffcent condtons for
More informationDurban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications
Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More information3.1 ML and Empirical Distribution
67577 Intro. to Machne Learnng Fall semester, 2008/9 Lecture 3: Maxmum Lkelhood/ Maxmum Entropy Dualty Lecturer: Amnon Shashua Scrbe: Amnon Shashua 1 In the prevous lecture we defned the prncple of Maxmum
More informationRandić Energy and Randić Estrada Index of a Graph
EUROPEAN JOURNAL OF PURE AND APPLIED MATHEMATICS Vol. 5, No., 202, 88-96 ISSN 307-5543 www.ejpam.com SPECIAL ISSUE FOR THE INTERNATIONAL CONFERENCE ON APPLIED ANALYSIS AND ALGEBRA 29 JUNE -02JULY 20, ISTANBUL
More informationAn (almost) unbiased estimator for the S-Gini index
An (almost unbased estmator for the S-Gn ndex Thomas Demuynck February 25, 2009 Abstract Ths note provdes an unbased estmator for the absolute S-Gn and an almost unbased estmator for the relatve S-Gn for
More informationParametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010
Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More information8 : Learning in Fully Observed Markov Networks. 1 Why We Need to Learn Undirected Graphical Models. 2 Structural Learning for Completely Observed MRF
10-708: Probablstc Graphcal Models 10-708, Sprng 2014 8 : Learnng n Fully Observed Markov Networks Lecturer: Erc P. Xng Scrbes: Meng Song, L Zhou 1 Why We Need to Learn Undrected Graphcal Models In the
More informationChapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of
Chapter 7 Generalzed and Weghted Least Squares Estmaton The usual lnear regresson model assumes that all the random error components are dentcally and ndependently dstrbuted wth constant varance. When
More information10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)
0-80: Advanced Optmzaton and Randomzed Methods Lecture : Convex functons (Jan 5, 04) Lecturer: Suvrt Sra Addr: Carnege Mellon Unversty, Sprng 04 Scrbes: Avnava Dubey, Ahmed Hefny Dsclamer: These notes
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More informationMAT 578 Functional Analysis
MAT 578 Functonal Analyss John Qugg Fall 2008 Locally convex spaces revsed September 6, 2008 Ths secton establshes the fundamental propertes of locally convex spaces. Acknowledgment: although I wrote these
More informationMaximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models
ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models
More informationCollege of Computer & Information Science Fall 2009 Northeastern University 20 October 2009
College of Computer & Informaton Scence Fall 2009 Northeastern Unversty 20 October 2009 CS7880: Algorthmc Power Tools Scrbe: Jan Wen and Laura Poplawsk Lecture Outlne: Prmal-dual schema Network Desgn:
More informationConvexity preserving interpolation by splines of arbitrary degree
Computer Scence Journal of Moldova, vol.18, no.1(52), 2010 Convexty preservng nterpolaton by splnes of arbtrary degree Igor Verlan Abstract In the present paper an algorthm of C 2 nterpolaton of dscrete
More informationErrors for Linear Systems
Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch
More informationThe Expectation-Maximization Algorithm
The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.
More informationSolutions Homework 4 March 5, 2018
1 Solutons Homework 4 March 5, 018 Soluton to Exercse 5.1.8: Let a IR be a translaton and c > 0 be a re-scalng. ˆb1 (cx + a) cx n + a (cx 1 + a) c x n x 1 cˆb 1 (x), whch shows ˆb 1 s locaton nvarant and
More informationChapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the
More informationMaximizing the number of nonnegative subsets
Maxmzng the number of nonnegatve subsets Noga Alon Hao Huang December 1, 213 Abstract Gven a set of n real numbers, f the sum of elements of every subset of sze larger than k s negatve, what s the maxmum
More informationx i1 =1 for all i (the constant ).
Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by
More informationLOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin
Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence
More informationLecture 21: Numerical methods for pricing American type derivatives
Lecture 21: Numercal methods for prcng Amercan type dervatves Xaoguang Wang STAT 598W Aprl 10th, 2014 (STAT 598W) Lecture 21 1 / 26 Outlne 1 Fnte Dfference Method Explct Method Penalty Method (STAT 598W)
More informationComputing Correlated Equilibria in Multi-Player Games
Computng Correlated Equlbra n Mult-Player Games Chrstos H. Papadmtrou Presented by Zhanxang Huang December 7th, 2005 1 The Author Dr. Chrstos H. Papadmtrou CS professor at UC Berkley (taught at Harvard,
More informationConjugacy and the Exponential Family
CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the
More informationBayesian Planning of Hit-Miss Inspection Tests
Bayesan Plannng of Ht-Mss Inspecton Tests Yew-Meng Koh a and Wllam Q Meeker a a Center for Nondestructve Evaluaton, Department of Statstcs, Iowa State Unversty, Ames, Iowa 5000 Abstract Although some useful
More informationFACTORIZATION IN KRULL MONOIDS WITH INFINITE CLASS GROUP
C O L L O Q U I U M M A T H E M A T I C U M VOL. 80 1999 NO. 1 FACTORIZATION IN KRULL MONOIDS WITH INFINITE CLASS GROUP BY FLORIAN K A I N R A T H (GRAZ) Abstract. Let H be a Krull monod wth nfnte class
More informationMATH 241B FUNCTIONAL ANALYSIS - NOTES EXAMPLES OF C ALGEBRAS
MATH 241B FUNCTIONAL ANALYSIS - NOTES EXAMPLES OF C ALGEBRAS These are nformal notes whch cover some of the materal whch s not n the course book. The man purpose s to gve a number of nontrval examples
More informationClassification as a Regression Problem
Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class
More informationA new construction of 3-separable matrices via an improved decoding of Macula s construction
Dscrete Optmzaton 5 008 700 704 Contents lsts avalable at ScenceDrect Dscrete Optmzaton journal homepage: wwwelsevercom/locate/dsopt A new constructon of 3-separable matrces va an mproved decodng of Macula
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationOn the Multicriteria Integer Network Flow Problem
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 5, No 2 Sofa 2005 On the Multcrtera Integer Network Flow Problem Vassl Vasslev, Marana Nkolova, Maryana Vassleva Insttute of
More informationHidden Markov Models
Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,
More informationP exp(tx) = 1 + t 2k M 2k. k N
1. Subgaussan tals Defnton. Say that a random varable X has a subgaussan dstrbuton wth scale factor σ< f P exp(tx) exp(σ 2 t 2 /2) for all real t. For example, f X s dstrbuted N(,σ 2 ) then t s subgaussan.
More information10) Activity analysis
3C3 Mathematcal Methods for Economsts (6 cr) 1) Actvty analyss Abolfazl Keshvar Ph.D. Aalto Unversty School of Busness Sldes orgnally by: Tmo Kuosmanen Updated by: Abolfazl Keshvar 1 Outlne Hstorcal development
More informationPerron Vectors of an Irreducible Nonnegative Interval Matrix
Perron Vectors of an Irreducble Nonnegatve Interval Matrx Jr Rohn August 4 2005 Abstract As s well known an rreducble nonnegatve matrx possesses a unquely determned Perron vector. As the man result of
More informationLecture 4: September 12
36-755: Advanced Statstcal Theory Fall 016 Lecture 4: September 1 Lecturer: Alessandro Rnaldo Scrbe: Xao Hu Ta Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer: These notes have not been
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More information