Counting Patterns in Degenerated Sequences

Size: px
Start display at page:

Download "Counting Patterns in Degenerated Sequences"

Transcription

1 Countng Patterns n Degenerate Sequences Grégory Nuel MAP5, CNRS 845, Unversty Pars Descartes, 45 rue es Sant-Pères, F Pars, France gregory.nuel@parsescartes.fr Abstract. Bologcal sequences lke DNA or protens, are always obtane through a sequencng process whch mght prouce some uncertanty. As a result, such sequences are usually wrtten n a egenerate alphabet where some symbols may correspon to several possble letters ex: IUPAC DNA alphabet. When countng patterns n such egenerate sequences, the queston that naturally arses s: how to eal wth egenerate postons? Snce most usually 99% of the postons are not egenerate, t s consere harmless to scar the egenerate postons n orer to get an observaton, but the exact consequences of such a practce are unclear. In ths paper, we ntrouce a rgorous metho to take nto account the uncertanty of sequencng for bologcal sequences DNA, Protens. We frst ntrouce a Forwar-Backwar approach to compute the margnal strbuton of the constrane sequence an use t both to perform a Expectaton-Maxmzaton estmaton of parameters, as well as ervng a heterogeneous Markov strbuton for the constrane sequence. hs strbuton s hence use along wth known DFA-base pattern approaches to obtan the exact strbuton of the pattern count uner the constrants. As an llustraton, we conser a ES ataset from the EMBL atabase. Despte the fact that only % of the postons n ths ataset are egenerate, we show that not takng nto account these postons mght lea to erroneous observatons, further provng the nterest of our approach. Keywors: Forwar-Backwar algorthm, Expectaton-Maxmzaton algorthmn, Markov chan embeng, Determnstc Fnte state Automaton. Introucton Bologcal sequences lke DNA or protens, are always obtane through a sequencng process whch mght prouce some uncertanty. As a result, such sequences are usually wrtten n a egenerate alphabet where some symbols may correspon to several possble letters. For example, the IUPAC [] proten alphabet nclues the followng egenerate symbols: X for any amno-ac, Z for glutamc ac or glutamne, an B for Aspartc ac or Asparagne. For DNA sequences, there s even more of such egenerate symbols whch exhaustve lst an meanng are gven n able along wth observe frequences n several atasets from the EMBL [2] atabase. V. Karkamanathan et al. Es.: PRIB 2009, LNBI 5780, pp , c Sprnger-Verlag Berln Heelberg 2009

2 Countng Patterns n Degenerate Sequences 223 able. Meanng an frequency of the IUPAC [] DNA symbols n several fles of the release 97 of the EMBL nucleote sequence atabase [2]. Degenerate symbols lowest part of the table contrbute to 0.5% to % of the ata. symbol meanng est pro 0 htg pro 0 htc fun 0 st hum 2 A Aenne C Cytosne G Guanne hymne U Uracl R Purne A or G Y Pyrmne C,, oru M C or A K, U, org W, U, ora S C or G B not A D not C H not G V not G, notu N any base When countng patterns n such egenerate sequences, the queston that naturally arse s: how to eal wth egenerate postons? Snce most usually 99% of the postons are not egenerate, t s usually consere harmless to scar the egenerate postons n orer to get an observaton. Another opton mght be to preprocess the ataset by replacng each specal letter by the most lkely compatble symbol at the poston n reference wth some backgroun moel. Fnally, one mght come up wth some a hoc countng rule lke: whenever the pattern mght occurs I a one to the observe count. However practcal, all these solutons reman qute unsatsfactory from the statstcan pont of vew an ther possble consequences lke ang or mssng occurrences reman unclear. In ths paper, we want to eal rgorously wth the problem of egenerate symbols n sequences by ntroucng the strbuton of sequences uner the uncertanty of ther sequencng, an then by usng ths strbuton to stuy the observe number of occurrences of a pattern of nterest. o o so we place ourself n a Markovan framework by assumng that the sequence X l = X...X l s a orer 2 homogeneous Markov chan over the fnte alphabet A. Weenotebyν ts startng strbuton an by π ts transton matrx. For all a Aan for all b Awe then have: P X = a = νa = an PX + = b X + = a =πa,bwth l. One mght also thnk to a a fracton of one whch correspon to the probablty to see the corresponng letter at the egenerate poston. 2 For the sake of smplcty, the partcular egenerate case where = 0 s left to the reaer.

3 224 G. Nuel For all l we enote by X Athesubset of all possble values taken by X accorng to the ata. For example, f we conser a IUPAC DNA sequence ANWY... wehavex = {A}, X 2 = {A, C, G, }, X 3 = {}, X 4 = {A, }, X 5 = {C, },... In a frst part we establsh the strbuton of X l uner the constrant that X l X l usng an aaptaton of the Baum-Welch algorthm [3]. We then emonstrate that the constrane sequence s strbute accorng to a heterogeneous Markov moel whch startng strbuton an transton functon have explct expressons. hs result hence allows to obtan the exact constrane strbuton of a pattern by the applcaton of known Markov chan embeng technques. he nterest of the metho s fnally llustrate wth ES ata an scusse. 2 Constrane Dstrbuton In orer to compute the constrane probablty PX X X we follow the sketch of the Baum-Welch algorthm [3] by ntroucng the Forwar an Backwar quanttes. Proposton Forwar. For all x l A l an, l we efne the forwar quantty F x + ef = P X + = x +,X + X + whch s computable by recurrence through: F x + = x + 2 π x + 2,x + x X F for 2 l + an wth the ntalzaton F x = ν x IX x where I s the ncatrx functon 3. We then obtan that: PX l X l = F l x l l π x l l,x l. 2 x l l X l l Proof. We prove Equaton by smply rewrtng F x + as: = F x + = PX + x X = x + 2,X + 2 X + 2 }{{} PX + 2 x X F x + 2 = x +,X + X + PX + = x +,X + X + X + 2 = x + 2,X + 2 X + 2. }{{} πx + 2,x + I X+ x + he proof of Equaton 2 s establshe n a smlar manner. 3 For any set E, subset A E an element a E, I Aa =fa A an I Aa =0 otherwse.

4 Countng Patterns n Degenerate Sequences 225 Proposton 2 Backwar. For all x l Al an, l we efne the backwar quantty B x + ef = P X l Xl X+ = x + whch s computable by recurrence through: B x + = x +,x + B+ x x + X + π for 2 l an wth the ntalzaton B l x l l = x l X l π x l l,x l I X l l x l l. We then obtan that: PX l X l = x X ν x B x. 4 Proof. he proof s very smlar to the one of Proposton an s hence omtte. heorem Margnal strbutons. For all x l Al we have the followng results: a P X = x,x l X l = ν x B x ; b P X + = x +,X l Xl = F x + π x +,x + B+ x + + ; c P Xl l = xl l,xl Xl = Fl x l l π x l l,x l ; P X + = x +,X l Xl = F x + B x +. Proof. a, b, an c are prove usng the same contonng mechansms use n the proofs of propostons an 2. One coul note that a s a rect consequence of Equaton 4, whle c coul be erve from Equaton 2. hanks to Equaton 3, t s also clear that b whch acheves the proof. From now on we enote by P C A ef = P A X l Xl the probablty of an event A uner the constrant that X l Xl. heorem 2 Heterogeneous Markov chan. X l s a orer heterogeneous Markov chan uner P C whch startng strbuton ν C s gven by: ν C x ef = P C X = x ν x B x 5 an the transton matrx π+ C towar poston + s gven by: π C + x ef = P C X + = x + X + = x + π x +,x + B+ x Proof. Equaton 5 s a rect consequence of heorem a an Equaton 4. For Equaton 6 we start by enotng P C X + = x + X + = x + = PA B,C,D wth A = {X + = x + }, B = {X + = x + }, C = {X X}, and = {X+ l X }. hanks to Bayes formula we get that PA B,C,D PD A, B, C PA B,C. We fnally use the Markov property to get PD A, B, C = B + x + + an PA B,C = π x +,x + whch acheves the proof.

5 226 G. Nuel One shoul note the reverse sequence Xl = X l...x s also a heterogeneous orer Markov moel whch parameters can be expresse through the Forwar quanttes. 3 Estmatng the Backgroun Moel Let us enote by θ ef = ν, π the parameters of our orer Markov moel. We enote by P θ all probablty computaton performe usng the parameter ef = logp θ X l X l may be e- θ. Snce the log-lkelhoo L θ X l Xl rve ether from the Forwar or Backwar quanttes, t s possble maxmze numercally ths lkelhoo to get the Maxmum Lkelhoo Estmator MLE θ ef =argmax θ L θ X l Xl. We suggest here an alternatve approach foune on the classcal Expectaton- Maxmzaton algorthm for maxmum lkelhoo estmaton from ncomplete ata [4]. o o so, we smply conser that X l Xl s the observe ata, whle X l = x l s the unobserve ata. We then get the followng result: ef Proposton 3 EM algorthm. For any startng parameter θ 0 = ν 0,π 0, ef we conser the sequence θ j j 0 efne for all j 0 by θ j+ = ν j+,π j+ wth: I ν j+ a {a = X } ν j a θ B j a P θj X l X l 7 a πj a,b B θj a 2 b π j+ a,b = l = I {a b X + l = I {a } F θj + X } F θj a B θ j + a 8 where the F θj an B θj enote respectvely the Forwar an Backwar quanttes compute wth the current value θ j of the parameter, an wth the conventon that B θj l +. he sequence θ j j 0 converge towars a local maxmum of L θ X l Xl. Proof. hs comes from a specal applcaton of the EM algorthm [4] where the Expectaton step Step E conssts n computng Q θ θ j ef = x l X l P θ X l = x l X l X l log Pθ X l = x l whle the Maxmzaton step Step M conssts n computng θ + =argmax θ Q θ θ j. Equatons 7 an 8 then smply come from a natural aaptaton of the classcal MLE of a orer Markov chans usng the pseuo counts that come rectly from heorem.

6 4 Countng Patterns Countng Patterns n Degenerate Sequences 227 Let us conser herew a fnte set of wors over A. We wantto countthe number N of postons where W occurs n our egenerate sequence. Unfortunately, snce the sequence tself s not observe, we stuy nstea the number N of matchng postons n the ranom sequence X l uner P C. hanks to heorem 2 we hence nee to establsh the strbuton of N over a heterogeneous orer Markov chan. o o so, we perform an optmal Markov chan embeng of the problem through a Determnstc Fnte Automaton DFA as t s suggeste n [5; 6; 7; 8]. We use here the notatons of [8]. Let A, Q, s,f, δ beamnmal DFA recognzng the language 4 A W of all texts over A enng wth an occurrence of W see Fgure for an example of such a mnmal DFA. Q s a fnte state space, s Qs the startng state, F Qs the subset of fnal states, an δ : Q A Q s the transton functon. We recursvely exten the efnton of δ over Q A thanks to the relaton δp, aw ef = δδp, a,w for all p Q,a A,w A.We atonally suppose that ths automaton s non -ambguous 5 whch means that for all q Q, δ p ef = { a A, p Q,δ } p, a = q s ether a sngleton, or the empty set. A,C,G A,C,G A,C,G 0 A,C A,C,G 3 A,C,G A,C 2 G 6 4 G G 5 7 A,C Fg.. Mnmal DFA recognzng the language of all DNA sequences enng wth an occurrence of the IUPAC pattern NG. hs DFA have a total of 8 states, s =0 beng the startng state, an F = {7} beng the subset of fnal states. hs DFA s -ambguous snce one can reach states 0 or state 3 wth more than one letter. heorem 3 Markov chan embeng. We conser the ranom sequence over Q efne by X ef 0 = s an X ef = δ X,X, l. UnerP C, 4 A enotes the set of all possbly empty texts over A. 5 A DFA havng ths property s also calle a -th orer DFA n [7].

7 228 G. Nuel X s a heterogeneous orer Markov chan over Q ef = δs, A A such as, for all p, q Q an l the startng strbuton μ p ef = P C X = p an transton matrx + p, q ef = P C X+ = q X + = p are gven by: ν C μ p ={ a f a A,δ s, a = p ; 0 else { μ C + p, q = + δ p,b f b A,δp, b =q 0 else. Snce Q + contans all countng transtons, we keep track of the number of occurrences by assocatng a ummy varable y to these transtons. hen computng the margnal strbuton at the en of the sequence woul gve us access to the moment generatng functon mgf of the ranom number of occurrences see[5;6;7;8]formoreetals: Corollary Moment generatng functon. he moment generatng functon F y of the ranom number N uner P C s gven by: [ + l ] F y ef = P C N = k y k = μ P + + yq + 9 k=0 where s a column vector of ones an where, for all l, + = P + + Q + wth P + p, q ef = I q/ F + p, q an Q + p, q ef = I q F + p, q for all p, q Q. 5 Algorthm he practcal mplementaton of ths results requres two man steps: compute Forwar an Backwar quanttes; 2 compute the mgf usng Corollary. For the frst step, the resultng complexty s Ol both n space an tme. For the secon step the space complexty s OD Q whered s the fference between the maxmum an the mnmum egree of F y, an the tme complexty s Ol D Q A we take here takng avantage of the sparse structure of +. Usng ths approach on a large ataset ex: l =5 0 6 or l =3 0 9 may then result nto hgh memory requrements an/or long runnng tme. Fortunately, t s possble to reuce ramatcally these complextes when conserng egenerate sequence where most postons are etermnstc lke t s the case wth bologcal sequences. Let us enote by I ef = { l, X > } the set of egenerate postons n the sequence. It s clear then that the ranom state X j s completely etermnstc for all j J ef = { l, j > + I}. hepostonsj J thus contrbute n a etermnstc way to N wth a fxe number of occurrence n. It =

8 Countng Patterns n Degenerate Sequences 229 hence remans only to take nto account the varable part N n = N N k where the N are nepenent contrbutons of each of the k segments of J the complementary of J n {,...,l}. If we enote by F y themgfofn,wegetthat F y =y n k F y whch ramatcally reuces the complexty of the problem. Snce each F y may be obtane by a smple applcaton of Corollary on the partcular short segment of nterest, an one only nee to compute the Forwar-Backwar quanttes for ths partcular segment. For example, let us conser that the observe IUPAC sequence s x l = AAYGCANGBAGGCACWAGR an that = 2. We have I = {3, 7, 9, 20, 24} an J =[3, 5] [7, ] [20, 22] [24, 25]. In orer to compute F y, F 2 y, F 3 y anf 4 y, one just nee to known the orer =2pastbeforeeachofthe corresponng segment: AA for the frst, CA for the secon, C for the thr, an G for the last one. 6 Dscusson Let us conser the ataset est pro 0 whch s escrbe n able. Here s the transton matrx over of a orer = homogeneous Markov moel over A = {A, C, G, } estmate on ths ataset usng MLE though the EM algorthm: π = Snce only % of the ataset s egenerate, we observe lttle fference between ths rgorous estmate an one obtane through a rough heurstc lke scarng all egenerate postons n the ata. However, ths result shoul not be taken as a rule, especally when conserng more egenerate sequences e. g. wth 0% egenerate postons an/or hgher orer Markov moels e. g. =4. Usng ths moel, t s possble to stuy the observe strbuton of a pattern n the ataset by computng though Corollary the strbuton of ts ranom number of occurrence N uner the constrane probablty P C. able 2 compares the number of occurrences obtane by scarng all egenerate postons n the ata Count to the observe strbuton. Despte the fact that only % of the ata are egenerate, we can see that there s great fferences between our nave approach an the real observe strbuton. For example, f we conser the smple pattern GCA we can see that the nave count of 75 occurrences les well outse the 90% creblty nterval [727, 740] an we have smlar results for the other consere patterns. =

9 230 G. Nuel able 2. Dstrbuton of patterns n the egenerate IUPAC sequences from est pro 0. Count s obtane by scarng all egenerate postons n the ataset, an Count2 by replacng each specal letter by the most lkely compatble symbol. Snce the observe strbuton s screte, percentles an mean are roune to the closest value. pattern Count Count2 mn 5%-tle mean 95%-tle max GCA AG NG RNANNNSM For more complex patterns lke NG the fference between the nave count an the observe strbuton s even more ramatc snce 839 oes not even belong to the support [853, 005] of the observe strbuton. hs s ue to the fact that the strng NG actually occurs = 4 tmes n the ataset. Snce our nave approach scars all postons n the ata where a symbol other than A, C, G or appears, these 4 occurrences are hence omtte. If we now preprocess the ataset by replacng all egenerate symbols by the most frequent letter n the corresponng subset we get the number of occurrences enote Count2. If ths heurstc seems to gve an nterestng result for Pattern GCA countng close to the mean, t s unfortunately not the case for the other ones for whch the metho results ether n uner-countng Pattern NG or over-countng patterns AG an RNANNNSM. As a general rule, t s usually ffcult to prect the bas ntrouce by a partcular heurstc snce t can ether lea to uner- or over-coutngs for example Count always result n uner-countngs an that ths may even epen on the pattern of nterest lke wth Count2. he rgorous metho we have here evelope may hence also prove a way to test the statstcal propertes of a partcular heurstc. Fnally, let us pont out that thanks to the optmal Markov chan embeng prove by the DFA-base approach presente above, we are here able to eal wth relatvely complex patterns lke RNANNNSM. 7 Concluson In ths paper, we prove a rgorous way to eal wth the strbuton of Markov chans over a fnte alphabet A uner the constrant that each poston X of the sequence belongs to restrcte subset X A. We prove a Forwar-Backwar framework to compute margnal strbutons an erve from t a EM estmaton proceure. We also prove that the resultng constrane strbuton s a heterogeneous Markov chans an prove explct formulas to recursvely compute ts transton matrx. hanks to ths result, t s possble to apply known DFA-base methos from pattern theory to stuy the strbuton of a pattern of nterest n ths constrane sequence, hence provng a trustful observe strbuton for

10 Countng Patterns n Degenerate Sequences 23 the pattern number of occurrences. hs nformaton may then be use to erve ap-valuep for a pattern by combnng p n the p-value of the observaton of n occurrences n a unconstrane ataset wth the observe strbuton through formulas lke p = n p np C N = n. One shoul note that the approach we ntrouce here may have more applcatons than just countng patterns n IUPAC sequences. For example, one mght use a smlar approach to take nto account the occurrences postons of known patterns of nterest thus allowng to erve strbuton of patterns contonally to a possbly complex set of other patterns. One shoul also pont out that the constrant X X shoul easly be complexfcate, for example by conserng a specfc strbuton over X. For nstance, such a strbuton may come from the posteror econg probabltes of a sequencng machne. From the computatonal pont of vew, t s essental to unerstan that the heterogeneous nature of the Markov chan we conser forb to use classcal computatonal trcks lke power computatons. he resultng complexty s hence lnear wth the sequence length l rather that logarthmc. However, one shoul expect a ramatc mprovement of the metho by restrctng the use of heterogeneous Markov moels only n the vcnty of egenerate postons lke t s suggeste n Secton 5. Wth such an approach, one mght rely on classcal pattern matchng for 99% of the ata, an the metho presente above woul be restrcte to the stuy of the % remanng ata. Usng ths computatonal trck, t hence seems possble to rely on the rgorous exact computaton ntrouce here rather than on a bas heurstc. Fnally, we have emonstrate wth our example that even a small amount of egenerate ata may have huge consequences n terms of pattern frequences, an thus possbly affect every subsequent analyss metho nvolvng these frequences lke Markov an hen Markov moel estmatons an pattern stues. Conserng the possble bas cause by egenerate letters n bologcal ata, an the reasonable complexty of the exact soluton we ntrouce n ths paper, our stuy suggests that the problem of egenerate ata n pattern relate analyss shoul no longer be gnore. References [] IUPAC: Internatonal Unon of Pure an Apple Chemstry 2009, [2] EMBL: European Molecular Bology Laboratory Nucleote Sequence Database 2009, [3] Baum, L.E., Petre,., Soules, G., Wess, N.: A maxmzaton technque occurrng n the statstcal analyss of probablstc functons of markov chans. Ann. Math. Statst. 4, [4] Dempster, A.P., Lar, N.M., Rubn, D.B.: Maxmum Lkelhoo from Incomplete Data va the EM Algorthm. Journal of the Royal Stat. Socety. Seres B 39, [5] Ncoème, P., Salvy, B., Flajolet, P.: Motf statstcs. heoretcal Com. Sc. 2872,

11 232 G. Nuel [6] Crochemore, M., Stefanov, V.: Watng tme an complexty for matchng patterns wth automata. Info. Proc. Letters 873, [7] Llaser, M.E.: Mnnal markov chan embengs of pattern problems. In: Informaton heory an Applcatons Workshop, pp [8] Nuel, G.: Pattern markov chans: optmal markov chan embeng through etermnstc fnte automata. J. of Apple Prob. 45,

p(z) = 1 a e z/a 1(z 0) yi a i x (1/a) exp y i a i x a i=1 n i=1 (y i a i x) inf 1 (y Ax) inf Ax y (1 ν) y if A (1 ν) = 0 otherwise

p(z) = 1 a e z/a 1(z 0) yi a i x (1/a) exp y i a i x a i=1 n i=1 (y i a i x) inf 1 (y Ax) inf Ax y (1 ν) y if A (1 ν) = 0 otherwise Dustn Lennon Math 582 Convex Optmzaton Problems from Boy, Chapter 7 Problem 7.1 Solve the MLE problem when the nose s exponentally strbute wth ensty p(z = 1 a e z/a 1(z 0 The MLE s gven by the followng:

More information

ENTROPIC QUESTIONING

ENTROPIC QUESTIONING ENTROPIC QUESTIONING NACHUM. Introucton Goal. Pck the queston that contrbutes most to fnng a sutable prouct. Iea. Use an nformaton-theoretc measure. Bascs. Entropy (a non-negatve real number) measures

More information

New Liu Estimators for the Poisson Regression Model: Method and Application

New Liu Estimators for the Poisson Regression Model: Method and Application New Lu Estmators for the Posson Regresson Moel: Metho an Applcaton By Krstofer Månsson B. M. Golam Kbra, Pär Sölaner an Ghaz Shukur,3 Department of Economcs, Fnance an Statstcs, Jönköpng Unversty Jönköpng,

More information

Explicit bounds for the return probability of simple random walk

Explicit bounds for the return probability of simple random walk Explct bouns for the return probablty of smple ranom walk The runnng hea shoul be the same as the ttle.) Karen Ball Jacob Sterbenz Contact nformaton: Karen Ball IMA Unversty of Mnnesota 4 Ln Hall, 7 Church

More information

Chapter 2 Transformations and Expectations. , and define f

Chapter 2 Transformations and Expectations. , and define f Revew for the prevous lecture Defnton: support set of a ranom varable, the monotone functon; Theorem: How to obtan a cf, pf (or pmf) of functons of a ranom varable; Eamples: several eamples Chapter Transformatons

More information

Difference Equations

Difference Equations Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Large-Scale Data-Dependent Kernel Approximation Appendix

Large-Scale Data-Dependent Kernel Approximation Appendix Large-Scale Data-Depenent Kernel Approxmaton Appenx Ths appenx presents the atonal etal an proofs assocate wth the man paper [1]. 1 Introucton Let k : R p R p R be a postve efnte translaton nvarant functon

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

On Liu Estimators for the Logit Regression Model

On Liu Estimators for the Logit Regression Model CESIS Electronc Workng Paper Seres Paper No. 59 On Lu Estmators for the Logt Regresson Moel Krstofer Månsson B. M. Golam Kbra October 011 The Royal Insttute of technology Centre of Excellence for Scence

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

On a one-parameter family of Riordan arrays and the weight distribution of MDS codes

On a one-parameter family of Riordan arrays and the weight distribution of MDS codes On a one-parameter famly of Roran arrays an the weght strbuton of MDS coes Paul Barry School of Scence Waterfor Insttute of Technology Irelan pbarry@wte Patrck Ftzpatrck Department of Mathematcs Unversty

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

A MULTIDIMENSIONAL ANALOGUE OF THE RADEMACHER-GAUSSIAN TAIL COMPARISON

A MULTIDIMENSIONAL ANALOGUE OF THE RADEMACHER-GAUSSIAN TAIL COMPARISON A MULTIDIMENSIONAL ANALOGUE OF THE RADEMACHER-GAUSSIAN TAIL COMPARISON PIOTR NAYAR AND TOMASZ TKOCZ Abstract We prove a menson-free tal comparson between the Euclean norms of sums of nepenent ranom vectors

More information

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede Fall Analyss of Expermental Measurements B. Esensten/rev. S. Erree Hypothess Testng, Lkelhoo Functons an Parameter Estmaton: We conser estmaton of (one or more parameters to be the expermental etermnaton

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Yukawa Potential and the Propagator Term

Yukawa Potential and the Propagator Term PHY304 Partcle Physcs 4 Dr C N Booth Yukawa Potental an the Propagator Term Conser the electrostatc potental about a charge pont partcle Ths s gven by φ = 0, e whch has the soluton φ = Ths escrbes the

More information

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1 On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool

More information

An efficient method for computing single parameter partial expected value of perfect information

An efficient method for computing single parameter partial expected value of perfect information An effcent metho for computng sngle parameter partal expecte value of perfect nformaton Mark Strong,, Jeremy E. Oakley 2. School of Health an Relate Research ScHARR, Unversty of Sheffel, UK. 2. School

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

Finding Dense Subgraphs in G(n, 1/2)

Finding Dense Subgraphs in G(n, 1/2) Fndng Dense Subgraphs n Gn, 1/ Atsh Das Sarma 1, Amt Deshpande, and Rav Kannan 1 Georga Insttute of Technology,atsh@cc.gatech.edu Mcrosoft Research-Bangalore,amtdesh,annan@mcrosoft.com Abstract. Fndng

More information

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14 APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

High-Order Hamilton s Principle and the Hamilton s Principle of High-Order Lagrangian Function

High-Order Hamilton s Principle and the Hamilton s Principle of High-Order Lagrangian Function Commun. Theor. Phys. Bejng, Chna 49 008 pp. 97 30 c Chnese Physcal Socety Vol. 49, No., February 15, 008 Hgh-Orer Hamlton s Prncple an the Hamlton s Prncple of Hgh-Orer Lagrangan Functon ZHAO Hong-Xa an

More information

Visualization of 2D Data By Rational Quadratic Functions

Visualization of 2D Data By Rational Quadratic Functions 7659 Englan UK Journal of Informaton an Computng cence Vol. No. 007 pp. 7-6 Vsualzaton of D Data By Ratonal Quaratc Functons Malk Zawwar Hussan + Nausheen Ayub Msbah Irsha Department of Mathematcs Unversty

More information

WHY NOT USE THE ENTROPY METHOD FOR WEIGHT ESTIMATION?

WHY NOT USE THE ENTROPY METHOD FOR WEIGHT ESTIMATION? ISAHP 001, Berne, Swtzerlan, August -4, 001 WHY NOT USE THE ENTROPY METHOD FOR WEIGHT ESTIMATION? Masaak SHINOHARA, Chkako MIYAKE an Kekch Ohsawa Department of Mathematcal Informaton Engneerng College

More information

2. High dimensional data

2. High dimensional data /8/00. Hgh mensons. Hgh mensonal ata Conser representng a ocument by a vector each component of whch correspons to the number of occurrences of a partcular wor n the ocument. The Englsh language has on

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

The characteristic landscape equation for an AR(2) landscape

The characteristic landscape equation for an AR(2) landscape Apple Mathematcs Letters 19 (2006) 260 266 www.elsever.com/locate/aml The characterstc lanscape equaton for an AR(2) lanscape B. Dmova, J.W. Barnes,E.Popova Grauate Program n Operatons Research an Inustral

More information

An Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation

An Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation An Experment/Some Intuton I have three cons n my pocket, 6.864 (Fall 2006): Lecture 18 The EM Algorthm Con 0 has probablty λ of heads; Con 1 has probablty p 1 of heads; Con 2 has probablty p 2 of heads

More information

= z 20 z n. (k 20) + 4 z k = 4

= z 20 z n. (k 20) + 4 z k = 4 Problem Set #7 solutons 7.2.. (a Fnd the coeffcent of z k n (z + z 5 + z 6 + z 7 + 5, k 20. We use the known seres expanson ( n+l ( z l l z n below: (z + z 5 + z 6 + z 7 + 5 (z 5 ( + z + z 2 + z + 5 5

More information

18.1 Introduction and Recap

18.1 Introduction and Recap CS787: Advanced Algorthms Scrbe: Pryananda Shenoy and Shjn Kong Lecturer: Shuch Chawla Topc: Streamng Algorthmscontnued) Date: 0/26/2007 We contnue talng about streamng algorthms n ths lecture, ncludng

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal

More information

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2) 1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons

More information

Computing MLE Bias Empirically

Computing MLE Bias Empirically Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.

More information

The Noether theorem. Elisabet Edvardsson. Analytical mechanics - FYGB08 January, 2016

The Noether theorem. Elisabet Edvardsson. Analytical mechanics - FYGB08 January, 2016 The Noether theorem Elsabet Evarsson Analytcal mechancs - FYGB08 January, 2016 1 1 Introucton The Noether theorem concerns the connecton between a certan kn of symmetres an conservaton laws n physcs. It

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

On the spectral norm of r-circulant matrices with the Pell and Pell-Lucas numbers

On the spectral norm of r-circulant matrices with the Pell and Pell-Lucas numbers Türkmen and Gökbaş Journal of Inequaltes and Applcatons (06) 06:65 DOI 086/s3660-06-0997-0 R E S E A R C H Open Access On the spectral norm of r-crculant matrces wth the Pell and Pell-Lucas numbers Ramazan

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

Computing Fault-Containment Times of Self-Stabilizing Algorithms Using Lumped Markov Chains

Computing Fault-Containment Times of Self-Stabilizing Algorithms Using Lumped Markov Chains algorthms Artcle Computng Fault-Contanment Tmes of Self-Stablzng Algorthms Usng Lumpe Markov Chans Volker Turau ID Insttute of Telematcs, Hamburg Unversty of Technology, 273 Hamburg, Germany; turau@tuhh.e

More information

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009 College of Computer & Informaton Scence Fall 2009 Northeastern Unversty 20 October 2009 CS7880: Algorthmc Power Tools Scrbe: Jan Wen and Laura Poplawsk Lecture Outlne: Prmal-dual schema Network Desgn:

More information

Module 9. Lecture 6. Duality in Assignment Problems

Module 9. Lecture 6. Duality in Assignment Problems Module 9 1 Lecture 6 Dualty n Assgnment Problems In ths lecture we attempt to answer few other mportant questons posed n earler lecture for (AP) and see how some of them can be explaned through the concept

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Expected Value and Variance

Expected Value and Variance MATH 38 Expected Value and Varance Dr. Neal, WKU We now shall dscuss how to fnd the average and standard devaton of a random varable X. Expected Value Defnton. The expected value (or average value, or

More information

ENGI9496 Lecture Notes Multiport Models in Mechanics

ENGI9496 Lecture Notes Multiport Models in Mechanics ENGI9496 Moellng an Smulaton of Dynamc Systems Mechancs an Mechansms ENGI9496 Lecture Notes Multport Moels n Mechancs (New text Secton 4..3; Secton 9.1 generalzes to 3D moton) Defntons Generalze coornates

More information

A MULTIDIMENSIONAL ANALOGUE OF THE RADEMACHER-GAUSSIAN TAIL COMPARISON

A MULTIDIMENSIONAL ANALOGUE OF THE RADEMACHER-GAUSSIAN TAIL COMPARISON A MULTIDIMENSIONAL ANALOGUE OF THE RADEMACHER-GAUSSIAN TAIL COMPARISON PIOTR NAYAR AND TOMASZ TKOCZ Abstract We prove a menson-free tal comparson between the Euclean norms of sums of nepenent ranom vectors

More information

Turing Machines (intro)

Turing Machines (intro) CHAPTER 3 The Church-Turng Thess Contents Turng Machnes defntons, examples, Turng-recognzable and Turng-decdable languages Varants of Turng Machne Multtape Turng machnes, non-determnstc Turng Machnes,

More information

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

Case A. P k = Ni ( 2L i k 1 ) + (# big cells) 10d 2 P k.

Case A. P k = Ni ( 2L i k 1 ) + (# big cells) 10d 2 P k. THE CELLULAR METHOD In ths lecture, we ntroduce the cellular method as an approach to ncdence geometry theorems lke the Szemeréd-Trotter theorem. The method was ntroduced n the paper Combnatoral complexty

More information

Statistical pattern recognition

Statistical pattern recognition Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve

More information

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there

More information

Hidden Markov Models

Hidden Markov Models CM229S: Machne Learnng for Bonformatcs Lecture 12-05/05/2016 Hdden Markov Models Lecturer: Srram Sankararaman Scrbe: Akshay Dattatray Shnde Edted by: TBD 1 Introducton For a drected graph G we can wrte

More information

Homological methods in Non-commutative Geometry Tokyo, 2007/2008 1

Homological methods in Non-commutative Geometry Tokyo, 2007/2008 1 Homologcal methos n Non-commutatve Geometry Tokyo, 2007/2008 1 Lecture 2. Secon bcomplex for cyclc homology. Connes fferental. Cyclc homology an the e Rham cohomology n the HKR case. Homology of small

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

The EM Algorithm (Dempster, Laird, Rubin 1977) The missing data or incomplete data setting: ODL(φ;Y ) = [Y;φ] = [Y X,φ][X φ] = X

The EM Algorithm (Dempster, Laird, Rubin 1977) The missing data or incomplete data setting: ODL(φ;Y ) = [Y;φ] = [Y X,φ][X φ] = X The EM Algorthm (Dempster, Lard, Rubn 1977 The mssng data or ncomplete data settng: An Observed Data Lkelhood (ODL that s a mxture or ntegral of Complete Data Lkelhoods (CDL. (1a ODL(;Y = [Y;] = [Y,][

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

Computing Correlated Equilibria in Multi-Player Games

Computing Correlated Equilibria in Multi-Player Games Computng Correlated Equlbra n Mult-Player Games Chrstos H. Papadmtrou Presented by Zhanxang Huang December 7th, 2005 1 The Author Dr. Chrstos H. Papadmtrou CS professor at UC Berkley (taught at Harvard,

More information

Design and Analysis of Algorithms

Design and Analysis of Algorithms Desgn and Analyss of Algorthms CSE 53 Lecture 4 Dynamc Programmng Junzhou Huang, Ph.D. Department of Computer Scence and Engneerng CSE53 Desgn and Analyss of Algorthms The General Dynamc Programmng Technque

More information

The Second Anti-Mathima on Game Theory

The Second Anti-Mathima on Game Theory The Second Ant-Mathma on Game Theory Ath. Kehagas December 1 2006 1 Introducton In ths note we wll examne the noton of game equlbrum for three types of games 1. 2-player 2-acton zero-sum games 2. 2-player

More information

Learning Theory: Lecture Notes

Learning Theory: Lecture Notes Learnng Theory: Lecture Notes Lecturer: Kamalka Chaudhur Scrbe: Qush Wang October 27, 2012 1 The Agnostc PAC Model Recall that one of the constrants of the PAC model s that the data dstrbuton has to be

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Solutions Homework 4 March 5, 2018

Solutions Homework 4 March 5, 2018 1 Solutons Homework 4 March 5, 018 Soluton to Exercse 5.1.8: Let a IR be a translaton and c > 0 be a re-scalng. ˆb1 (cx + a) cx n + a (cx 1 + a) c x n x 1 cˆb 1 (x), whch shows ˆb 1 s locaton nvarant and

More information

Information Geometry of Gibbs Sampler

Information Geometry of Gibbs Sampler Informaton Geometry of Gbbs Sampler Kazuya Takabatake Neuroscence Research Insttute AIST Central 2, Umezono 1-1-1, Tsukuba JAPAN 305-8568 k.takabatake@ast.go.jp Abstract: - Ths paper shows some nformaton

More information

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES SVANTE JANSON Abstract. We gve explct bounds for the tal probabltes for sums of ndependent geometrc or exponental varables, possbly wth dfferent

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecture 3 Contnuous Systems an Fels (Chapter 3) Where Are We Now? We ve fnshe all the essentals Fnal wll cover Lectures through Last two lectures: Classcal Fel Theory Start wth wave equatons

More information

Search sequence databases 2 10/25/2016

Search sequence databases 2 10/25/2016 Search sequence databases 2 10/25/2016 The BLAST algorthms Ø BLAST fnds local matches between two sequences, called hgh scorng segment pars (HSPs). Step 1: Break down the query sequence and the database

More information

Summary. Introduction

Summary. Introduction Sesmc reflecton stuy n flu-saturate reservor usng asymptotc Bot s theory Yangun (Kevn) Lu* an Gennay Goloshubn Unversty of Houston Dmtry Sln Lawrence Bereley Natonal Laboratory Summary It s well nown that

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.65/15.070J Fall 013 Lecture 1 10/1/013 Martngale Concentraton Inequaltes and Applcatons Content. 1. Exponental concentraton for martngales wth bounded ncrements.

More information

Chapter 7 Clustering Analysis (1)

Chapter 7 Clustering Analysis (1) Chater 7 Clusterng Analyss () Outlne Cluster Analyss Parttonng Clusterng Herarchcal Clusterng Large Sze Data Clusterng What s Cluster Analyss? Cluster: A collecton of ata obects smlar (or relate) to one

More information

CHAPTER 14 GENERAL PERTURBATION THEORY

CHAPTER 14 GENERAL PERTURBATION THEORY CHAPTER 4 GENERAL PERTURBATION THEORY 4 Introducton A partcle n orbt around a pont mass or a sphercally symmetrc mass dstrbuton s movng n a gravtatonal potental of the form GM / r In ths potental t moves

More information

CHAPTER 4 HYDROTHERMAL COORDINATION OF UNITS CONSIDERING PROHIBITED OPERATING ZONES A HYBRID PSO(C)-SA-EP-TPSO APPROACH

CHAPTER 4 HYDROTHERMAL COORDINATION OF UNITS CONSIDERING PROHIBITED OPERATING ZONES A HYBRID PSO(C)-SA-EP-TPSO APPROACH 77 CHAPTER 4 HYDROTHERMAL COORDINATION OF UNITS CONSIDERING PROHIBITED OPERATING ZONES A HYBRID PSO(C)-SA-EP-TPSO APPROACH 4.1 INTRODUCTION HTC consttutes the complete formulaton of the hyrothermal electrc

More information

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore 8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø

More information

Perron Vectors of an Irreducible Nonnegative Interval Matrix

Perron Vectors of an Irreducible Nonnegative Interval Matrix Perron Vectors of an Irreducble Nonnegatve Interval Matrx Jr Rohn August 4 2005 Abstract As s well known an rreducble nonnegatve matrx possesses a unquely determned Perron vector. As the man result of

More information

Bézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0

Bézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0 Bézer curves Mchael S. Floater September 1, 215 These notes provde an ntroducton to Bézer curves. 1 Bernsten polynomals Recall that a real polynomal of a real varable x R, wth degree n, s a functon of

More information

Algorithms for Molecular Biology

Algorithms for Molecular Biology Algorthms for Molecular Bology BoMed Central Research Effectve p-value computatons usng Fnte Markov Chan Imbeddng (FMCI): applcaton to local score and to pattern statstcs Grégory Nuel* Open Access Address:

More information

Power law and dimension of the maximum value for belief distribution with the max Deng entropy

Power law and dimension of the maximum value for belief distribution with the max Deng entropy Power law and dmenson of the maxmum value for belef dstrbuton wth the max Deng entropy Bngy Kang a, a College of Informaton Engneerng, Northwest A&F Unversty, Yanglng, Shaanx, 712100, Chna. Abstract Deng

More information

Text S1: Detailed proofs for The time scale of evolutionary innovation

Text S1: Detailed proofs for The time scale of evolutionary innovation Text S: Detaled proofs for The tme scale of evolutonary nnovaton Krshnendu Chatterjee Andreas Pavloganns Ben Adlam Martn A. Nowak. Overvew and Organzaton We wll present detaled proofs of all our results.

More information

Chapter 6. Supplemental Text Material

Chapter 6. Supplemental Text Material Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.

More information

Lecture Space-Bounded Derandomization

Lecture Space-Bounded Derandomization Notes on Complexty Theory Last updated: October, 2008 Jonathan Katz Lecture Space-Bounded Derandomzaton 1 Space-Bounded Derandomzaton We now dscuss derandomzaton of space-bounded algorthms. Here non-trval

More information

Hidden Markov Models

Hidden Markov Models Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS

A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS HCMC Unversty of Pedagogy Thong Nguyen Huu et al. A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS Thong Nguyen Huu and Hao Tran Van Department of mathematcs-nformaton,

More information

Bounds for Spectral Radius of Various Matrices Associated With Graphs

Bounds for Spectral Radius of Various Matrices Associated With Graphs 45 5 Vol.45, No.5 016 9 AVANCES IN MATHEMATICS (CHINA) Sep., 016 o: 10.11845/sxjz.015015b Bouns for Spectral Raus of Varous Matrces Assocate Wth Graphs CUI Shuyu 1, TIAN Guxan, (1. Xngzh College, Zhejang

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Beyond Zudilin s Conjectured q-analog of Schmidt s problem

Beyond Zudilin s Conjectured q-analog of Schmidt s problem Beyond Zudln s Conectured q-analog of Schmdt s problem Thotsaporn Ae Thanatpanonda thotsaporn@gmalcom Mathematcs Subect Classfcaton: 11B65 33B99 Abstract Usng the methodology of (rgorous expermental mathematcs

More information

A GENERALIZATION OF JUNG S THEOREM. M. Henk

A GENERALIZATION OF JUNG S THEOREM. M. Henk A GENERALIZATION OF JUNG S THEOREM M. Henk Abstract. The theorem of Jung establshes a relaton between crcumraus an ameter of a convex boy. The half of the ameter can be nterprete as the maxmum of crcumra

More information

Using the estimated penetrances to determine the range of the underlying genetic model in casecontrol

Using the estimated penetrances to determine the range of the underlying genetic model in casecontrol Georgetown Unversty From the SelectedWorks of Mark J Meyer 8 Usng the estmated penetrances to determne the range of the underlyng genetc model n casecontrol desgn Mark J Meyer Neal Jeffres Gang Zheng Avalable

More information

ECE559VV Project Report

ECE559VV Project Report ECE559VV Project Report (Supplementary Notes Loc Xuan Bu I. MAX SUM-RATE SCHEDULING: THE UPLINK CASE We have seen (n the presentaton that, for downlnk (broadcast channels, the strategy maxmzng the sum-rate

More information

Société de Calcul Mathématique SA

Société de Calcul Mathématique SA Socété de Calcul Mathématque SA Outls d'ade à la décson Tools for decson help Probablstc Studes: Normalzng the Hstograms Bernard Beauzamy December, 202 I. General constructon of the hstogram Any probablstc

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Expectation Maximization Mixture Models HMMs

Expectation Maximization Mixture Models HMMs -755 Machne Learnng for Sgnal Processng Mture Models HMMs Class 9. 2 Sep 200 Learnng Dstrbutons for Data Problem: Gven a collecton of eamples from some data, estmate ts dstrbuton Basc deas of Mamum Lelhood

More information

LECTURE 9 CANONICAL CORRELATION ANALYSIS

LECTURE 9 CANONICAL CORRELATION ANALYSIS LECURE 9 CANONICAL CORRELAION ANALYSIS Introducton he concept of canoncal correlaton arses when we want to quantfy the assocatons between two sets of varables. For example, suppose that the frst set of

More information