Analysis of Noisy Evolutionary Optimization When Sampling Fails

Size: px
Start display at page:

Download "Analysis of Noisy Evolutionary Optimization When Sampling Fails"

Transcription

1 Aalyss of Nosy Evolutoary Optmzato Whe Samplg Fals Chao Qa, Member, IEEE, Chao Ba, Studet Member, IEEE, Yag Yu, Member, IEEE, Ke Tag, Seor Member, IEEE, X Yao, Fellow, IEEE arxv: v [cs.ne] Oct 08 Abstract I osy evolutoary optmzato, samplg s a commo strategy to deal wth ose. By the samplg strategy, the ftess of a soluto s evaluated multple tmes called sample sze depedetly, ad ts true ftess s the approxmated by the average of these evaluatos. Prevous studes o samplg are maly emprcal. I ths paper, we frst vestgate the effect of sample sze from a theoretcal perspectve. By aalyzg the +-EA o the osy LeadgOes problem, we show that as the sample sze creases, the rug tme ca reduce from expoetal to polyomal, but the retur to expoetal. Ths suggests that a proper sample sze s crucal practce. The, we vestgate what strateges ca work whe samplg wth ay fxed sample sze fals. By two llustratve examples, we prove that usg paret or offsprg populatos ca be better. Fally, we costruct a artfcal osy example to show that whe usg ether samplg or populatos s effectve, adaptve samplg.e., samplg wth a adaptve sample sze ca work. Ths, for the frst tme, provdes a theoretcal support for the use of adaptve samplg. Idex Terms Nosy optmzato, evolutoary algorthms, samplg, populato, rug tme aalyss. I. INTRODUCTION Evolutoary algorthms EAs are a type of geeral-purpose radomzed optmzato algorthms, spred by atural evoluto. They have bee wdely appled to solve real-world optmzato problems, whch are ofte subject to ose. Samplg s a popular strategy for dealg wth ose: to estmate the ftess of a soluto, t evaluates the ftess multple m tmes called sample sze depedetly ad the uses the sample average to approxmate the true ftess. Samplg reduces the varace of ose by a factor of m, but also creases the computato tme for the ftess estmato of a soluto by m tmes. Prevous studes maly focused o the emprcal desg of effcet samplg methods, e.g., adaptve samplg [4], [5], whch dyamcally decdes the sample sze m for each soluto each geerato. The theoretcal aalyss o samplg was rarely touched. Due to ther sophstcated behavors of mmckg atural pheomea, the theoretcal aalyss of EAs s dffcult. Much effort thus has bee devoted to uderstadg the behavor C. Qa ad C. Ba are wth the School of Computer Scece ad Techology, Uversty of Scece ad Techology of Cha, Hefe, Cha e-mals: chaoqa@ustc.edu.c, bacht@mal.ustc.edu.c Y. Yu s wth the Natoal Key Laboratory for Novel Software Techology, Najg Uversty, Najg, Cha e-mal: yuy@ju.edu.c K. Tag ad X. Yao are wth the Departmet of Computer Scece ad Egeerg, Souther Uversty of Scece ad Techology, Shezhe, Cha e-mals: tagk3, xy}@sustc.edu.c Ths work s exteded from [5]. of EAs from a theoretcal vewpot [], [7], but most of them focus o ose-free optmzato. The presece of ose further creases the radomess of optmzato, ad thus also creases the dffculty of aalyss. For rug tme aalyss oe essetal theoretcal aspect osy evolutoary optmzato, oly a few results have bee reported. The classc +-EA algorthm was frst studed o the OeMax ad LeadgOes problems uder varous ose models [3], [7], [0], [4], [], [7]. The results showed that the +-EA s effcet oly uder low ose levels, e.g., for the +-EA solvg OeMax the presece of oebt ose, the maxmal ose level of allowg a polyomal rug tme s Olog/, where the ose level s characterzed by the ose probablty p [0,] ad s the problem sze. Later studes maly proved the robustess of dfferet strateges to ose, cludg usg populatos [6], [7], [4], [], [7], samplg [], [3] ad threshold selecto [4]. For example, the µ+-ea wth µ log5 [4], the +λ- EA wth λ 4log [4], the +-EA usg samplg wth m = 3 [3] or the +-EA usg threshold selecto wth threshold τ = [4] ca solve OeMax polyomal tme eve f the probablty of oe-bt ose reaches. Note that there was also a sequece of papers aalyzg the rug tme of the compact geetc algorthm [3] ad at coloy optmzato algorthms [8], [], [], [6] solvg osy problems, cludg OeMax as well as the combatoral optmzato problem sgle destato shortest paths. The very few rug tme aalyses volvg samplg [], [3] maly showed the effectveess of samplg wth a large eough fxed sample sze m. For example, for the +-EA solvg OeMax uder oe-bt ose wth p = ωlog/, usg samplg wth m = 4 3 ca reduce the rug tme expoetally. I addto, Akmoto et al. [] proved that usg samplg wth a large eough m ca make optmzato uder addtve ubased ose behave as oseless optmzato. However, there are stll may fudametal theoretcal ssues that have ot bee addressed, e.g., how the sample sze ca affect the effectveess of samplg, ad what strateges ca work whe samplg fals. I ths paper, we frst theoretcally vestgate the effect of sample sze. It may be beleved that oce the sample sze m reaches a effectve value, the rug tme wll always be polyomal as m cotues to crease. We gve a couterexample,.e., the +-EA solvg LeadgOes uder oe-bt ose wth p =. Qa et al. [] have show that the rug tme wll reduce from expoetal to polyomal whe m = 4 4 log/5. We prove that the rug tme wll retur

2 Algorthm +-EA Gve a pseudo-boolea fucto f to be maxmzed, the procedure of the +-EA: : Let x be a uformly chose soluto from 0,}. : Repeat utl some termato codto s met 3: x := flp each bt of x depedetly wth prob. /. 4: f fx fx the x := x. to expoetal whe m 5. Our aalyss suggests that the selecto of sample sze should be careful practce. The, we theoretcally compare the two strateges of usg populatos ad samplg o the robustess to ose. Prevous studes have show that both of them are effectve for solvg OeMax uder oe-bt ose [4], [], [3], whle usg samplg s better for solvg OeMax uder addtve Gaussa ose [3]. Here, we complemet ths comparso by costructg two specfc osy OeMax problems. For oe of them, usg paret populatos s better tha usg samplg, whle for the other, usg offsprg populatos s better. I both cases, we prove that the employed paret ad offsprg populato szes are almost tght. We also gve a artfcal osy OeMax problem where usg ether populatos or samplg s effectve. For ths case, we further prove that usg adaptve samplg ca reduce the rug tme expoetally, whch provdes some theoretcal justfcato for the good emprcal performace of adaptve samplg [8], [3]. Ths paper exteds our prelmary work [5]. Whe comparg samplg wth populatos, we oly cosdered paret populatos [5]. To get a complete uderstadg, we add the aalyss of usg offsprg populatos. We costruct a ew osy example to show that usg offsprg populatos ca be better tha usg samplg.e., Theorems 9 ad 0 Secto V. For the osy example Secto VI, where we prevously proved that usg ether samplg or paret populatos s effectve whle adaptve samplg ca work, we ow prove that usg offsprg populatos s also effectve.e., Theorem 4 Secto VI. To show that usg paret populatos s better tha usg samplg, we oly gave a effectve paret populato sze [5]. We ow add the aalyss of the tghtess of the effectve paret populato sze.e., Theorem 8 Secto IV as well as the effectve offsprg populato sze.e., Theorem Secto V. The rest of ths paper s orgazed as follows. Secto II troduces some prelmares. Secto III aalyzes the effect of sample sze. The effectveess of usg paret ad offsprg populatos whe samplg fals s proved Sectos IV ad V, respectvely. Secto VI the shows that whe usg ether samplg or populatos s effectve, adaptve samplg ca work. Fally, Secto VII cocludes the paper. II. PRELIMINARIES I ths secto, we frst troduce the EAs ad the samplg strategy, ad the preset the aalyss tools that wll be used ths paper. Algorthm µ+-ea Gve a pseudo-boolea fucto f to be maxmzed, the procedure of the µ+-ea: : Let P be a set of µ uformly chose solutos from 0,}. : Repeat utl some termato codto s met 3: x := uformly selected from P at radom. 4: x := flp each bt of x depedetly wth prob. /. 5: Let z argm z P fz; tes are broke radomly. 6: f fx fz the P := P \z} x }. Algorthm 3 +λ-ea Gve a pseudo-boolea fucto f to be maxmzed, the procedure of the +λ-ea: : Let x be a uformly chose soluto from 0,}. : Repeat utl some termato codto s met 3: Let Q :=. 4: for = to λ do 5: x := flp each bt of x depedetly wth prob./. 6: Q := Q x }. 7: Let z argmax z Q fz; tes are broke radomly. 8: f fz fx the x := z. A. Evolutoary Algorthms The +-EA.e., Algorthm matas oly oe soluto, ad teratvely tres to produce oe better soluto by btwse mutato ad selecto. The µ+-ea.e., Algorthm uses a paret populato sze µ. I each terato, t also geerates oe ew soluto x, ad the uses x to replace the worst soluto the populato P f x s ot worse. The +λ-ea.e., Algorthm 3 uses a offsprg populato sze λ. I each terato, t geerates λ offsprg solutos depedetly by mutatg the paret solutox, ad the uses the best offsprg soluto to replace the paret soluto f t s ot worse. Whe µ = ad λ =, both the µ+-ea ad +λ-ea degeerate to the +-EA. Note that for the µ+-ea, a slghtly dfferet updatg rule s also used [3], [30]: x s smply added to P ad the the worst soluto P x } s deleted. Our results about the µ+-ea derved the paper also apply to ths settg. I osy optmzato, oly a osy ftess value f x stead of the exact oe fx ca be accessed. Note that our aalyss, the algorthms are assumed to use the reevaluato strategy as [8], [0], [4]. That s, besdes evaluatg the osy ftess f x of offsprg solutos, the osy ftess values of paret solutos wll be reevaluated each terato. The rug tme of EAs s usually measured by the umber of ftess evaluatos utl fdg a optmal soluto w.r.t. the true ftess fucto f for the frst tme [], [0], [4]. B. Samplg Samplg as descrbed Defto s a commo strategy to deal wth ose. It approxmates the true ftess fx usg the average of a umber of radom evaluatos. The umber

3 3 m of radom evaluatos s called the sample sze. Note that m = mples that samplg s ot used. Qa et al. [], [3] have theoretcally show the robustess of samplg to ose. Partcularly, they proved that by usg samplg wth some fxed sample sze, the rug tme of the +-EA for solvg OeMax ad LeadgOes uder ose ca reduce from expoetal to polyomal. Defto Samplg. Samplg frst evaluates the ftess of a soluto m tmes depedetly ad obtas the osy ftess values f x,f x,...,f mx, ad the outputs ther average,.e., ˆfx = m = f x/m. Adaptve samplg dyamcally decdes the sample sze for each soluto the optmzato process, stead of usg a fxed sze. For example, oe popular strategy [4], [5] s to frst estmate the ftess of two solutos by a small umber of samples, ad the sequetally crease samples utl the dfferece ca be sgfcatly dscrmated. It has bee foud well useful may applcatos [8], [3], whle there has bee o theoretcal work supportg ts effectveess. C. Aalyss Tools EAs ofte geerate offsprg solutos oly based o the curret populato, thus, a EA ca be modeled as a Markov cha ξ t } + t=0 e.g., [6], [3] by takg the EA s populato space X as the cha s state space.e., ξ t X ad takg the set X of all optmal populatos as the cha s target state space. Note that the populato space X cossts of all possble populatos, ad a optmal populato cotas at least oe optmal soluto. Gve a Markov cha ξ t } + t=0 ad ξˆt, we defe ts frst httg tme as τ = mt ξˆt+t X,t 0}. The expectato of τ, Eτ ξˆt = + =0 Pτ = ξˆt, s called the expected frst httg tme EFHT. If ξ 0 s draw from a dstrbuto π 0, Eτ ξ 0 π 0 = ξ π 0 X 0ξ 0 Eτ ξ 0 s called the EFHT of the cha over the tal dstrbuto π 0. Thus, the expected rug tme of the µ+-ea startg from ξ 0 π 0 s µ+µ+ Eτ ξ 0 π 0, where the frst µ s the cost of evaluatg the tal populato, ad µ+ s the cost of oe terato, where t eeds to evaluate the offsprg solutox ad reevaluate the µ paret solutos. Smlarly, the expected rug tme of the +λ-ea startg from ξ 0 π 0 s ++λ Eτ ξ 0 π 0, where the frst s the cost of evaluatg the tal soluto, ad +λ s the cost of oe terato, where t eeds to evaluate the λ offsprg solutos ad reevaluate the paret soluto. For the +-EA, the expected rug tme s calculated by settg µ = or λ =,.e., + Eτ ξ 0 π 0. For the +-EA wth samplg, t becomes m + m Eτ ξ 0 π 0, because the ftess estmato of a soluto eeds m depedet evaluatos. Note that ths paper, we cosder the expected rug tme of a EA startg from a uform tal dstrbuto. The, we troduce several drft theorems whch wll be used to aalyze the EFHT of Markov chas ths paper. The multplcatve drft theorem.e., Theorem [9] s for dervg upper bouds o the EFHT. Frst, a dstace fucto Vx satsfyg that Vx X = 0 ad Vx / X > 0 eeds to be desged to measure the dstace of a state x to the target state space X. The, we eed to aalyze the drft towards X each step,.e., EVξ t Vξ t+ ξ t. If the drft each step s roughly proportoal to the curret dstace to the optmum, we ca derve a upper boud o the EFHT accordgly. Theorem Multplcatve Drft [9]. Gve a Markov cha ξ t } + t=0 ad a dstace fucto V over X, f for ay t 0 ad ay ξ t wth Vξ t > 0, there exsts c > 0 such that EVξ t Vξ t+ ξ t c Vξ t, the t holds that Eτ ξ 0 +logvξ0/vm c, where V m deotes the mmum amog all possble postve values of V. The smplfed egatve drft theorem.e., Theorem [8], [9] s for provg expoetal lower bouds o the EFHT of Markov chas, where X t s ofte represeted by a mappg of ξ t. From Theorem, we ca see that two codtos are requred: a costat egatve drft ad expoetally decayg probabltes of jumpg towards or away from the target state. By buldg a relatoshp betwee the jumpg dstace ad the legth of the drft terval, a more geeral theorem smplfed egatve drft wth scalg [0] as preseted Theorem 3 has bee proposed. Theorem 4 gves the orgal egatve drft theorem [5], whch s stroger because both the two smplfed versos are proved by usg ths orgal theorem. Theorem Smplfed Negatve Drft [8], [9]. Let X t, t 0, be real-valued radom varables descrbg a stochastc process over some state space. Suppose there exsts a terval [a,b] R, two costats δ,ǫ > 0 ad, possbly depedg o l := b a, a fucto rl satsfyg rl = ol/logl such that for all t 0: EX t X t+ a < X t < b ǫ, j N + : P X t+ X t j X t > a rl +δ j. The there exsts a costatc > 0 such that for T := mt 0 : X t a X 0 b} t holds PT cl/rl = Ωl/rl. Theorem 3 Smplfed Negatve Drft wth Scalg [0]. Let X t, t 0, be real-valued radom varables descrbg a stochastc process over some state space. Suppose there exsts a terval [a,b] R ad, possbly depedg o l := b a, a drft boud ǫ := ǫl > 0 as well as a scalg factor r := rl such that for all t 0: EX t X t+ a < X t < b ǫ, j N + : P X t+ X t jr X t > a e j, 3 r mǫ l, ǫl/3logǫl}. The t holds for the frst httg tme T := mt 0 : X t a X 0 b} that PT e ǫl/3r = Oe ǫl/3r. Theorem 4 Negatve Drft [5]. Let X t,t 0, be realvalued radom varables descrbg a stochastc process over some state space. Pck two real umbers al ad bl depedg o a parameter l such that al < bl holds. Let Tl be the radom varable deotg the earlest tme t 0 such that

4 4 X t al holds. Suppose there exsts λl > 0 ad pl > 0 such that for all t 0: E e λl Xt+ Xt al < X t < bl pl. The t holds that for all tme bouds Ll 0, PTl Ll X 0 bl e λl bl al Ll Dl pl, where Dl = max,e e λl Xt+ bl X t bl }. III. THE EFFECT OF SAMPLE SIZE Prevous studes [], [3] have show that for osy evolutoary optmzato, samplg wth some fxed sample sze m ca decrease the rug tme expoetally some stuatos. For example, for the +-EA solvg the OeMax problem uder oe-bt ose wth the ose probabltyp = ωlog/, the expected rug tme s super-polyomal [0]; whle by usg samplg wth m = 4 3, the rug tme reduces to polyomal []. The, a atural questo s that whether the rug tme wll always be polyomal by usg ay polyomally bouded sample sze larger tha the effectve m. It may be beleved that the aswer s yes, sce the sample sze m has bee effectve ad usg a larger sample sze wll make the ftess estmato more accurate. For example, for the +-EA solvg OeMax uder oe-bt ose, t s easy to see from Lemma 3 [] that usg a larger sample sze tha 4 3 wll make the probablty of acceptg a true worse soluto the comparso cotue to decrease ad the rug tme wll obvously stay polyomal. I ths secto, we gve a couterexample by cosderg the +- EA solvg the LeadgOes problem uder oe-bt ose, whch suggests that the selecto of sample sze should be careful practce. As preseted Defto, the goal of the LeadgOes problem s to maxmze the umber of cosecutve -bts coutg from the left of a soluto. We ca easly see that the optmal soluto s the strg wth all s deoted as. As preseted Defto 3, the oe-bt ose model flps a radom bt of a soluto before evaluato wth probablty p. Whe p =, t was kow [] that the expected rug tme of the +-EA s expoetal, whle the rug tme wll reduce to polyomal by usg samplg wthm = 4 4 log/5. We prove Theorem 5 that the rug tme of the +-EA wll retur to expoetal f m 5. Defto LeadgOes. The LeadgOes Problem s to fd a bary strg x 0,} that maxmses fx = = j= x j. Defto 3 Oe-bt Nose. Gve a parameter p [0, ], let f x ad fx deote the osy ad true ftess of a soluto x 0,}, respectvely, the f fx wth prob. p, x = fx wth prob. p, where x s geerated by flppg a radomly chose bt of x. From Lemma 6 [], we ca fd the reaso why samplg s effectve oly wth a moderate sample sze. I most cases, f fx > fy, the expected gap betwee f x adf y s postve, whch mples that a larger sample sze s better sce t wll decrease Pˆfx ˆfy. However, whe x = ad y s close to the optmum, the expectato of f f y ca be egatve, whch mples that a larger sample sze s worse sce t wll crease Pˆf ˆfy. Thus, ether a small sample sze or a large sample sze s effectve. The sample sze of m = 4 4 log/5 just makes a good tradeoff, whch ca lead to a ot too large probablty of ˆf ˆfy ad a suffcetly small probablty of ˆfx ˆfy for two solutos x ad y wth fx > fy ad Ef x f y > 0. Theorem 5. For the +-EA solvg LeadgOes uder oe-bt ose wth p =, the expected rug tme s expoetal []; f usg samplg wth m = 4 4 log/5, the expected rug tme s polyomal []; f usg samplg wth m 5, the expected rug tme s expoetal. Proof. We oly eed to prove the case m 5. Our ma dea s to show that before reachg the optmal soluto, the algorthm wll frst fd the soluto 0 or 0 wth a probablty of at least + ; whle the probablty of leavg 0 or 0 s expoetally small. Combg these two pots, the theorem holds. Let a Markov cha ξ t } + t=0 model the aalyzed evolutoary process. Let LOx deote the true umber of leadg -bts of a soluto x. For ay t, let C t deote the evet that at tme t, the +-EA fds a soluto wth at least leadg -bts for the frst tme,.e., LOξ t ad t < t : LOξ t < ; let A t ad B t deote the subsets of C t, whch requre that ξ t 0, 0} ad ξ t, 0 }, respectvely. Thus, before reachg the optmal soluto, the +-EA ca fd a soluto 0, 0} wth probablty at least + t= PA t C t PC t. We the show that PA t C t /+. Assume that ξ t = x, where LOx <. Let P mut x,y deote the probablty that x s mutated to y by bt-wse mutato. The, PA t C t = P mut x, 0 Pˆf 0 ˆfx 3 +P mut x, 0 Pˆf 0 ˆfx /PC t. For Pˆf 0 ˆfx ad Pˆf 0 ˆfx, we apply Hoeffdg s equalty to get a lower boud e /. By the defto of oe-bt ose, we get, for 0 k, Ef k 0 k = k j= j + k + k. The, we have, for k, Ef k 0 k Ef k 0 k k =. 4 Thus, for k 3, Ef 0 Ef k 0 k Ef 0 Ef 3 0 = /. Sce LOx 3 ad Ef x Ef LOx 0 LOx, we have Ef 0 Ef x /.

5 5 Let r = Eˆfx ˆf 0. Sce the ˆf value by samplg s the average of m depedet evaluatos, r = Ef x Ef 0 /. The, we have Pˆfx ˆf 0 = Pˆfx ˆf 0 r r exp m r m e /, where the frst equalty s by Hoeffdg s equalty ad f x f 0, ad the last s by r / ad m 5. It s easy to see from Eq. 4 that Ef 0 = Ef 0. Thus, we ca smlarly get 5 Pˆfx ˆf 0 e /. 6 By applyg Eqs. 5 ad 6 to Eq. 3, we get PA t C t e / Pmutx, 0+P mut x, 0. PC t Sce PB t C t P mut x, +P mut x, 0 /PC t, PA t C t PB t C t e / Pmutx, 0+P mut x, 0 P mut x, +P mut x, 0. If x = x = 0 or x = x =, PA t C t PB t C t e / + +. If x +x =, we ca smlarly derve that PAt Ct PB t C t. Sce PA t C t + PB t C t =, our clam that PA t C t /+ holds. Thus, the probablty that the +-EA frst fds a soluto 0, 0} before reachg the optmum s at least + t= PA t C t PC t + + PC t t= = + PLOξ 0 < = +, where the frst equalty s because the uo of the evets C t wth t mples that the tme of fdg a soluto wth at least leadg -bts s at least, whch s equvalet to that the tal soluto ξ 0 has less tha leadg -bts; ad the last equalty s due to the uform tal dstrbuto. We the show that after fdg 0 or 0, the probablty of the +-EA leavg ths state each terato s expoetally small. From Eqs. 5 ad 6, we kow that for ay x wth LOx < ad y 0, 0}, Pˆfx ˆfy e /. For x 0, } ad y 0, 0}, t s easy to verfy that Ef y f x = j= j + j= j =. Usg the same aalyss as Eq. 5, we ca get, for x 0, } ad y 0, 0}, Pˆfx ˆfy e /. Combg the above two cases, we get, for x / 0, 0} ad y 0, 0}, Pˆfx ˆfy e /. Thus, our clam that the probablty of leavg 0, 0} each step s expoetally small holds. IV. PARENT POPULATIONS CAN WORK ON SOME TASKS WHERE SAMPLING FAILS Prevous studes [4], [], [3] have show that both usg populatos ad samplg ca brg robustess to ose. For example, for the OeMax problem uder oe-bt ose wth p = ωlog /, the +-EA eeds expoetal tme to fd the optmum [0], whle usg a paret populato sze µ log5/p [4], a offsprg populato sze λ max/p,4}log [4] or a sample sze m = 4 3 [] ca all reduce the rug tme to polyomal. The, a atural questo s that whether there exst cases where oly oe of these two strateges.e., populatos ad samplg s effectve. Ths questo has bee partally addressed. For the OeMax problem uder addtve Gaussa ose wth large varaces, t was show that the µ+-ea wth µ = ω eeds super-polyomal tme to fd the optmum [3], whle the +-EA usg samplg ca fd the optmum polyomal tme [3]. Now, we try to solve the other part of ths questo. That s, we are to prove that usg populatos ca be better tha usg samplg. I ths secto, we show that compared wth usg samplg, usg paret populatos ca be more robust to ose. Partcularly, we compare the +-EA usg samplg wth the µ+-ea for solvg OeMax uder symmetrc ose. As preseted Defto 4, the goal of the OeMax problem s to maxmze the umber of -bts, ad the optmal soluto s. As preseted Defto 5, symmetrc ose returs a false ftess fx wth probablty /. It s easy to see that uder ths ose model, the dstrbuto of f x for ay x s symmetrc about. Defto 4 OeMax. The OeMax Problem s to fd a bary strg x 0,} that maxmses fx = = x. Defto 5 Symmetrc Nose. Let f x ad fx deote the osy ad true ftess of a soluto x, respectvely, the f fx wth prob. /, x = fx wth prob. /. We prove Theorem 6 that the expected rug tme of the +-EA usg samplg wth ay sample sze m s expoetal. From the proof, we ca fd the reaso why usg samplg fals. Uder symmetrc ose, the dstrbuto of f x for ay x s symmetrc about. Thus, for ay two solutos x ad y, the dstrbuto of f x f y s symmetrc about 0. By samplg, the dstrbuto of ˆfx ˆfy s stll symmetrc about 0, whch mples that the offsprg soluto wll always be accepted wth probablty at least / each terato of the +-EA. Such a behavor s aalogous to radom walk, ad thus the optmzato s effcet. Theorem 6. For the +-EA solvg OeMax uder symmetrc ose, f usg samplg, the expected rug tme s expoetal. Proof. We apply the smplfed egatve drft theorem.e., Theorem to prove t. Let X t = x 0 deote the umber

6 6 of 0-bts of the soluto x mataed by the +-EA after rug t teratos. We cosder the terval [0, /0],.e., the parameters a = 0 ad b = /0 Theorem. We the aalyze EX t X t+ X t = for < /0. The drft s dvded to two parts: E + ad E. That s, EX t X t+ X t = = E + E, where 7 E + = P mut x,x Pˆfx ˆfx x 0, x : x 0< E = x : x 0> P mut x,x Pˆfx ˆfx x 0. To aalyze E +, we use a trval upper boud o Pˆfx ˆfx. The, we have E + x : x 0< P mutx,x x 0, 8 where the last equalty s drectly derved from Eq. 7 the proof of Theorem 9 []. For E, we have to cosder that the umber of 0-bts s creased. We aalyze the cases where oly oe -bt s flpped.e., x 0 = +, whose probablty s e. Let Z = f x f x. By the defto of symmetrc ose, the value of Z ca be,, ad +, each wth probablty /4. It s easy to see that the dstrbuto of Z s symmetrc about 0,.e., Z has the same dstrbuto as Z. Sce ˆfx ˆfx s the average of m depedet radom varables, whch have the same dstrbuto as Z, the dstrbuto of ˆfx ˆfx s also symmetrc about 0, ad thus Pˆfx ˆfx /. The, E e + = e. By calculatg E + E, we get EX t X t+ X t = e 0.05, where the last equalty s by < /0. Thus, codto of Theorem holds wth ǫ = To make X t+ X t j, t s ecessary to flp at least j bts of x. Thus, we get P X t+ X t j X t j j j! j. That s, codto of Theorem holds wth δ = ad rl =. Note that l = b a = /0. By Theorem, we ca coclude that the expected rug tme s expoetal. We prove Theorem 7 that the µ+-ea wth µ = 3log ca fd the optmum Olog 3 tme. The reaso for the effectveess of usg paret populatos s that the true best soluto wll be dscarded oly f t appears worse tha all the other solutos the populato, the probablty of whch ca be very small by usg a logarthmc paret populato sze. Note that ths fdg s cosstet wth that [4]. Theorem 7. For the µ+-ea solvg OeMax uder symmetrc ose, f µ = 3log, the expected rug tme s Olog 3. Proof. We apply the multplcatve drft theorem.e., Theorem to prove t. Note that the state of the correspodg Markov cha s curretly a populato,.e., a set of µ solutos. We frst desg a dstace fuctov: for ay populato P, VP = m x P x 0,.e., the mmum umber of 0-bts of the soluto P. It s easy to see that VP = 0 ff P X,.e., P cotas the optmum. The, we vestgate EVξ t Vξ t+ ξ t =P for ay P wth VP>0.e., P / X. Assume that curretly VP =, where. We also dvde the drft to two parts: EVξ t Vξ t+ ξ t = P = E + E, where E + = Pξ t+ = P ξ t = P VP, E = P :VP < P :VP > Pξ t+ = P ξ t = P VP. For E +, we eed to cosder that the best soluto P s mproved. Let x argm x P x 0, the x 0 =. I oe terato of the µ+-ea, a soluto x wth x 0 = ca be geerated by selectg x ad flppg oly oe 0-bt mutato, whose probablty s µ eµ. If x s ot added to P, t must hold that f x < f x for ay x P, whch happes wth probablty / µ sce f x < f x ff f x = fx. Thus, the probablty that x s added to P whch mples that VP = s / µ. We the get E + eµ µ = eµ µ. For E, f there are at least two solutos x,y P such that x 0 = y 0 =, t obvously holds that E = 0. Otherwse, VP > VP = mples that for the uque best soluto x P ad ay x P \x }, f x f x, whch happes wth probablty/ µ scef x f x ff f x = fx. Thus, PVP > / µ. Furthermore, VP ca crease by at most. Thus, E / µ. By calculatg E + E, we get EVξ t Vξ t+ ξ t eµ eµ µ µ 0log = 0log Vξ t, where the secod equalty holds wth large eough. Note that µ = 3log. Thus, by Theorem, Eτ ξ 0 0log+log = Olog, whch mples that the expected rug tme s Olog 3, sce the algorthm eeds to evaluate the offsprg soluto ad reevaluate the µ paret solutos each terato. I the followg, we show that the paret populato sze µ = 3log s almost tght for makg the µ+-ea effcet. Partcularly, we prove that µ = O s suffcet. Note that the proof s fshed by applyg the orgal egatve drft theorem.e., Theorem 4 stead of the smplfed versos.e., Theorems ad 3. To apply the smplfed egatve drft theorems, we have to show that the probablty of jumpg towards ad away from the target s expoetally decayg. However, the probablty of jumpg away from the target s at least a costat ths studed case. To jump away from the

7 7 target, t s suffcet that oe o-best soluto the curret populato s cloed by mutato ad the the best soluto s deleted the process of updatg the populato. The former evet happes wth probablty µ µ = Θ, ad the latter happes wth probablty, whch s Θ for µ µ = O. The orgal egatve drft theorem s stroger tha the smplfed oes, ad ca be appled here to prove the expoetal rug tme. Theorem 8. For the µ+-ea solvg OeMax uder symmetrc ose, f µ = O, the expected rug tme s expoetal. Proof. We apply the orgal egatve drft theorem.e., Theorem 4 to prove t. Let X t = Y t hz t, where Y t = m x P x 0 deotes the mmum umber of 0-bts of the soluto the populato P after t teratos of the µ+-ea, Z t = x P x 0 = Y t } deotes the umber of solutos P that have the mmum 0-bts Y t, ad for,,...,µ}, h = dµ d µ d µ wth d = µ+4. Note that 0 = h < h <... < hµ <, ad X t 0 ff Y t = 0,.e., P cotas at least oe optmum. We set l =, λl = ad cosder the terval [0,c ], where c = 3d,.e., the parameters al = 0 ad bl = c µ Theorem 4. The, we aalyze Eq.. It s easy to verfy that Eq. s equvalet to the followg equato: e X t r pl. PX t+ =r al<x t <bl r X t 9 We dvde the term the left sde of Eq. 9 to two parts: r < X t.e., X t+ < X t ad r > X t.e., X t+ > X t. We frst cosder X t+ < X t. Sce X t+ = Y t+ hz t+, X t = Y t hz t ad 0 hz t+,hz t <, we have X t+ < X t ff Y t+ Y t < 0 or Y t+ = Y t hz t+ > hz t. We cosder these two cases separately. Y t+ Y t = j. It mples that a ew soluto x wth x 0 = Y t j s geerated the t+-th terato of the algorthm. Suppose that x s geerated from some soluto x whch must satsfy that x 0 Y t selected from P, the x : x 0=Y t j P mut x,x x : x 0=Y t j Yt j j P mut x Y t,x j Yt < c j, where x j deotes ay soluto wth j 0-bts, the secod equalty s because t s ecessary to flp at least j 0-bts, ad the last equalty s by Y t = X t +hz t < bl+ = c. Meawhle, we have X t X t+ = Y t hz t Y t+ +hz t+ = j hz t j, where the secod equalty s by hz t+ = h = 0. Y t+ = Y t hz t+ > hz t. It mples that Z t < µ ad a ew soluto x wth x 0 = Y t s geerated. Suppose that the t + -th terato, the soluto selected from P for mutato s x. If x 0 > Y t, x : x 0=Y t P mut x,x x : x 0=Y t P mut x Yt+,x Y t+ = Yt+. If x 0 = Y t, x : x 0=Y t P mut x,x + Y t Yt j= j j e + Y t j= Yt j e + Yt/ Y. Sce Y t/ t = X t + hz t < bl+ = c ad c = 3d =, we have µ 3 µµ+4 P mut x,x. x : x 0=Y t Meawhle, t must hold that Z t+ = Z t +, thus we have X t X t+ = hz t+ hz t = hz t + hz t. By combg the above two cases, we get PX t+ = r al < X t < bl e X t r 0 r<x t Y t c j e j e + hzt+ hzt, Z t < µ 0, Z t = µ j= Y t ce j hzt + hz + t, Z t < µ 0, Z t = µ j= ce ce + hzt + hz t, Z t < µ 0, Z t = µ, where the secod equalty s by 0 < hz t + hz t < ad e s +s for 0 < s <. The, we cosder X t+ > X t. It s easy to verfy that X t+ > X t ff the t+-th terato, the ewly geerated soluto x satsfes that x 0 > Y t ad oe soluto x P wth x 0 = Y t s deleted. We frst aalyze the probablty of geeratg a ew soluto x wth x 0 > Y t. Suppose that the soluto selected from P for mutato s x. If x 0 > Y t, t s suffcet that all bts of x are ot flpped, thus x : x 0>Y t P mut x,x e. If x 0 = Y t, t s suffcet that oly oe -bt ofxs flpped, thus x : x 0>Y t P mut x,x Yt Yt e. Note that Y t = X t +hz t < bl + = c ad c = = 3 µµ+4 Θ for µ = O. Thus, we have P mut x,x c. e x : x 0>Y t We the aalyze the probablty of deletg oe soluto x P wth x 0 = Y t. Sce t s suffcet that the ftess evaluato of all solutos P x } wth more tha Y t 0- bts s affected by ose, the probablty s at least / µ. We fally aalyze X t X t+. If Z t =, Y t+ Y t +, thus X t X t+ = Y t Y t+ +hz t+ hz t hµ. If Z t, we have Y t+ = Y t ad Z t+ = Z t, thus X t X t+ = hz t+ hz t = hz t hz t. Note that for X t+ > X t, e Xt Xt+ < 0. Thus, we have PX t+ = r al < X t < bl e Xt r r>x t µ c e c µ+ d e hµ, Z t = e hzt hzt, Z t hµ, Zt = hz t hz t, Z t, e hµ, Zt = hz t hz t, Z t,

8 8 where the secod equalty s by e s s+s / = s+ s/ s/ for < s < 0, ad the last s by d = µ+4 ad c =. 3 By combg µµ+4 Eq. 0 ad Eq., we ca get PX t+ =r al<x t <bl e X t r ce ce + r X t hz t + hz t + d hµ, Z t = hz t + hz t + d hz t hz t, < Z t < µ. d hz t hz t, Z t = µ hµ If Z t =, hz = dµ d µ d t+ hz t d µ µ d µ d = d, ad µ we have hz t + hz t + d hµ = hz t + hz t d d hµ hµ. If < Z t < µ, hz t hz t hz = dµ Z t+ d µ Z t t+ hz t = d, ad smlarly we have d µ Z t d µ Z t hz t + hz t + d hz t hz t = hz t hz t + hµ hµ. If Z t = µ, d hz t hz t = hµ hµ. Thus, the above equato cotues wth d ce ce + d hµ hµ = /ce + d d µ 3 d µ = d µ, d d µ where the secod equalty s by c = 3d ad d 4. The µ codto of Theorem 4.e., Eq. or equvaletly Eq. 9 thus holds wth pl = d µ. Now we vestgate Dl = max,e e λl Xt+ bl X t bl } = max,e e bl Xt+ X t bl } Eq.. To derve a upper bod o Dl, we oly eed to aalyze E e bl Xt+ X t bl. E e bl Xt+ X t bl = PY t+ = r X t bl r bl + r<bl E e bl Xt+ X t bl,y t+ = r PY t+ = r X t bl E e bl Xt+ X t bl,y t+ = r. Whe Y t+ = r bl, we have bl X t+ = bl Y t+ + hz t+ hz t+ <. The we cosder the case that Y t+ < bl. Sce X t = Y t hz t bl, we have Y t bl > Y t+, whch mples that Y t bl ad Y t+ bl. To make Y t+ = r bl, t s ecessary that a ew soluto x wth x 0 = r bl s geerated by mutato. Let x deote the soluto selected from the populato P for mutato. Note that x 0 Y t bl. The, for r bl, PY t+ = r X t bl x : x P 0=r mutx,x x : x P 0=r mutx bl,x bl bl r bl r bl bl r. Furthermore, for Y t+ < Y t, t must hold that Z t+ =, ad thusbl X t+ = bl Y t+ +hz t+ = bl Y t+. Thus, the above equato cotues wth e+ r bl bl r bl e bl r e+ = e+ bl j= bl j e j e+ e bl / e bl / /e bl e+ /ce e+, where the fourth equalty s by bl bl+ = c ad the last equalty s by c = 3d. Thus, µ } Dl = max,e e bl Xt+ X t bl e+. Let Ll = e c/ Theorem 4. Note that c = 3d µ = 3 µµ+4 = Θ for µ = O. The, by Theorem 4, we get PTl e c/ X 0 bl e c e c/ e+ d µ = e Ω. By Cheroff bouds, for ay x chose from 0,} u.a.r., P x 0 < c = e Ω. By the uo boud, PY 0 < c µ e Ω = e Ω, whch mples that PX 0 < bl = PY 0 hz 0 < bl PY 0 < bl+ = PY 0 < c = e Ω. Thus, the expected rug tme s expoetal. V. OFFSPRING POPULATIONS CAN WORK ON SOME TASKS WHERE SAMPLING FAILS I the above secto, we have show that usg paret populatos ca be better tha usg samplg. We the show the superorty of usg offsprg populatos over samplg o the robustess to ose. Partcularly, we compare the +-EA usg samplg wth the +λ-ea, o the OeMax problem uder reverse ose. As preseted Defto 6, reverse ose returs a reverse ftess fx wth probablty /. Defto 6 Reverse Nose. Let f x ad fx deote the osy ad true ftess of a soluto x, respectvely, the f fx wth prob. /, x = fx wth prob. /. For OeMax uder reverse ose, t s easy to see that for ay two solutos x ad y, the dstrbuto of f x f y s symmetrc about 0. Thus, as we have foud uder symmetrc ose, the algorthm behavor by usg samplg s aalogous to radom walk, ad thus the optmzato s effcet. The proof of Theorem 9 s very smlar to that of Theorem 6, whch proves the expoetal rug tme requred by the +-EA usg samplg to solve OeMax uder symmetrc ose. Theorem 9. For the +-EA solvg OeMax uder reverse ose, f usg samplg, the expected rug tme s expoetal. Proof. The proof ca be accomplshed the same way as that of Theorem 6. The proof of Theorem 6 apples the smplfed egatve drft theorem.e., Theorem, ad aalyzes the drft EX t X t+ X t = by dvdg t to two parts: E + ad E. I the proof of Theorem 6, the aalyss of the postve drft E + does ot rely o the ose model, ad thus the upper boud E + / stll holds here. For the lower boud e of the egatve drft E, t reles o the property that for ay two solutos x wth x 0 = ad x wth x 0 = +, the

9 9 dstrbuto of f x f x s symmetrc about 0. By the defto of reverse ose, f x f x ca be +,, ad, each wth probablty /4; thus ts dstrbuto s stll symmetrc about 0. The, t stll holds that E e. Accordg to Theorem, we ca get that the expected rug tme s expoetal. By usg offsprg populatos, the probablty of losg the curret ftess becomes very small. Ths s because a far umber of offsprg solutos wth ftess ot worse tha the curret ftess wll be geerated wth a hgh probablty the reproducto of each terato of the +λ-ea, ad the curret ftess becomes worse oly f all these good offsprg solutos ad the paret soluto are evaluated correctly, the probablty of whch ca be very small by usg a logarthmc offsprg populato sze. Thus, usg offsprg populatos ca lead to a effcet optmzato, as show Theorem 0. Note that the reaso for the effectveess of usg offsprg populatos foud here s cosstet wth that [4]. Theorem 0. For the +λ-ea solvg OeMax uder reverse ose, f λ = 8 log, the expected rug tme s Olog. Proof. We apply Theorem to prove t. Each state of the correspodg Markov cha ξ t } + t=0 s just a soluto here. That s, ξ t correspods to the soluto after rugtteratos of the +λ-ea. We desg the dstace fucto as for x 0,}, Vx = x 0. Assume that curretly x 0 =, where. To aalyze EVξ t Vξ t+ ξ t = x, we dvde t to two parts as the proof of Theorem 7. That s, EVξ t Vξ t+ ξ t = x = E + E, where E + = Pξ t+ = y ξ t = x y 0, y: y 0< E = y: y 0> Pξ t+ = y ξ t = x y 0. For E +, sce y 0 <, we have y 0. Thus, E + Pξ t+ = y ξ t = x = P ξ t+ 0 < ξ t = x. y: y 0< To make ξ t+ 0 <, t requres that at least oe soluto x wth x 0 < s geerated the reproducto ad at least oe of them s evaluated correctly. To geerate a soluto x wth x 0 < by mutatg x, t s suffcet that oly oe 0-bt of x s flpped, whose probablty s e. Thus, each terato of the +λ-ea, the probablty of geeratg at least oe offsprg soluto x wth x 0 < s at least λ e λ e e. +λ e If λ e λ e +λ e >, e λ ; otherwse, e λ λ e. Thus, e λ m, λ e } = m, 4log e }, where the equalty s by λ = 8 log. Sce each soluto s evaluated correctly wth probablty, P ξ t+ 0 < ξ t = x m, 4log e }. Thus, E + m, 4log } e = m 4, log } e 4. For E, sce y 0, we have E P ξ t+ 0 > ξ t = x. Let q = x : x P 0 mutx,x deote the probablty of geeratg a offsprg solutox wth at most0-bts by mutatg x. Sce t s suffcet that o bt s flpped or oly oe 0-bt s flpped mutato,q + e. Now we aalyze P ξ t+ 0 > ξ t = x. Assume that the reproducto, exactly k offsprg solutos wth at most 0-bts are geerated, where 0 k λ; t happes wth probablty λ k q k q λ k. If k < λ, the soluto the ext geerato has more tha 0-bts.e., ξ t+ 0 > ff the ftess evaluato of these k offsprg solutos ad the paret soluto x s all affected by ose, whose probablty s. If k = λ, the soluto the ext geerato must have k+ at most 0-bts.e., ξ t+ 0. Thus, λ λ P ξ t+ 0 > ξ t = x = k k=0 q q k q λ k k+ λ λ, e where the last equalty s by q e. We the get E 8log e.3.3, where the base of the logarthm s. By calculatg E + E, we have EVξ t Vξ t+ ξ t = 5 Vξ t, where the secod equalty holds wth large eough. Thus, by Theorem, Eτ ξ 0 5+log = Olog, whch mples that the expected rug tme s Olog, sce t eeds to reevaluate the paret soluto ad evaluate the λ = 8log offsprg solutos each terato. The, we prove that a costat offsprg populato sze λ = O s ot suffcet to allow solvg the osy problem polyomal tme. Ths also mples that the effectve value λ = 8log derved the above theorem s early tght. From the proof, we ca fd that λ = O caot guaratee a suffcetly small probablty of losg the curret ftess, ad thus the optmzato s effcet. Theorem. For the +λ-ea solvg OeMax uder reverse ose, f λ = O, the expected rug tme s expoetal. Proof. We apply Theorem to prove t. Let X t = x 0 deote the umber of 0-bts of the soluto x mataed by the +λ-ea after rug t teratos. We cosder the terval [0, ],.e., a = 0 ad b = Theorem. 6e λ 6e λ We the aalyze EX t X t+ X t = for < 6e. We dvde the drft as follows: λ EX t X t+ X t = = E + E, where

10 0 E + = PX t+ = j X t = j, E = j=+ PX t+ = j X t = j. For E +, we eed to derve a upper boud o PX t+ = j X t = for j <. Note that X t+ = j mples that at least oe offsprg soluto x wth x 0 = j s geerated by mutatg x the reproducto. Thus, we have PX t+ =j X t = λ P mut x,x λ The, we get E + λ = λ x : x 0< x : x 0=j x : x 0=j x : x 0=j P mut x,x. P mut x,x j P mut x,x x 0 λ, where the last equalty s drectly derved by Eq. 8. For E, we easly have E PX t+ = j X t = = PX t+ > X t =. j=+ Let q = x : x P 0 mutx,x, where x s ay soluto wth 0-bts. Usg the same aalyss as Eq., we ca get λ λ PX t+ > X t = = q k q λ k k k+ k=0 = q λ q λ = q λ q λ + q λ q λ q λ 8e λ, where the last equalty s by q e ad q x : x P 0=+ mutx,x e 4. Thus, E λ/8e λ. By calculatg E + E, we have EX t X t+ X t = λ λ 8e λ λ 6e λ, where the last equalty s by < 6e. Thus, codto λ of Theorem holds wth ǫ = λ, whch s a costat for 6e λ λ = O. To make X t+ X t j, t s ecessary that at least oe offsprg soluto geerated by mutatg x flps at least j bts of x. Let pj deote the probablty that at least j bts of x are flpped mutato. We easly have pj j. Thus, j P X t+ X t j X t pj λ λ pj λ j j λ 3 j,.e., codto of Theorem holds wth δ= ad rl= λ=o. Note thatl=b a= =Θ. By Theorem, 6e λ we get that the expected rug tme s expoetal. VI. ADAPTIVE SAMPLING CAN WORK ON SOME TASKS WHERE BOTH SAMPLING AND POPULATIONS FAIL I ths secto, we frst theoretcally vestgate whether there exst cases where usg ether populatos or samplg s effectve. We gve a postve aswer by cosderg OeMax uder segmeted ose. The, we prove that such a stuato, usg adaptve samplg ca be effectve, whch provdes some theoretcal justfcato for the good emprcal performace of adaptve samplg practce [8], [3]. As preseted Defto 7, the OeMax problem s dvded to four segmets. I oe segmet, the ftess s evaluated correctly, whle the other three segmets, the ftess s dsturbed by dfferet oses. We prove Theorem that the expected rug tme of the +-EA usg samplg wth ay sample sze m s expoetal. From the proof, we ca fd the reaso for the effectveess of samplg. For two solutosx adx wth x 0 = x 0 +.e.,fx = fx +, the expected gaps betwee f x ad f x are postve ad egatve, respectvely, the segmets of 00 < x 0 50 ad 00 < x Thus, the former segmet, a larger sample sze s better sce t wll decrease Pˆfx ˆfx, whle the latter segmet, a larger sample sze s worse sce t wll crease Pˆfx ˆfx. Furthermore, there s o moderate sample sze whch ca make a good tradeoff. Thus, samplg fals ths case. Defto 7 OeMax uder Segmeted Nose. For ay x 0,}, the osy ftess value f x s calculated as: f x 0 > 50, f x = x 0 ; f 00 < x 0 50, f x 0 wth prob. /+/, x = 3+ x 0 wth prob. / /; 3 f 00 < x 0 00, f 4 x 0 wth prob. /, x = + x 0 3 wth prob. /; 4 f x 0 00, f x = 4 x 0 wth prob. /5, 4 δ wth prob. 4/5, where δ s radomly draw from a cotuous uform dstrbuto U[0,], ad /00 N +. Theorem. For the +-EA solvg OeMax uder segmeted ose, f usg samplg, the expected rug tme s expoetal. Proof. We dvde the proof to two parts accordg to the rage of m. Let X t = x 0 deote the umber of 0-bts of the soluto x mataed by the +-EA after rug t teratos. Whe m 4 400, we apply Theorem to prove that startg from X 0 50, the expected umber of teratos utl X t 4 00 s expoetal. Whem > 400, we apply Theorem to prove that startg from X 0 00, the expected umber of teratos utl X t 00 s expoetal. Due to the uform tal dstrbuto, both X 0 50 ad X 0 00 hold wth a hgh probablty. Thus, for ay m, the expected rug tme

11 utl fdg the optmum s expoetal. For the proof of each part, codto of Theorem trvally holds, ad we oly eed to show that EX t X t+ X t s upper bouded by a egatve costat. [Part I: m 4 ] We cosder the terval [ ]. As , 50 the proof of Theorem 6, we compute the drft EX t X t+ X t = where 00 < < 50 by E+ E.e., Eq. 7. For E, we cosder the cases where oly oe -bt of x s flpped mutato. That s, x 0 = +. We the show that the offsprg soluto x s accepted wth probablty at least 0.07.e., Pˆfx ˆfx 0.07 by cosderg two subcases for m. m 4. For 00 < k 50, let xk deote a soluto wth k 0-bts. Accordg to case of Defto 7, we have Ef x k = + k + = k ; Varf x k = + k + 3+k 4 3+k k 0 +k +4k 4. Let Y = f x f x. Note that x 0 = 00, 50 ad x 0 = +. The, we get that µ := EY = ad σ := VarY. Let Z = Y µ. The, we have EZ = 0, VarZ = σ ad ρ := E Z , where the last equalty holds wth large eough. Note that ˆfx ˆfx µ s the average of m depedet radom varables, whch have the same dstrbuto as Z. By Berry- Essee equalty [9], P ˆfx ˆfx µ m x σ Φx ρ σ 3 m, where Φx deotes the cumulatve dstrbuto fucto of the stadard ormal dstrbuto. Thus, Pˆfx ˆfx 0 = Pˆfx ˆfx µ µ ˆfx = P ˆfx µ m σ Φ µ m σ ρ σ 3 m 0.07, µ m σ where the last equalty s derved by µ =, 4 m 4 400, σ ad ρ 9 3. m 3. It holds that Pˆfx ˆfx 3 0., sce t s suffcet thatf x s always evaluated to3++ m depedet evaluatos. Combg the above two cases, our clam that Pˆfx ˆfx 0.07 holds. Note that < /50. Thus, we have E For E +, we ca smlarly get E + 50 as the proof of Theorem 6, because the offsprg soluto x s optmstcally assumed to be always accepted. Thus, the drft satsfes that EX t X t+ X t = = E + E 0./50. [Part II: m > ] We cosder the terval [ 00, 00 ], ad compute the drft EX t X t+ X t = where 00 < < 00 by E+ E.e., Eq. 7. For the egatve drft, we show that the probablty of acceptg the offsprg soluto x wth x 0 = + s at least 0.9. Let x k deote a soluto wth k 0-bts. Accordg to case 3 of Defto 7, we have, for 00 < k < 00, Ef x k f x k+ = 4 3+k +3+k+ 8; ad for 00 < k 00, Varf x k = +k6 + 4 k Ef x k / The, µ := Eˆfx ˆfx 8 ad σ := Varˆfx ˆfx = Varf x f x /m 8 5 /m. By Chebyshev s equalty ad m > 4 400, we have Pˆfx ˆfx P ˆfx ˆfx µ µ σ /µ 0., where the last equalty holds wth large eough. Thus, E e For E+, we stll have E Thus, the drft satsfes that EX t X t+ X t = = E + E 0.3. To prove the effectveess of paret populatos, we derve a suffcet codto for the expoetal rug tme of the µ+-ea requred to solve OeMax uder ose, whch s spred from Theorem 4 [3]. We geeralze ther result from addtve ose to arbtrary ose. As show Lemma, the codto tutvely meas that whe the soluto s close to the optmum, the probablty of dscardg t from the populato decreases learly w.r.t. the populato sze µ, whch s, however, ot small eough to make a effcet optmzato. Note that for the case where usg paret populatos works Secto IV, the probablty of dscardg the best soluto from the populato decreases expoetally w.r.t. µ. Let poly dcate ay polyomal of. Lemma. For the µ+-ea where µ poly solvg OeMax uder ose, f for ay y wth y > ad ay set of µ solutos Q = x,x,...,x µ }, Pf y < m x Qf x 3/5µ+, 5 the the expected rug tme s expoetal.

12 Proof. Let ξ t deote the populato after t teratos of the algorthm. Let X t deote the umber of solutos wth - bts ξ t. Let a = ad b = 0. We frst use a ductve proof to show that t 0, > a : EX t µb a. 6 For t = 0, due to the uform tal dstrbuto, we have EX 0 = µ /. Note that for j 3, j+ / j = j /3 j+ /3+. Thus, for > a, / / / b a, whch mples that > a,ex 0 µba. We the assume that 0 t k, > a : EX t µba, ad aalyze EX k+ for > a. Let X k = X0 k,xk,...,xk, l = l 0,l,...,l, l = =0 l 3 ad p = 5µ+. Let x deote the offsprg soluto geerated the t + -th terato of the algorthm, ad let x deote ay soluto wth -bts. Let P mut x,y deote the probablty that x s mutated to y by bt-wse mutato. We use P mut x j,x = y: y P = mutx j,y to deote the probablty of geeratg a soluto wth -bts by mutatg ay soluto wth j -bts. The, we have EX k+ X k = EEX k+ X k X k = PX k = l l =µ P x =,x ad ay x ξ k are ot deleted X k =l P x, oe x ξ k s deleted X k = l PX k =l P x = X k =l l +p l =µ = l =µ l =µ P x = X k = l l p PX k =l P x = X k =l p l p = PX k =l l j µ P mutx j,x p l p = p P mut x j,x l =µ = p PX k = ll p P mut x j,x µ PX k = l l p l =0 = p µ l =µ µ l PX k = l l j µ PX k j = l j l j µ P mut x j,x EXj p EX k k, where the secod equalty s because X k+ X k = ff x = ad x s added to the populato meawhle the solutos wth -bts ξ k are ot deleted; X k+ X k = ff x = ad oe soluto wth -bts ξ k s deleted, the frst equalty s because ay soluto wth -bts s deleted wth probablty at least p = 3 5µ+ by the codto Eq. 5, ad the fourth equalty s sce a paret soluto s uformly selected from ξ k for mutato. We further derve a upper boud o µ P mut x j,x EXj k as follows: µ P mut x j,x EXj k = a µ + j=a+ + + j= P mut x j,x EXj k j=+ a a j j + b a j a j j=a+ +b a l + + b a j l l= j=+ a a j +b a a b j j=a+ + e + l a + b j l= j=+ a b a + b a b a + e + a + b b a /, where the frst equalty s derved by applyg j a : P mut x j,x P mut x a,x a a a, EXk j = E Xk j = µ, j > a : EXk j µba j ad some smple upper bouds o P mut x j,x for j > a, the thrd equalty s by 0 < c < : + l= cl = c c = /c, ad the last s by a = , b = 0 ad > a. Combg the above two formulas, we get EX k+ X k p b a / p EX k, whch mples that EX k+ p b a /+ p EX k µ + 5µ+ 5µ+ µba µb a, where the secod equalty s by p = 3 5µ+ ad EXk µb a, ad the last equalty holds wth µ. Thus, our clam that t 0, > a : EX t µba holds. Based o Eq. 6 ad Markov s equalty, we get, for ay t 0, PX t EXt µba. Note that X t s the umber of optmal solutos the populato after t teratos. Let T = b a/. The, the probablty of fdg the optmal soluto T teratos s P t T,X t T t=0 PXt T µb a = µ b a /, whch s expoetally small for µ poly. Ths mples that the expected rug tme for fdg the optmal soluto s expoetal.

13 3 By verfyg the codto of Lemma, we prove Theorem 3 that the µ+-ea wth µ poly eeds expoetal tme for solvg OeMax uder segmeted ose. Theorem 3. For the µ+-ea where µ poly solvg OeMax uder segmeted ose, the expected rug tme s expoetal. Proof. We apply Lemma to prove t. For ay soluto y wth y 0 /00 ad Q = x,...,x µ }, let A deote the evet that f y < m x Qf x. We wll show that PA 4 5µ+, whch mples that the codto Eq. 5 holds sce y 0 /00 covers the requred rage of y > 599/600. Let B l 0 l µ deote the evet that l solutos Q are evaluated to have egatve osy ftess values. Note that for ay x, f x < 0 mples that x 0 /00, ad f x = 4 δ where δ U[0,]. For 0 l µ, PA B l Pf y < 0 B l PA f y < 0,B l. Uder the codtos f y < 0 ad B l, the osy ftess values of y ad the correspodg l solutos Q satsfy the same cotuous dstrbuto 4 δ where δ U[0,], thus PA f y < 0,B l l+ µ+. The, we get PA B l 4 5 µ+ ad PA = µ l=0 PA B l PB l 4 5µ+. By Lemma, the theorem holds. We the show Theorem 4 that usg offsprg populatos s also effectve ths case. By usg offsprg populatos, the probablty of mprovg the curret ftess becomes very small whe the soluto s the d segmet.e., 00 < x Ths s because a far umber of offsprg solutos wth ftess ot better tha the curret ftess wll be geerated wth a hgh probablty, ad the curret ftess becomes better oly f all these bad offsprg solutos ad the paret soluto are evaluated correctly, the probablty of whch almost decreases expoetally w.r.t. λ. Note that for the +λ-ea solvg OeMax uder reverse ose.e., Theorem 0, the effectveess of usg offsprg populatos s due to the small probablty of losg the curret ftess, sce t requres a far umber of offsprg solutos wth ftess ot worse tha the curret ftess to be evaluated correctly. Therefore, we ca see that usg offsprg populatos ca geerate a far umber of good ad bad offsprg solutos smultaeously, ad whether t wll be effectve depeds o the cocrete osy problem. Theorem 4. For the +λ-ea where λ poly solvg OeMax uder segmeted ose, the expected rug tme s expoetal. Proof. We apply the smplfed egatve drft theorem wth scalg.e., Theorem 3 to prove t. Let X t = x 0 deote the umber of 0-bts of the soluto x mataed by the +λ- EA after rugt teratos. We cosder the terval[ 75, 50 ],.e., a = 75 ad b = 50 Theorem 3. Frst, we aalyze EX t X t+ X t = for 75 < < 50. As the proof of Theorem, the drft s dvded to two parts: E + = PX t+ = j X t = j ad E = j=+ PX t+ = j X t = j. For E +, we cosder that the umber of 0-bts s decreased. Let q = x : x P 0,+} mutx,x,.e., the probablty of geeratg a soluto wthor+ 0-bts by mutatgx. Sce t s suffcet to flp o bts or flp oly oe -bt, q +. Now we aalyzepx t+ = j X t = for < j <. Assume that the reproducto, exactly k 00 offsprg solutos wth or + 0-bts are geerated, where 0 k λ; t happes wth probablty λ k q k q λ k. For k = λ, the soluto the ext geerato must have at least 0-bts.e., X t+. For 0 k < λ, each of the remag λ k solutos has j 0-bts wth probablty pj q, where pj := x : x P 0=j mutx,x. Thus, uder the codto that exactly k offsprg solutos wth or + 0-bts are geerated, the probablty that at least oe offsprg soluto has j 0-bts s pj q λ k. Furthermore, to make the soluto the ext geerato have j 0-bts.e., X t+ = j, t s ecessary that the ftess evaluato of these k offsprg solutos ad the paret soluto x s ot affected by ose, the probablty of whch s + k+. Thus, we have, for 00 < j <, PX t+ = j X t = λ λ q k q λ k pj λ k k+ k q + k=0 λ λ q k q λ k pj λ k k q + k+ k=0 λ k λ = pjλ + q k + q λ k k=0 λ λ = pjλ + q pjλ, 3 where the last equalty s by q + e For λ, λ+ 3 λ+ /λ 3 λ = λ+ λ 3, ad ote that 3, 3. Thus, we have, for 00 < j <, PX t+ = j X t = pj = P mut x,x. 7 For 0 j 00, we have x : x 0=j PX t+ = j X t = pj λ λ pj λ λ 8 λ j j j /300, where the frst equalty s because to make X t+ = j, t s ecessary that at least oe offsprg soluto wth j 0-bts s geerated, ad the last equalty s by > 75 ad j 00. By applyg Eqs. 7 ad 8 to E +, we get E + P mut x,x j /00<j< x : x 0=j + 0 j /00 λ j /300 + λ / ,

Discrete Mathematics and Probability Theory Fall 2016 Seshia and Walrand DIS 10b

Discrete Mathematics and Probability Theory Fall 2016 Seshia and Walrand DIS 10b CS 70 Dscrete Mathematcs ad Probablty Theory Fall 206 Sesha ad Walrad DIS 0b. Wll I Get My Package? Seaky delvery guy of some compay s out delverg packages to customers. Not oly does he had a radom package

More information

Runtime analysis RLS on OneMax. Heuristic Optimization

Runtime analysis RLS on OneMax. Heuristic Optimization Lecture 6 Rutme aalyss RLS o OeMax trals of {,, },, l ( + ɛ) l ( ɛ)( ) l Algorthm Egeerg Group Hasso Platter Isttute, Uversty of Potsdam 9 May T, We wat to rgorously uderstad ths behavor 9 May / Rutme

More information

Econometric Methods. Review of Estimation

Econometric Methods. Review of Estimation Ecoometrc Methods Revew of Estmato Estmatg the populato mea Radom samplg Pot ad terval estmators Lear estmators Ubased estmators Lear Ubased Estmators (LUEs) Effcecy (mmum varace) ad Best Lear Ubased Estmators

More information

{ }{ ( )} (, ) = ( ) ( ) ( ) Chapter 14 Exercises in Sampling Theory. Exercise 1 (Simple random sampling): Solution:

{ }{ ( )} (, ) = ( ) ( ) ( ) Chapter 14 Exercises in Sampling Theory. Exercise 1 (Simple random sampling): Solution: Chapter 4 Exercses Samplg Theory Exercse (Smple radom samplg: Let there be two correlated radom varables X ad A sample of sze s draw from a populato by smple radom samplg wthout replacemet The observed

More information

Special Instructions / Useful Data

Special Instructions / Useful Data JAM 6 Set of all real umbers P A..d. B, p Posso Specal Istructos / Useful Data x,, :,,, x x Probablty of a evet A Idepedetly ad detcally dstrbuted Bomal dstrbuto wth parameters ad p Posso dstrbuto wth

More information

A tighter lower bound on the circuit size of the hardest Boolean functions

A tighter lower bound on the circuit size of the hardest Boolean functions Electroc Colloquum o Computatoal Complexty, Report No. 86 2011) A tghter lower boud o the crcut sze of the hardest Boolea fuctos Masak Yamamoto Abstract I [IPL2005], Fradse ad Mlterse mproved bouds o the

More information

Lecture 3. Sampling, sampling distributions, and parameter estimation

Lecture 3. Sampling, sampling distributions, and parameter estimation Lecture 3 Samplg, samplg dstrbutos, ad parameter estmato Samplg Defto Populato s defed as the collecto of all the possble observatos of terest. The collecto of observatos we take from the populato s called

More information

PTAS for Bin-Packing

PTAS for Bin-Packing CS 663: Patter Matchg Algorthms Scrbe: Che Jag /9/00. Itroducto PTAS for B-Packg The B-Packg problem s NP-hard. If we use approxmato algorthms, the B-Packg problem could be solved polyomal tme. For example,

More information

Lecture 02: Bounding tail distributions of a random variable

Lecture 02: Bounding tail distributions of a random variable CSCI-B609: A Theorst s Toolkt, Fall 206 Aug 25 Lecture 02: Boudg tal dstrbutos of a radom varable Lecturer: Yua Zhou Scrbe: Yua Xe & Yua Zhou Let us cosder the ubased co flps aga. I.e. let the outcome

More information

Point Estimation: definition of estimators

Point Estimation: definition of estimators Pot Estmato: defto of estmators Pot estmator: ay fucto W (X,..., X ) of a data sample. The exercse of pot estmato s to use partcular fuctos of the data order to estmate certa ukow populato parameters.

More information

CS286.2 Lecture 4: Dinur s Proof of the PCP Theorem

CS286.2 Lecture 4: Dinur s Proof of the PCP Theorem CS86. Lecture 4: Dur s Proof of the PCP Theorem Scrbe: Thom Bohdaowcz Prevously, we have prove a weak verso of the PCP theorem: NP PCP 1,1/ (r = poly, q = O(1)). Wth ths result we have the desred costat

More information

Feature Selection: Part 2. 1 Greedy Algorithms (continued from the last lecture)

Feature Selection: Part 2. 1 Greedy Algorithms (continued from the last lecture) CSE 546: Mache Learg Lecture 6 Feature Selecto: Part 2 Istructor: Sham Kakade Greedy Algorthms (cotued from the last lecture) There are varety of greedy algorthms ad umerous amg covetos for these algorthms.

More information

Lecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model

Lecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model Lecture 7. Cofdece Itervals ad Hypothess Tests the Smple CLR Model I lecture 6 we troduced the Classcal Lear Regresso (CLR) model that s the radom expermet of whch the data Y,,, K, are the outcomes. The

More information

Lecture 4 Sep 9, 2015

Lecture 4 Sep 9, 2015 CS 388R: Radomzed Algorthms Fall 205 Prof. Erc Prce Lecture 4 Sep 9, 205 Scrbe: Xagru Huag & Chad Voegele Overvew I prevous lectures, we troduced some basc probablty, the Cheroff boud, the coupo collector

More information

TESTS BASED ON MAXIMUM LIKELIHOOD

TESTS BASED ON MAXIMUM LIKELIHOOD ESE 5 Toy E. Smth. The Basc Example. TESTS BASED ON MAXIMUM LIKELIHOOD To llustrate the propertes of maxmum lkelhood estmates ad tests, we cosder the smplest possble case of estmatg the mea of the ormal

More information

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS Postpoed exam: ECON430 Statstcs Date of exam: Jauary 0, 0 Tme for exam: 09:00 a.m. :00 oo The problem set covers 5 pages Resources allowed: All wrtte ad prted

More information

8.1 Hashing Algorithms

8.1 Hashing Algorithms CS787: Advaced Algorthms Scrbe: Mayak Maheshwar, Chrs Hrchs Lecturer: Shuch Chawla Topc: Hashg ad NP-Completeess Date: September 21 2007 Prevously we looked at applcatos of radomzed algorthms, ad bega

More information

Simulation Output Analysis

Simulation Output Analysis Smulato Output Aalyss Summary Examples Parameter Estmato Sample Mea ad Varace Pot ad Iterval Estmato ermatg ad o-ermatg Smulato Mea Square Errors Example: Sgle Server Queueg System x(t) S 4 S 4 S 3 S 5

More information

Class 13,14 June 17, 19, 2015

Class 13,14 June 17, 19, 2015 Class 3,4 Jue 7, 9, 05 Pla for Class3,4:. Samplg dstrbuto of sample mea. The Cetral Lmt Theorem (CLT). Cofdece terval for ukow mea.. Samplg Dstrbuto for Sample mea. Methods used are based o CLT ( Cetral

More information

hp calculators HP 30S Statistics Averages and Standard Deviations Average and Standard Deviation Practice Finding Averages and Standard Deviations

hp calculators HP 30S Statistics Averages and Standard Deviations Average and Standard Deviation Practice Finding Averages and Standard Deviations HP 30S Statstcs Averages ad Stadard Devatos Average ad Stadard Devato Practce Fdg Averages ad Stadard Devatos HP 30S Statstcs Averages ad Stadard Devatos Average ad stadard devato The HP 30S provdes several

More information

BIOREPS Problem Set #11 The Evolution of DNA Strands

BIOREPS Problem Set #11 The Evolution of DNA Strands BIOREPS Problem Set #11 The Evoluto of DNA Strads 1 Backgroud I the md 2000s, evolutoary bologsts studyg DNA mutato rates brds ad prmates dscovered somethg surprsg. There were a large umber of mutatos

More information

Introduction to local (nonparametric) density estimation. methods

Introduction to local (nonparametric) density estimation. methods Itroducto to local (oparametrc) desty estmato methods A slecture by Yu Lu for ECE 66 Sprg 014 1. Itroducto Ths slecture troduces two local desty estmato methods whch are Parze desty estmato ad k-earest

More information

Chapter 5 Properties of a Random Sample

Chapter 5 Properties of a Random Sample Lecture 6 o BST 63: Statstcal Theory I Ku Zhag, /0/008 Revew for the prevous lecture Cocepts: t-dstrbuto, F-dstrbuto Theorems: Dstrbutos of sample mea ad sample varace, relatoshp betwee sample mea ad sample

More information

Chapter 9 Jordan Block Matrices

Chapter 9 Jordan Block Matrices Chapter 9 Jorda Block atrces I ths chapter we wll solve the followg problem. Gve a lear operator T fd a bass R of F such that the matrx R (T) s as smple as possble. f course smple s a matter of taste.

More information

9 U-STATISTICS. Eh =(m!) 1 Eh(X (1),..., X (m ) ) i.i.d

9 U-STATISTICS. Eh =(m!) 1 Eh(X (1),..., X (m ) ) i.i.d 9 U-STATISTICS Suppose,,..., are P P..d. wth CDF F. Our goal s to estmate the expectato t (P)=Eh(,,..., m ). Note that ths expectato requres more tha oe cotrast to E, E, or Eh( ). Oe example s E or P((,

More information

Chapter 8. Inferences about More Than Two Population Central Values

Chapter 8. Inferences about More Than Two Population Central Values Chapter 8. Ifereces about More Tha Two Populato Cetral Values Case tudy: Effect of Tmg of the Treatmet of Port-We tas wth Lasers ) To vestgate whether treatmet at a youg age would yeld better results tha

More information

Estimation of Stress- Strength Reliability model using finite mixture of exponential distributions

Estimation of Stress- Strength Reliability model using finite mixture of exponential distributions Iteratoal Joural of Computatoal Egeerg Research Vol, 0 Issue, Estmato of Stress- Stregth Relablty model usg fte mxture of expoetal dstrbutos K.Sadhya, T.S.Umamaheswar Departmet of Mathematcs, Lal Bhadur

More information

The Occupancy and Coupon Collector problems

The Occupancy and Coupon Collector problems Chapter 4 The Occupacy ad Coupo Collector problems By Sarel Har-Peled, Jauary 9, 08 4 Prelmares [ Defto 4 Varace ad Stadard Devato For a radom varable X, let V E [ X [ µ X deote the varace of X, where

More information

Chapter 3 Sampling For Proportions and Percentages

Chapter 3 Sampling For Proportions and Percentages Chapter 3 Samplg For Proportos ad Percetages I may stuatos, the characterstc uder study o whch the observatos are collected are qualtatve ature For example, the resposes of customers may marketg surveys

More information

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS Exam: ECON430 Statstcs Date of exam: Frday, December 8, 07 Grades are gve: Jauary 4, 08 Tme for exam: 0900 am 00 oo The problem set covers 5 pages Resources allowed:

More information

CIS 800/002 The Algorithmic Foundations of Data Privacy October 13, Lecture 9. Database Update Algorithms: Multiplicative Weights

CIS 800/002 The Algorithmic Foundations of Data Privacy October 13, Lecture 9. Database Update Algorithms: Multiplicative Weights CIS 800/002 The Algorthmc Foudatos of Data Prvacy October 13, 2011 Lecturer: Aaro Roth Lecture 9 Scrbe: Aaro Roth Database Update Algorthms: Multplcatve Weghts We ll recall aga) some deftos from last tme:

More information

Lecture 9: Tolerant Testing

Lecture 9: Tolerant Testing Lecture 9: Tolerat Testg Dael Kae Scrbe: Sakeerth Rao Aprl 4, 07 Abstract I ths lecture we prove a quas lear lower boud o the umber of samples eeded to do tolerat testg for L dstace. Tolerat Testg We have

More information

A New Family of Transformations for Lifetime Data

A New Family of Transformations for Lifetime Data Proceedgs of the World Cogress o Egeerg 4 Vol I, WCE 4, July - 4, 4, Lodo, U.K. A New Famly of Trasformatos for Lfetme Data Lakhaa Watthaacheewakul Abstract A famly of trasformatos s the oe of several

More information

CHAPTER 4 RADICAL EXPRESSIONS

CHAPTER 4 RADICAL EXPRESSIONS 6 CHAPTER RADICAL EXPRESSIONS. The th Root of a Real Number A real umber a s called the th root of a real umber b f Thus, for example: s a square root of sce. s also a square root of sce ( ). s a cube

More information

For combinatorial problems we might need to generate all permutations, combinations, or subsets of a set.

For combinatorial problems we might need to generate all permutations, combinations, or subsets of a set. Addtoal Decrease ad Coquer Algorthms For combatoral problems we mght eed to geerate all permutatos, combatos, or subsets of a set. Geeratg Permutatos If we have a set f elemets: { a 1, a 2, a 3, a } the

More information

X ε ) = 0, or equivalently, lim

X ε ) = 0, or equivalently, lim Revew for the prevous lecture Cocepts: order statstcs Theorems: Dstrbutos of order statstcs Examples: How to get the dstrbuto of order statstcs Chapter 5 Propertes of a Radom Sample Secto 55 Covergece

More information

Bounds on the expected entropy and KL-divergence of sampled multinomial distributions. Brandon C. Roy

Bounds on the expected entropy and KL-divergence of sampled multinomial distributions. Brandon C. Roy Bouds o the expected etropy ad KL-dvergece of sampled multomal dstrbutos Brado C. Roy bcroy@meda.mt.edu Orgal: May 18, 2011 Revsed: Jue 6, 2011 Abstract Iformato theoretc quattes calculated from a sampled

More information

1 Solution to Problem 6.40

1 Solution to Problem 6.40 1 Soluto to Problem 6.40 (a We wll wrte T τ (X 1,...,X where the X s are..d. wth PDF f(x µ, σ 1 ( x µ σ g, σ where the locato parameter µ s ay real umber ad the scale parameter σ s > 0. Lettg Z X µ σ we

More information

CHAPTER VI Statistical Analysis of Experimental Data

CHAPTER VI Statistical Analysis of Experimental Data Chapter VI Statstcal Aalyss of Expermetal Data CHAPTER VI Statstcal Aalyss of Expermetal Data Measuremets do ot lead to a uque value. Ths s a result of the multtude of errors (maly radom errors) that ca

More information

THE ROYAL STATISTICAL SOCIETY 2016 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5

THE ROYAL STATISTICAL SOCIETY 2016 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5 THE ROYAL STATISTICAL SOCIETY 06 EAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5 The Socety s provdg these solutos to assst cadtes preparg for the examatos 07. The solutos are teded as learg ads ad should

More information

Introduction to Probability

Introduction to Probability Itroducto to Probablty Nader H Bshouty Departmet of Computer Scece Techo 32000 Israel e-mal: bshouty@cstechoacl 1 Combatorcs 11 Smple Rules I Combatorcs The rule of sum says that the umber of ways to choose

More information

ENGI 4421 Joint Probability Distributions Page Joint Probability Distributions [Navidi sections 2.5 and 2.6; Devore sections

ENGI 4421 Joint Probability Distributions Page Joint Probability Distributions [Navidi sections 2.5 and 2.6; Devore sections ENGI 441 Jot Probablty Dstrbutos Page 7-01 Jot Probablty Dstrbutos [Navd sectos.5 ad.6; Devore sectos 5.1-5.] The jot probablty mass fucto of two dscrete radom quattes, s, P ad p x y x y The margal probablty

More information

Homework 1: Solutions Sid Banerjee Problem 1: (Practice with Asymptotic Notation) ORIE 4520: Stochastics at Scale Fall 2015

Homework 1: Solutions Sid Banerjee Problem 1: (Practice with Asymptotic Notation) ORIE 4520: Stochastics at Scale Fall 2015 Fall 05 Homework : Solutos Problem : (Practce wth Asymptotc Notato) A essetal requremet for uderstadg scalg behavor s comfort wth asymptotc (or bg-o ) otato. I ths problem, you wll prove some basc facts

More information

Part 4b Asymptotic Results for MRR2 using PRESS. Recall that the PRESS statistic is a special type of cross validation procedure (see Allen (1971))

Part 4b Asymptotic Results for MRR2 using PRESS. Recall that the PRESS statistic is a special type of cross validation procedure (see Allen (1971)) art 4b Asymptotc Results for MRR usg RESS Recall that the RESS statstc s a specal type of cross valdato procedure (see Alle (97)) partcular to the regresso problem ad volves fdg Y $,, the estmate at the

More information

Unimodality Tests for Global Optimization of Single Variable Functions Using Statistical Methods

Unimodality Tests for Global Optimization of Single Variable Functions Using Statistical Methods Malaysa Umodalty Joural Tests of Mathematcal for Global Optmzato Sceces (): of 05 Sgle - 5 Varable (007) Fuctos Usg Statstcal Methods Umodalty Tests for Global Optmzato of Sgle Varable Fuctos Usg Statstcal

More information

Analysis of Variance with Weibull Data

Analysis of Variance with Weibull Data Aalyss of Varace wth Webull Data Lahaa Watthaacheewaul Abstract I statstcal data aalyss by aalyss of varace, the usual basc assumptos are that the model s addtve ad the errors are radomly, depedetly, ad

More information

Bayes (Naïve or not) Classifiers: Generative Approach

Bayes (Naïve or not) Classifiers: Generative Approach Logstc regresso Bayes (Naïve or ot) Classfers: Geeratve Approach What do we mea by Geeratve approach: Lear p(y), p(x y) ad the apply bayes rule to compute p(y x) for makg predctos Ths s essetally makg

More information

Pseudo-random Functions

Pseudo-random Functions Pseudo-radom Fuctos Debdeep Mukhopadhyay IIT Kharagpur We have see the costructo of PRG (pseudo-radom geerators) beg costructed from ay oe-way fuctos. Now we shall cosder a related cocept: Pseudo-radom

More information

The Mathematical Appendix

The Mathematical Appendix The Mathematcal Appedx Defto A: If ( Λ, Ω, where ( λ λ λ whch the probablty dstrbutos,,..., Defto A. uppose that ( Λ,,..., s a expermet type, the σ-algebra o λ λ λ are defed s deoted by ( (,,...,, σ Ω.

More information

best estimate (mean) for X uncertainty or error in the measurement (systematic, random or statistical) best

best estimate (mean) for X uncertainty or error in the measurement (systematic, random or statistical) best Error Aalyss Preamble Wheever a measuremet s made, the result followg from that measuremet s always subject to ucertaty The ucertaty ca be reduced by makg several measuremets of the same quatty or by mprovg

More information

Summary of the lecture in Biostatistics

Summary of the lecture in Biostatistics Summary of the lecture Bostatstcs Probablty Desty Fucto For a cotuos radom varable, a probablty desty fucto s a fucto such that: 0 dx a b) b a dx A probablty desty fucto provdes a smple descrpto of the

More information

X X X E[ ] E X E X. is the ()m n where the ( i,)th. j element is the mean of the ( i,)th., then

X X X E[ ] E X E X. is the ()m n where the ( i,)th. j element is the mean of the ( i,)th., then Secto 5 Vectors of Radom Varables Whe workg wth several radom varables,,..., to arrage them vector form x, t s ofte coveet We ca the make use of matrx algebra to help us orgaze ad mapulate large umbers

More information

Ordinary Least Squares Regression. Simple Regression. Algebra and Assumptions.

Ordinary Least Squares Regression. Simple Regression. Algebra and Assumptions. Ordary Least Squares egresso. Smple egresso. Algebra ad Assumptos. I ths part of the course we are gog to study a techque for aalysg the lear relatoshp betwee two varables Y ad X. We have pars of observatos

More information

Derivation of 3-Point Block Method Formula for Solving First Order Stiff Ordinary Differential Equations

Derivation of 3-Point Block Method Formula for Solving First Order Stiff Ordinary Differential Equations Dervato of -Pot Block Method Formula for Solvg Frst Order Stff Ordary Dfferetal Equatos Kharul Hamd Kharul Auar, Kharl Iskadar Othma, Zara Bb Ibrahm Abstract Dervato of pot block method formula wth costat

More information

Lecture 3 Probability review (cont d)

Lecture 3 Probability review (cont d) STATS 00: Itroducto to Statstcal Iferece Autum 06 Lecture 3 Probablty revew (cot d) 3. Jot dstrbutos If radom varables X,..., X k are depedet, the ther dstrbuto may be specfed by specfyg the dvdual dstrbuto

More information

Parameter, Statistic and Random Samples

Parameter, Statistic and Random Samples Parameter, Statstc ad Radom Samples A parameter s a umber that descrbes the populato. It s a fxed umber, but practce we do ot kow ts value. A statstc s a fucto of the sample data,.e., t s a quatty whose

More information

F. Inequalities. HKAL Pure Mathematics. 進佳數學團隊 Dr. Herbert Lam 林康榮博士. [Solution] Example Basic properties

F. Inequalities. HKAL Pure Mathematics. 進佳數學團隊 Dr. Herbert Lam 林康榮博士. [Solution] Example Basic properties 進佳數學團隊 Dr. Herbert Lam 林康榮博士 HKAL Pure Mathematcs F. Ieualtes. Basc propertes Theorem Let a, b, c be real umbers. () If a b ad b c, the a c. () If a b ad c 0, the ac bc, but f a b ad c 0, the ac bc. Theorem

More information

Functions of Random Variables

Functions of Random Variables Fuctos of Radom Varables Chapter Fve Fuctos of Radom Varables 5. Itroducto A geeral egeerg aalyss model s show Fg. 5.. The model output (respose) cotas the performaces of a system or product, such as weght,

More information

L5 Polynomial / Spline Curves

L5 Polynomial / Spline Curves L5 Polyomal / Sple Curves Cotets Coc sectos Polyomal Curves Hermte Curves Bezer Curves B-Sples No-Uform Ratoal B-Sples (NURBS) Mapulato ad Represetato of Curves Types of Curve Equatos Implct: Descrbe a

More information

22 Nonparametric Methods.

22 Nonparametric Methods. 22 oparametrc Methods. I parametrc models oe assumes apror that the dstrbutos have a specfc form wth oe or more ukow parameters ad oe tres to fd the best or atleast reasoably effcet procedures that aswer

More information

Chapter 4 Multiple Random Variables

Chapter 4 Multiple Random Variables Revew for the prevous lecture: Theorems ad Examples: How to obta the pmf (pdf) of U = g (, Y) ad V = g (, Y) Chapter 4 Multple Radom Varables Chapter 44 Herarchcal Models ad Mxture Dstrbutos Examples:

More information

Solving Constrained Flow-Shop Scheduling. Problems with Three Machines

Solving Constrained Flow-Shop Scheduling. Problems with Three Machines It J Cotemp Math Sceces, Vol 5, 2010, o 19, 921-929 Solvg Costraed Flow-Shop Schedulg Problems wth Three Maches P Pada ad P Rajedra Departmet of Mathematcs, School of Advaced Sceces, VIT Uversty, Vellore-632

More information

Dimensionality Reduction and Learning

Dimensionality Reduction and Learning CMSC 35900 (Sprg 009) Large Scale Learg Lecture: 3 Dmesoalty Reducto ad Learg Istructors: Sham Kakade ad Greg Shakharovch L Supervsed Methods ad Dmesoalty Reducto The theme of these two lectures s that

More information

(b) By independence, the probability that the string 1011 is received correctly is

(b) By independence, the probability that the string 1011 is received correctly is Soluto to Problem 1.31. (a) Let A be the evet that a 0 s trasmtted. Usg the total probablty theorem, the desred probablty s P(A)(1 ɛ ( 0)+ 1 P(A) ) (1 ɛ 1)=p(1 ɛ 0)+(1 p)(1 ɛ 1). (b) By depedece, the probablty

More information

Algorithms Design & Analysis. Hash Tables

Algorithms Design & Analysis. Hash Tables Algorthms Desg & Aalyss Hash Tables Recap Lower boud Order statstcs 2 Today s topcs Drect-accessble table Hash tables Hash fuctos Uversal hashg Perfect Hashg Ope addressg 3 Symbol-table problem Symbol

More information

Qualifying Exam Statistical Theory Problem Solutions August 2005

Qualifying Exam Statistical Theory Problem Solutions August 2005 Qualfyg Exam Statstcal Theory Problem Solutos August 5. Let X, X,..., X be d uform U(,),

More information

Chapter 4 (Part 1): Non-Parametric Classification (Sections ) Pattern Classification 4.3) Announcements

Chapter 4 (Part 1): Non-Parametric Classification (Sections ) Pattern Classification 4.3) Announcements Aoucemets No-Parametrc Desty Estmato Techques HW assged Most of ths lecture was o the blacboard. These sldes cover the same materal as preseted DHS Bometrcs CSE 90-a Lecture 7 CSE90a Fall 06 CSE90a Fall

More information

arxiv:math/ v1 [math.gm] 8 Dec 2005

arxiv:math/ v1 [math.gm] 8 Dec 2005 arxv:math/05272v [math.gm] 8 Dec 2005 A GENERALIZATION OF AN INEQUALITY FROM IMO 2005 NIKOLAI NIKOLOV The preset paper was spred by the thrd problem from the IMO 2005. A specal award was gve to Yure Boreko

More information

The Selection Problem - Variable Size Decrease/Conquer (Practice with algorithm analysis)

The Selection Problem - Variable Size Decrease/Conquer (Practice with algorithm analysis) We have covered: Selecto, Iserto, Mergesort, Bubblesort, Heapsort Next: Selecto the Qucksort The Selecto Problem - Varable Sze Decrease/Coquer (Practce wth algorthm aalyss) Cosder the problem of fdg the

More information

MATH 247/Winter Notes on the adjoint and on normal operators.

MATH 247/Winter Notes on the adjoint and on normal operators. MATH 47/Wter 00 Notes o the adjot ad o ormal operators I these otes, V s a fte dmesoal er product space over, wth gve er * product uv, T, S, T, are lear operators o V U, W are subspaces of V Whe we say

More information

Mu Sequences/Series Solutions National Convention 2014

Mu Sequences/Series Solutions National Convention 2014 Mu Sequeces/Seres Solutos Natoal Coveto 04 C 6 E A 6C A 6 B B 7 A D 7 D C 7 A B 8 A B 8 A C 8 E 4 B 9 B 4 E 9 B 4 C 9 E C 0 A A 0 D B 0 C C Usg basc propertes of arthmetc sequeces, we fd a ad bm m We eed

More information

å 1 13 Practice Final Examination Solutions - = CS109 Dec 5, 2018

å 1 13 Practice Final Examination Solutions - = CS109 Dec 5, 2018 Chrs Pech Fal Practce CS09 Dec 5, 08 Practce Fal Examato Solutos. Aswer: 4/5 8/7. There are multle ways to obta ths aswer; here are two: The frst commo method s to sum over all ossbltes for the rak of

More information

Lecture Notes Types of economic variables

Lecture Notes Types of economic variables Lecture Notes 3 1. Types of ecoomc varables () Cotuous varable takes o a cotuum the sample space, such as all pots o a le or all real umbers Example: GDP, Polluto cocetrato, etc. () Dscrete varables fte

More information

Pseudo-random Functions. PRG vs PRF

Pseudo-random Functions. PRG vs PRF Pseudo-radom Fuctos Debdeep Muhopadhyay IIT Kharagpur PRG vs PRF We have see the costructo of PRG (pseudo-radom geerators) beg costructed from ay oe-way fuctos. Now we shall cosder a related cocept: Pseudo-radom

More information

Analysis of System Performance IN2072 Chapter 5 Analysis of Non Markov Systems

Analysis of System Performance IN2072 Chapter 5 Analysis of Non Markov Systems Char for Network Archtectures ad Servces Prof. Carle Departmet of Computer Scece U Müche Aalyss of System Performace IN2072 Chapter 5 Aalyss of No Markov Systems Dr. Alexader Kle Prof. Dr.-Ig. Georg Carle

More information

( ) = ( ) ( ) Chapter 13 Asymptotic Theory and Stochastic Regressors. Stochastic regressors model

( ) = ( ) ( ) Chapter 13 Asymptotic Theory and Stochastic Regressors. Stochastic regressors model Chapter 3 Asmptotc Theor ad Stochastc Regressors The ature of eplaator varable s assumed to be o-stochastc or fed repeated samples a regresso aalss Such a assumpto s approprate for those epermets whch

More information

Chapter 11 The Analysis of Variance

Chapter 11 The Analysis of Variance Chapter The Aalyss of Varace. Oe Factor Aalyss of Varace. Radomzed Bloc Desgs (ot for ths course) NIPRL . Oe Factor Aalyss of Varace.. Oe Factor Layouts (/4) Suppose that a expermeter s terested populatos

More information

Multivariate Transformation of Variables and Maximum Likelihood Estimation

Multivariate Transformation of Variables and Maximum Likelihood Estimation Marquette Uversty Multvarate Trasformato of Varables ad Maxmum Lkelhood Estmato Dael B. Rowe, Ph.D. Assocate Professor Departmet of Mathematcs, Statstcs, ad Computer Scece Copyrght 03 by Marquette Uversty

More information

STATISTICAL INFERENCE

STATISTICAL INFERENCE (STATISTICS) STATISTICAL INFERENCE COMPLEMENTARY COURSE B.Sc. MATHEMATICS III SEMESTER ( Admsso) UNIVERSITY OF CALICUT SCHOOL OF DISTANCE EDUCATION CALICUT UNIVERSITY P.O., MALAPPURAM, KERALA, INDIA -

More information

Random Variables and Probability Distributions

Random Variables and Probability Distributions Radom Varables ad Probablty Dstrbutos * If X : S R s a dscrete radom varable wth rage {x, x, x 3,. } the r = P (X = xr ) = * Let X : S R be a dscrete radom varable wth rage {x, x, x 3,.}.If x r P(X = x

More information

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE THE ROYAL STATISTICAL SOCIETY 00 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER I STATISTICAL THEORY The Socety provdes these solutos to assst caddates preparg for the examatos future years ad for the

More information

Lecture 07: Poles and Zeros

Lecture 07: Poles and Zeros Lecture 07: Poles ad Zeros Defto of poles ad zeros The trasfer fucto provdes a bass for determg mportat system respose characterstcs wthout solvg the complete dfferetal equato. As defed, the trasfer fucto

More information

MA 524 Homework 6 Solutions

MA 524 Homework 6 Solutions MA 524 Homework 6 Solutos. Sce S(, s the umber of ways to partto [] to k oempty blocks, ad c(, s the umber of ways to partto to k oempty blocks ad also the arrage each block to a cycle, we must have S(,

More information

18.413: Error Correcting Codes Lab March 2, Lecture 8

18.413: Error Correcting Codes Lab March 2, Lecture 8 18.413: Error Correctg Codes Lab March 2, 2004 Lecturer: Dael A. Spelma Lecture 8 8.1 Vector Spaces A set C {0, 1} s a vector space f for x all C ad y C, x + y C, where we take addto to be compoet wse

More information

Chapter 8: Statistical Analysis of Simulated Data

Chapter 8: Statistical Analysis of Simulated Data Marquette Uversty MSCS600 Chapter 8: Statstcal Aalyss of Smulated Data Dael B. Rowe, Ph.D. Departmet of Mathematcs, Statstcs, ad Computer Scece Copyrght 08 by Marquette Uversty MSCS600 Ageda 8. The Sample

More information

1 Convergence of the Arnoldi method for eigenvalue problems

1 Convergence of the Arnoldi method for eigenvalue problems Lecture otes umercal lear algebra Arold method covergece Covergece of the Arold method for egevalue problems Recall that, uless t breaks dow, k steps of the Arold method geerates a orthogoal bass of a

More information

Multiple Linear Regression Analysis

Multiple Linear Regression Analysis LINEA EGESSION ANALYSIS MODULE III Lecture - 4 Multple Lear egresso Aalyss Dr. Shalabh Departmet of Mathematcs ad Statstcs Ida Isttute of Techology Kapur Cofdece terval estmato The cofdece tervals multple

More information

3. Basic Concepts: Consequences and Properties

3. Basic Concepts: Consequences and Properties : 3. Basc Cocepts: Cosequeces ad Propertes Markku Jutt Overvew More advaced cosequeces ad propertes of the basc cocepts troduced the prevous lecture are derved. Source The materal s maly based o Sectos.6.8

More information

Idea is to sample from a different distribution that picks points in important regions of the sample space. Want ( ) ( ) ( ) E f X = f x g x dx

Idea is to sample from a different distribution that picks points in important regions of the sample space. Want ( ) ( ) ( ) E f X = f x g x dx Importace Samplg Used for a umber of purposes: Varace reducto Allows for dffcult dstrbutos to be sampled from. Sestvty aalyss Reusg samples to reduce computatoal burde. Idea s to sample from a dfferet

More information

PPCP: The Proofs. 1 Notations and Assumptions. Maxim Likhachev Computer and Information Science University of Pennsylvania Philadelphia, PA 19104

PPCP: The Proofs. 1 Notations and Assumptions. Maxim Likhachev Computer and Information Science University of Pennsylvania Philadelphia, PA 19104 PPCP: The Proofs Maxm Lkhachev Computer ad Iformato Scece Uversty of Pesylvaa Phladelpha, PA 19104 maxml@seas.upe.edu Athoy Stetz The Robotcs Isttute Carege Mello Uversty Pttsburgh, PA 15213 axs@rec.r.cmu.edu

More information

UNIT 2 SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS

UNIT 2 SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS Numercal Computg -I UNIT SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS Structure Page Nos..0 Itroducto 6. Objectves 7. Ital Approxmato to a Root 7. Bsecto Method 8.. Error Aalyss 9.4 Regula Fals Method

More information

Chapter 14 Logistic Regression Models

Chapter 14 Logistic Regression Models Chapter 4 Logstc Regresso Models I the lear regresso model X β + ε, there are two types of varables explaatory varables X, X,, X k ad study varable y These varables ca be measured o a cotuous scale as

More information

Complete Convergence and Some Maximal Inequalities for Weighted Sums of Random Variables

Complete Convergence and Some Maximal Inequalities for Weighted Sums of Random Variables Joural of Sceces, Islamc Republc of Ira 8(4): -6 (007) Uversty of Tehra, ISSN 06-04 http://sceces.ut.ac.r Complete Covergece ad Some Maxmal Iequaltes for Weghted Sums of Radom Varables M. Am,,* H.R. Nl

More information

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA THE ROYAL STATISTICAL SOCIETY 3 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA PAPER I STATISTICAL THEORY & METHODS The Socety provdes these solutos to assst caddates preparg for the examatos future years ad

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Marquette Uverst Maxmum Lkelhood Estmato Dael B. Rowe, Ph.D. Professor Departmet of Mathematcs, Statstcs, ad Computer Scece Coprght 08 b Marquette Uverst Maxmum Lkelhood Estmato We have bee sag that ~

More information

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ " 1

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ  1 STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS Recall Assumpto E(Y x) η 0 + η x (lear codtoal mea fucto) Data (x, y ), (x 2, y 2 ),, (x, y ) Least squares estmator ˆ E (Y x) ˆ " 0 + ˆ " x, where ˆ

More information

1 0, x? x x. 1 Root finding. 1.1 Introduction. Solve[x^2-1 0,x] {{x -1},{x 1}} Plot[x^2-1,{x,-2,2}] 3

1 0, x? x x. 1 Root finding. 1.1 Introduction. Solve[x^2-1 0,x] {{x -1},{x 1}} Plot[x^2-1,{x,-2,2}] 3 Adrew Powuk - http://www.powuk.com- Math 49 (Numercal Aalyss) Root fdg. Itroducto f ( ),?,? Solve[^-,] {{-},{}} Plot[^-,{,-,}] Cubc equato https://e.wkpeda.org/wk/cubc_fucto Quartc equato https://e.wkpeda.org/wk/quartc_fucto

More information

Multiple Regression. More than 2 variables! Grade on Final. Multiple Regression 11/21/2012. Exam 2 Grades. Exam 2 Re-grades

Multiple Regression. More than 2 variables! Grade on Final. Multiple Regression 11/21/2012. Exam 2 Grades. Exam 2 Re-grades STAT 101 Dr. Kar Lock Morga 11/20/12 Exam 2 Grades Multple Regresso SECTIONS 9.2, 10.1, 10.2 Multple explaatory varables (10.1) Parttog varablty R 2, ANOVA (9.2) Codtos resdual plot (10.2) Trasformatos

More information

Laboratory I.10 It All Adds Up

Laboratory I.10 It All Adds Up Laboratory I. It All Adds Up Goals The studet wll work wth Rema sums ad evaluate them usg Derve. The studet wll see applcatos of tegrals as accumulatos of chages. The studet wll revew curve fttg sklls.

More information

D KL (P Q) := p i ln p i q i

D KL (P Q) := p i ln p i q i Cheroff-Bouds 1 The Geeral Boud Let P 1,, m ) ad Q q 1,, q m ) be two dstrbutos o m elemets, e,, q 0, for 1,, m, ad m 1 m 1 q 1 The Kullback-Lebler dvergece or relatve etroy of P ad Q s defed as m D KL

More information