Beating the Adaptive Bandit with High Probability

Size: px
Start display at page:

Download "Beating the Adaptive Bandit with High Probability"

Transcription

1 Beaing he Adapive Bandi wih High Probabiliy Jacob Abernehy Compuer Science Division UC Berkeley Alexander Rakhlin Deparmen of Saisics Universiy of Pennsylvania Absrac We provide a principled way of proving Õ ) high-probabiliy guaranees for parial-informaion bandi) problems over arbirary convex decision ses. Firs, we prove a regre guaranee for he full-informaion problem in erms of local norms, boh for enropy and self-concordan barrier regularizaion, unifying hese mehods. Given one of such algorihms as a black-box, we can conver a bandi problem ino a full-informaion problem using a sampling scheme. he main resul saes ha a high-probabiliy Õ ) bound holds whenever he black-box, he sampling scheme, and he esimaes of missing informaion saisfy a number of condiions, which are relaively easy o check. A he hear of he mehod is a consrucion of linear upper bounds on confidence inervals. As applicaions of he main resul, we provide he firs known efficien algorihm for he sphere wih an Õ ) high-probabiliy bound. We also derive he resul for he n-simplex, improving he O p n logn )) bound of Auer e al [3] by replacing he log erm wih log log and closing he gap o he lower bound of Ω n ). While Õ ) high-probabiliy bounds should hold for general decision ses hrough our main resul, consrucion of linear upper bounds depends on he paricular geomery of he se; we believe ha he sphere example already exhibis he necessary ingrediens. he guaranees we obain hold for adapive adversaries unlike he in-expecaion resuls of [1]) and he algorihms are efficien, given ha he linear upper bounds on confidence can be compued. I. INRODUCION he problem of Online Convex Opimizaion, in which a player aemps o minimize his regre agains a possibly adversarial sequence of convex cos funcions, is now quie well-undersood. he more recen research rend has been o consider various limied-informaion versions of his problem. In paricular, he bandi version of Online Linear Opimizaion OLO) has received much aenion in he pas few years. o be precise, he problem we are ineresed in can be phrased as a repeaed game beween he player and he adversary. A each round, he player picks a decision from he allowed convex se of moves, and he adversary simulaneously picks a linear cos funcion from her allowed se of moves. Unlike he well-undersood OLO game, in he bandi version only he cos of he decision is revealed o he player, no he cos funcion iself. he adversary, on he oher hand, is aware of he complee hisory. he aim of he player is o minimize regre, he cumulaive cos incurred over he course of he game minus he cumulaive cos of he bes fixed decision. he scarciy of informaion revealed o he player makes he problem difficul. he firs efficien algorihm wih an Õ ) guaranee on he regre for opimizaion over arbirary convex ses was recenly obained in [1]. his guaranee was shown o hold in expecaion and he quesion of obaining guaranees in high probabiliy was lef open. In his paper, we develop a general framework for obaining high-probabiliy saemens for bandi problems. We aim o provide a clean picure, building upon he mechanism employed in [3], [5]. We also simplify he proof of [1] for he regre of regularizaion wih a self-concordan barrier and pu i ino he conex of a general class of regre bounds based on local norms. A reader surveying he lieraure on bandi opimizaion can easily ge confused rying o disinguish beween he resuls. hus, we firs iemize some recen papers according o he following crieria: a) efficien algorihm vs inefficien algorihm, b) arbirary convex se vs simplex or he se of flows in a graph, c) opimal Õ ) vs subopimal e.g. O 2/3) ) guaranee, d) in-expecaion vs high-probabiliy guaranee, and e) wheher he resul holds for an adapive adversary or only an oblivious one. For all he resuls we are aware of including he ones in his paper), a high-probabiliy guaranee on he regre naurally covers he case of an adapive adversary. his is no necessarily rue for he in-expecaion resuls. Wih respec o hese parameers, Auer e al [3] obained an efficien algorihm for he simplex, wih an opimal guaranee which holds in high probabiliy. McMahan and Blum [13] and Flaxman e al [11] obained efficien algorihms for an arbirary convex se wih subopimal guaranees which hold in expecaion agains an adapive adversary. Awerbuch and Kleinberg [4] obained an efficien algorihm for he se of flows wih a subopimal guaranee which holds in expecaion agains an adapive adversary. György e al [12] obained an efficien algorihm for he se of flows wih a subopimal guaranee which holds in high probabiliy. 1 Dani e al [9] obained an inefficien algorihm for an arbirary se, wih an opimal guaranee which holds in expecaion agains an oblivious adversary. he algorihm can be implemened efficienly for he se of flows. Barle e al [5] exended he resul of [9] o obain an inefficien algorihm for an arbirary se, wih an opimal 1 he auhors also obained an opimal guaranee for he se of flows in he seing where he lenghs of all edges on he chosen pah are revealed. his does no mach he bandi problem considered in his paper.

2 guaranee which holds in high probabiliy. he algorihm canno be in a sraighforward way) implemened efficienly for he se of flows. Abernehy e al [1] exhibied an efficien algorihm for an arbirary convex se, wih an opimal guaranee which holds in expecaion agains an oblivious adversary. In his paper, we obain an efficien algorihm for a sphere and simplex wih an opimal guaranee which holds in high probabiliy and, hus, agains an adapive adversary). Analogous resuls can be obained for oher convex ses; however, such resuls would have o be considered on he per-case basis, as he specific geomery of he se plays an imporan role for obaining an efficien algorihm wih an opimal high-probabiliy guaranee. his paper is organized as follows. In Secion II, we discuss full-informaion algorihms which will be used as blackboxes for bandi opimizaion. In Secion II-B we prove he known regre guaranees which arise from regularizaion wih a srongly convex funcion. We argue ha hese guaranees are no srong enough o be used for bandi opimizaion and, in Secion II-C, we inroduce a noion of local norms. We prove general regre guaranees wih respec o hese norms for regularizaion wih a self-concordan barrier and, for he case of he n-simplex, wih he enropy funcion. his allows us o have a unified analysis of bandi opimizaion wih eiher of hese wo mehods as a black-box. Secion III discusses he mehod of using a randomized algorihm for convering a full-informaion algorihm ino a bandi algorihm. We discuss he advanages of high-probabiliy resuls over he in-expecaion resuls and explain why he sraighforward way of applying concenraion inequaliies does no work. Secion IV conains he main resuls of he paper. We sae he main resul, heorem 4.1, and hen apply i o various seings in he subsequen secions. he muliarmed bandi seing he simplex case) is considered in Secion V-A, and we improve upon he resul of Auer e al [3] by removing he log facor. We provide a soluion for he sphere in Secion V- B. In passing, we menion how he in-expecaion resul for general convex ses of [1] immediaely follows heorem 2.3. Anoher sampling scheme for general bodies is suggesed, alhough we do no go ino he deails. he proof of our main resul, heorem 4.1, is given in Secion VI. I is based on lemmas whose proofs can be found in he echnical repor [2]. II. FULL-INFORMAION ALGORIHMS In his paper, we srive o obain he mos general resuls possible. o his end, bandi algorihms in Secion IV will ake as a sub-rouine an absrac full-informaion black-box for regre minimizaion. We devoe he presen secion o describing known guaranees for some full-informaion algorihms, as well as o developing a new family of guaranees under local norms. he laer are suied o he sudy of bandi opimizaion. o make hings concree, he full-informaion seing is ha of online linear opimizaion, which is phrased as he following game beween he learner player) and he environmen adversary). Le K R n be a closed convex se. A each ime sep = 1 o, Player chooses K Adversary independenly chooses f R n Player suffers loss f and observes f he aim of he player algorihm) is o minimize he regre agains any comparaor u K A. Algorihms R u) := f f u. Le Rx) be a convex funcion. We consider he following family wih respec o he choice of R) of Follow he Regularized Leader algorihms: Algorihm 1 Follow he Regularized Leader FRL) Inpu: η > 0. On he firs round, play x 1 := arg min x K Rx). On round + 1, play [ ] +1 := arg min x K η fs x + Rx) s=1. 1) Wihou loss of generaliy, we assume ha R akes is minimum a 0, since arg min is he same modulo consan shifs of R. We begin wih a well-known fac, whose easy inducion proof can be found e.g. in [16]. Proposiion 2.1: he regre of Algorihm 1, relaive o a comparaor u K, can be upper bounded as R u) f +1 ) + η 1 Ru). 2) he FRL algorihm is closely relaed o he following Mirror Descen-syle algorihm [8], [16]. Algorihm 2 Mirror Descen wih Projecions On he firs round, play x 1 := arg min x K Rx). On round + 1, compue +1 := arg min x R n ηf x + D R x, ) and hen play he projeced poin +1 := arg min x K D Rx, +1 ) his algorihm is given in wo seps alhough i can be described in one. Indeed, he poin +1 can simply be obained as he soluion o arg min x K ηf x + D R x, ). However, we emphasize he unprojeced poin +1 as i gives us an occasionally more useful regre bound:

3 Proposiion 2.2: he regre of Algorihm 2, relaive o a comparaor u K, can be upper bounded as R u) f +1 ) + η 1 Ru). 3) he analogue of Proposiion 2.1 also holds: R u) f +1 ) + η 1 Ru). 4) We also noe ha he wo algorihms coincide if R is a barrier. We refer o [16] for he proofs of hese facs. B. Regre Bounds wih Respec o Fixed Norms he regre bounds saed in Proposiions 2.1 and 2.2 are no ulimaely saisfying. In paricular, i is no immediaely obvious wheher he erms f +1 ) are small. Noice ha he poin +1 depends on boh f as well as on he behavior of R. I would be much more appealing if we could remove he dependence on he poins and have he regre depend solely on he Adversary s choices f and our choice of regularizer. his can indeed be achieved if we require cerain condiions on our regularizer. he ypical approach is o require ha R is srongly convex wih respec o some norm, which implies ha +1 2 R ) R+1 ), +1 5) R ) R+1 ) +1. where is he norm dual o, and he las sep follows by Hölder s Inequaliy. Hence, srong convexiy of R implies +1 R ) R+1 ), making possible he following resul. Proposiion 2.3: When R is srongly convex wih respec o he norm, hen for Algorihms 1 and 2 we have he following regre bound 2 : R u) η f 2 + η 1 Ru). Proof: For he case of FRL Algorihm 1), when R is a barrier funcion and hus is always aained on he inerior of K) i is a convenien fac ha R ) R+1 ) = ηf. Applying Hölder s inequaliy in he saemen of Proposiion 2.1 leads o he desired resul. If R is no a barrier, an applicaion of he Kolmogorov crierion see [7], heorem 2.4.2) for generalized projecions a sep 5) yields he saemen of he Proposiion. For Algorihm 2, he proof is a bi more involved, bu is well-known see e.g. [6]). Again, we refer he reader o [16], [17] for deails. he easies way o see Proposiion 2.3 a work is o assume ha f B p and K B q, he uni zero-cenered balls wih respec o l p and l q norms, where p, q) is a dual pair. When faced wih he paricular choice of l, l 1 ) pair of norms, he 2 We also menion ha a more refined proof leads o a consan of 1 2 insead of 1 in fron of he η P f 2 erm. naural choice of regularizaion is he unnormalized negaive enropy funcion Rx) = i x[i] log x[i] x[i]) log n), 6) defined over he posiive orhan. Here he 1 + log n erm ensures ha min R = 0 over he n-simplex K. I is easy o see ha his regularizaion funcion leads o he so-called exponenial weighs : exp η ) s=1 f [i] +1 [i] = n j=1 η exp ), s=1 f [j] and indeed his is rue for boh Algorihm 1 and Algorihm 2. For he fuure, i is useful o noe ha he unprojeced updaed +1 has he very simple unnormalized form : +1 [i] = [i] exp ηf [i]). 7) I is well-known ha he enropy funcion has he useful propery of srong convexiy wih respec o he l 1 norm. We can hus apply Proposiion 2.3 o obain: R u) η f 2 + η 1 log N. where he log N arises by aking R ) a any corner of he n-simplex. In he exper seing i is ypical o assume ha f 1, and so seing η = log N)/ appropriaely we obain R u) η + η 1 log N = 2 log N. C. Regre Bounds wih Respec o Local Norms he analysis of Proposiion 2.3 is he ypical approach, and indeed i can be shown ha he above bound for exponenial weighs is very igh, i.e. wihin a small consan facor from opimal. On he oher hand, here are imes when we canno make he assumpion ha f is bounded wih respec o a fixed norm. his is paricularly relevan in he bandi seing, when we will be esimaing he funcions f ye our esimaes will blow up depending on he locaion of he poin. In such cases, o obain igher bounds, i will be necessary o measure he size of f wih respec o a changing norm. While i may no be obvious a presen, he ideal choice of norm is he inverse Hessian of R a he poin. From now on, define z x := z 2 Rx)z, where z R n is arbirary and where R is assumed o be he regularizer in quesion. he dual of his norm z x is idenically he norm wih respec o he inverse Hessian, i.e. z x := z 2 Rx) 1 z. Our goal will now be o obain bounds of he form R u) η f ) 2 + η 1 Ru). 8) Le us inroduce he following shorhand: z := z x for he norm defined wih respec o.

4 For he case when Rx) = x 2 2 leading o he online gradien descen algorihm), his bound is easy: since 2 Rx) = I n, and R is srongly convex wih respec o he l 2 norm, we already know ha R u) η f 2 2+η 1 Ru) = η f ) 2 +η 1 Ru). 1) Regre guaranee for he enropy regularizer.: For he enropic regularizaion case menioned above, proving a regre bound wih respec o he local norm x requires a lile bi more work. Firs noice ha 2 Rx) = diagx[1] 1,..., x[n] 1 ), and ha 1 e x x for all real x. Nex, using Eq. 7), +1 = n [i] +1 [i]) 2 / [i] = n [i]1 e ηf[i] ) 2 η n [i]f [i] 2 = η f. Now we make special use of Proposiion 2.2. By Hölder s Inequaliy, R u) η f ) 2 + η 1 Ru). I can be verified ha Algorihms 1 and 2 produce he same when R is he enropy funcion and K is he simplex. hus, we have proved he following heorem. heorem 2.1: he exponenial weighs algorihm eiher Algorihm 1 or Algorihm 2) enjoys he following bound in erms of local norms: R u) η f ) 2 + η 1 Ru). As a side remark, we menion ha one can prove he same guaranee wih a slighly worse consan) by saring from Eq. 2) insead of Eq. 3). A lemma which can be found in he Appendix of [2], implies ha +1 2 = = β [i] 1 n [i] e ηf[i]) 2 n ) 2 e ηf[i] n j=1 [j]e ηf[j] ) 2 1 j=1 [j]e ηf[j] [i]ηf [i]) 2 = β ηf ) 2 for a small consan β. 2) Regre guaranee for he self-concordan regularizer.: I was shown in [1] ha, for he case of linear bandi opimizaion, he regularizaion funcion mus have he propery ha i curves srongly near he boundary. Indeed, i was observed ha he Hessian of R mus behave roughly as inverse disance 1/d, or even inverse squared disance 1/d 2, o he boundary. Indeed, he enropy funcion discussed above possesses he former propery on he n-simplex, bu funcions wih his 1/d growh propery are no readily available for general convex ses. o obain a funcion whose Hessian grows as 1/d 2 is much easier: he self-concordan barrier, commonly known as log barrier, is he cenral objec of sudy in Inerior Poin Mehods. In paricular, self-concordan barriers always exis and can be efficienly compued for many known bodies see, e.g., [14]). For a convex se wih linear consrains, he ypical choice of a self-concordan barrier is simply he sum of negaive log disance o each boundary. ha is, if he se is defined by Ax b, hen we would le R ) = i logb i e i Ax). I is rue ha, up o a consan, R is srongly convex wih respec o he l 2 norm, and we can hen easily prove a bound in erms of f 2 2. On he oher hand, i is precisely he case of bandi linear opimizaion for which i is useful o bound he regre in erms of he local norms f as in 8). I was shown in [1] ha he Hessian of a self-concordan barrier no only plays a crucial role in bounding he regre, bu also gives a handle on he local geomery hrough he noion of a Dikin ellipsoid. We refer he reader o [1] for more informaion on he Dikin ellipsoid and is relaion o sampling. As before, we can use Hölder s in equaliy o bound f +1 ) f +1, and now, as in he previous secion, we would like o replace +1 wih he dual norm η f. While i is no immediaely obvious how his should be accomplished, we can appeal o several nice resuls abou self-concordan funcions which makes our job easy. Define he objecive of Algorihm 1 as Φ x) = η s=1 f x + Rx). Since he barrier R goes o infiniy a he boundary of he se K, we have ha +1 is he unconsrained minimizer of Φ. o begin our shor journey o he land of Inerior Poin Mehods, define he Newon decremen for Φ as λx, Φ ) := Φ x) x = 2 Φ x) 1 Φ x) x and noe ha since R is self-concordan hen so is Φ. he above quaniy can be used o measure roughly how far a poin is from he global opimum: heorem 2.2 e.g. [14]): For any self-concordan funcion g, whenever λx, g) 1/2, we have x arg min g x 2λx, g) where he local norm x is defined wih respec o g, i.e. y x := y 2 gx))y. We can immediaely apply his heorem using he objecive Φ and he poin. Recalling ha 2 Φ = 2 R, we see ha, under he condiions of he heorem, +1 = arg min Φ 2λ, Φ ) = 2η f he las equaliy holds because, as is easy o check, Φ ) = ηf. We herefore have

5 heorem 2.3: Suppose for all {1... } we have η f 1 2, and R ) is self-concordan. hen R u) 2η [ f ] 2 + η 1 Ru). Given heorem 2.3, he resul of Abernehy, Hazan, and Rakhlin [1] follows immediaely, as we show in Secion V-C. III. BANDI FEEDBACK In he bandi version of online linear opimizaion, he funcion f is no revealed o us excep for is value a. he mechanism employed by all algorihms known o he auhors is o consruc a biased or unbiased esimae f of he vecor f from he single number revealed o us and feed i o he black box full-informaion algorihm. In order o consruc f, he algorihm has o randomly sample y around insead of deerminisically playing. Hence, he emplae bandi algorihm is: a round o predic y such ha Ey, obain f y, consruc f, feed i ino he black box, and obain he new +1. he paricular mehod for sampling y and consrucing f will be called he sampling scheme. he regre of he above procedure, relaive o a comparaor u, is R u) = f y u). However, he guaranees for he black-box are for a differen quaniy, which we denoe as R u) = f u). Le E denoe he condiional expecaion, given he random variables for ime seps If i is he case ha E f = f and E y =, hen for any fixed u, ] E [ f u) = E [f y u)]. 9) We conclude ha ER u) = E Ru). Hence, expeced regre agains a fixed u can be bounded hrough he expeced regre of he black-box. here are wo downsides o he above argumen. he firs is ha an in expecaion resul is much weaker han he corresponding high probabiliy saemen as he variance of he quaniies involved can be and, in fac, is) very large. I is no very saisfying o say ha he regre is of he correc order in expecaion bu has flucuaions of a higher order of magniude. he second weakness is in he fac ha u is fixed and, herefore, canno depend on he random moves of he player; in oher words, he adversary mus be oblivious. Boh of he downsides are overcome by proving a high probabiliy guaranee. I is emping o use he following incorrec) argumen for proving a high-probabiliy bound on R u) given an Õ ) bound on E R u): o obain a high-probabiliy bound, fix a u K and use Azuma-Hoeffding inequaliy o show an O ) concenraion of R u) around ER u). Nex, replace ER u) by E R u), which is Õ ), and ake a union bound over a discreizaion of u. he las sep only inroduces a log facor ino he bound, as we discuss laer. his approach fails 3 for he simple reason ha hrough he maringale difference argumen R u) is concenraed around he sum of condiional expecaions E f u), no he full expecaion E f u). he sum of condiional expecaions of f y u) erms is indeed equal o he sum of condiional expecaions of f u) erms. However, we do no know how o bound he laer: he regre guaranee for he black-box comes for he expeced regre, no he sum of condiional expecaions, hus breaking he argumen. Indeed, for proving high probabiliy bounds, a more refined analysis is needed. We ry o convey he big picure in he nex secion and illusrae i by proving a high-probabiliy bound for he sphere and he simplex, using he regularizaion wih selfconcordan barrier and enropy, respecively, as black-boxes. IV. HIGH PROBABILIY BOUNDS We now presen a emplae algorihm for bandi opimizaion. We assume ha a full-informaion black-box algorihm for linear opimizaion is available o us. A each ime sep = 1 o, Decide on he sampling scheme for his round, i.e. consruc a disribuion for y wih E y. Draw a sample y K from he disribuion and observe he loss f y. Consruc f such ha E f = f. Consruc a linear bias-funcion g u) = g u + µ. Feed f α g ino he black-box and receive +1. he algorihm requires wo parameers, α and η, which in urn depend on various aspecs of he problem. he following is he main resul of he paper. heorem 4.1: Suppose f B p for all and K B q, where log2 log )/δ p and q are dual. Le α = ) n. Suppose we can find c 1, c 2, c 3, c 4, c 5, c 6 0, such ha for all {1,..., } all of he following hold: A) he black-box full informaion algorihm enjoys a regre bound of he form R u) c 1 η [ f ] 2 + η 1 Ru) wih he local norm defined by 2 R ). B) E y q c 2 n. C) f u c 3 n for all u K. D) We can consruc a linear funcion g u) = g u + µ such ha u) E f f u) g u) u K and g ) c 4 n. [ ] 2 E) he consrucion saisfies f α g c5. [ ] 2 F ) On average, he norm is small: E f α g c6. 3 We hank Ambuj ewari for very helpful discussions in undersanding his.

6 G) Condiions for he regre bound in A) o hold are saisfied e.g. η f α g for log-barrier ) 1 2 hen, for any fixed u K, wih probabiliy a leas 1 δ + δ + δ ) where and f y u) η 1 Ru) + η A 1 + A 2, ) A 1 = c 1 c 6 + c 5 8 log1/δ ) A 2 = 8 log1/δ)+c 2 n+2c3 +c 4 +2) n log2 log )/δ ). Remark 4.1: As long as c 1,..., c 6 depend only weakly e.g. logarihmically) on, we obain he opimal Õ ) dependence by seing η 1/2. he growh of he bound in erms of n depends on he problem a hand and he sampling mehod. Remark 4.2: o obain a saemen wih probabiliy a leas 1 δ, for all u he guaranee holds, a union bound needs o be aken. For a se K, which can be represened as a convex hull of a number of is verices, he union bound inroduces an exra logarihm of his number of verices see he simplex example below). For a se such as sphere, an exra sep of discreizing he se ino a fine grid and aking a union over his exponenial) discreizaion is required. his echnique can inroduce an exra n log ino he bound see [10], [5] for deails). Since his sep depends on he paricular K a hand, we leave i ou of he main resul. Remark 4.3: he requiremen B) is a relaxaion of E y =. his slack is absoluely crucial for D) o be even possible. In he simplex case he slack corresponds o mixing in a uniform disribuion, which Auer e al [3] inerpre as an exploraion sep. For he sphere case, i corresponds o saying O 1/2 ) away from he boundary. From he poin of view of he proof, he relaxaion allows us o consruc g, i.e. o conrol he sum of condiional variances of f u). We noe ha he slack is no necessary for bounding he expeced regre only. his poins o he large variance of he esimaes and he weakness of he in-expecaion resuls. A. A Proof Skech Le us skech he mechanism for proving high-probabiliy bounds, which is applicable o a wide variey of ses and assumpions. We already menioned ha R u) is concenraed, for a fixed u K around he sum of condiional expecaions E f y u) wih ypical deviaions of O ). he laer is equal o he sum of condiional expecaions E f u). he ricky par is in proving ha R u) is concenraed around his sum. he ypical flucuaions of Ru) are more han, as he magniude of f depends on. hus, he only saemen we can make is ha, wih high probabiliy, E f u) f u) + c Var, where Var is he sum of condiional variances, growing faser han linear in. he magic comes from spliing he Var erm ino erms by he arihmeic-geomeric mean inequaliy and absorbing each of hese erms ino f, hereby biasing he esimaes. A a high level, we are adding he sandard deviaion a each ime sep o he esimaes f. Since his confidence inerval is a concave funcion, he black-box opimizaion over he modified f s will no work; he second magic sep due o his paper) is o find a linear funcion which uniformly bounds he confidence over he whole se K. If his can be done, he modified linear funcions are fed o he black-box, which enjoys an upper bound of η f ) 2, wih he norms of modified funcions. Finally, we show ha his quaniy is concenraed around he sum of condiional expecaions of he erms wih he ypical deviaions of O ), and he sum of condiional expecaions iself is bounded by O ) if f s have been consruced carefully. he las resul criically depends on availabiliy of a regre guaranee wih local norms, which have been exhibied earlier in he paper. he above paragraph is an informal descripion of he proof, which can be found in Secion VI. We refer o [2] for he deails. V. APPLICAIONS: HEOREM 4.1 A WORK For he sampling schemes below, we show ha our consrucion saisfies condiions of heorem 4.1, implying a highprobabiliy guaranee of Õ ). For each scheme, we provide a visual depicion of he disribuion from which we draw y. he size of he dos represens he relaive probabiliy mass, while he doed ellipsoid represens a sphere in he local norm a. Noe ha in he case of self-concordan R, his ellipsoid he Dikin ellipsoid) is conained in he se, which allows us o sample from is eigenvecors see [1]). A. Example 1: Soluion for he simplex his case corresponds o he non-sochasic muliarmed bandi problem [3]. We assume ha K is he simplex i.e. q = 1) and 0 f [i] 1 p = ). Regularizer R: We se our regularizaion funcion o be he enropy 6) and use Algorihm 1 or 2 as he black-box. Sampling of y : Le = n. Given he poin in he simplex, sample y = e i wih prob. p [i] := 1 ) [i] + /n. Consrucion of f : Given he above sampling scheme, we define our esimaes f he usual way: f = f e i )e i p [i] = f [i]e i p [i] when y = e i. 10)

7 Consrucion of g : he following g is appropriae for his problem: e i g u) := 2 + u p [i]. Before we ge sared, we noe a couple of useful facs ha we use several imes below: [i] p [i] p [i] 1 n 1 Now we check he condiions of he heorem. A) Since we are using enropy as our regularizaion, we have already shown in heorem 2.1 how o obain he necessary bound wih c 1 = 1. B) Noice ha Ey = 1 ) +unifn) and hus Ey 1 = unifn) 1 2 n, i.e. c 2 = 2. C) Since u is in he simplex, we see ha c 3 = 1: f u max i f [i] max p [i] 1 n i = n. D) We check ha g does indeed bound he variance. We can firs compue ) 2 f [i] E f f = p [i] e i e i p [i] 1 e i e i. p [i] We can now upper bound he variance of he esimaed losses, bu we need o do his on he enire simplex. Forunaely, since we are upper bounding a quadraic a convex funcion) i suffices o check he corners u = e i : x e i ) [j] 1[i = j]) 2 E f f e i ) p [j] j=1 < 1 p [i] + [j] 2 1 p [j] p [i] + p [j]) 2 1 ) 2 p [j] j i j i p [i] = g e i ). where we use he fac ha 1 ) 2 1/2 when 16n. Addiionally, we see ha c 4 = 3: g ) = 2 + n p [i] p [i] 2 + [i] 1 p [i] 3n. E) We now check ha, in he -norm, he biased esimae is no oo big. I is easy o check ha 2 R = [i] 1 e i e i 2 R 1 = [i]e i e i. Now assume y = e j, we can bound: ) 2 1[i = j] α f α g 2 = [i] p [i] ) 2 ) 1[i = j] α p [i] p [i] 1 2 α2 n 2 n1 α)2 + 2 Subsiuing = n, we obain c 5 = 2α 2 n 3/2 + 2 n1 α) 2. F ) We also mus check ha, in expecaion, he biased esimae is of consan order in he -norm: and [ ] 2 [ E f α g E 2 f 2 + 2α 2 g 2 n ) 1 2 p [i] p [i] 2 [i] + α 2 1 p [i] 2 [i] 2 n ) n + α n + α2 n 2 ) 1 p [i] = 4n + α 2 )n 3/2 ) =: c 6. We conclude ha ] A 1 = 4n + α 2 )n 3/2 ) + 2α 2 n 3/2 + 2 n1 α) 2) 8 log1/δ ) A 2 = 8 log1/δ) + 2 n + 7 n log2 log )/δ ). Now we swich o he Big-O noaion o elucidae he dependence on and n. Recalling ha α 2 = Olog log /n )), we observe ha A 1 = On) and A 2 = O n log log ). heorem 4.1 now saes ha wih probabiliy a leas 1 δ + δ + δ ), f y u) η 1 Ru) + η A 1 + A 2 for any fixed u. Since he regre is a linear funcional, i aains is maximum a one of he verices of he simplex. Hence, unlike in he nex secion, we only need o ake a union bound over hese verices o arrive a a saemen for all u K. We hus se δ = δ = δ = δ /n. Observe ha A 1 s asympoic dependence on n does no change, while A 2 now becomes O n logn log )). For any verex u, he shifed) enropy is Ru) = log n. log n Seing η = n, we conclude ha, wih high probabiliy, u K, f y u) = O n logn log )). his bound improves upon he resul of Auer e al [3], who obained an O n logn )) bound for he problem Algorihm EXP3.P). Our bound replaces he log erm wih log log, closing he gap o he lower bound of Ω n ). We conjecure ha log log growh in erms of is he mos sharp bound possible, due o he Law of he Ieraed Logarihm. In he full version of he paper, we will use sharper concenraion inequaliies o keep log log under he square roo.

8 B. Example 2: Soluion for he Euclidean sphere Suppose ha K = B 2 R n and ha he choices of he adversary are also l 2 -bounded by 1, i.e. p = q = 2. We poin ou ha wih he sampling scheme of [1] i is impossible o consruc g o saisfy he requiremens of heorem 4.1 see also Secion V-C). he modified sampling procedure below is key o reducing he variance of he esimaes. Regularizer R: We se our regularizaion funcion o be he sandard log-barrier Rx) = log1 x 2 ) for he sphere and use Algorihm 1 as he black-box. Sampling of y : We can assume wihou loss of generaliy ha 0, so define z := / 2. owards he goal of keeping our sampled poin y away from he boundary, define := max1 2, n ). Now consruc some n 1 orhonormal basis of he subspace perpendicular o z, which we will call Perpz ). Sample our predicion y as follows: z w.p y = z w.p. 4 ±w Perpz ) w.p. 4n 1) 11) Consrucion of f : Given he above sampling scheme, we define our esimaes f as follows, f = f y )y 2 Pry ) 12) where he probabiliies Pr ) are defined in equaion 11). I is sraighforward o check ha E f = f. Consrucion of g : he following choice of g will be shown o saisfy he requiremens: g u) := 4n ) u 2z. We now check he condiions of he heorem o verify ha his consrucion leads o a high-probabiliy bound. A) Since we are using a self-concordan regularizer, we already showed in heorem 2.3 how o obain he necessary bound wih c 1 = 2. B) Noice ha Ey = 1 )z = 1 )x 2 and because 1 ) 2 n i follows ha Ey n. Hence, c 2 = 1. C) Since we assume ha u 2 1, we see ha c 3 = 2: f u f 2 u 2 f 2 2 Pry )) 1 2n 2 n. D) We check ha g does indeed bound he variance. We firs upper bound he marix E f f : E f f = ) f 2 Pry ) y y y 2 Pry y ) 1 2 max y Pry ) 1 )I n, since he range of y is over ± vecors from an orhonormal basis. By consrucion each of hese probabiliies is /4n). Now we can bound: u) E f f u) 2n u 2 2 2n 2 E y u E y 2 2) 4n 1 )z u n 4n 2 2u z + 2u z ) + 4n g u). Addiionally, we check ha he bias is no large a. Recalling ha z = / and since 1 by consrucion, g ) = 4n x ) 20n, i.e. c 4 = 20. E) We now check ha, in he -norm, he biased esimae is no oo big. We can roughly lower bound 2 2 R = 1 2 I ) 2 x 1 1 I ) 2 z z 1 where we used ha 1 21 ) whenever [0, 1]. his ells us ha he eigenvalues of 2 R are bounded from below 1 ) 1 in all ) direcions orhogonal o, and by x 2 1 in he direcion of. hus, 1 1 = 2 1 )1+ ) 2 R 1 1 )I z z ) + 1 ) z z 1 )I z z ) + 21 ) 2 z z where he las inequaliy holds since 1 x + x 2 > 1/2 when x [0, 1]. Now ha we have conrol of he norm 2 R 1, we can bound g 2 g 2 R 1 g ) 2 8n = z 2 R 1 z 64n2 21 ) n 2 If y = z or z, f 2 4 )2 z 2 R 1 z ) 2 32,

9 If ±y Perpz ), ) 2 2n 1) f 2 y 2 x R 1 y 4n 1)2 2 1 ) 4n 3/2 1/2. hese las wo bounds give us, for large enough : f α g 2 8n 3/2 1/ n 2 α 2 i.e. c 5 = 8n 3/ n2 α 2. F ) We also mus check ha, in expecaion, he biased esimae is of consan order in he -norm: [ ] 2 [ ] E f α g E 2 f 2 + 2α 2 g 2 [ ] 2 Pry) f 2 y = y + 128n 2 α 2 y {±z,perpz )} < n n 2 α 2 =: c 6. G) heorem 2.3 comes wih he requiremen ha η f α g 1/2. From E), f α g = O 1/4 ). By aking η = O 1/2 ), he requiremen is saisfied for large enough. We conclude ha and A 1 = n n 2 α 2 + 8n 3/ n2 α 2 ) ) 8 log1/δ ) A 2 = 8 log1/δ) + n + 26 n log2 log )/δ ). log2 log )/δ Recalling ha α = ) n, we observe ha A 1 = On 2 ) and A 2 = O n log log ). heorem 4.1 hen gives us, wih η = log n, f y u) η 1 Ru)+η A 1 + A 2 = On log ) wih high probabiliy for any u K which is 1/2 away from he boundary. he asympoic behavior in erms of n and exacly maches he in-expecaion resul of [1], as he self-concordance parameer ϑ = 1 for he sphere. Now, o make he resul uniform for any u, we discreize he se K ino a grid of size n/2 and ake a union bound for all u in his se see [9], [5] for deails). Seing δ = δ = δ = δ leads n/2 o replacing all hree log 1/δ erms by n log + log 1/δ. Inspecing A 1, we observe ha his subsiuion inroduces n log in fron of 8n 3/2, which, when balanced wih η, exhibis η A 1 = On log ) behavior. However, A 2 = On 3/2 log ) now becomes he dominaing erm, as he log log /δ is no under he square roo. We conclude ha, wih high probabiliy, u K, f y u) = On 3/2 log ). A more careful analysis, involving a sharper inequaliy in one of he seps of he proof of heorem 4.1 see [2]), should reduce he dependence on n o linear. his will be carried ou in he full version of his paper. C. Example 3: Recovering he resul of [1] While i does no require heorem 4.1, for he sake of compleeness we show ha he in-expecaion resul of [1] immediaely follows from heorem 2.3. For any convex se, he sampling procedure proposed in ha paper is Regularizer R: he regularizaion funcion is a ϑ-selfconcordan barrier for K, whose exisence is guaraneed see [15], [14]). We use Algorihm 1 as he black-box. Sampling of y : Le {e 1,..., e n } and {λ 1,..., λ n } be he se of eigenvecors and eigenvalues of 2 R ). Choose i uniformly a random from {1,..., n} and ε = ±1 wih probabiliy 1/2. Sample y = + ε λ 1/2 i e i. Consrucion of f : Define f := n f y ) ε λ 1/2 i e i. Since here we are no ineresed in high-probabiliy bounds, we do no need o consruc g. Appealing o 9) and heorem 2.3, i only remains o bound f. By consrucion, f ) 2 = f 1 R ) f n 2. For any u which is 1/2 away from he boundary, Ru) 2ϑ log see [1]). hus, wih η = ϑ log n, we obain ER u) 4n ϑ log, which recovers he in-expecaion resul wih a slighly beer consan. he sampling scheme presened here does no saisfy condiions of heorem 4.1. Indeed, following he discussion in Remark 4.3, i is easy o prove even for K = [0, 1]) ha i is impossible o consruc g wih he desired properies. In oher words, he variance of he esimaes is larger han he desired regime. his realizaion was indeed he main moivaion for his paper. D. Example 4: Sampling schemes for general bodies We remark ha, while R has o be fixed hroughou he game, he sampling scheme does no. As long as he requiremens of heorem 4.1 are saisfied a each sep, he high probabiliy bound holds rue. he main difficuly in obaining a resul for general convex bodies K is in consrucion of g u), an upper-bound on he variance. Such a funcion heavily depends on he geomery and mus be consruced on per-case basis. We conjecure ha he following wo sampling schemes, one for he curved boundary similar o he spherical case) and one for he fla boundary similar o he simplex case), should be enough o deal wih mos nice ses K.

10 Pu large mass e.g. O 1 n ) on n 1 poins along he fla boundary, and pu a small probabiliy mass on a far away poin. As in he spherical case, pu large mass close o 1) on a single poin close o and small mass on 2n 1 oher poins far away. VI. PROOFS We sae four lemmas whose proof can be found in he echnical repor [2]. Lemma 6.1: Wih probabiliy a leas 1 δ, f y u) f u) + 8 log1/δ) + c 2 n. he following lemma is based on a resul proved in [5]. Lemma 6.2: For any δ < e 1 and 4, wih probabiliy a leas 1 2 log )δ, f u) R u) + 2 max 2 u) E f f u), 1 + 2c 3 n ) log1/δ) } log1/δ). Lemma 6.3: For any δ < e 1 and 4, wih probabiliy a leas 1 δ, ) f u) f α g x u) + [ 2c 3 + c 4 + 2) ] n log2 log )/δ ). he final ingredien is he following concenraion resul. Lemma 6.4: Wih probabiliy a leas 1 δ, [ ] 2 η f [ ] 2 α g η E f α g + η c 5 8 log1/δ ). Combining he above lemmas, we now prove he heorem. Proof: [Proof of heorem 4.1] Combining Lemma 6.1 and Lemma 6.3, we obain ha ) f y u) f α g x u) + 8 log1/δ) + c 2 n + 2c3 + c 4 + 2) n log2 log /δ ) wih probabiliy a leas 1 δ+δ ) ). By he black-box guaranee applied o funcions f α g, for any fixed u K, ) f α g x u) η 1 2 Ru)+c 1 η f α g. Combining he resuls, wih probabiliy a leas 1 δ + δ ), f y u) η 1 2 Ru) + c 1 η f α g + c 2 n + 8 log1/δ) + 2c 3 + c 4 + 2) n log2 log /δ ). Finally, by Lemma 6.4 and our assumpion F ), wih a bi of algebra we arrive a he saemen of heorem 4.1. REFERENCES [1] J. Abernehy, E. Hazan, and A. Rakhlin. Compeing in he dark: An efficien algorihm for bandi linear opimizaion. In Proceedings of he weny Firs Annual Conference on Learning heory, [2] J. Abernehy and A. Rakhlin. Beaing he adapive bandi wih high probabiliy. echnical Repor UCB/EECS , EECS Deparmen, Universiy of California, Berkeley, Jan [3] Peer Auer, Nicolò Cesa-Bianchi, Yoav Freund, and Rober E. Schapire. he nonsochasic muliarmed bandi problem. SIAM J. Compu., 321):48 77, [4] Baruch Awerbuch and Rober D. Kleinberg. Adapive rouing wih end-o-end feedback: disribued learning and geomeric approaches. In SOC 04: Proceedings of he hiry-sixh annual ACM symposium on heory of compuing, pages 45 53, New York, NY, USA, ACM. [5] P. L. Barle, V. Dani,. Hayes, S. Kakade, A. Rakhlin, and A. ewari. High probabiliy regre bounds for online opimizaion. In Proceedings of he weny Firs Annual Conference on Learning heory, [6] Amir Beck and Marc eboulle. Mirror descen and nonlinear projeced subgradien mehods for convex opimizaion. Oper. Res. Le., 313): , [7] Y. Censor and S. A. Zenios. Parallel Opimizaion: heory, Algorihms, and Applicaions. Oxford Universiy Press, [8] Nicolò Cesa-Bianchi and Gábor Lugosi. Predicion, Learning, and Games. Cambridge Universiy Press, [9] Varsha Dani, homas Hayes, and Sham Kakade. he price of bandi informaion for online opimizaion. In J.C. Pla, D. Koller, Y. Singer, and S. Roweis, ediors, Advances in Neural Informaion Processing Sysems 20. MI Press, Cambridge, MA, [10] Varsha Dani and homas P. Hayes. Robbing he bandi: less regre in online geomeric opimizaion agains an adapive adversary. In SODA 06: Proceedings of he seveneenh annual ACM-SIAM symposium on Discree algorihm, pages , New York, NY, USA, ACM. [11] Abraham D. Flaxman, Adam auman Kalai, and H. Brendan McMahan. Online convex opimizaion in he bandi seing: gradien descen wihou a gradien. In SODA 05: Proceedings of he sixeenh annual ACM- SIAM symposium on Discree algorihms, pages , Philadelphia, PA, USA, Sociey for Indusrial and Applied Mahemaics. [12] A. György,. Linder, G. Lugosi, and G. Oucsák. he on-line shores pah problem under parial monioring. Journal of Machine Learning Research, 8: , [13] H. Brendan McMahan and Avrim Blum. Online geomeric opimizaion in he bandi seing agains an adapive adversary. In COL, pages , [14] A. Nemirovski and M. odd. Inerior-poin mehods for opimizaion. Aca Numerica, pages , [15] Y. E. Neserov and A. S. Nemirovskii. Inerior Poin Polynomial Algorihms in Convex Programming. SIAM, Philadelphia, [16] A. Rakhlin and A. ewari. Lecure noes on online learning, Available a hp://wwwsa.wharon.upenn.edu/ rakhlin/papers/online learning.pdf. [17] Shai Shalev-Shwarz. Online Learning: heory, Algorihms, and Applicaions. PhD hesis, Hebrew Universiy, 2007.

Notes on online convex optimization

Notes on online convex optimization Noes on online convex opimizaion Karl Sraos Online convex opimizaion (OCO) is a principled framework for online learning: OnlineConvexOpimizaion Inpu: convex se S, number of seps T For =, 2,..., T : Selec

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal

More information

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence Supplemen for Sochasic Convex Opimizaion: Faser Local Growh Implies Faser Global Convergence Yi Xu Qihang Lin ianbao Yang Proof of heorem heorem Suppose Assumpion holds and F (w) obeys he LGC (6) Given

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

Matrix Versions of Some Refinements of the Arithmetic-Geometric Mean Inequality

Matrix Versions of Some Refinements of the Arithmetic-Geometric Mean Inequality Marix Versions of Some Refinemens of he Arihmeic-Geomeric Mean Inequaliy Bao Qi Feng and Andrew Tonge Absrac. We esablish marix versions of refinemens due o Alzer ], Carwrigh and Field 4], and Mercer 5]

More information

Approximation Algorithms for Unique Games via Orthogonal Separators

Approximation Algorithms for Unique Games via Orthogonal Separators Approximaion Algorihms for Unique Games via Orhogonal Separaors Lecure noes by Konsanin Makarychev. Lecure noes are based on he papers [CMM06a, CMM06b, LM4]. Unique Games In hese lecure noes, we define

More information

Lecture Notes 2. The Hilbert Space Approach to Time Series

Lecture Notes 2. The Hilbert Space Approach to Time Series Time Series Seven N. Durlauf Universiy of Wisconsin. Basic ideas Lecure Noes. The Hilber Space Approach o Time Series The Hilber space framework provides a very powerful language for discussing he relaionship

More information

We just finished the Erdős-Stone Theorem, and ex(n, F ) (1 1/(χ(F ) 1)) ( n

We just finished the Erdős-Stone Theorem, and ex(n, F ) (1 1/(χ(F ) 1)) ( n Lecure 3 - Kövari-Sós-Turán Theorem Jacques Versraëe jacques@ucsd.edu We jus finished he Erdős-Sone Theorem, and ex(n, F ) ( /(χ(f ) )) ( n 2). So we have asympoics when χ(f ) 3 bu no when χ(f ) = 2 i.e.

More information

Two Coupled Oscillators / Normal Modes

Two Coupled Oscillators / Normal Modes Lecure 3 Phys 3750 Two Coupled Oscillaors / Normal Modes Overview and Moivaion: Today we ake a small, bu significan, sep owards wave moion. We will no ye observe waves, bu his sep is imporan in is own

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

Lecture 4: November 13

Lecture 4: November 13 Compuaional Learning Theory Fall Semeser, 2017/18 Lecure 4: November 13 Lecurer: Yishay Mansour Scribe: Guy Dolinsky, Yogev Bar-On, Yuval Lewi 4.1 Fenchel-Conjugae 4.1.1 Moivaion Unil his lecure we saw

More information

Online Learning with Partial Feedback. 1 Online Mirror Descent with Estimated Gradient

Online Learning with Partial Feedback. 1 Online Mirror Descent with Estimated Gradient Avance Course in Machine Learning Spring 2010 Online Learning wih Parial Feeback Hanous are joinly prepare by Shie Mannor an Shai Shalev-Shwarz In previous lecures we alke abou he general framework of

More information

Notes for Lecture 17-18

Notes for Lecture 17-18 U.C. Berkeley CS278: Compuaional Complexiy Handou N7-8 Professor Luca Trevisan April 3-8, 2008 Noes for Lecure 7-8 In hese wo lecures we prove he firs half of he PCP Theorem, he Amplificaion Lemma, up

More information

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals

More information

MATH 5720: Gradient Methods Hung Phan, UMass Lowell October 4, 2018

MATH 5720: Gradient Methods Hung Phan, UMass Lowell October 4, 2018 MATH 5720: Gradien Mehods Hung Phan, UMass Lowell Ocober 4, 208 Descen Direcion Mehods Consider he problem min { f(x) x R n}. The general descen direcions mehod is x k+ = x k + k d k where x k is he curren

More information

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,

More information

SOLUTIONS TO ECE 3084

SOLUTIONS TO ECE 3084 SOLUTIONS TO ECE 384 PROBLEM 2.. For each sysem below, specify wheher or no i is: (i) memoryless; (ii) causal; (iii) inverible; (iv) linear; (v) ime invarian; Explain your reasoning. If he propery is no

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

Random Walk with Anti-Correlated Steps

Random Walk with Anti-Correlated Steps Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and

More information

Hamilton- J acobi Equation: Weak S olution We continue the study of the Hamilton-Jacobi equation:

Hamilton- J acobi Equation: Weak S olution We continue the study of the Hamilton-Jacobi equation: M ah 5 7 Fall 9 L ecure O c. 4, 9 ) Hamilon- J acobi Equaion: Weak S oluion We coninue he sudy of he Hamilon-Jacobi equaion: We have shown ha u + H D u) = R n, ) ; u = g R n { = }. ). In general we canno

More information

Non-Stochastic Bandit Slate Problems

Non-Stochastic Bandit Slate Problems Non-Sochasic Bandi Slae Problems Sayen Kale Yahoo! Research Sana Clara, CA skale@yahoo-inccom Lev Reyzin Georgia Ins of echnology Alana, GA lreyzin@ccgaechedu Absrac Rober E Schapire Princeon Universiy

More information

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H. ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple

More information

Some Ramsey results for the n-cube

Some Ramsey results for the n-cube Some Ramsey resuls for he n-cube Ron Graham Universiy of California, San Diego Jozsef Solymosi Universiy of Briish Columbia, Vancouver, Canada Absrac In his noe we esablish a Ramsey-ype resul for cerain

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes Some common engineering funcions 2.7 Inroducion This secion provides a caalogue of some common funcions ofen used in Science and Engineering. These include polynomials, raional funcions, he modulus funcion

More information

Solutions from Chapter 9.1 and 9.2

Solutions from Chapter 9.1 and 9.2 Soluions from Chaper 9 and 92 Secion 9 Problem # This basically boils down o an exercise in he chain rule from calculus We are looking for soluions of he form: u( x) = f( k x c) where k x R 3 and k is

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t... Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

SOME MORE APPLICATIONS OF THE HAHN-BANACH THEOREM

SOME MORE APPLICATIONS OF THE HAHN-BANACH THEOREM SOME MORE APPLICATIONS OF THE HAHN-BANACH THEOREM FRANCISCO JAVIER GARCÍA-PACHECO, DANIELE PUGLISI, AND GUSTI VAN ZYL Absrac We give a new proof of he fac ha equivalen norms on subspaces can be exended

More information

18 Biological models with discrete time

18 Biological models with discrete time 8 Biological models wih discree ime The mos imporan applicaions, however, may be pedagogical. The elegan body of mahemaical heory peraining o linear sysems (Fourier analysis, orhogonal funcions, and so

More information

11!Hí MATHEMATICS : ERDŐS AND ULAM PROC. N. A. S. of decomposiion, properly speaking) conradics he possibiliy of defining a counably addiive real-valu

11!Hí MATHEMATICS : ERDŐS AND ULAM PROC. N. A. S. of decomposiion, properly speaking) conradics he possibiliy of defining a counably addiive real-valu ON EQUATIONS WITH SETS AS UNKNOWNS BY PAUL ERDŐS AND S. ULAM DEPARTMENT OF MATHEMATICS, UNIVERSITY OF COLORADO, BOULDER Communicaed May 27, 1968 We shall presen here a number of resuls in se heory concerning

More information

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits DOI: 0.545/mjis.07.5009 Exponenial Weighed Moving Average (EWMA) Char Under The Assumpion of Moderaeness And Is 3 Conrol Limis KALPESH S TAILOR Assisan Professor, Deparmen of Saisics, M. K. Bhavnagar Universiy,

More information

A Primal-Dual Type Algorithm with the O(1/t) Convergence Rate for Large Scale Constrained Convex Programs

A Primal-Dual Type Algorithm with the O(1/t) Convergence Rate for Large Scale Constrained Convex Programs PROC. IEEE CONFERENCE ON DECISION AND CONTROL, 06 A Primal-Dual Type Algorihm wih he O(/) Convergence Rae for Large Scale Consrained Convex Programs Hao Yu and Michael J. Neely Absrac This paper considers

More information

5. Stochastic processes (1)

5. Stochastic processes (1) Lec05.pp S-38.45 - Inroducion o Teleraffic Theory Spring 2005 Conens Basic conceps Poisson process 2 Sochasic processes () Consider some quaniy in a eleraffic (or any) sysem I ypically evolves in ime randomly

More information

Games Against Nature

Games Against Nature Advanced Course in Machine Learning Spring 2010 Games Agains Naure Handous are joinly prepared by Shie Mannor and Shai Shalev-Shwarz In he previous lecures we alked abou expers in differen seups and analyzed

More information

Finish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details!

Finish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details! MAT 257, Handou 6: Ocober 7-2, 20. I. Assignmen. Finish reading Chaper 2 of Spiva, rereading earlier secions as necessary. handou and fill in some missing deails! II. Higher derivaives. Also, read his

More information

Introduction to Probability and Statistics Slides 4 Chapter 4

Introduction to Probability and Statistics Slides 4 Chapter 4 Inroducion o Probabiliy and Saisics Slides 4 Chaper 4 Ammar M. Sarhan, asarhan@mahsa.dal.ca Deparmen of Mahemaics and Saisics, Dalhousie Universiy Fall Semeser 8 Dr. Ammar Sarhan Chaper 4 Coninuous Random

More information

Essential Microeconomics : OPTIMAL CONTROL 1. Consider the following class of optimization problems

Essential Microeconomics : OPTIMAL CONTROL 1. Consider the following class of optimization problems Essenial Microeconomics -- 6.5: OPIMAL CONROL Consider he following class of opimizaion problems Max{ U( k, x) + U+ ( k+ ) k+ k F( k, x)}. { x, k+ } = In he language of conrol heory, he vecor k is he vecor

More information

Let us start with a two dimensional case. We consider a vector ( x,

Let us start with a two dimensional case. We consider a vector ( x, Roaion marices We consider now roaion marices in wo and hree dimensions. We sar wih wo dimensions since wo dimensions are easier han hree o undersand, and one dimension is a lile oo simple. However, our

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number

More information

Lecture 9: September 25

Lecture 9: September 25 0-725: Opimizaion Fall 202 Lecure 9: Sepember 25 Lecurer: Geoff Gordon/Ryan Tibshirani Scribes: Xuezhi Wang, Subhodeep Moira, Abhimanu Kumar Noe: LaTeX emplae couresy of UC Berkeley EECS dep. Disclaimer:

More information

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing Applicaion of a Sochasic-Fuzzy Approach o Modeling Opimal Discree Time Dynamical Sysems by Using Large Scale Daa Processing AA WALASZE-BABISZEWSA Deparmen of Compuer Engineering Opole Universiy of Technology

More information

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon 3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of

More information

The Asymptotic Behavior of Nonoscillatory Solutions of Some Nonlinear Dynamic Equations on Time Scales

The Asymptotic Behavior of Nonoscillatory Solutions of Some Nonlinear Dynamic Equations on Time Scales Advances in Dynamical Sysems and Applicaions. ISSN 0973-5321 Volume 1 Number 1 (2006, pp. 103 112 c Research India Publicaions hp://www.ripublicaion.com/adsa.hm The Asympoic Behavior of Nonoscillaory Soluions

More information

Convergence of the Neumann series in higher norms

Convergence of the Neumann series in higher norms Convergence of he Neumann series in higher norms Charles L. Epsein Deparmen of Mahemaics, Universiy of Pennsylvania Version 1.0 Augus 1, 003 Absrac Naural condiions on an operaor A are given so ha he Neumann

More information

EXERCISES FOR SECTION 1.5

EXERCISES FOR SECTION 1.5 1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

INDEPENDENT SETS IN GRAPHS WITH GIVEN MINIMUM DEGREE

INDEPENDENT SETS IN GRAPHS WITH GIVEN MINIMUM DEGREE INDEPENDENT SETS IN GRAPHS WITH GIVEN MINIMUM DEGREE JAMES ALEXANDER, JONATHAN CUTLER, AND TIM MINK Absrac The enumeraion of independen ses in graphs wih various resricions has been a opic of much ineres

More information

Optimality Conditions for Unconstrained Problems

Optimality Conditions for Unconstrained Problems 62 CHAPTER 6 Opimaliy Condiions for Unconsrained Problems 1 Unconsrained Opimizaion 11 Exisence Consider he problem of minimizing he funcion f : R n R where f is coninuous on all of R n : P min f(x) x

More information

Ordinary Differential Equations

Ordinary Differential Equations Ordinary Differenial Equaions 5. Examples of linear differenial equaions and heir applicaions We consider some examples of sysems of linear differenial equaions wih consan coefficiens y = a y +... + a

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information

ACE 562 Fall Lecture 4: Simple Linear Regression Model: Specification and Estimation. by Professor Scott H. Irwin

ACE 562 Fall Lecture 4: Simple Linear Regression Model: Specification and Estimation. by Professor Scott H. Irwin ACE 56 Fall 005 Lecure 4: Simple Linear Regression Model: Specificaion and Esimaion by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Simple Regression: Economic and Saisical Model

More information

GMM - Generalized Method of Moments

GMM - Generalized Method of Moments GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................

More information

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

Expert Advice for Amateurs

Expert Advice for Amateurs Exper Advice for Amaeurs Ernes K. Lai Online Appendix - Exisence of Equilibria The analysis in his secion is performed under more general payoff funcions. Wihou aking an explici form, he payoffs of he

More information

A Stochastic View of Optimal Regret through Minimax Duality

A Stochastic View of Optimal Regret through Minimax Duality A Sochasic View of Opimal Regre hrough Minimax Dualiy Jacob Abernehy Compuer Science Division UC Berkeley Alekh Agarwal Compuer Science Division UC Berkeley Peer L. Barle Compuer Science Division Deparmen

More information

23.5. Half-Range Series. Introduction. Prerequisites. Learning Outcomes

23.5. Half-Range Series. Introduction. Prerequisites. Learning Outcomes Half-Range Series 2.5 Inroducion In his Secion we address he following problem: Can we find a Fourier series expansion of a funcion defined over a finie inerval? Of course we recognise ha such a funcion

More information

Christos Papadimitriou & Luca Trevisan November 22, 2016

Christos Papadimitriou & Luca Trevisan November 22, 2016 U.C. Bereley CS170: Algorihms Handou LN-11-22 Chrisos Papadimiriou & Luca Trevisan November 22, 2016 Sreaming algorihms In his lecure and he nex one we sudy memory-efficien algorihms ha process a sream

More information

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H.

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H. ACE 56 Fall 5 Lecure 8: The Simple Linear Regression Model: R, Reporing he Resuls and Predicion by Professor Sco H. Irwin Required Readings: Griffihs, Hill and Judge. "Explaining Variaion in he Dependen

More information

Stability and Bifurcation in a Neural Network Model with Two Delays

Stability and Bifurcation in a Neural Network Model with Two Delays Inernaional Mahemaical Forum, Vol. 6, 11, no. 35, 175-1731 Sabiliy and Bifurcaion in a Neural Nework Model wih Two Delays GuangPing Hu and XiaoLing Li School of Mahemaics and Physics, Nanjing Universiy

More information

Predator - Prey Model Trajectories and the nonlinear conservation law

Predator - Prey Model Trajectories and the nonlinear conservation law Predaor - Prey Model Trajecories and he nonlinear conservaion law James K. Peerson Deparmen of Biological Sciences and Deparmen of Mahemaical Sciences Clemson Universiy Ocober 28, 213 Ouline Drawing Trajecories

More information

2. Nonlinear Conservation Law Equations

2. Nonlinear Conservation Law Equations . Nonlinear Conservaion Law Equaions One of he clear lessons learned over recen years in sudying nonlinear parial differenial equaions is ha i is generally no wise o ry o aack a general class of nonlinear

More information

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS NA568 Mobile Roboics: Mehods & Algorihms Today s Topic Quick review on (Linear) Kalman Filer Kalman Filering for Non-Linear Sysems Exended Kalman Filer (EKF)

More information

Chapter 3 Boundary Value Problem

Chapter 3 Boundary Value Problem Chaper 3 Boundary Value Problem A boundary value problem (BVP) is a problem, ypically an ODE or a PDE, which has values assigned on he physical boundary of he domain in which he problem is specified. Le

More information

Lecture 2 October ε-approximation of 2-player zero-sum games

Lecture 2 October ε-approximation of 2-player zero-sum games Opimizaion II Winer 009/10 Lecurer: Khaled Elbassioni Lecure Ocober 19 1 ε-approximaion of -player zero-sum games In his lecure we give a randomized ficiious play algorihm for obaining an approximae soluion

More information

Technical Report Doc ID: TR March-2013 (Last revision: 23-February-2016) On formulating quadratic functions in optimization models.

Technical Report Doc ID: TR March-2013 (Last revision: 23-February-2016) On formulating quadratic functions in optimization models. Technical Repor Doc ID: TR--203 06-March-203 (Las revision: 23-Februar-206) On formulaing quadraic funcions in opimizaion models. Auhor: Erling D. Andersen Convex quadraic consrains quie frequenl appear

More information

An Introduction to Malliavin calculus and its applications

An Introduction to Malliavin calculus and its applications An Inroducion o Malliavin calculus and is applicaions Lecure 5: Smoohness of he densiy and Hörmander s heorem David Nualar Deparmen of Mahemaics Kansas Universiy Universiy of Wyoming Summer School 214

More information

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions Muli-Period Sochasic Models: Opimali of (s, S) Polic for -Convex Objecive Funcions Consider a seing similar o he N-sage newsvendor problem excep ha now here is a fixed re-ordering cos (> 0) for each (re-)order.

More information

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t Exercise 7 C P = α + β R P + u C = αp + βr + v (a) (b) C R = α P R + β + w (c) Assumpions abou he disurbances u, v, w : Classical assumions on he disurbance of one of he equaions, eg. on (b): E(v v s P,

More information

Multiarmed Bandits With Limited Expert Advice

Multiarmed Bandits With Limited Expert Advice uliarmed Bandis Wih Limied Exper Advice Sayen Kale Yahoo Labs ew York sayen@yahoo-inc.com Absrac We consider he problem of minimizing regre in he seing of advice-efficien muliarmed bandis wih exper advice.

More information

In this chapter the model of free motion under gravity is extended to objects projected at an angle. When you have completed it, you should

In this chapter the model of free motion under gravity is extended to objects projected at an angle. When you have completed it, you should Cambridge Universiy Press 978--36-60033-7 Cambridge Inernaional AS and A Level Mahemaics: Mechanics Coursebook Excerp More Informaion Chaper The moion of projeciles In his chaper he model of free moion

More information

Essential Maps and Coincidence Principles for General Classes of Maps

Essential Maps and Coincidence Principles for General Classes of Maps Filoma 31:11 (2017), 3553 3558 hps://doi.org/10.2298/fil1711553o Published by Faculy of Sciences Mahemaics, Universiy of Niš, Serbia Available a: hp://www.pmf.ni.ac.rs/filoma Essenial Maps Coincidence

More information

Matlab and Python programming: how to get started

Matlab and Python programming: how to get started Malab and Pyhon programming: how o ge sared Equipping readers he skills o wrie programs o explore complex sysems and discover ineresing paerns from big daa is one of he main goals of his book. In his chaper,

More information

Fishing limits and the Logistic Equation. 1

Fishing limits and the Logistic Equation. 1 Fishing limis and he Logisic Equaion. 1 1. The Logisic Equaion. The logisic equaion is an equaion governing populaion growh for populaions in an environmen wih a limied amoun of resources (for insance,

More information

RANDOM LAGRANGE MULTIPLIERS AND TRANSVERSALITY

RANDOM LAGRANGE MULTIPLIERS AND TRANSVERSALITY ECO 504 Spring 2006 Chris Sims RANDOM LAGRANGE MULTIPLIERS AND TRANSVERSALITY 1. INTRODUCTION Lagrange muliplier mehods are sandard fare in elemenary calculus courses, and hey play a cenral role in economic

More information

15. Vector Valued Functions

15. Vector Valued Functions 1. Vecor Valued Funcions Up o his poin, we have presened vecors wih consan componens, for example, 1, and,,4. However, we can allow he componens of a vecor o be funcions of a common variable. For example,

More information

A DELAY-DEPENDENT STABILITY CRITERIA FOR T-S FUZZY SYSTEM WITH TIME-DELAYS

A DELAY-DEPENDENT STABILITY CRITERIA FOR T-S FUZZY SYSTEM WITH TIME-DELAYS A DELAY-DEPENDENT STABILITY CRITERIA FOR T-S FUZZY SYSTEM WITH TIME-DELAYS Xinping Guan ;1 Fenglei Li Cailian Chen Insiue of Elecrical Engineering, Yanshan Universiy, Qinhuangdao, 066004, China. Deparmen

More information

Single and Double Pendulum Models

Single and Double Pendulum Models Single and Double Pendulum Models Mah 596 Projec Summary Spring 2016 Jarod Har 1 Overview Differen ypes of pendulums are used o model many phenomena in various disciplines. In paricular, single and double

More information

Cash Flow Valuation Mode Lin Discrete Time

Cash Flow Valuation Mode Lin Discrete Time IOSR Journal of Mahemaics (IOSR-JM) e-issn: 2278-5728,p-ISSN: 2319-765X, 6, Issue 6 (May. - Jun. 2013), PP 35-41 Cash Flow Valuaion Mode Lin Discree Time Olayiwola. M. A. and Oni, N. O. Deparmen of Mahemaics

More information

u(x) = e x 2 y + 2 ) Integrate and solve for x (1 + x)y + y = cos x Answer: Divide both sides by 1 + x and solve for y. y = x y + cos x

u(x) = e x 2 y + 2 ) Integrate and solve for x (1 + x)y + y = cos x Answer: Divide both sides by 1 + x and solve for y. y = x y + cos x . 1 Mah 211 Homework #3 February 2, 2001 2.4.3. y + (2/x)y = (cos x)/x 2 Answer: Compare y + (2/x) y = (cos x)/x 2 wih y = a(x)x + f(x)and noe ha a(x) = 2/x. Consequenly, an inegraing facor is found wih

More information

Lecture 20: Riccati Equations and Least Squares Feedback Control

Lecture 20: Riccati Equations and Least Squares Feedback Control 34-5 LINEAR SYSTEMS Lecure : Riccai Equaions and Leas Squares Feedback Conrol 5.6.4 Sae Feedback via Riccai Equaions A recursive approach in generaing he marix-valued funcion W ( ) equaion for i for he

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

Instructor: Barry McQuarrie Page 1 of 5

Instructor: Barry McQuarrie Page 1 of 5 Procedure for Solving radical equaions 1. Algebraically isolae one radical by iself on one side of equal sign. 2. Raise each side of he equaion o an appropriae power o remove he radical. 3. Simplify. 4.

More information

On Boundedness of Q-Learning Iterates for Stochastic Shortest Path Problems

On Boundedness of Q-Learning Iterates for Stochastic Shortest Path Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. 38, No. 2, May 2013, pp. 209 227 ISSN 0364-765X (prin) ISSN 1526-5471 (online) hp://dx.doi.org/10.1287/moor.1120.0562 2013 INFORMS On Boundedness of Q-Learning Ieraes

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning

More information

Families with no matchings of size s

Families with no matchings of size s Families wih no machings of size s Peer Franl Andrey Kupavsii Absrac Le 2, s 2 be posiive inegers. Le be an n-elemen se, n s. Subses of 2 are called families. If F ( ), hen i is called - uniform. Wha is

More information

Robust estimation based on the first- and third-moment restrictions of the power transformation model

Robust estimation based on the first- and third-moment restrictions of the power transformation model h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,

More information

1. An introduction to dynamic optimization -- Optimal Control and Dynamic Programming AGEC

1. An introduction to dynamic optimization -- Optimal Control and Dynamic Programming AGEC This documen was generaed a :45 PM 8/8/04 Copyrigh 04 Richard T. Woodward. An inroducion o dynamic opimizaion -- Opimal Conrol and Dynamic Programming AGEC 637-04 I. Overview of opimizaion Opimizaion is

More information

Stochastic Linear Optimization under Bandit Feedback

Stochastic Linear Optimization under Bandit Feedback Sochasic Linear Opimizaion under Bandi Feedback Varsha Dani and homas P. Hayes and Sham M. Kakade Absrac In he classical sochasic k-armed bandi problem, in each of a sequence of rounds, a decision maker

More information

The following report makes use of the process from Chapter 2 in Dr. Cumming s thesis.

The following report makes use of the process from Chapter 2 in Dr. Cumming s thesis. Zaleski 1 Joseph Zaleski Mah 451H Final Repor Conformal Mapping Mehods and ZST Hele Shaw Flow Inroducion The Hele Shaw problem has been sudied using linear sabiliy analysis and numerical mehods, bu a novel

More information