Online Boosting Algorithms for Multi-label Ranking

Size: px
Start display at page:

Download "Online Boosting Algorithms for Multi-label Ranking"

Transcription

1 Onlne Boosng Algorhms for Mul-label Rankng Young Hun Jung Deparmen of Sascs Unversy of Mchgan Ann Arbor, MI 4809 Ambuj Tewar Deparmen of Sascs Unversy of Mchgan Ann Arbor, MI 4809 arxv: v [sa.ml] 23 Oc 207 Absrac We consder he mul-label rankng approach o mullabel learnng. Boosng s a naural mehod for mullabel rankng as aggregaes weak predcons hrough majory voes, whch can be drecly used as scores o produce a rankng of he labels. We desgn onlne boosng algorhms wh provable loss bounds for mul-label rankng. We show ha our frs algorhm s opmal n erms of he number of learners requred o aan a desred accuracy, bu requres knowledge of he edge of he weak learners. We also desgn an adapve algorhm ha does no requre hs knowledge and s hence more praccal. Expermenal resuls on real daa ses demonsrae ha our algorhms are a leas as good as exsng bach boosng algorhms.. INTRODUCTION Mul-label learnng has mporan praccal applcaons (e.g., Schapre and Snger [2000]) and s heorecal properes connue o be suded (e.g., Koyejo e al. [205]). In conras o sandard mul-class classfcaons, mul-label learnng problems allow mulple correc answers. In oher words, we have a fxed se of basc labels and he acual label s a subse of he basc labels. Snce he number of subses ncreases exponenally as he number of basc labels grows, hnkng of each subse as a dfferen class leads o nracably. I s que common n applcaons for he mul-label learner o smply oupu a rankng of he labels on a new es nsance. For example, he popular MULAN lbrary desgned by Tsoumakas e al. [20] allows he oupu of mul-label learnng o be a mul-label ranker. In hs paper, we herefore focus on he mul-label rankng (MLR) seng. Tha s o say, he learner produces a score vecor such ha a label wh a hgher score wll be ranked above a label wh a lower score. We are parcularly neresed n onlne MLR sengs where he labeled daa arrve sequenally. The onlne framework s desgned o handle a large volume of daa ha accumulaes rapdly. In conras o a classcal bach learner, Ocober 24, 207 whch observes he enre ranng se, an onlne learner does no requre sorage of large amoun of daa n memory and can also adap o non-saonary n he daa because keeps updang s nernal sae as new nsances arrve. Boosng, frs proposed by Freund and Schapre [995], aggregaes mldly powerful learners no a srong learner. I has been used o produce sae-of-he-ar resuls n a wde range of felds (e.g., Korykowsk e al. [206] and Zhang and Wang [204]). Boosng algorhms ake weghed majory voes among weak learners predcons, and he cumulave voes can be nerpreed as a score vecor. Ths feaure makes boosng very well sued o MLR problems. The heory of boosng emerged n bach bnary sengs and became arguably complee (cf. Schapre and Freund [202]), bu s exenson o an onlne seng s relavely new. To he bes of our knowledge, Chen e al. [202] frs nroduced an onlne boosng algorhm wh heorecal jusfcaons, and Beygelzmer e al. [205] pushed he sae-of-he-ar n onlne bnary sengs furher by proposng wo onlne algorhms and provng opmaly of one. Recen work by Jung e al. [207] has exended he heory o mul-class sengs, bu her scope remaned lmed o sngle-label problems. In hs paper, we presen he frs onlne MLR boosng algorhms along wh her heorecal jusfcaons. Our work s manly nspred by he onlne sngle-label work (Jung e al. [207]). The man conrbuon s o allow general forms of weak predcons whereas he prevous onlne boosng algorhms only consdered homogeneous predcon formas. By nroducng a general way o encode weak predcons, our algorhms can combne bnary, sngle-label, and MLR predcons. Afer nroducng problem sengs, we defne an edge of an onlne learner over a random learner (Defnon ). Under he assumpon ha every weak learner has a known posve edge, we desgn an opmal way o combne her predcons (Secon 3.). In order o deal wh praccal sengs where such an assumpon s unenable, we presen an adapve algorhm ha can

2 aggregae learners wh arbrary edges (Secon 3.2). In Secon 4, we es our wo algorhms on real daa ses, and fnd ha her performance s ofen comparable wh, and somemes beer han, ha of exsng bach boosng algorhms for MLR. 2. PRELIMINARIES 2.. Problem Seng and Noaons The number of canddae labels s fxed o be k, whch s known o he learner. Whou loss of generaly, we may wre he labels usng negers n [k] := {,,k}. We are allowng mulple correc answers, and he label Y s a subse of [k]. The labels n Y s called relevan, and hose n Y c, rrelevan. A me =,,T, an adversary sequenally chooses a labeled example (x,y ) X 2 [k], where X s some doman. Only he nsance x s shown o he learner, and he label Y s revealed once he learner makes a predcon ŷ. As we are neresed n MLR sengs, ŷ s a k dmensonal score vecor. The learner suffers a loss (ŷ ) where he loss funcon wll be specfed laer n Secon 3.. In our boosng framework, we assume ha he learner consss of a booser and a fxed N number of weak learners. Ths resembles a manager-worker framework n ha booser dsrbues asks by specfyng losses, and each weak learner makes a predcon o mnmze he loss. Booser makes he fnal decson by aggregang weak predcons. Once he rue label s revealed, he booser shares hs nformaon so ha weak learners can updae her parameers for he nex example Onlne Weak Learners and Cos Vecor We wll keep he form of weak predcons h general n ha we only assume s a dsrbuon over [k]. Ths can n fac represen varous ypes of predcons. For example, a sngle-label predcon l can be encoded as a sandard bass vecor e l, or a mul-label predcon {l,,l n } by n n = e l. Due o hs general forma, our boosng algorhm can even combne weak predcons of dfferen formas. Ths mples ha f a researcher has a srong famly of bnary learners, he can smply boos hem whou ransformng hem no mul-class learners hrough well known echnques such as one-vs-all or one-vs-one. We exend he cos marx framework, frs proposed by Mukherjee and Schapre [203] and hen adoped n onlne sengs by Jung e al. [207], as a means of communcaon beween booser and weak learners. A eraon, booser compues a cos vecor c for he h weak learner WL whose predcon h wll suffer he cos c h. The cos vecor s unknown o WL unl produces h, whch s usual n onlne sengs. Oherwse, WL can rvally mnmze he cos. Fnally we assume ha weak learners can ake an mporance wegh as an npu. Ths s allowed for many onlne algorhms such as decson rees (Domngos and Hulen [2000]) or SVMs (Pedregosa e al. [20]) General Onlne Boosng Schema We nroduce a general algorhm schema shared by our boosng algorhms. α denoes he wegh for WL a eraon. We wll keep rack of weghed cumulave voes hrough s j := j = α h. Tha s o say, we can gve more cred for well performng weak learners by seng larger weghs. Furhermore, allowng negave weghs, we can avod poor learner s predcons. We call s j a predcon made by exper j. In he end, he booser makes he fnal decson by followng one of hese expers. The schema s summarzed n Algorhm. We wan o emphasze ha he rue labely s only avalable once he fnal predcon ŷ s made. Compuaon of weghs and cos vecors requres he knowledge of Y, and hus happens afer he fnal decson s made. To keep our heory general, we are no specfyng weak learners (lne 4 and 2). The specfc ways o calculae oher varables such as α, c, and wll depend on algorhms, whch wll be nroduced n he nex secon. Algorhm Onlne Boosng Schema : Inalze: α for [N] 2: for =,,T do 3: Receve example x 4: Gaher weak predcons h = WL (x ), 5: Record exper predcons s j := j = α h 6: Choose an ndex [N] 7: Make a fnal decson ŷ = s 8: Ge he rue label Y 9: Compue weghs α +, 0: Compue cos vecors c, : Weak learners suffer he loss c h 2: Weak learners updae he nernal parameers 3: Updae booser s parameers, f any 4: end for 3. ALGORITHMS WITH THEORETI- CAL LOSS BOUNDS An mporan facor for he performance of boosng algorhms s he performance of he ndvdual weak learners. For example, f weak learners are makng compleely random predcons, hey canno produce meanngful oucomes n accordance wh he booser s nenon. We deal wh hs maer n wo dfferen ways. One way s o defne an edge of a weak leaner over a 2

3 compleely random learner and assume all weak learners have posve edges. Anoher way s o measure each learner s emprcal edge and manpulae he wegh α o maxmze he accuracy of he fnal predcon. Even a learner ha s worse han random guessng can conrbue posvely f we allow negave weghs. The frs mehod leads o OnlneBMR (Secon 3.), and he second o Ada.OLMR (Secon 3.2). 3.. Opmal Algorhm We wll rgorously defne he edge of a weak learner. Recall ha weak learners suffer losses deermned by cos vecors. Gven he rue label Y, we resrc he booser s choce of cos vecors o C eor 0 := {c [0,] k maxc[l] mn c[r], mnc[l] = 0 and maxc[l] = }, l l where he name C0 eor s used by Jung e al. [207] and eor sands for edge-over-random. Snce booser wans weak learners o pu hgher scores a he relevan labels, coss a he relevan labels should be less han hose a he rrelevan ones. Resrcon o[0,] k makes sure ha he learner s cos s bounded. Along wh cos vecors, he booser also passes he mporance weghs w [0,] so ha he learner s cos becomes w c h. We also consruc a baselne learner ha has edge γ. Is predcon u Y γ s also a dsrbuon over[k] ha pus γ more probably for he relevan labels. Tha s o say, we can wre { u Y γ [l] = a+γ f l Y a f l / Y, where he value of a s deermned by he sze, Y. Now we sae our onlne weak learnng condon. Defnon. (OnlneWLC) For parameers γ, δ (0,), and S > 0, a par of an onlne learner and an adversary s sad o sasfy OnlneWLC (δ,γ,s) f for any T, wh probably a leas δ, he learner can generae predcons ha sasfy w c h w c u Y γ +S. γ s called an edge, and S an excess loss. Ths exends he condon made by Jung e al. [207, Defnon ]. The probablsc saemen s needed as many onlne learners produce randomzed predcons. The excess loss can be nerpreed as a warm-up perod. Throughou hs secon, we assume our learners sasfy OnlneWLC (δ,γ,s) wh a fxed adversary. Cos Vecors The opmal desgn of a cos vecor depends on he choce of loss. We wll use L Y (s) o denoe he loss whou specfyng where s s he predced score vecor. The only consran ha we mpose on our loss s ha s proper, whch mples ha s decreasng n s[l] for l Y, and ncreasng n s[r] for r / Y (readers should noe ha proper loss has a leas one oher defnon). Then we nroduce poenal funcon, a well known concep n game heory whch s frs nroduced o boosng by Schapre [200]: φ 0 (s) := LY (s) φ (s) := E l u Y γ φ (s+e l ). () φ (s) ams o esmae booser s fnal loss when more weak learners are lef unl he fnal predcon and s s he curren sae. I can be easly shown by nducon ha many arbues of L are nhered by poenals. Beng proper or convex s a good example. Essenally, we wan o se c [l] := φn (s + e l ), (2) where s s he predcon of exper. The proper propery nhered by poenals ensures he relevan labels have less coss han he rrelevan. To sasfy he boundedness condon of C0 eor, we normalze (2) o ge d [l] := c [l] mn r c [r] w, (3) [] where w [] := max r c [r] mn r c [r]. Snce Defnon s assumng w [0,], we should furher normalze w []. Ths requres he knowledge of w := max w []. Noe ha hs s unavalable unl we observe all he nsances, whch s fne because we only need hs value n provng he loss bound. Algorhm Deals The algorhm s named by OnlneBMR (onlne boos-by-majory for mul-label rankng) as s poenal deas sem from classcal boos-bymajory algorhm (Schapre [200]). In OnlneBMR, we smply se α =, or n oher words, he booser akes smple cumulave voes. Cos vecors are compued usng (2), and he booser always follows he las exper N, or = N. These daals are summarzed n Algorhm 2. Algorhm 2 OnlneBMR Deals : Inalze: α = for [N] 6: Se = N 9: Se he weghs α + =, [N] 0: Compue c [l] = φn (s + e l ), l [k] 3: No exra parameers o be updaed The followng heorem holds eher f weak learners are sngle-label learners or f he loss L s convex. 3

4 Theorem 2. (BMR, General Loss Bound) For any T and N δ, he fnal loss suffered by OnlneBMR sasfes he followng nequaly wh probably N δ: (ŷ ) N φ N (0)+S w. (4) Proof. From () and (2), we can wre φ N + (s ) = E Y l u = c uy γ φn γ = (s + e l ) = c (uy γ h )+c h c (u Y γ h )+φ N (s ), where he las nequaly s n fac equaly f weak learners are sngle-label learners, or holds by Jensen s nequaly f he loss s convex (whch mples he convexy of poenals). Also noe ha s = s + h. Snce boh u Y γ and h have l norm, we can subrac common numbers from every enry of c whou changng he value of c (uy γ h ). Ths mples we can plug n w []d a he place of c. Then we have φ N + (s ) φ N (s ) w []d uy γ w []d h. By summng hs over, we have φ N + (s ) w []d u Y γ φ N (s ) w []d h. OnlneWLC (δ, γ, S) provdes, wh probably δ, w [] w d h w Pluggng hs n (5), we ge φ N + (s ) w []d u Y γ +S. φ N (s ) Sw. (5) Now summng hs over, we ge wh probably Nδ (due o unon bound), N φ N (0)+S w = whch complees he proof. T φ 0 (sn ) = (ŷ ), Now we evaluae he effcency of OnlneBMR by fxng a loss. Unforunaely, here s no canoncal loss n MLR sengs, bu followng rank loss s a srong canddae (cf. Cheng e al. [200] and Gao and Zhou [20]): L Y rnk(s) := w Y ½(s[l] < s[r])+ ½(s[l] = s[r]), 2 where w Y = Y Y c s a normalzaon consan ha ensures he loss les n [0,]. Noe ha hs loss s no convex. In case weak learners are n fac sngle-label learners, we can smply use rank loss o compue poenals, bu n more general case, we wll use he followng hnge loss o compue poenals: L Y hnge (s) := w Y (+s[r] s[l]) +, where ( ) + := max(,0). I s convex and always greaer han rank loss, and hus Theorem 2 can be used o bound rank loss. In Appendx A, we bound wo erms n he RHS of (4) when he poenals are bul upon rank and hnge losses. Here we record he resuls. Table : Upper Bounds for φ N loss φ N rank loss hnge loss (0) and w (0) w e γ2 N 2 O( N ) (N +)e γ2 N 2 2 For he case ha we use rank loss, we can check N N w O( ) O( N). N = = Combnng hese resuls wh Theorem 2, we ge he followng corollary. Corollary 3. (BMR, Rank Loss Bound) For any T and N δ, OnlneBMR sasfes followng rank loss bounds wh probably N δ. Wh sngle-label learners, we have 2 N rnk (ŷ ) e γ 2 T +O( NS), (6) and wh general learners, we have rnk (ŷ 2 N ) (N +)e γ 2 T +2NS. (7) Remark. When we dvde boh sdes by T, we fnd he average loss s asympocally bounded by he frs erm. The second erm wll deermne he sample complexy. In boh cases, he frs erm decreases exponenally as N grows, whch means he algorhm does no requre oo many learners o acheve a desred loss bound. 4

5 Machng Lower Bounds From (6), we can deduce ha o aan average loss less han ǫ, OnlneBMR needs N = Ω( γ ln 2 ǫ ) weak learners and T = Ω( S ǫγ ) samples. A naural queson s wheher hese numbers are opmal. In fac he followng heorem consrucs a crcumsance ha maches hese bounds up o logarhmc facors. Throughou he proof, we consder k as a fxed consan. Theorem 4. For any γ (0, 2k ), δ,ǫ (0,), and S kln( δ ) γ, here exss an adversary wh a famly of weak learners sasfyng OnlneWLC (δ,γ,s), such ha o acheve asympoc error rae ǫ, an onlne boosng algorhm requres a leas Ω( γ ln 2 ǫ ) weak learners and a sample complexy of Ω( S ǫγ ). Proof. We nroduce a skech here and pospone he complee dscusson o Appendx B. We assume ha an adversary draws a label Y unformly a random from 2 [k] {,[k]}, and he weak learners generae snglelabel predcons w.r.. p [k]. We wll manpulae p such ha weak learners sasfy OnlneWLC (δ,γ,s) bu he bes possble performance s close o (6). Boundedness condons n C0 eor and he Azuma- Hoeffdng nequaly provde ha wh probably δ, w c [l ] w c p + γ w k + kln( δ ). For he opmaly of he number of weak learners, we le p = u Y for all. The above nequaly guaranees OnlneWLC s me. Then a smlar argumen of Schapre and Freund [202, Secon 3.2.6] can show ha he opmal choce of weghs over he learners s ( N,, N ). Fnally adopng he argumens n he proof of Jung e al. [207, Theorem 4], we can show EL Y rnk (ŷ ) Ω(e 4Nk2 γ 2 ). Seng hs value equal o ǫ, we have N Ω( γ ln 2 ǫ ), consderng k as a fxed consan. Ths proves he frs par of he heorem. For he second par, le T 0 := S 4γ and defne p = uy 0 for T 0 and p = u Y for > T 0. Then OnlneWLC can be shown o be me n a smlar fashon. Observng ha weak learners do no provde meanngful nformaon for T 0, we can clam any onlne boosng algorhm suffers a loss a leas Ω(T 0 ). Therefore o oban he ceran accuracy ǫ, he number of nsances T should be a leas Ω( T0 ǫ ) = Ω( S ǫγ ), whch complees he second par of he proof Adapve Algorhm Despe he opmal loss bound, OnlneBMR has a few drawbacks when s appled n pracce. Frsly, poenals do no have a closed form, and s compuaon becomes a major boleneck (cf. Table 3). Furhermore, he edge γ becomes an exra unng parameer, whch ncreases he runme even more. Fnally, s possble ha learners have dfferen edges, and assumng a consan edge can lead o neffcency. To overcome hese drawbacks, raher han assumng posve edges for weak learners, our second algorhm chooses he wegh α adapvely o handle varable edges. Surrogae Loss Lke oher adapve boosng algorhms (e.g., Beygelzmer e al. [205] and Freund e al. [999]), our algorhm needs a surrogae loss. The choce of loss s broadly dscussed by Jung e al. [207], and logsc loss seems o be a vald choce n onlne sengs as s graden s unformly bounded. In hs regard, we wll use he followng logsc loss: L Y log (s) := w Y log(+exp(s[r] s[l])). Noe ha hs loss s proper and convex. We wan o emphasze ha n he end he booser s predcon wll suffer he rank loss, and hs surrogae only plays an nermedae role n opmzng parameers. Algorhm Deals The algorhm s nspred by Jung e al. [207, Adaboos.OLM], and we wll call by Ada.OLMR. Snce nernally ams o mnmze he logsc loss, we wll se he cos vecor o be he graden of he surrogae: c := LY log (s ). (8) Nex we presen how o se he weghs α. Afer nalzng α equals o 0, we wll use onlne graden descen mehod, proposed by Znkevch [2003], o compue he nex weghs. Recall ha Ada.OLMR wans o choose α o mnmze he cumulave logsc loss: log (s +α h ). If we wre f (α) := LY log (s +αh ), we need an onlne algorhm ha chooses α sasfyng f (α ) mn α F f (α)+r (T), where F s some feasble se, and R (T) s a sublnear regre. To apply he resul by Znkevch [2003, Theorem ],f needs o be convex, andf should be compac. The former condon s me by our choce of logsc loss, and we wll use F = [ 2,2] for he feasble se. Snce he booser s loss s nvaran under he scalng of weghs, we can shrnk he weghs o f n F. Onlne, Logsc, Mul-label, and Rankng 5

6 Takng dervave, we can check f (α). Now le Π( ) denoe a projecon ono F: Π( ) := max{ 2, mn{2, }}. By seng α + = Π(α η f (α )) where η =, we ger (T) 9 T. Consderng ha s = s +α h, we can also wre f (α ) = c + h. Fnally, remans o address how o choose. In conras o OnlneBMR, we canno show ha he las exper s relably sophscaed. Insead, wha can be shown s ha a leas one of he expers s good enough. Thus we wll use classcal Hedge algorhm (cf. Freund and Schapre [995] and Llesone and Warmuh [989]) o randomly choose an exper a each eraon wh adapve probably weghs dependng on each exper s predcon hsory. In parcular, we nroduce new varables v, whch are nalzed as v =,. A each eraon, s randomly drawn such ha P( = ) v, and hen v s updaed based on he exper s rank loss: v + := v e LY rnk (s ). The deals are summarzed n Algorhm 3. Algorhm 3 Ada.OLMR Deals : Inalze: α = 0 and v =, [N] 6: Randomly draw s.. P( = ) v 9: Compue α + = Π(α f (α )) 0: Compue c = LY log (s ) 3: Updae v + = v e LY rnk (s ) Emprcal Edges As we are no mposng OnlneWLC, we need oher measures of he learners predcve powers o prove he loss bound. From (8), can be observed ha he relevan labels have negave coss and he rrelevan ones have posve cos. Furhermore, he summaon of enres of c s exacly 0. Ths observaon suggess a new defnon of wegh: w [] := w Y +exp(s l Y r/ Y [l] s [r]) = c [l] = c [r] = c. 2 l Y r/ Y (9) Ths does no drecly correspond o he wegh used n (3), bu plays a smlar role. Then we defne he emprcal edge: T γ := c h w. (0) Noe ha he baselne learner u Y γ has hs value exacly γ, whch suggess ha s a good proxy for he edge defned n Defnon. Now we presen he loss bound of Ada.OLMR. Theorem 5. (Ada.OLMR, Rank loss bound) For any T and N, wh probably δ, he rank loss suffered by Ada.OLMR s bounded as follows: rnk (ŷ 8 N2 ) γ T +Õ( γ ), () where Õ noaon suppresses dependence on log δ. Proof. We sar he proof by defnng he rank loss suffered by exper as below: M := rnk (s ). Accordng o he formula, here s no harm o defne M 0 = T 2 snce s0 = 0. As he booser chooses an exper hrough he Hedge algorhm, a sandard analyss (cf. [Cesa-Banch and Lugos, 2006, Corollary 2.3]) along wh he Azuma-Hoeffdng nequaly provdes wh probably δ, rnk (ŷ ) 2mn M +2logN +Õ( T), (2) where Õ noaon suppresses dependence on log δ. I s no hard o check ha +exp(a b) 2½(a b), from whch we can nfer w [] 2 LY rnk (s ) and w M, (3) 2 where w s defned n (9). Noe ha hs relaon holds for he case = as well. Now le denoe he dfference of he cumulave logsc loss beween wo consecuve expers: := = log (s ) LY log (s ) log (s +α h ) log (s ). Then he onlne graden descen algorhm provdes mn α [ 2,2] [ log (s +αh ) LY log (s )] +9 T. Here we record an unvarae nequaly: log(+e s+α ) log(+e s ) = log(+ eα +e s) +e s(eα ). (4) 6

7 We expand he dfference o ge = [ log (s l Y +αh ) LY log (s )] log +es r/ Y l Y r/ Y +e s =: f(α). [r] s [l]+α(h [r] h [l]) +e s [l] s [r] s [l] [r] (eα(h [r] h [l]) ) (5) We clam ha mn α [ 2,2] f(α) γ 2 w. Le us rewre w n (9) and γ n (0) as followng. w = γ = l Y r/ Y +e s [l] s [r] l Y r/ Y w h [l] h [r] +e s [l] s [r]. (6) For he ease of noaon, le j denoe an ndex ha moves hrough all uples of (,l,r) [T] Y Y c, and a j and b j denoe followng erms. a j = w +e s b j = h [l] h [r]. [l] s [r] Then from (6), we have j a j = and j a jb j = γ. Now we express f(α) n erms of a j and b j as below. f(α) w = j a j (e αbj ) e α j ajbj = e αγ, where he nequaly holds by Jensen s nequaly. From hs, we can deduce ha mn α [ 2,2] f(α) w e 2 γ γ 2, where he las nequaly can be checked by nvesgang γ = 0, and observng he convexy of he exponenal funcon. Ths proves our clam ha mn f(α) γ α [ 2,2] 2 w. (7) Combnng (3), (4), (5) and (7), we have γ 4 M +9 T. Summng over, we ge by elescopng rule log (sn ) 4 4 log (0) N γ M +9N T = N = γ mn M +9N T. Noe ha log (0) = log2 and LY log (sn ) 0. Therefore we have mnm 4log2 γ T + 36N T γ. Pluggng hs n (2), we ge wh probably δ, rnk (ŷ ) 8log2 γ T +Õ( N T γ +logn) 8 N2 γ T +Õ( γ ), where he las nequaly holds from AM-GM nequaly: cn T c2 N 2 +T 2. Ths complees our proof. Comparson wh OnlneBMR We fnsh hs secon by comparng our wo algorhms. For a far comparson, assume ha all learners are havng edge γ. Snce he baselne learner u Y γ has emprcal edge γ, for suffcenly larget, we can deduce ha γ γ wh hgh probably. Usng hs relaon, () can be wren as rnk (ŷ ) 8 Nγ T +Õ(N γ ). Comparng hs o eher (6) or (7), we can see ha OnlneBMR ndeed has beer asympoc loss bound and sample complexy. Despe hs sub-opmaly, Ada.OLMR shows comparable resuls n real daa ses due o s adapve naure. 4. EXPERIMENTS We performed an expermen on benchmark daa ses aken from MULAN 2. We parcularly chose four daa ses because Dembczynsk and Hüllermeer [202] already provded performances of bach seng boosng algorhms, gvng us a benchmark o compare wh. The auhors n fac used fve daa ses, bu mage daa se became no longer avalable from he source. Table 2 summarzes he basc sascs of daa ses, ncludng ranng and es se szes, number of feaures and labels, 2 Tsoumakas e al. [20],hp://mulan.sourceforge.ne/daases.hml 7

8 and hree sascs of he szes of relevan ses. The daa se m-reduced s a reduced verson of medamll obaned by random samplng whou replacemen. We kep he orgnal spl for ranng and es ses o provde more relevan comparsons. Table 2: Summary of Daa Ses daa #ran #es dm k mn mean max emoons scene yeas medamll m-reduced VFDT algorhms presened by Domngos and Hulen [2000] were used as weak learners. Every algorhm used 00 rees whose parameers were randomly chosen. VFDT s raned usng sngle-label daa, and we fed ndvdual relevan labels along wh mporance weghs ha were compued as max l c c [l]. Insead of usng all covaraes, he booser fed o rees randomly chosen 20 covaraes o make weak predcons less correlaed. All compuaons were carred ou on a Nehalem archecure 0-core 2.27 GHz Inel Xeon E processors wh 25 GB RAM per core. Each algorhm was raned a leas en mes 3 wh dfferen random seeds, and he resuls were aggregaed hrough mean. Predcons were evaluaed by rank loss. Noe ha he algorhm s loss was only recorded for es ses, bu kep updang s parameers whle explorng es ses as well. Snce VFDT oupus a condonal dsrbuon, whch s no of a sngle-label forma, we used hnge loss o compue poenals. Furhermore, OnlneBMR has an addonal parameer of edge γ. We red four dfferen values 4, and he bes resul s recorded as bes BMR. The resuls are summarzed n Table 3. Table 3: Average Loss and Runme n seconds daa bach 5 Ada.OLMR bes BMR emoons scene yeas medamll m-reduced OnlneBMR for m-reduced was esed 0 mes due o long runmes, and ohers were esed 20 mes 4 {.2,.,.0,.00} for small k and {.05,.0,.005,.00} for large k Two algorhms average losses are comparable o each oher and o bach seng resuls, bu OnlneBMR requred much longer runmes. Based on he fac ha bes BMR s performance s repored on he bes edge parameer ou of four rals, Ada.OLMR s far more favorable n pracce. Wh large number of labels, runme for OnlneBMR grows rapdly, and was even mpossble o run medamll daa whn a week, and hs was why we produced he reduced verson. The man boleneck s he compuaon of poenals as hey do no have closed form. 5. CONCLUSION In hs paper, we presened wo onlne boosng algorhms ha make mul-label rankng (MLR) predcons. The algorhms are que flexble n her choce of weak learners n ha varous ypes of learners can be combned o produce a srong learner. OnlneBMR s bul upon he assumpon ha all weak learners are srcly beer han random guessng, and s loss bound s shown o be gh under ceran condons. Ada.OLMR adapvely chooses he weghs over he learners so ha learners wh arbrary (even negave) edges can be boosed. Despe s subopmal loss bound, produces comparable resuls wh OnlneBMR and runs much faser. Onlne MLR boosng provdes several opporunes for furher research. A major ssue n MLR problems s ha here does no exs a canoncal loss. Forunaely, Theorem 2 holds for any proper loss, bu Ada.OLMR only has a rank loss bound. An adapve algorhm ha can handle more general losses wll be desrable. The exsence of an opmal adapve algorhm s anoher neresng open queson. Acknowledgmens We acknowledge he suppor of NSF under CAREER gran IIS and CIF References Rober E Schapre and Yoram Snger. Boosexer: A boosng-based sysem for ex caegorzaon. Machne learnng, 39(2-3):35 68, Oluwasanm O Koyejo, Nagarajan Naarajan, Pradeep K Ravkumar, and Inderj S Dhllon. Conssen mullabel classfcaon. In Advances n Neural Informaon Processng Sysems, pages , 205. Grgoros Tsoumakas, Elefheros Spyromros-Xoufs, Jozef Vlcek, and Ioanns Vlahavas. Mulan: A java 5 The bes resul from bach boosng algorhms n Dembczynsk and Hüllermeer [202] 8

9 lbrary for mul-label learnng. Journal of Machne Learnng Research, 2:24 244, 20. Yoav Freund and Rober E Schapre. A desconheorec generalzaon of on-lne learnng and an applcaon o boosng. In European conference on compuaonal learnng heory, pages Sprnger, 995. Marcn Korykowsk, Leszek Rukowsk, and Rafał Scherer. Fas mage classfcaon by boosng fuzzy classfers. Informaon Scences, 327:75 82, 206. Xao-Le Zhang and DeLang Wang. Boosed deep neural neworks and mul-resoluon cochleagram feaures for voce acvy deecon. In INTERSPEECH, pages , 204. Rober E Schapre and Yoav Freund. Boosng: Foundaons and algorhms. MIT press, 202. Yoav Freund, Rober Schapre, and N Abe. A shor nroducon o boosng. Journal-Japanese Socey For Arfcal Inellgence, 4(77-780):62, 999. Marn Znkevch. Onlne convex programmng and generalzed nfnesmal graden ascen. In Proceedngs of 20h ICML, Nck Llesone and Manfred K Warmuh. The weghed majory algorhm. In Foundaons of Compuer Scence, 989., 30h Annual Symposum on, pages IEEE, 989. Ncolo Cesa-Banch and Gábor Lugos. Predcon, learnng, and games. Cambrdge unversy press, Krzyszof Dembczynsk and Eyke Hüllermeer. Conssen mullabel rankng hrough unvarae loss mnmzaon. In Proceedngs of he 29h ICML, 202. Shang-Tse Chen, Hsuan-Ten Ln, and Ch-Jen Lu. An onlne boosng algorhm wh heorecal jusfcaons. ICML, 202. Alna Beygelzmer, Sayen Kale, and Hapeng Luo. Opmal and adapve algorhms for onlne boosng. ICML, 205. Young Hun Jung, Jack Goez, and Ambuj Tewar. Onlne mulclass boosng. In Advances n Neural Informaon Processng Sysems, 207. Indraneel Mukherjee and Rober E Schapre. A heory of mulclass boosng. Journal of Machne Learnng Research, 4(Feb): , 203. Pedro Domngos and Geoff Hulen. Mnng hgh-speed daa sreams. In Proceedngs of he sxh ACM SIGKDD nernaonal conference on Knowledge dscovery and daa mnng, pages ACM, F. Pedregosa, G. Varoquaux, A. Gramfor, V. Mchel, B. Thron, O. Grsel, M. Blondel, P. Preenhofer, R. Wess, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perro, and E. Duchesnay. Sck-learn: Machne learnng n Pyhon. Journal of Machne Learnng Research, 2: , 20. Rober E Schapre. Drfng games. Machne Learnng, 43(3):265 29, 200. Wewe Cheng, Eyke Hüllermeer, and Krzyszof J Dembczynsk. Bayes opmal mullabel classfcaon va probablsc classfer chans. In Proceedngs of he 27h ICML, 200. We Gao and Zh-Hua Zhou. On he conssency of mullabel learnng. In Proceedngs of he 24h annual conference on learnng heory, pages , 20. 9

10 Appendx A. SPECIFIC BOUNDS FOR OnlneBMR We wll begn hs secon by nroducng a random walk framework o compue poenals. Suppose X := (X,,X k ) s a random vecor ha racks he number of draws of each label among..d. random draws w.r.. u Y γ. Then accordng o (), we may wre φ (s) = E (s+x). Ths framework wll appear frequenly hroughou he proofs. We sar from rank loss. Lemma 6. Under he same seng as n Theorem 2 bu wh poenals bul upon rank loss, we may bound φ N (0) as followng: φ N (0) e γ2 N 2. Proof. For smplcy, we wll drop n he proof. Le X N be he aforemenoned random vecor. Then we may wre he poenal by φ N (0) = EL Y rnk(x N ) w Y E½(X r X l ) = w Y P(X r X l 0). Fx l Y and r / Y. By defnon of u Y γ, we have a := u Y γ [l] = u Y γ [r]+γ =: b. Now suppose we draw wh probably a, wh probably b, and 0 oherwse. Then P(X r X l 0) equals he probably ha he summaon of N..d. random numbers s non-negave. Then we can apply Hoeffdng s nequaly o ge P(X r X l 0) e γ2 N 2. Snce w Y s he nverse of he number of pars (l,r), hs proves our asseron. Lemma 7. Under he same seng as n Theorem 2 bu wh poenals bul upon rank loss, we can show ha, w O( N ). Proof. Frs we fx and. We also fx l Y and r Y c. Then wre s := s +e l and s 2 := s +e r. Agan we nroduce X N. Then we may wre c [r ] c [l ] = φ N (s 2 ) φ N (s ) = E[ rnk (s 2 + X N ) rnk (s + X N )] w Y f(r,l), r/ Y l Y where f(r,l) := E[½(s 2 [r]+x r s 2 [l]+x l ) ½(s [r]+x r > s [l]+x l )]. Here we nenonally nclude and exclude equaly for he ease of compuaon. Changng he order of erms, we can derve f(r,l) P(s [l] s [r] X r X l s 2 [l] s 2 [r]) 3max n P(X r X l = n), where he las nequaly s deduced from he fac ha (s [l] s [r]) (s 2 [l] s 2 [r]) {0,,2}. Usng Berry-Esseen heorem, s shown by Jung e al. [207, Lemma 0] ha max n P(X r X l = n) O( N ), whch mples ha c [r ] c [l ] O( ). N Snce l and r are arbrary, and he bound does no depend on, he las nequaly proves our asseron. Now we provde smlar bound when he poenals are compued from hnge loss. Lemma 8. Under he same seng as n Theorem 2 bu wh poenals bul upon hnge loss, we may bound φ N (0) as followng: φ N (0) (N +)e γ2 N 2. Proof. Agan we wll drop n he proof and nroduce X N. Then we may wre he poenal by φ N (0) = EL Y hnge(x N ) = w Y E(+X r X l ) + N = w Y P(X r X l n) n=0 w Y (N +)P(X r X l 0). We already checked n Lemma 6 ha whch concludes he proof. P(X r X l 0) e γ2 N 2, Lemma 9. Under he same seng as n Theorem 2 bu wh poenals bul upon hnge loss, we can show ha, w 2. 0

11 Proof. Frs we fx and. We also fx l Y and r Y c. Then wre s := s +e l and s 2 := s +e r. Agan wh X N, we may wre c [r ] c [l ] = φ N (s 2 ) φ N (s ) where = E[ hnge (s 2 + X N ) hnge (s + X N )] = w Y f(r,l), r/ Y l Y f(r,l) := E[(+(s 2 + X N )[r] (s 2 + X N )[l]) + (+(s + X N )[r] (s + X N )[l]) + ]. I s no hard o check ha he erm nsde he expecaon s always bounded above by 2. Ths fac along wh he defnon ofw Y provdes ha c [r ] c [l ] 2. Snce our choce of l and r are arbrary, hs proves w [] 2, whch complees he proof. Appendx B. COMPLETE PROOF OF THEOREM 4 Proof. We assume ha an adversary draws a label Y unformly a random from 2 [k] {,[k]}, and he weak learners generae sngle-label predcons w.r.. p [k]. Any boosng algorhm can only make a fnal decson by weghed cumulave voes of N weak learners. We wll manpulae p such ha weak learners sasfy OnlneWLC (δ,γ,s) bu he bes possble performance s close o (6). As we are assumng sngle-label predcons, h = e l for some l [k] and c h = c [l ]. Furhermore, he bounded condon of C0 eor ensures c [l ] s conaned n [0, ]. The Azuma-Hoeffdng nequaly provdes ha wh probably δ, w c [l ] w c p + 2 w 2 2 ln( δ ) w c p + γ w 2 2 k w c p + γ w k + kln( δ ) + kln( δ ), (8) where he second nequaly holds by arhmec mean and geomerc mean relaon and he las nequaly holds due o w [0,]. We sar from provdng a lower bound on he number of weak learners. Le p = u Y for all. Ths can be done by he consran γ < 4k. From he condon of C0 eor ha mn l c[l] = 0,max l c = along wh he fac ha Y {,[k]}, we can show ha c (u Y γ u Y ) γ k. Then he las lne of (8) becomes w c u Y + γ w + kln( δ ) k (w c u Y γ γw k )+ γ w k w c u Y γ +S, + kln( δ ) whch valdaes ha weak learners ndeed sasfy OnlneWLC (δ, γ, S). Followng he argumen of Schapre and Freund [202, Secon 3.2.6], we can also prove ha he opmal choce of weghs over he learners s ( N,, N ). Now we compue a lower bound for he booser s loss. Le X := (X,,X k ) be a random vecor ha racks he number of labels drawn from N..d. random draws w.r.. u Y. Then he expeced rank loss for he booser can be wren as: EL Y rnk (X) w Y P(X l < X r ). Adopng he argumens n he proof by Jung e al. [207, Theorem 4], we can show ha P(X l < X r ) Ω(e 4Nk2 γ 2 ). Ths showsel Y rnk (X) γ 2 ). Seng hs value Ω(e 4Nk2 equal o ǫ, we have N Ω( γ ln 2 ǫ ), consderng k as a fxed consan. Ths proves he frs par of he heorem. Now we move on o he opmaly of sample complexy. We record anoher nequaly ha can be checked from he condons of C0 eor : c (u Y 0 uy γ ) γ. Le T 0 := S 4γ and defne p = uy 0 for T 0 and p = u Y for > T 0. Then for T T 0, (8) mples w c [l ] w c u Y 0 + γ w + kln( δ ) k w c u Y γ +γ(+ k ) w + kln( δ ) w c u Y γ +S. (9) where he las nequaly holds because w T 0 = S 4γ.

12 For T > T 0, agan (8) mples T 0 T w c [l ] w c u Y γ w k =T 0+ + kln( δ ) w c u Y w c u Y γ + k + k γt 0 + kln( δ ) w c u Y γ +S. (20) (9) and (20) prove ha he weak learners ndeed sasfy OnlneWLC (δ, γ, S). Observng ha weak learners do no provde meanngful nformaon for T 0, we can clam any onlne boosng algorhm suffers a loss a leas Ω(T 0 ). Therefore o ge he ceran accuracy, he number of nsancest should be a leasω( T0 ǫ ) = Ω( S ǫγ ), whch complees he second par of he proof. 2

Variants of Pegasos. December 11, 2009

Variants of Pegasos. December 11, 2009 Inroducon Varans of Pegasos SooWoong Ryu bshboy@sanford.edu December, 009 Youngsoo Cho yc344@sanford.edu Developng a new SVM algorhm s ongong research opc. Among many exng SVM algorhms, we wll focus on

More information

Introduction to Boosting

Introduction to Boosting Inroducon o Boosng Cynha Rudn PACM, Prnceon Unversy Advsors Ingrd Daubeches and Rober Schapre Say you have a daabase of news arcles, +, +, -, -, +, +, -, -, +, +, -, -, +, +, -, + where arcles are labeled

More information

CS286.2 Lecture 14: Quantum de Finetti Theorems II

CS286.2 Lecture 14: Quantum de Finetti Theorems II CS286.2 Lecure 14: Quanum de Fne Theorems II Scrbe: Mara Okounkova 1 Saemen of he heorem Recall he las saemen of he quanum de Fne heorem from he prevous lecure. Theorem 1 Quanum de Fne). Le ρ Dens C 2

More information

V.Abramov - FURTHER ANALYSIS OF CONFIDENCE INTERVALS FOR LARGE CLIENT/SERVER COMPUTER NETWORKS

V.Abramov - FURTHER ANALYSIS OF CONFIDENCE INTERVALS FOR LARGE CLIENT/SERVER COMPUTER NETWORKS R&RATA # Vol.) 8, March FURTHER AALYSIS OF COFIDECE ITERVALS FOR LARGE CLIET/SERVER COMPUTER ETWORKS Vyacheslav Abramov School of Mahemacal Scences, Monash Unversy, Buldng 8, Level 4, Clayon Campus, Wellngon

More information

An introduction to Support Vector Machine

An introduction to Support Vector Machine An nroducon o Suppor Vecor Machne 報告者 : 黃立德 References: Smon Haykn, "Neural Neworks: a comprehensve foundaon, second edon, 999, Chaper 2,6 Nello Chrsann, John Shawe-Tayer, An Inroducon o Suppor Vecor Machnes,

More information

CHAPTER 10: LINEAR DISCRIMINATION

CHAPTER 10: LINEAR DISCRIMINATION CHAPER : LINEAR DISCRIMINAION Dscrmnan-based Classfcaon 3 In classfcaon h K classes (C,C,, C k ) We defned dscrmnan funcon g j (), j=,,,k hen gven an es eample, e chose (predced) s class label as C f g

More information

GENERATING CERTAIN QUINTIC IRREDUCIBLE POLYNOMIALS OVER FINITE FIELDS. Youngwoo Ahn and Kitae Kim

GENERATING CERTAIN QUINTIC IRREDUCIBLE POLYNOMIALS OVER FINITE FIELDS. Youngwoo Ahn and Kitae Kim Korean J. Mah. 19 (2011), No. 3, pp. 263 272 GENERATING CERTAIN QUINTIC IRREDUCIBLE POLYNOMIALS OVER FINITE FIELDS Youngwoo Ahn and Kae Km Absrac. In he paper [1], an explc correspondence beween ceran

More information

Solution in semi infinite diffusion couples (error function analysis)

Solution in semi infinite diffusion couples (error function analysis) Soluon n sem nfne dffuson couples (error funcon analyss) Le us consder now he sem nfne dffuson couple of wo blocks wh concenraon of and I means ha, n a A- bnary sysem, s bondng beween wo blocks made of

More information

( ) () we define the interaction representation by the unitary transformation () = ()

( ) () we define the interaction representation by the unitary transformation () = () Hgher Order Perurbaon Theory Mchael Fowler 3/7/6 The neracon Represenaon Recall ha n he frs par of hs course sequence, we dscussed he chrödnger and Hesenberg represenaons of quanum mechancs here n he chrödnger

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 4

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 4 CS434a/54a: Paern Recognon Prof. Olga Veksler Lecure 4 Oulne Normal Random Varable Properes Dscrmnan funcons Why Normal Random Varables? Analycally racable Works well when observaon comes form a corruped

More information

Advanced Machine Learning & Perception

Advanced Machine Learning & Perception Advanced Machne Learnng & Percepon Insrucor: Tony Jebara SVM Feaure & Kernel Selecon SVM Eensons Feaure Selecon (Flerng and Wrappng) SVM Feaure Selecon SVM Kernel Selecon SVM Eensons Classfcaon Feaure/Kernel

More information

Graduate Macroeconomics 2 Problem set 5. - Solutions

Graduate Macroeconomics 2 Problem set 5. - Solutions Graduae Macroeconomcs 2 Problem se. - Soluons Queson 1 To answer hs queson we need he frms frs order condons and he equaon ha deermnes he number of frms n equlbrum. The frms frs order condons are: F K

More information

Lecture 11 SVM cont

Lecture 11 SVM cont Lecure SVM con. 0 008 Wha we have done so far We have esalshed ha we wan o fnd a lnear decson oundary whose margn s he larges We know how o measure he margn of a lnear decson oundary Tha s: he mnmum geomerc

More information

( t) Outline of program: BGC1: Survival and event history analysis Oslo, March-May Recapitulation. The additive regression model

( t) Outline of program: BGC1: Survival and event history analysis Oslo, March-May Recapitulation. The additive regression model BGC1: Survval and even hsory analyss Oslo, March-May 212 Monday May 7h and Tuesday May 8h The addve regresson model Ørnulf Borgan Deparmen of Mahemacs Unversy of Oslo Oulne of program: Recapulaon Counng

More information

In the complete model, these slopes are ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL. (! i+1 -! i ) + [(!") i+1,q - [(!

In the complete model, these slopes are ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL. (! i+1 -! i ) + [(!) i+1,q - [(! ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL The frs hng o es n wo-way ANOVA: Is here neracon? "No neracon" means: The man effecs model would f. Ths n urn means: In he neracon plo (wh A on he horzonal

More information

SOME NOISELESS CODING THEOREMS OF INACCURACY MEASURE OF ORDER α AND TYPE β

SOME NOISELESS CODING THEOREMS OF INACCURACY MEASURE OF ORDER α AND TYPE β SARAJEVO JOURNAL OF MATHEMATICS Vol.3 (15) (2007), 137 143 SOME NOISELESS CODING THEOREMS OF INACCURACY MEASURE OF ORDER α AND TYPE β M. A. K. BAIG AND RAYEES AHMAD DAR Absrac. In hs paper, we propose

More information

Comparison of Differences between Power Means 1

Comparison of Differences between Power Means 1 In. Journal of Mah. Analyss, Vol. 7, 203, no., 5-55 Comparson of Dfferences beween Power Means Chang-An Tan, Guanghua Sh and Fe Zuo College of Mahemacs and Informaon Scence Henan Normal Unversy, 453007,

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Lnear Response Theory: The connecon beween QFT and expermens 3.1. Basc conceps and deas Q: ow do we measure he conducvy of a meal? A: we frs nroduce a weak elecrc feld E, and hen measure

More information

Robust and Accurate Cancer Classification with Gene Expression Profiling

Robust and Accurate Cancer Classification with Gene Expression Profiling Robus and Accurae Cancer Classfcaon wh Gene Expresson Proflng (Compuaonal ysems Bology, 2005) Auhor: Hafeng L, Keshu Zhang, ao Jang Oulne Background LDA (lnear dscrmnan analyss) and small sample sze problem

More information

Econ107 Applied Econometrics Topic 5: Specification: Choosing Independent Variables (Studenmund, Chapter 6)

Econ107 Applied Econometrics Topic 5: Specification: Choosing Independent Variables (Studenmund, Chapter 6) Econ7 Appled Economercs Topc 5: Specfcaon: Choosng Independen Varables (Sudenmund, Chaper 6 Specfcaon errors ha we wll deal wh: wrong ndependen varable; wrong funconal form. Ths lecure deals wh wrong ndependen

More information

General Weighted Majority, Online Learning as Online Optimization

General Weighted Majority, Online Learning as Online Optimization Sascal Technques n Robocs (16-831, F10) Lecure#10 (Thursday Sepember 23) General Weghed Majory, Onlne Learnng as Onlne Opmzaon Lecurer: Drew Bagnell Scrbe: Nahanel Barshay 1 1 Generalzed Weghed majory

More information

Outline. Probabilistic Model Learning. Probabilistic Model Learning. Probabilistic Model for Time-series Data: Hidden Markov Model

Outline. Probabilistic Model Learning. Probabilistic Model Learning. Probabilistic Model for Time-series Data: Hidden Markov Model Probablsc Model for Tme-seres Daa: Hdden Markov Model Hrosh Mamsuka Bonformacs Cener Kyoo Unversy Oulne Three Problems for probablsc models n machne learnng. Compung lkelhood 2. Learnng 3. Parsng (predcon

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecure 9 Hamlonan Equaons of Moon (Chaper 8) Wha We Dd Las Tme Consruced Hamlonan formalsm H ( q, p, ) = q p L( q, q, ) H p = q H q = p H = L Equvalen o Lagrangan formalsm Smpler, bu

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecure 0 Canoncal Transformaons (Chaper 9) Wha We Dd Las Tme Hamlon s Prncple n he Hamlonan formalsm Dervaon was smple δi δ Addonal end-pon consrans pq H( q, p, ) d 0 δ q ( ) δq ( ) δ

More information

. The geometric multiplicity is dim[ker( λi. number of linearly independent eigenvectors associated with this eigenvalue.

. The geometric multiplicity is dim[ker( λi. number of linearly independent eigenvectors associated with this eigenvalue. Lnear Algebra Lecure # Noes We connue wh he dscusson of egenvalues, egenvecors, and dagonalzably of marces We wan o know, n parcular wha condons wll assure ha a marx can be dagonalzed and wha he obsrucons

More information

Department of Economics University of Toronto

Department of Economics University of Toronto Deparmen of Economcs Unversy of Torono ECO408F M.A. Economercs Lecure Noes on Heeroskedascy Heeroskedascy o Ths lecure nvolves lookng a modfcaons we need o make o deal wh he regresson model when some of

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecure 9 Hamlonan Equaons of Moon (Chaper 8) Wha We Dd Las Tme Consruced Hamlonan formalsm Hqp (,,) = qp Lqq (,,) H p = q H q = p H L = Equvalen o Lagrangan formalsm Smpler, bu wce as

More information

Computing Relevance, Similarity: The Vector Space Model

Computing Relevance, Similarity: The Vector Space Model Compung Relevance, Smlary: The Vecor Space Model Based on Larson and Hears s sldes a UC-Bereley hp://.sms.bereley.edu/courses/s0/f00/ aabase Managemen Sysems, R. Ramarshnan ocumen Vecors v ocumens are

More information

Lecture 6: Learning for Control (Generalised Linear Regression)

Lecture 6: Learning for Control (Generalised Linear Regression) Lecure 6: Learnng for Conrol (Generalsed Lnear Regresson) Conens: Lnear Mehods for Regresson Leas Squares, Gauss Markov heorem Recursve Leas Squares Lecure 6: RLSC - Prof. Sehu Vjayakumar Lnear Regresson

More information

. The geometric multiplicity is dim[ker( λi. A )], i.e. the number of linearly independent eigenvectors associated with this eigenvalue.

. The geometric multiplicity is dim[ker( λi. A )], i.e. the number of linearly independent eigenvectors associated with this eigenvalue. Mah E-b Lecure #0 Noes We connue wh he dscusson of egenvalues, egenvecors, and dagonalzably of marces We wan o know, n parcular wha condons wll assure ha a marx can be dagonalzed and wha he obsrucons are

More information

UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 2017 EXAMINATION

UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 2017 EXAMINATION INTERNATIONAL TRADE T. J. KEHOE UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 27 EXAMINATION Please answer wo of he hree quesons. You can consul class noes, workng papers, and arcles whle you are workng on he

More information

Robustness Experiments with Two Variance Components

Robustness Experiments with Two Variance Components Naonal Insue of Sandards and Technology (NIST) Informaon Technology Laboraory (ITL) Sascal Engneerng Dvson (SED) Robusness Expermens wh Two Varance Componens by Ana Ivelsse Avlés avles@ns.gov Conference

More information

On One Analytic Method of. Constructing Program Controls

On One Analytic Method of. Constructing Program Controls Appled Mahemacal Scences, Vol. 9, 05, no. 8, 409-407 HIKARI Ld, www.m-hkar.com hp://dx.do.org/0.988/ams.05.54349 On One Analyc Mehod of Consrucng Program Conrols A. N. Kvko, S. V. Chsyakov and Yu. E. Balyna

More information

Epistemic Game Theory: Online Appendix

Epistemic Game Theory: Online Appendix Epsemc Game Theory: Onlne Appendx Edde Dekel Lucano Pomao Marcano Snscalch July 18, 2014 Prelmnares Fx a fne ype srucure T I, S, T, β I and a probably µ S T. Le T µ I, S, T µ, βµ I be a ype srucure ha

More information

Online Supplement for Dynamic Multi-Technology. Production-Inventory Problem with Emissions Trading

Online Supplement for Dynamic Multi-Technology. Production-Inventory Problem with Emissions Trading Onlne Supplemen for Dynamc Mul-Technology Producon-Invenory Problem wh Emssons Tradng by We Zhang Zhongsheng Hua Yu Xa and Baofeng Huo Proof of Lemma For any ( qr ) Θ s easy o verfy ha he lnear programmng

More information

Dynamic Team Decision Theory. EECS 558 Project Shrutivandana Sharma and David Shuman December 10, 2005

Dynamic Team Decision Theory. EECS 558 Project Shrutivandana Sharma and David Shuman December 10, 2005 Dynamc Team Decson Theory EECS 558 Proec Shruvandana Sharma and Davd Shuman December 0, 005 Oulne Inroducon o Team Decson Theory Decomposon of he Dynamc Team Decson Problem Equvalence of Sac and Dynamc

More information

Fall 2010 Graduate Course on Dynamic Learning

Fall 2010 Graduate Course on Dynamic Learning Fall 200 Graduae Course on Dynamc Learnng Chaper 4: Parcle Flers Sepember 27, 200 Byoung-Tak Zhang School of Compuer Scence and Engneerng & Cognve Scence and Bran Scence Programs Seoul aonal Unversy hp://b.snu.ac.kr/~bzhang/

More information

Appendix to Online Clustering with Experts

Appendix to Online Clustering with Experts A Appendx o Onlne Cluserng wh Expers Furher dscusson of expermens. Here we furher dscuss expermenal resuls repored n he paper. Ineresngly, we observe ha OCE (and n parcular Learn- ) racks he bes exper

More information

Testing a new idea to solve the P = NP problem with mathematical induction

Testing a new idea to solve the P = NP problem with mathematical induction Tesng a new dea o solve he P = NP problem wh mahemacal nducon Bacground P and NP are wo classes (ses) of languages n Compuer Scence An open problem s wheher P = NP Ths paper ess a new dea o compare he

More information

Appendix H: Rarefaction and extrapolation of Hill numbers for incidence data

Appendix H: Rarefaction and extrapolation of Hill numbers for incidence data Anne Chao Ncholas J Goell C seh lzabeh L ander K Ma Rober K Colwell and Aaron M llson 03 Rarefacon and erapolaon wh ll numbers: a framewor for samplng and esmaon n speces dversy sudes cology Monographs

More information

Lecture VI Regression

Lecture VI Regression Lecure VI Regresson (Lnear Mehods for Regresson) Conens: Lnear Mehods for Regresson Leas Squares, Gauss Markov heorem Recursve Leas Squares Lecure VI: MLSC - Dr. Sehu Vjayakumar Lnear Regresson Model M

More information

John Geweke a and Gianni Amisano b a Departments of Economics and Statistics, University of Iowa, USA b European Central Bank, Frankfurt, Germany

John Geweke a and Gianni Amisano b a Departments of Economics and Statistics, University of Iowa, USA b European Central Bank, Frankfurt, Germany Herarchcal Markov Normal Mxure models wh Applcaons o Fnancal Asse Reurns Appendx: Proofs of Theorems and Condonal Poseror Dsrbuons John Geweke a and Gann Amsano b a Deparmens of Economcs and Sascs, Unversy

More information

Tight results for Next Fit and Worst Fit with resource augmentation

Tight results for Next Fit and Worst Fit with resource augmentation Tgh resuls for Nex F and Wors F wh resource augmenaon Joan Boyar Leah Epsen Asaf Levn Asrac I s well known ha he wo smple algorhms for he classc n packng prolem, NF and WF oh have an approxmaon rao of

More information

Lecture 2 L n i e n a e r a M od o e d l e s

Lecture 2 L n i e n a e r a M od o e d l e s Lecure Lnear Models Las lecure You have learned abou ha s machne learnng Supervsed learnng Unsupervsed learnng Renforcemen learnng You have seen an eample learnng problem and he general process ha one

More information

Math 128b Project. Jude Yuen

Math 128b Project. Jude Yuen Mah 8b Proec Jude Yuen . Inroducon Le { Z } be a sequence of observed ndependen vecor varables. If he elemens of Z have a on normal dsrbuon hen { Z } has a mean vecor Z and a varancecovarance marx z. Geomercally

More information

Li An-Ping. Beijing , P.R.China

Li An-Ping. Beijing , P.R.China A New Type of Cpher: DICING_csb L An-Png Bejng 100085, P.R.Chna apl0001@sna.com Absrac: In hs paper, we wll propose a new ype of cpher named DICING_csb, whch s derved from our prevous sream cpher DICING.

More information

ON THE WEAK LIMITS OF SMOOTH MAPS FOR THE DIRICHLET ENERGY BETWEEN MANIFOLDS

ON THE WEAK LIMITS OF SMOOTH MAPS FOR THE DIRICHLET ENERGY BETWEEN MANIFOLDS ON THE WEA LIMITS OF SMOOTH MAPS FOR THE DIRICHLET ENERGY BETWEEN MANIFOLDS FENGBO HANG Absrac. We denfy all he weak sequenal lms of smooh maps n W (M N). In parcular, hs mples a necessary su cen opologcal

More information

THE PREDICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS

THE PREDICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS THE PREICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS INTROUCTION The wo dmensonal paral dfferenal equaons of second order can be used for he smulaon of compeve envronmen n busness The arcle presens he

More information

Existence and Uniqueness Results for Random Impulsive Integro-Differential Equation

Existence and Uniqueness Results for Random Impulsive Integro-Differential Equation Global Journal of Pure and Appled Mahemacs. ISSN 973-768 Volume 4, Number 6 (8), pp. 89-87 Research Inda Publcaons hp://www.rpublcaon.com Exsence and Unqueness Resuls for Random Impulsve Inegro-Dfferenal

More information

Lecture 18: The Laplace Transform (See Sections and 14.7 in Boas)

Lecture 18: The Laplace Transform (See Sections and 14.7 in Boas) Lecure 8: The Lalace Transform (See Secons 88- and 47 n Boas) Recall ha our bg-cure goal s he analyss of he dfferenal equaon, ax bx cx F, where we emloy varous exansons for he drvng funcon F deendng on

More information

Boosted LMS-based Piecewise Linear Adaptive Filters

Boosted LMS-based Piecewise Linear Adaptive Filters 016 4h European Sgnal Processng Conference EUSIPCO) Boosed LMS-based Pecewse Lnear Adapve Flers Darush Kar and Iman Marvan Deparmen of Elecrcal and Elecroncs Engneerng Blken Unversy, Ankara, Turkey {kar,

More information

Comb Filters. Comb Filters

Comb Filters. Comb Filters The smple flers dscussed so far are characered eher by a sngle passband and/or a sngle sopband There are applcaons where flers wh mulple passbands and sopbands are requred Thecomb fler s an example of

More information

Volatility Interpolation

Volatility Interpolation Volaly Inerpolaon Prelmnary Verson March 00 Jesper Andreasen and Bran Huge Danse Mares, Copenhagen wan.daddy@danseban.com brno@danseban.com Elecronc copy avalable a: hp://ssrn.com/absrac=69497 Inro Local

More information

Clustering (Bishop ch 9)

Clustering (Bishop ch 9) Cluserng (Bshop ch 9) Reference: Daa Mnng by Margare Dunham (a slde source) 1 Cluserng Cluserng s unsupervsed learnng, here are no class labels Wan o fnd groups of smlar nsances Ofen use a dsance measure

More information

FTCS Solution to the Heat Equation

FTCS Solution to the Heat Equation FTCS Soluon o he Hea Equaon ME 448/548 Noes Gerald Reckenwald Porland Sae Unversy Deparmen of Mechancal Engneerng gerry@pdxedu ME 448/548: FTCS Soluon o he Hea Equaon Overvew Use he forward fne d erence

More information

WiH Wei He

WiH Wei He Sysem Idenfcaon of onlnear Sae-Space Space Baery odels WH We He wehe@calce.umd.edu Advsor: Dr. Chaochao Chen Deparmen of echancal Engneerng Unversy of aryland, College Par 1 Unversy of aryland Bacground

More information

A decision-theoretic generalization of on-line learning. and an application to boosting. AT&T Labs. 180 Park Avenue. Florham Park, NJ 07932

A decision-theoretic generalization of on-line learning. and an application to boosting. AT&T Labs. 180 Park Avenue. Florham Park, NJ 07932 A decson-heorec generalzaon of on-lne learnng and an applcaon o boosng Yoav Freund Rober E. Schapre AT&T Labs 80 Park Avenue Florham Park, NJ 07932 fyoav, schapreg@research.a.com December 9, 996 Absrac

More information

TSS = SST + SSE An orthogonal partition of the total SS

TSS = SST + SSE An orthogonal partition of the total SS ANOVA: Topc 4. Orhogonal conrass [ST&D p. 183] H 0 : µ 1 = µ =... = µ H 1 : The mean of a leas one reamen group s dfferen To es hs hypohess, a basc ANOVA allocaes he varaon among reamen means (SST) equally

More information

( ) [ ] MAP Decision Rule

( ) [ ] MAP Decision Rule Announcemens Bayes Decson Theory wh Normal Dsrbuons HW0 due oday HW o be assgned soon Proec descrpon posed Bomercs CSE 90 Lecure 4 CSE90, Sprng 04 CSE90, Sprng 04 Key Probables 4 ω class label X feaure

More information

New M-Estimator Objective Function. in Simultaneous Equations Model. (A Comparative Study)

New M-Estimator Objective Function. in Simultaneous Equations Model. (A Comparative Study) Inernaonal Mahemacal Forum, Vol. 8, 3, no., 7 - HIKARI Ld, www.m-hkar.com hp://dx.do.org/.988/mf.3.3488 New M-Esmaor Objecve Funcon n Smulaneous Equaons Model (A Comparave Sudy) Ahmed H. Youssef Professor

More information

Online Appendix for. Strategic safety stocks in supply chains with evolving forecasts

Online Appendix for. Strategic safety stocks in supply chains with evolving forecasts Onlne Appendx for Sraegc safey socs n supply chans wh evolvng forecass Tor Schoenmeyr Sephen C. Graves Opsolar, Inc. 332 Hunwood Avenue Hayward, CA 94544 A. P. Sloan School of Managemen Massachuses Insue

More information

A decision-theoretic generalization of on-line learning. and an application to boosting. AT&T Bell Laboratories. 600 Mountain Avenue

A decision-theoretic generalization of on-line learning. and an application to boosting. AT&T Bell Laboratories. 600 Mountain Avenue A decson-heorec generalzaon of on-lne learnng and an applcaon o boosng Yoav Freund Rober E. Schapre AT&T Bell Laboraores 600 Mounan Avenue Room f2b-428, 2A-424g Murray Hll, NJ 07974-0636 fyoav, schapreg@research.a.com

More information

A Generalized Online Mirror Descent with Applications to Classification and Regression

A Generalized Online Mirror Descent with Applications to Classification and Regression Journal of Machne Learnng Research 1 000 1-48 Submed 4/00; Publshed 10/00 A Generalzed Onlne Mrror Descen wh Applcaons o Classfcaon and Regresson Francesco Orabona Toyoa Technologcal Insue a Chcago 60637

More information

Time-interval analysis of β decay. V. Horvat and J. C. Hardy

Time-interval analysis of β decay. V. Horvat and J. C. Hardy Tme-nerval analyss of β decay V. Horva and J. C. Hardy Work on he even analyss of β decay [1] connued and resuled n he developmen of a novel mehod of bea-decay me-nerval analyss ha produces hghly accurae

More information

Advanced Macroeconomics II: Exchange economy

Advanced Macroeconomics II: Exchange economy Advanced Macroeconomcs II: Exchange economy Krzyszof Makarsk 1 Smple deermnsc dynamc model. 1.1 Inroducon Inroducon Smple deermnsc dynamc model. Defnons of equlbrum: Arrow-Debreu Sequenal Recursve Equvalence

More information

Chapter 6: AC Circuits

Chapter 6: AC Circuits Chaper 6: AC Crcus Chaper 6: Oulne Phasors and he AC Seady Sae AC Crcus A sable, lnear crcu operang n he seady sae wh snusodal excaon (.e., snusodal seady sae. Complee response forced response naural response.

More information

Ordinary Differential Equations in Neuroscience with Matlab examples. Aim 1- Gain understanding of how to set up and solve ODE s

Ordinary Differential Equations in Neuroscience with Matlab examples. Aim 1- Gain understanding of how to set up and solve ODE s Ordnary Dfferenal Equaons n Neuroscence wh Malab eamples. Am - Gan undersandng of how o se up and solve ODE s Am Undersand how o se up an solve a smple eample of he Hebb rule n D Our goal a end of class

More information

CH.3. COMPATIBILITY EQUATIONS. Continuum Mechanics Course (MMC) - ETSECCPB - UPC

CH.3. COMPATIBILITY EQUATIONS. Continuum Mechanics Course (MMC) - ETSECCPB - UPC CH.3. COMPATIBILITY EQUATIONS Connuum Mechancs Course (MMC) - ETSECCPB - UPC Overvew Compably Condons Compably Equaons of a Poenal Vecor Feld Compably Condons for Infnesmal Srans Inegraon of he Infnesmal

More information

FI 3103 Quantum Physics

FI 3103 Quantum Physics /9/4 FI 33 Quanum Physcs Aleander A. Iskandar Physcs of Magnesm and Phooncs Research Grou Insu Teknolog Bandung Basc Conces n Quanum Physcs Probably and Eecaon Value Hesenberg Uncerany Prncle Wave Funcon

More information

Should Exact Index Numbers have Standard Errors? Theory and Application to Asian Growth

Should Exact Index Numbers have Standard Errors? Theory and Application to Asian Growth Should Exac Index umbers have Sandard Errors? Theory and Applcaon o Asan Growh Rober C. Feensra Marshall B. Rensdorf ovember 003 Proof of Proposon APPEDIX () Frs, we wll derve he convenonal Sao-Vara prce

More information

Relative controllability of nonlinear systems with delays in control

Relative controllability of nonlinear systems with delays in control Relave conrollably o nonlnear sysems wh delays n conrol Jerzy Klamka Insue o Conrol Engneerng, Slesan Techncal Unversy, 44- Glwce, Poland. phone/ax : 48 32 37227, {jklamka}@a.polsl.glwce.pl Keywor: Conrollably.

More information

Cubic Bezier Homotopy Function for Solving Exponential Equations

Cubic Bezier Homotopy Function for Solving Exponential Equations Penerb Journal of Advanced Research n Compung and Applcaons ISSN (onlne: 46-97 Vol. 4, No.. Pages -8, 6 omoopy Funcon for Solvng Eponenal Equaons S. S. Raml *,,. Mohamad Nor,a, N. S. Saharzan,b and M.

More information

Let s treat the problem of the response of a system to an applied external force. Again,

Let s treat the problem of the response of a system to an applied external force. Again, Page 33 QUANTUM LNEAR RESPONSE FUNCTON Le s rea he problem of he response of a sysem o an appled exernal force. Agan, H() H f () A H + V () Exernal agen acng on nernal varable Hamlonan for equlbrum sysem

More information

e-journal Reliability: Theory& Applications No 2 (Vol.2) Vyacheslav Abramov

e-journal Reliability: Theory& Applications No 2 (Vol.2) Vyacheslav Abramov June 7 e-ournal Relably: Theory& Applcaons No (Vol. CONFIDENCE INTERVALS ASSOCIATED WITH PERFORMANCE ANALYSIS OF SYMMETRIC LARGE CLOSED CLIENT/SERVER COMPUTER NETWORKS Absrac Vyacheslav Abramov School

More information

Attribute Reduction Algorithm Based on Discernibility Matrix with Algebraic Method GAO Jing1,a, Ma Hui1, Han Zhidong2,b

Attribute Reduction Algorithm Based on Discernibility Matrix with Algebraic Method GAO Jing1,a, Ma Hui1, Han Zhidong2,b Inernaonal Indusral Informacs and Compuer Engneerng Conference (IIICEC 05) Arbue educon Algorhm Based on Dscernbly Marx wh Algebrac Mehod GAO Jng,a, Ma Hu, Han Zhdong,b Informaon School, Capal Unversy

More information

@FMI c Kyung Moon Sa Co.

@FMI c Kyung Moon Sa Co. Annals of Fuzzy Mahemacs and Informacs Volume 8, No. 2, (Augus 2014), pp. 245 257 ISSN: 2093 9310 (prn verson) ISSN: 2287 6235 (elecronc verson) hp://www.afm.or.kr @FMI c Kyung Moon Sa Co. hp://www.kyungmoon.com

More information

Multiclass Boosting for Weak Classifiers

Multiclass Boosting for Weak Classifiers Journal of Machne Learnng Research 6 2005) 89 20 Submed 6/03; Revsed 7/04; Publshed 2/05 Mulclass Boosng for Weak Classfers Günher Ebl Karl Peer Pfeffer Deparmen of Bosascs Unversy of Innsbruck Schöpfsrasse

More information

HEAT CONDUCTION PROBLEM IN A TWO-LAYERED HOLLOW CYLINDER BY USING THE GREEN S FUNCTION METHOD

HEAT CONDUCTION PROBLEM IN A TWO-LAYERED HOLLOW CYLINDER BY USING THE GREEN S FUNCTION METHOD Journal of Appled Mahemacs and Compuaonal Mechancs 3, (), 45-5 HEAT CONDUCTION PROBLEM IN A TWO-LAYERED HOLLOW CYLINDER BY USING THE GREEN S FUNCTION METHOD Sansław Kukla, Urszula Sedlecka Insue of Mahemacs,

More information

Hidden Markov Models Following a lecture by Andrew W. Moore Carnegie Mellon University

Hidden Markov Models Following a lecture by Andrew W. Moore Carnegie Mellon University Hdden Markov Models Followng a lecure by Andrew W. Moore Carnege Mellon Unversy www.cs.cmu.edu/~awm/uorals A Markov Sysem Has N saes, called s, s 2.. s N s 2 There are dscree meseps, 0,, s s 3 N 3 0 Hdden

More information

DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL

DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL Sco Wsdom, John Hershey 2, Jonahan Le Roux 2, and Shnj Waanabe 2 Deparmen o Elecrcal Engneerng, Unversy o Washngon, Seale, WA, USA

More information

Second-Order Non-Stationary Online Learning for Regression

Second-Order Non-Stationary Online Learning for Regression Second-Order Non-Saonary Onlne Learnng for Regresson Nna Vas, Edward Moroshko, and Koby Crammer, Fellow, IEEE arxv:303040v cslg] Mar 03 Absrac he goal of a learner, n sandard onlne learnng, s o have he

More information

Survival Analysis and Reliability. A Note on the Mean Residual Life Function of a Parallel System

Survival Analysis and Reliability. A Note on the Mean Residual Life Function of a Parallel System Communcaons n Sascs Theory and Mehods, 34: 475 484, 2005 Copyrgh Taylor & Francs, Inc. ISSN: 0361-0926 prn/1532-415x onlne DOI: 10.1081/STA-200047430 Survval Analyss and Relably A Noe on he Mean Resdual

More information

Performance Analysis for a Network having Standby Redundant Unit with Waiting in Repair

Performance Analysis for a Network having Standby Redundant Unit with Waiting in Repair TECHNI Inernaonal Journal of Compung Scence Communcaon Technologes VOL.5 NO. July 22 (ISSN 974-3375 erformance nalyss for a Nework havng Sby edundan Un wh ang n epar Jendra Sngh 2 abns orwal 2 Deparmen

More information

THEORETICAL AUTOCORRELATIONS. ) if often denoted by γ. Note that

THEORETICAL AUTOCORRELATIONS. ) if often denoted by γ. Note that THEORETICAL AUTOCORRELATIONS Cov( y, y ) E( y E( y))( y E( y)) ρ = = Var( y) E( y E( y)) =,, L ρ = and Cov( y, y ) s ofen denoed by whle Var( y ) f ofen denoed by γ. Noe ha γ = γ and ρ = ρ and because

More information

CHAPTER 5: MULTIVARIATE METHODS

CHAPTER 5: MULTIVARIATE METHODS CHAPER 5: MULIVARIAE MEHODS Mulvarae Daa 3 Mulple measuremens (sensors) npus/feaures/arbues: -varae N nsances/observaons/eamples Each row s an eample Each column represens a feaure X a b correspons o he

More information

Bayes rule for a classification problem INF Discriminant functions for the normal density. Euclidean distance. Mahalanobis distance

Bayes rule for a classification problem INF Discriminant functions for the normal density. Euclidean distance. Mahalanobis distance INF 43 3.. Repeon Anne Solberg (anne@f.uo.no Bayes rule for a classfcaon problem Suppose we have J, =,...J classes. s he class label for a pxel, and x s he observed feaure vecor. We can use Bayes rule

More information

Dual Approximate Dynamic Programming for Large Scale Hydro Valleys

Dual Approximate Dynamic Programming for Large Scale Hydro Valleys Dual Approxmae Dynamc Programmng for Large Scale Hydro Valleys Perre Carpener and Jean-Phlppe Chanceler 1 ENSTA ParsTech and ENPC ParsTech CMM Workshop, January 2016 1 Jon work wh J.-C. Alas, suppored

More information

Notes on the stability of dynamic systems and the use of Eigen Values.

Notes on the stability of dynamic systems and the use of Eigen Values. Noes on he sabl of dnamc ssems and he use of Egen Values. Source: Macro II course noes, Dr. Davd Bessler s Tme Seres course noes, zarads (999) Ineremporal Macroeconomcs chaper 4 & Techncal ppend, and Hamlon

More information

GMM parameter estimation. Xiaoye Lu CMPS290c Final Project

GMM parameter estimation. Xiaoye Lu CMPS290c Final Project GMM paraeer esaon Xaoye Lu M290c Fnal rojec GMM nroducon Gaussan ure Model obnaon of several gaussan coponens Noaon: For each Gaussan dsrbuon:, s he ean and covarance ar. A GMM h ures(coponens): p ( 2π

More information

arxiv: v1 [cs.sy] 2 Sep 2014

arxiv: v1 [cs.sy] 2 Sep 2014 Noname manuscrp No. wll be nsered by he edor Sgnalng for Decenralzed Roung n a Queueng Nework Y Ouyang Demoshens Tenekezs Receved: dae / Acceped: dae arxv:409.0887v [cs.sy] Sep 04 Absrac A dscree-me decenralzed

More information

AT&T Labs Research, Shannon Laboratory, 180 Park Avenue, Room A279, Florham Park, NJ , USA

AT&T Labs Research, Shannon Laboratory, 180 Park Avenue, Room A279, Florham Park, NJ , USA Machne Learnng, 43, 65 91, 001 c 001 Kluwer Acadec Publshers. Manufacured n The Neherlands. Drfng Gaes ROBERT E. SCHAPIRE schapre@research.a.co AT&T Labs Research, Shannon Laboraory, 180 Park Avenue, Roo

More information

Hidden Markov Models

Hidden Markov Models 11-755 Machne Learnng for Sgnal Processng Hdden Markov Models Class 15. 12 Oc 2010 1 Admnsrva HW2 due Tuesday Is everyone on he projecs page? Where are your projec proposals? 2 Recap: Wha s an HMM Probablsc

More information

NATIONAL UNIVERSITY OF SINGAPORE PC5202 ADVANCED STATISTICAL MECHANICS. (Semester II: AY ) Time Allowed: 2 Hours

NATIONAL UNIVERSITY OF SINGAPORE PC5202 ADVANCED STATISTICAL MECHANICS. (Semester II: AY ) Time Allowed: 2 Hours NATONAL UNVERSTY OF SNGAPORE PC5 ADVANCED STATSTCAL MECHANCS (Semeser : AY 1-13) Tme Allowed: Hours NSTRUCTONS TO CANDDATES 1. Ths examnaon paper conans 5 quesons and comprses 4 prned pages.. Answer all

More information

ABSTRACT KEYWORDS. Bonus-malus systems, frequency component, severity component. 1. INTRODUCTION

ABSTRACT KEYWORDS. Bonus-malus systems, frequency component, severity component. 1. INTRODUCTION EERAIED BU-MAU YTEM ITH A FREQUECY AD A EVERITY CMET A IDIVIDUA BAI I AUTMBIE IURACE* BY RAHIM MAHMUDVAD AD HEI HAAI ABTRACT Frangos and Vronos (2001) proposed an opmal bonus-malus sysems wh a frequency

More information

Bundling with Customer Self-Selection: A Simple Approach to Bundling Low Marginal Cost Goods On-Line Appendix

Bundling with Customer Self-Selection: A Simple Approach to Bundling Low Marginal Cost Goods On-Line Appendix Bundlng wh Cusomer Self-Selecon: A Smple Approach o Bundlng Low Margnal Cos Goods On-Lne Appendx Lorn M. H Unversy of Pennsylvana, Wharon School 57 Jon M. Hunsman Hall Phladelpha, PA 94 lh@wharon.upenn.edu

More information

Anomaly Detection. Lecture Notes for Chapter 9. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar

Anomaly Detection. Lecture Notes for Chapter 9. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Anomaly eecon Lecure Noes for Chaper 9 Inroducon o aa Mnng, 2 nd Edon by Tan, Senbach, Karpane, Kumar 2/14/18 Inroducon o aa Mnng, 2nd Edon 1 Anomaly/Ouler eecon Wha are anomales/oulers? The se of daa

More information

Introduction ( Week 1-2) Course introduction A brief introduction to molecular biology A brief introduction to sequence comparison Part I: Algorithms

Introduction ( Week 1-2) Course introduction A brief introduction to molecular biology A brief introduction to sequence comparison Part I: Algorithms Course organzaon Inroducon Wee -2) Course nroducon A bref nroducon o molecular bology A bref nroducon o sequence comparson Par I: Algorhms for Sequence Analyss Wee 3-8) Chaper -3, Models and heores» Probably

More information

Machine Learning Linear Regression

Machine Learning Linear Regression Machne Learnng Lnear Regresson Lesson 3 Lnear Regresson Bascs of Regresson Leas Squares esmaon Polynomal Regresson Bass funcons Regresson model Regularzed Regresson Sascal Regresson Mamum Lkelhood (ML)

More information

J i-1 i. J i i+1. Numerical integration of the diffusion equation (I) Finite difference method. Spatial Discretization. Internal nodes.

J i-1 i. J i i+1. Numerical integration of the diffusion equation (I) Finite difference method. Spatial Discretization. Internal nodes. umercal negraon of he dffuson equaon (I) Fne dfference mehod. Spaal screaon. Inernal nodes. R L V For hermal conducon le s dscree he spaal doman no small fne spans, =,,: Balance of parcles for an nernal

More information

Political Economy of Institutions and Development: Problem Set 2 Due Date: Thursday, March 15, 2019.

Political Economy of Institutions and Development: Problem Set 2 Due Date: Thursday, March 15, 2019. Polcal Economy of Insuons and Developmen: 14.773 Problem Se 2 Due Dae: Thursday, March 15, 2019. Please answer Quesons 1, 2 and 3. Queson 1 Consder an nfne-horzon dynamc game beween wo groups, an ele and

More information