On-line Reinforcement Learning Using Incremental Kernel-Based Stochastic Factorization

Size: px
Start display at page:

Download "On-line Reinforcement Learning Using Incremental Kernel-Based Stochastic Factorization"

Transcription

1 On-lne Renforcement Lernng Usng Incrementl Kernel-Bsed Stochstc Fctorzton André M. S. Brreto School of Computer Scence McGll Unversty Montrel, Cnd Don Precup School of Computer Scence McGll Unversty Montrel, Cnd Abstrct Joelle Pneu School of Computer Scence McGll Unversty Montrel, Cnd Kernel-bsed stochstc fctorzton KBSF s n lgorthm for solvng renforcement lernng tsks wth contnuous stte spces whch bulds Mrkov decson process MDP bsed on set of smple trnstons. Wht sets KBSF prt from other kernel-bsed pproches s the fct tht the sze of ts MDP s ndependent of the number of trnstons, whch mkes t possble to control the trde-off between the qulty of the resultng pproxmton nd the ssocted computtonl cost. However, KBSF s memory usge grows lnerly wth the number of trnstons, precludng ts pplcton n scenros where lrge mount of dt must be processed. In ths pper we show tht t s possble to construct KBSF s MDP n fully ncrementl wy, thus freeng the spce complexty of ths lgorthm from ts dependence on the number of smple trnstons. The ncrementl verson of KBSF s ble to process n rbtrry mount of dt, whch results n model-bsed renforcement lernng lgorthm tht cn be used to solve contnuous MDPs n both off-lne nd on-lne regmes. We present theoretcl results showng tht KBSF cn pproxmte the vlue functon tht would be computed by conventonl kernel-bsed lernng wth rbtrry precson. We emprclly demonstrte the effectveness of the proposed lgorthm n the chllengng threepole blncng tsk, n whch the blty to process lrge number of trnstons s crucl for success. Introducton The tsk of lernng polcy for sequentl decson problem wth contnuous stte spce s long-stndng chllenge tht hs ttrcted the ttenton of the renforcement lernng communty for yers. Among the mny pproches tht hve been proposed to solve ths problem, kernel-bsed renforcement lernng KBRL stnds out for ts good theoretcl gurntees [, 2]. KBRL solves contnuous stte-spce Mrkov decson process MDP usng fnte model constructed bsed on smple trnstons only. By cstng the problem s non-prmetrc pproxmton, t provdes sttstclly consstent wy of pproxmtng n MDP s vlue functon. Moreover, snce t comes down to the soluton of fnte model, KBRL lwys converges to unque soluton. Unfortuntely, the good theoretcl propertes of kernel-bsed lernng come t prce: snce the model constructed by KBRL grows wth the mount of smple trnstons, the number of opertons performed by ths lgorthm quckly becomes prohbtvely lrge s more dt become vlble. Such computtonl burden severely lmts the pplcblty of KBRL to rel renforcement lernng RL problems. Relzng tht, mny reserchers hve proposed wys of turnng KBRL nto more prctcl tool [3, 4, 5]. In ths pper we focus on our own pproch to leverge KBRL, n lgorthm clled kernel-bsed stochstc fctorzton KBSF [4]. KBSF uses KBRL s kernel-bsed strtegy to perform soft ggregton of the sttes of ts MDP. By dong so, our lgorthm s ble to summrze the nformton contned n KBRL s model n n MDP whose sze s ndependent of the number of smple trnstons. KBSF enjoys good theoretcl

2 gurntees nd hs shown excellent performnce on severl tsks [4]. The mn lmtton of the lgorthm s the fct tht, n order to construct ts model, t uses n mount of memory tht grows lnerly wth the number of smple trnstons. Although ths s sgnfcnt mprovement over KBRL, t stll hnders the pplcton of KBSF n scenros n whch lrge mount of dt must be processed, such s n complex domns or n on-lne renforcement lernng. In ths pper we show tht t s possble to construct KBSF s MDP n fully ncrementl wy, thus freeng the spce complexty of ths lgorthm from ts dependence on the number of smple trnstons. In order to dstngush t from ts orgnl, btch counterprt, we cll ths new verson of our lgorthm ncrementl KBSF, or KBSF for short. As wll be seen, KBSF s ble to process n rbtrry number of smple trnstons. Ths results n model-bsed RL lgorthm tht cn be used to solve contnuous MDPs n both off-lne nd on-lne regmes. A second mportnt contrbuton of ths pper s theoretcl nlyss showng tht t s possble to control the error n the vlue-functon pproxmton performed by KBSF. In our prevous experments wth KBSF, we defned the model used by ths lgorthm by clusterng the smple trnstons nd then usng the clusters s centers s the representtve sttes n the reduced MDP [4]. However, we dd not provde theoretcl justfcton for such strtegy. In ths pper we fll ths gp by showng tht we cn pproxmte KBRL s vlue functon t ny desred level of ccurcy by mnmzng the dstnce from smpled stte to the nerest representtve stte. Besdes ts theoretcl nterest, the bound s lso relevnt from prctcl pont of vew, snce t cn be used n KBSF to gude the on-lne selecton of representtve sttes. Fnlly, thrd contrbuton of ths pper s n emprcl demonstrton of the performnce of KBSF n new, chllengng control problem: the trple pole-blncng tsk, n extenson of the well-known double pole-blncng domn. Here, KBSF s blty to process lrge number of trnstons s crucl for chevng hgh success rte, whch cnnot be esly replcted wth btch methods. 2 Bckground In renforcement lernng, n gent ntercts wth n envronment n order to fnd polcy tht mxmzes the dscounted sum of rewrds [6]. As usul, we ssume tht such n ntercton cn be modeled s Mrkov decson process MDP, [7]. An MDP s tuple M S,A,P,r,γ, where S s the stte spce nd A s the fnte cton set. In ths pper we re mostly concerned wth MDPs wth contnuous stte spces, but our strtegy wll be to pproxmte such models s fnte MDPs. In fnte MDP the mtrx P R S S gves the trnston probbltes ssocted wth cton A nd the vector r R S stores the correspondng expected rewrds. The dscount fctor γ [0, s used to gve smller weghts to rewrds receved further n the future. Consder n MDP M wth contnuous stte spce S [0,] d. Kernel-bsed renforcement lernng KBRL uses smple trnstons to derve fnte MDP tht pproxmtes the contnuous model [, 2]. Let S {s k,r k,ŝ k k,2,...,n } be smple trnstons ssocted wth cton A, where s k,ŝ k S nd r k R. Let φ : R+ R + be Lpschtz contnuous functon nd let k τ s,s be kernel functon defned s k τ s,s φ s s /τ, where s norm n R d nd τ > 0. Fnlly, defne the normlzed kernel functon ssocted wth cton s κτ s,s k τs,s / n j k τ s,s j. The model constructed by KBRL hs the followng trnston nd rewrd functons: ˆP s s{ κ τ s,s, f s ŝ, { nd ˆR 0, otherwse s,s r, f s ŝ, 0, otherwse. Snce only trnstons endng n the sttes ŝ hve non-zero probblty of occurrence, one cn defne fnte MDP ˆM composed solely of these n n sttes [2, 3]. After ˆV, the optml vlue functon of ˆM, hs been computed, the vlue of ny stte-cton pr cn be determned s: Qs, n κ τ s,s [ r + γ ˆV ŝ ], where s S nd A. Ormonet nd Sen [] proved tht, f n for ll A nd the wdths of the kernels τ shrnk t n dmssble rte, the probblty of choosng suboptml cton bsed on Qs, converges to zero. Usng dynmc progrmmng, one cn compute the optml vlue functon of ˆM, but the tme nd spce requred to do so grow fst wth the number of sttes n [7, 8]. Therefore, the use of KBRL leds to dlemm: on the one hnd, one wnts to use s mny trnstons s possble to cpture the dynmcs of M, but on the other hnd one wnts to hve n MDP ˆM of mngeble sze. 2

3 Kernel-bsed stochstc fctorzton KBSF provdes prctcl wy of weghng these two conflctng objectves [4]. Our lgorthm compresses the nformton contned n KBRL s model ˆM n n MDP M whose sze s ndependent of the number of trnstons n. The fundmentl de behnd KBSF s the stochstc-fctorzton trck, whch we now summrze. Let P R n n be trnston-probblty mtrx nd let P DK be fctorzton n whch D R n m nd K R m n re stochstc mtrces. Then, swppng the fctors D nd K yelds nother trnston mtrx P KD tht retns the bsc topology of P tht s, the number of recurrent clsses nd ther respectve reducbltes nd perodctes [9]. The nsght s tht, n some cses, one cn work wth P nsted of P; when m n, ths replcement ffects sgnfcntly the memory usge nd computng tme. KBSF results from the pplcton of the stochstc-fctorzton trck to ˆM. Let S { s, s 2,..., s m } be set of representtve sttes n S. KBSF computes mtrces Ḋ R n m nd K R m n wth elements d j κ τŝ, s j nd k j κ τ s,s j, where κ τ s defned s κ τ s, s k τ s, s / m j k τs, s j. The bsc de of the lgorthm s to replce the MDP ˆM wth M S,A, P, r,γ, where P K Ḋ nd r K r r R n s the vector composed of smple rewrds r. Thus, nsted of solvng n MDP wth n sttes, one solves model wth m sttes only. Let D [ Ḋ Ḋ2 Ḋ A... ] R m n nd let K [ K K 2... K A ] R m n. Bsed on Q R m A, the optml cton-vlue functon of M, one cn obtn n pproxmte vlue functon for ˆM s ṽ ΓD Q, where Γ s the mx opertor ppled row wse, tht s, ṽ mx D Q. We hve showed tht the error n ṽ s bounded by: ˆv ṽ γ mx ˆr D r + γ 2 Cmx mx d j +Ĉγ j 2 mx ˆP DK, 2 where s the nfnty norm, ˆv R n s the optml vlue functon of KBRL s MDP, Ĉ mx, ˆr mn, ˆr, C mx, r mn, r, nd K s mtrx K wth ll elements equl to zero except for those correspondng to mtrx K see [4] for detls. 3 Incrementl kernel-bsed stochstc fctorzton In the btch verson of KBSF, descrbed n Secton 2, the mtrces P nd vectors r re determned usng ll the trnstons n the correspondng sets S smultneously. Ths hs two undesrble consequences. Frst, the constructon of the MDP M requres n mount of memory of On mx m, where n mx mx n. Although ths s sgnfcnt mprovement over KBRL s memory usge, whch s On 2 mx, n more chllengng domns even lner dependence on n mx my be mprctcl. Second, wth btch KBSF the only wy to ncorporte new dt nto the model M s to recompute the multplcton P K Ḋ for ll ctons for whch there re new smple trnstons vlble. Even f we gnore the ssue of memory usge, ths s clerly neffcent n terms of computton. In ths secton we present n ncrementl verson of KBSF tht crcumvents these mportnt lmttons. Suppose we splt the set of smple trnstons S n two subsets S nd S 2 such tht S S 2 /0 nd S S 2 S. Wthout loss of generlty, suppose tht the smple trnstons re ndexed so tht S {s k,r k,ŝ k k,2,...,n } nd S 2 {s k,r k,ŝ k k n +,n + 2,...,n + n 2 n }. Let P S nd r S be mtrx P nd vector r computed by KBSF usng only the n trnstons n S f n 0, we defne P S 0 R m m nd r S 0 R m for ll A. We wnt to compute P S S 2 nd r S S 2 from P S, r S, nd S 2, wthout usng the set of smple trnston S. We strt wth the trnston mtrces P. We know tht n p S j t k t d tj n t k τ s,st k τŝ t, s j l k τ s,s l m l k τŝ t, s l n l k τ s,s l n n t k τ s,st k τ ŝt, s j m l k τŝ t., s l To smplfy the notton, defne n l k τ s,s l,ws 2 n +n 2 ln + k τ s,s l, nd ct j k τ s,st k τ ŝt, s j m l k τ ŝt, s, wth t {,2,...,n + n 2 }. Then, l p S S 2 j n t ct j + n +n 2 tn + ct j 3 p S j ws + n +n 2 tn + ct j.

4 Now, defnng b S 2 j n +n 2 tn + ct j, we hve the smple updte rule: p S S 2 j b S 2 j + ps j. 3 We cn pply smlr resonng to derve n updte rule for the rewrds r. We know tht n r S n l k τ s,s l k τ s,st rt n t k τ s,st rt. t Let h t k τ s,st rt, wth t {,2,...,n + n 2 }. Then, r S S 2 n t ht + n +n 2 tn + ht r S + n +n 2 tn + ht. Defnng e S 2 n +n 2 tn + ht, we hve the followng updte rule: r S S 2 e S 2 + r S. 4 Snce b S 2 j, es 2, nd w S 2 cn be computed bsed on S 2 only, we cn dscrd the smple trnstons n S fter computng P S nd r S. To do tht, we only hve to keep the vrbles. These vrbles cn be stored n A vectors w R m, resultng n modest memory overhed. Note tht we cn pply the des bove recursvely, further splttng the sets S nd S 2 n subsets of smller sze. Thus, we hve fully ncrementl wy of computng KBSF s MDP whch requres lmost no extr memory. Algorthm shows step-by-step descrpton of how to updte M bsed on set of smple trnstons. Usng ths method to updte ts model, KBSF s spce complexty drops from Onm to Om 2. Snce the mount of memory used by KBSF s now ndependent of n, t cn process n rbtrry number of smple trnstons. Algorthm Updte KBSF s MDP Input: P, r, w for ll A S for ll A Output: Updted M nd w for A do for t,...,n do z t m l k τŝ t, s l n S for,2,...,m do w n t k τ s,st for j,2,...,m do b n t k τ s,st k τ ŝt, s j /z t p j w b + p +w j w e n t k τ s,s t r t r w +w e + r w w w + w Algorthm 2 Incrementl KBSF KBSF s Representtve sttes,,2,...,m t Input: m Intervl to updte model t v Intervl to updte vlue functon n Totl number of smple trnstons Output: Approxmte vlue functon Qs, Q rbtrry mtrx n R m A P 0 R m m, r 0 R m, w 0 R m, A for t,2,...,n do Select bsed on Qs t, m κ τs t, s q Execute n s t nd observe r t nd ŝ t S S {s t,r t,ŝ t } f t mod t m 0 then Add new representtve sttes to M usng S Updte M nd w usng Algorthm nd S S /0 for ll A f t mod t v 0 updte Q Insted of ssumng tht S nd S 2 re prtton of fxed dtset S, we cn consder tht S 2 ws generted bsed on the polcy lerned by KBSF usng the trnstons n S. Thus, Algorthm provdes flexble frmework for ntegrtng lernng nd plnnng wthn KBSF. A generl descrpton of the ncrementl verson of KBSF s gven n Algorthm 2. KBSF updtes the model M nd the vlue functon Q t fxed ntervls t m nd t v, respectvely. When t m t v n, we recover the btch verson of KBSF; when t m t v, we hve n on-lne method whch stores no smple trnstons. Note tht Algorthm 2 lso llows for the ncluson of new representtve sttes to the model M. Usng Algorthm ths s esy to do: gven new representtve stte s m+, t suffces to set w m+ 0, r m+ 0, nd p m+, j p j,m+ 0 for j,2,...,m + nd ll A. Then, n the followng pplctons of Eqns 3 nd 4, the dynmcs of M wll nturlly reflect the exstence of stte s m+. 4

5 4 Theoretcl Results Our prevous experments wth KBSF suggest tht, t lest emprclly, the lgorthm s performnce mproves s m n [4]. In ths secton we present theoretcl results tht confrm ths property. The results below re prtculrly useful for KBSF becuse they provde prctcl gudnce towrds where nd when to dd new representtve sttes. Suppose we hve fxed set of smple trnstons S. We wll show tht, f we re free to defne the representtve sttes, then we cn use KBSF to pproxmte KBRL s soluton to ny desred level of ccurcy. To be more precse, let d mx, mn j ŝ s j, tht s, d s the mxmum dstnce from smpled stte ŝ to the closest representtve stte. We wll show tht, by mnmzng d, we cn mke ˆv ṽ s smll s desred cf. Eqn 2. Let ŝ ŝ k wth k rgmx mn j ŝ s j nd s s h where h rgmn j ŝ s j, tht s, ŝ s the smpled stte n S whose dstnce to the closest representtve stte s mxml, nd s s the representtve stte tht s closest to ŝ. Usng these defntons, we cn select the pr ŝ, s tht mxmzes ŝ s : ŝ ŝ b nd s s b where b rgmx ŝ s. Obvously, ŝ s d. We mke the followng smple ssumptons: ŝ nd s re unque for ll A, 0 φxdx L φ <, φx φy f x < y, v A φ,λ φ > 0, B φ 0 such tht A φ exp x φx λ φ A φ exp x f x B φ. Assumpton v mples tht the kernel functon φ wll eventully decy exponentlly. We strt by ntroducng the followng defnton: Defnton. Gven α 0,] nd s,s S, the α-rdus of k τ wth respect to s nd s s defned s ρk τ,s,s,αmx{x R + φ x/ταk τ s,s }. The exstence of ρk τ,s,s,α s gurnteed by ssumptons nd nd the fct tht φ s contnuous []. To provde some ntuton on the menng of the α-rdus of k τ, suppose tht φ s strctly decresng nd let c φ s s /τ. Then, there s s S such tht φ s s /ταc. The rdus of k τ n ths cse s s s. It should be thus obvous tht ρk τ,s,s,α s s. We cn show tht ρ hs the followng propertes proved n the supplementry mterl: Property. If s s < s s, then ρk τ,s,s,α ρk τ,s,s,α. Property 2. If α < α, then ρk τ,s,s,α > ρk τ,s,s,α. Property 3. For α 0, nd ε > 0, there s δ > 0 such tht ρk τ,s,s,α s s < ε f τ < δ. We now ntroduce noton of dssmlrty between two sttes s,s S whch s nduced by specfc set of smple trnstons S nd the choce of kernel functon: Defnton 2. Gven β > 0, theβ-dssmlrty between s nd s wth respect to κτ s defned s { ψκτ,s,s n,β k κ τ s,s k κ τ s,s k, f s s β, 0, otherwse. The prmeter β defnes the volume of the bll wthn whch we wnt to compre sttes. As we wll see, ths prmeter lnks Defntons nd 2. Note tht ψκ τ,s,s,β [0,2]. It s possble to show tht ψ stsfes the followng property see supplementry mterl: Property 4. For β > 0 nd ε > 0, there s δ > 0 such tht ψκ τ,s,s,β < ε f s s < δ. Defntons nd 2 llow us to enuncte the followng result: Lemm. For ny α 0,] nd ny t m, let ρ ρk τ,ŝ, s,α/t, let ψρ ψκτ,ŝ, s j,ρ, nd let ψmx mx ψκτ,ŝ, s j,. Then,, j mx, j Proof. See supplementry mterl. P DK + α ψ ρ + α + α ψ mx. 5 Snce ψ mx ψ ρ, one mght thnk t frst tht the rght-hnd sde of Eqn 5 decreses monotonclly s α 0. Ths s not necessrly true, though, becuse ψ ρ ψ mx s α 0 see Property 2. We re fnlly redy to prove the mn result of ths secton. 5

6 Proposton. For ny ε > 0, there re δ,δ 2 > 0 such tht ˆv ṽ < ε f d < δ nd τ < δ 2. Proof. Let ř [r,r 2,...,r A ] R n. From Eqn nd the defnton of r, we cn wrte ˆr D r ˆP ř D K r ˆP ř DK ř ˆP DK ř ˆP DK ř. 6 Thus, pluggng Eqn 6 bck nto Eqn 2, t s cler tht there s η > 0 such tht ˆv ṽ < ε f mx ˆP DK < η nd mx mx j d j < η. We strt by showng tht f d nd τ re smll enough, then mx ˆP DK < η. From Lemm we know tht, for ny set of m n representtve sttes, nd for ny α 0,], the followng must hold: mx P DK + α ψ ρ + α + α ψ MAX, where ψ MAX mx,,s ψk τ,ŝ,s, nd ψ ρ mx ψρ mx,, j ψκτ,ŝ, s j,ρ, wth ρ ρk τ,ŝ, s,α/n. Note tht ψ MAX s ndependent of the representtve sttes. Defne α such tht α/+αψ MAX < η. We hve to show tht, f we defne the representtve sttes n such wy tht d s smll enough, nd set τ ccordngly, then we cn mke ψ ρ < αη αψ MAX η. From Property 4 we know tht there s δ > 0 such tht ψ ρ < η f ρ < δ for ll A. From Property we know tht ρ ρk τ,ŝ, s,α/n for ll A. From Property 3 we know tht, for ny ε > 0, there s δ > 0 such tht ρk τ,ŝ, s,α/n < d + ε f τ < δ. Therefore, f d < δ, we cn tke ny ε < δ d to hve n upper bound δ for τ. It remns to show tht there s δ > 0 such tht mn mx j d j > η f τ < δ. Recllng tht d j k τŝ, s j/ m k k τŝ, s k, let h rgmx j k τ ŝ, s j, nd let y k τ ŝ, s h nd ˇy mx j h k τ ŝ, s j. Then, for ny, mx j d j y / y + j h k τ ŝ, s j y /y +m ˇy. From Assump. nd Prop. 3 we know tht there s δ > 0 such tht y > m η ˇy /η f τ < δ. Thus, by mkng δ mn, δ, we cn gurntee tht mn mx j d j > η. If we tke δ 2 mnδ,δ, the result follows. Proposton tells us tht, regrdless of the specfc renforcement-lernng problem t hnd, f the dstnces between smpled sttes nd the respectve nerest representtve sttes re smll enough, then we cn mke KBSF s pproxmton of KBRL s vlue functon s ccurte s desred by settng τ to smll vlue. How smll d nd τ should be depends on the prtculr choce of kernel k τ nd on the chrcterstcs of the sets of trnstons S. Of course, fxed number m of representtve sttes mposes mnmum possble vlue for d, nd f ths vlue s not smll enough decresng τ my ctully hurt the pproxmton. Agn, the optml vlue for τ n ths cse s problem-dependent. Our result supports the use of locl pproxmton bsed on representtve sttes spred over the stte spce S. Ths s n lne wth the quntzton strteges used n btch-mode kernel-bsed renforcement lernng to defne the sttes s j [4, 5]. In the cse of on-lne lernng, we hve to dptvely defne the representtve sttes s j s the smple trnstons come n. One cn thnk of severl wys of dong so [0]. In the next secton we show smple strtegy for ddng representtve sttes whch s bsed on the theoretcl results presented n ths secton. 5 Emprcl Results We now nvestgte the emprcl performnce of the ncrementl verson of KBSF. We strt wth smple tsk n whch KBSF s contrsted wth btch KBSF. Next we explot the sclblty of KBSF to solve dffcult control tsk tht, to the best of our knowledge, hs never been solved before. We use the puddle world problem s proof of concept []. In ths frst experment we show tht KBSF s ble to recover the model tht would be computed by ts btch counterprt. In order to do so, we ppled Algorthm 2 to the puddle-world tsk usng rndom polcy to select ctons. Fgure shows the result of such n experment when we vry the prmeters t m nd t v. Note tht the cse n whch t m t v 8000 corresponds to the btch verson of KBSF. As expected, the performnce of KBSF decson polces mproves grdully s the lgorthm goes through more smple trnstons, nd n generl the ntensty of the mprovement s proportonl to the mount of dt processed. More mportnt, the performnce of the decson polces fter ll smple trnstons hve been processed s essentlly the sme for ll vlues of t m nd t v, whch shows tht KBSF cn be used s tool to crcumvent KBSF s memory demnd whch s lner n n. Thus, f one hs btch of smple trnstons tht does not ft n the vlble memory, t s possble to splt the dt n chunks of smller szes nd stll get the sme vlue-functon pproxmton tht would 6

7 be computed f the entre dt set were processed t once. As shown n Fgure b, there s only smll computtonl overhed ssocted wth such strtegy ths results from unnormlzng nd normlzng the elements of P nd r severl tmes through updte rules 3 nd 4. Return ι 000 ι 2000 ι 4000 ι 8000 Seconds ι 000 ι 2000 ι 4000 ι Number of smple trnstons Performnce Number of smple trnstons b Run tmes Fgure : Results on the puddle-world tsk verged over 50 runs. KBSF used 00 representtve sttes evenly dstrbuted over the stte spce nd t m t v ι see legends. Smple trnstons were collected by rndom polcy. The gents were tested on two sets of sttes surroundng the puddles : 3 3 grd over [0., 0.3] [0.3, 0.5] nd the four sttes {0., 0.3} {0.9,.0}. But KBSF s more thn just tool for vodng the memory lmttons ssocted wth btch lernng. We llustrte ths fct wth more chllengng RL tsk. Pole blncng hs long hstory s benchmrk problem becuse t represents rch clss of unstble systems [2, 3, 4]. The objectve n ths tsk s to pply forces to wheeled crt movng long lmted trck n order to keep one or more poles hnged to the crt from fllng over [5]. There re severl vrtons of the problem wth dfferent levels of dffculty; mong them, blncng two poles t the sme tme s prtculrly hrd [6]. In ths pper we rse the br, nd dd thrd pole to the pole-blncng tsk. We performed our smultons usng the prmeters usully dopted wth the double pole tsk, except tht we dded thrd pole wth the sme length nd mss s the longer pole [5]. Ths results n problem wth n 8-dmensonl stte spce S. In our experments wth the double-pole tsk, we used 200 representtve sttes nd 0 6 smple trnstons collected by rndom polcy [4]. Here we strt our experment wth trple pole-blncng usng exctly the sme confgurton, nd then we let KBSF refne ts model M by ncorportng more smple trnstons through updte rules 3 nd 4. Specfclly, we used Algorthm 2 wth 0.3-greedy polcy, t m t v 0 6, nd n 0 7. Polcy terton ws used to compute Q t ech vluefuncton updte. As for the kernels, we dopted Gussn functons wth wdths τ 00 nd τ to mprove effcency, we used KD-tree to only compute the 50 lrgest vlues of k τ s, nd the 0 lrgest vlues of k τ ŝ,. Representtve sttes were dded to the model on-lne every tme the gent encountered smple stte ŝ for whch k τ ŝ, s j < 0.0 for ll j,2,...,m ths corresponds to settng the mxmum llowed dstnce d from smpled stte to the closest representtve stte. We compre KBSF wth ftted Q-terton usng n ensemble of 30 trees generted by Ernst et l. s extr-trees lgorthm [7]. We chose ths lgorthm becuse t hs shown excellent performnce n both benchmrk nd rel-world renforcement-lernng tsks [7, 8]. Snce ths s btch-mode lernng method, we used ts result on the ntl set of 0 6 smple trnstons s bselne for our emprcl evluton. To buld the trees, the number of cut-drectons evluted t ech node ws fxed t dms 8, nd the mnmum number of elements requred to splt node, denoted here by η mn, ws frst set to 000 nd then to 00. The lgorthm ws run for 50 tertons, wth the structure of the trees fxed fter the 0 th terton. As shown n Fgure 2, both ftted Q-terton nd btch KBSF perform poorly n the trple poleblncng tsk, wth verge success rtes below 55%. Ths suggests tht the mount of dt used Another reson for choosng ftted Q-terton ws tht some of the most nturl compettors of KBSF hve lredy been tested on the smpler double pole-blncng tsk, wth dsppontng results [9, 4]. 7

8 by these lgorthms s nsuffcent to descrbe the dynmcs of the control tsk. Of course, we could gve more smple trnstons to ftted Q-terton nd btch KBSF. Note however tht, snce they re btch-lernng methods, there s n nherent lmt on the mount of dt tht these lgorthms cn use to construct ther pproxmton. In contrst, the mount of memory requred by KBSF s ndependent of the number of smple trnstons n. Ths fct together wth the fct tht KBSF s computtonl complexty s only lner n n llow our lgorthm to process lrge mount of dt wthn resonble tme. Ths cn be observed n Fgure 2b, whch shows tht KBSF cn buld n pproxmton usng 0 7 trnstons n under 20 mnutes. As reference for comprson, ftted Q-terton usng η mn 000 took n verge of hour nd 8 mnutes to process 0 tmes less dt. Successful epsodes KBSF TREE 000 TREE 00 Btch KBSF 2e+06 4e+06 6e+06 8e+06 e+07 Number of smple trnstons Seconds log Btch KBSF KBSF TREE 000 TREE 00 2e+06 4e+06 6e+06 8e+06 e+07 Number of smple trnstons Number of representtve sttes e+06 6e+06 e+07 Number of smple trnstons Performnce b Run tmes c Sze of KBSF s MDP Fgure 2: Results on the trple pole-blncng tsk verged over 50 runs. The vlues correspond to the frcton of epsodes ntted from the test sttes n whch the 3 poles could be blnced for 3000 steps one mnute of smulted tme. The test set ws composed of 256 sttes eqully dstrbuted over the hypercube defned by ±[.2m,0.24m/s,8 o,75 o /s,8 o,50 o /s,8 o,75 o /s]. Shdowed regons represent 99% confdence ntervls. As shown n Fgure 2, the blty of KBSF to process lrge number of smple trnstons llows our lgorthm to cheve success rte of pproxmtely 80%. Ths s smlr to the performnce of btch KBSF on the double-pole verson of the problem [4]. The good performnce of KBSF on the trple pole-blncng tsk s especlly mpressve when we recll tht the decson polces were evluted on set of test sttes representng ll possble drectons of nclnton of the three poles. In order to cheve the sme level of performnce wth KBSF, pproxmtely 2 Gb of memory would be necessry, even usng sprse kernels, wheres KBSF used less thn 0.03 Gb of memory. To conclude, observe n Fgure 2c how the number of representtve sttes m grows s functon of the number of smple trnstons processed by KBSF. As expected, n the begnnng of the lernng process m grows fst, reflectng the fct tht some relevnt regons of the stte spce hve not been vsted yet. As more nd more dt come n, the number of representtve sttes strts to stblze. 6 Concluson Ths pper presented two contrbutons, one prctcl nd one theoretcl. The prctcl contrbuton s KBSF, the ncrementl verson of KBSF. KBSF retns ll the nce propertes of ts precursor: t s smple, fst, nd enjoys good theoretcl gurntees. However, snce ts memory complexty s ndependent of the number of smple trnstons, KBSF cn be ppled to dtsets of ny sze, nd t cn lso be used on-lne. To show how KBSF s blty to process lrge mounts of dt cn be useful n prctce, we used the proposed lgorthm to lern how to smultneously blnce three poles, dffcult control tsk tht hd never been solved before. As for the theoretcl contrbuton, we showed tht KBSF cn pproxmte KBRL s vlue functon t ny level of ccurcy by mnmzng the dstnce between smpled sttes nd the closest representtve stte. Ths supports the quntzton strteges usully dopted n kernel-bsed RL, nd lso offers gudnce towrds where nd when to dd new representtve sttes n on-lne lernng. Acknowledgments The uthors would lke to thnk Amr mssoud Frhmnd for helpful dscussons regrdng ths work. Fundng for ths reserch ws provded by the Ntonl Insttutes of Helth grnt R2 DA09800 nd the NSERC Dscovery Grnt progrm. 8

9 References [] D. Ormonet nd S. Sen. Kernel-bsed renforcement lernng. Mchne Lernng, : 6 78, [2] D. Ormonet nd P. Glynn. Kernel-bsed renforcement lernng n verge-cost problems. IEEE Trnsctons on Automtc Control, 470: , [3] N. Jong nd P. Stone. Kernel-bsed models for renforcement lernng n contnuous stte spces. In Proceedngs of the Interntonl Conference on Mchne Lernng ICML Workshop on Kernel Mchnes nd Renforcement Lernng, [4] A. M. S. Brreto, D. Precup, nd J. Pneu. Renforcement lernng usng kernel-bsed stochstc fctorzton. In Advnces n Neurl Informton Processng Systems NIPS, pges , 20. [5] B. Kveton nd G. Theochrous. Kernel-bsed renforcement lernng on representtve sttes. In Proceedngs of the AAAI Conference on Artfcl Intellgence AAAI, pges 24 3, 202. [6] R. S. Sutton nd A. G. Brto. Renforcement Lernng: An Introducton. MIT Press, 998. [7] M. L. Putermn. Mrkov Decson Processes Dscrete Stochstc Dynmc Progrmmng. John Wley & Sons, Inc., 994. [8] M. L. Lttmn, T. L. Den, nd L. P. Kelblng. On the complexty of solvng Mrkov decson problems. In Proceedngs of the Conference on Uncertnty n Artfcl Intellgence UAI, pges , 995. [9] A. M. S. Brreto nd M. D. Frgoso. Computng the sttonry dstrbuton of fnte Mrkov chn through stochstc fctorzton. SIAM Journl on Mtrx Anlyss nd Applctons, 32: , 20. [0] Y. Engel, S. Mnnor, nd R. Mer. The kernel recursve lest squres lgorthm. IEEE Trnsctons on Sgnl Processng, 52: , [] R. S. Sutton. Generlzton n renforcement lernng: Successful exmples usng sprse corse codng. In Advnces n Neurl Informton Processng Systems NIPS, pges , 996. [2] D. Mche nd R. Chmbers. BOXES: An experment n dptve control. Mchne Intellgence 2, pges 25 33, 968. [3] C. W. Anderson. Lernng nd Problem Solvng wth Multlyer Connectonst Systems. PhD thess, Computer nd Informton Scence, Unversty of Msschusetts, 986. [4] A. G. Brto, R. S. Sutton, nd C. W. Anderson. Neuronlke dptve elements tht cn solve dffcult lernng control problems. IEEE Trnsctons on Systems, Mn, nd Cybernetcs, 3: , 983. [5] F. J. Gomez. Robust non-lner control through neuroevoluton. PhD thess, The Unversty of Texs t Austn, [6] A. P. Welnd. Evolvng neurl network controllers for unstble systems. In Proceedngs of the Interntonl Jont Conference on Neurl Networks IJCNN, pges , 99. [7] D. Ernst, P. Geurts, nd L. Wehenkel. Tree-bsed btch mode renforcement lernng. Journl of Mchne Lernng Reserch, 6: , [8] D. Ernst, G. B. Stn, J. Gonçlves, nd L. Wehenkel. Clncl dt bsed optml STI strteges for HIV: renforcement lernng pproch. In Proceedngs of the IEEE Conference on Decson nd Control CDC, pges 24 3, [9] F. Gomez, J. Schmdhuber, nd R. Mkkulnen. Effcent non-lner control through neuroevoluton. In Proceedngs of the Europen Conference on Mchne Lernng ECML, pges ,

Partially Observable Systems. 1 Partially Observable Markov Decision Process (POMDP) Formalism

Partially Observable Systems. 1 Partially Observable Markov Decision Process (POMDP) Formalism CS294-40 Lernng for Rootcs nd Control Lecture 10-9/30/2008 Lecturer: Peter Aeel Prtlly Oservle Systems Scre: Dvd Nchum Lecture outlne POMDP formlsm Pont-sed vlue terton Glol methods: polytree, enumerton,

More information

Rank One Update And the Google Matrix by Al Bernstein Signal Science, LLC

Rank One Update And the Google Matrix by Al Bernstein Signal Science, LLC Introducton Rnk One Updte And the Google Mtrx y Al Bernsten Sgnl Scence, LLC www.sgnlscence.net here re two dfferent wys to perform mtrx multplctons. he frst uses dot product formulton nd the second uses

More information

Dennis Bricker, 2001 Dept of Industrial Engineering The University of Iowa. MDP: Taxi page 1

Dennis Bricker, 2001 Dept of Industrial Engineering The University of Iowa. MDP: Taxi page 1 Denns Brcker, 2001 Dept of Industrl Engneerng The Unversty of Iow MDP: Tx pge 1 A tx serves three djcent towns: A, B, nd C. Ech tme the tx dschrges pssenger, the drver must choose from three possble ctons:

More information

Applied Statistics Qualifier Examination

Applied Statistics Qualifier Examination Appled Sttstcs Qulfer Exmnton Qul_june_8 Fll 8 Instructons: () The exmnton contns 4 Questons. You re to nswer 3 out of 4 of them. () You my use ny books nd clss notes tht you mght fnd helpful n solvng

More information

Chapter Newton-Raphson Method of Solving a Nonlinear Equation

Chapter Newton-Raphson Method of Solving a Nonlinear Equation Chpter.4 Newton-Rphson Method of Solvng Nonlner Equton After redng ths chpter, you should be ble to:. derve the Newton-Rphson method formul,. develop the lgorthm of the Newton-Rphson method,. use the Newton-Rphson

More information

Principle Component Analysis

Principle Component Analysis Prncple Component Anlyss Jng Go SUNY Bufflo Why Dmensonlty Reducton? We hve too mny dmensons o reson bout or obtn nsghts from o vsulze oo much nose n the dt Need to reduce them to smller set of fctors

More information

International Journal of Pure and Applied Sciences and Technology

International Journal of Pure and Applied Sciences and Technology Int. J. Pure Appl. Sc. Technol., () (), pp. 44-49 Interntonl Journl of Pure nd Appled Scences nd Technolog ISSN 9-67 Avlle onlne t www.jopst.n Reserch Pper Numercl Soluton for Non-Lner Fredholm Integrl

More information

Remember: Project Proposals are due April 11.

Remember: Project Proposals are due April 11. Bonformtcs ecture Notes Announcements Remember: Project Proposls re due Aprl. Clss 22 Aprl 4, 2002 A. Hdden Mrov Models. Defntons Emple - Consder the emple we tled bout n clss lst tme wth the cons. However,

More information

Two Coefficients of the Dyson Product

Two Coefficients of the Dyson Product Two Coeffcents of the Dyson Product rxv:07.460v mth.co 7 Nov 007 Lun Lv, Guoce Xn, nd Yue Zhou 3,,3 Center for Combntorcs, LPMC TJKLC Nnk Unversty, Tnjn 30007, P.R. Chn lvlun@cfc.nnk.edu.cn gn@nnk.edu.cn

More information

UNIVERSITY OF IOANNINA DEPARTMENT OF ECONOMICS. M.Sc. in Economics MICROECONOMIC THEORY I. Problem Set II

UNIVERSITY OF IOANNINA DEPARTMENT OF ECONOMICS. M.Sc. in Economics MICROECONOMIC THEORY I. Problem Set II Mcroeconomc Theory I UNIVERSITY OF IOANNINA DEPARTMENT OF ECONOMICS MSc n Economcs MICROECONOMIC THEORY I Techng: A Lptns (Note: The number of ndctes exercse s dffculty level) ()True or flse? If V( y )

More information

Definition of Tracking

Definition of Tracking Trckng Defnton of Trckng Trckng: Generte some conclusons bout the moton of the scene, objects, or the cmer, gven sequence of mges. Knowng ths moton, predct where thngs re gong to project n the net mge,

More information

Improving Anytime Point-Based Value Iteration Using Principled Point Selections

Improving Anytime Point-Based Value Iteration Using Principled Point Selections In In Proceedngs of the Twenteth Interntonl Jont Conference on Artfcl Intellgence (IJCAI-7) Improvng Anytme Pont-Bsed Vlue Iterton Usng Prncpled Pont Selectons Mchel R. Jmes, Mchel E. Smples, nd Dmtr A.

More information

Chapter Newton-Raphson Method of Solving a Nonlinear Equation

Chapter Newton-Raphson Method of Solving a Nonlinear Equation Chpter 0.04 Newton-Rphson Method o Solvng Nonlner Equton Ater redng ths chpter, you should be ble to:. derve the Newton-Rphson method ormul,. develop the lgorthm o the Newton-Rphson method,. use the Newton-Rphson

More information

Variable time amplitude amplification and quantum algorithms for linear algebra. Andris Ambainis University of Latvia

Variable time amplitude amplification and quantum algorithms for linear algebra. Andris Ambainis University of Latvia Vrble tme mpltude mplfcton nd quntum lgorthms for lner lgebr Andrs Ambns Unversty of Ltv Tlk outlne. ew verson of mpltude mplfcton;. Quntum lgorthm for testng f A s sngulr; 3. Quntum lgorthm for solvng

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science Updte Rules for Weghted Non-negtve FH*G Fctorzton Peter Peers Phlp Dutré Report CW 440, Aprl 006 Ktholeke Unverstet Leuven Deprtment of Computer Scence Celestjnenln 00A B-3001 Heverlee (Belgum) Updte Rules

More information

Statistics and Probability Letters

Statistics and Probability Letters Sttstcs nd Probblty Letters 79 (2009) 105 111 Contents lsts vlble t ScenceDrect Sttstcs nd Probblty Letters journl homepge: www.elsever.com/locte/stpro Lmtng behvour of movng verge processes under ϕ-mxng

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 9

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 9 CS434/541: Pttern Recognton Prof. Olg Veksler Lecture 9 Announcements Fnl project proposl due Nov. 1 1-2 prgrph descrpton Lte Penlt: s 1 pont off for ech d lte Assgnment 3 due November 10 Dt for fnl project

More information

18.7 Artificial Neural Networks

18.7 Artificial Neural Networks 310 18.7 Artfcl Neurl Networks Neuroscence hs hypotheszed tht mentl ctvty conssts prmrly of electrochemcl ctvty n networks of brn cells clled neurons Ths led McCulloch nd Ptts to devse ther mthemtcl model

More information

arxiv: v2 [cs.lg] 9 Nov 2017

arxiv: v2 [cs.lg] 9 Nov 2017 Renforcement Lernng under Model Msmtch Aurko Roy 1, Hun Xu 2, nd Sebstn Pokutt 2 rxv:1706.04711v2 cs.lg 9 Nov 2017 1 Google Eml: urkor@google.com 2 ISyE, Georg Insttute of Technology, Atlnt, GA, USA. Eml:

More information

Math 497C Sep 17, Curves and Surfaces Fall 2004, PSU

Math 497C Sep 17, Curves and Surfaces Fall 2004, PSU Mth 497C Sep 17, 004 1 Curves nd Surfces Fll 004, PSU Lecture Notes 3 1.8 The generl defnton of curvture; Fox-Mlnor s Theorem Let α: [, b] R n be curve nd P = {t 0,...,t n } be prtton of [, b], then the

More information

Online Appendix to. Mandating Behavioral Conformity in Social Groups with Conformist Members

Online Appendix to. Mandating Behavioral Conformity in Social Groups with Conformist Members Onlne Appendx to Mndtng Behvorl Conformty n Socl Groups wth Conformst Members Peter Grzl Andrze Bnk (Correspondng uthor) Deprtment of Economcs, The Wllms School, Wshngton nd Lee Unversty, Lexngton, 4450

More information

4. Eccentric axial loading, cross-section core

4. Eccentric axial loading, cross-section core . Eccentrc xl lodng, cross-secton core Introducton We re strtng to consder more generl cse when the xl force nd bxl bendng ct smultneousl n the cross-secton of the br. B vrtue of Snt-Vennt s prncple we

More information

In this Chapter. Chap. 3 Markov chains and hidden Markov models. Probabilistic Models. Example: CpG Islands

In this Chapter. Chap. 3 Markov chains and hidden Markov models. Probabilistic Models. Example: CpG Islands In ths Chpter Chp. 3 Mrov chns nd hdden Mrov models Bontellgence bortory School of Computer Sc. & Eng. Seoul Ntonl Unversty Seoul 5-74, Kore The probblstc model for sequence nlyss HMM (hdden Mrov model)

More information

Computing a complete histogram of an image in Log(n) steps and minimum expected memory requirements using hypercubes

Computing a complete histogram of an image in Log(n) steps and minimum expected memory requirements using hypercubes Computng complete hstogrm of n mge n Log(n) steps nd mnmum expected memory requrements usng hypercubes TAREK M. SOBH School of Engneerng, Unversty of Brdgeport, Connectcut, USA. Abstrct Ths work frst revews

More information

The Number of Rows which Equal Certain Row

The Number of Rows which Equal Certain Row Interntonl Journl of Algebr, Vol 5, 011, no 30, 1481-1488 he Number of Rows whch Equl Certn Row Ahmd Hbl Deprtment of mthemtcs Fcult of Scences Dmscus unverst Dmscus, Sr hblhmd1@gmlcom Abstrct Let be X

More information

DCDM BUSINESS SCHOOL NUMERICAL METHODS (COS 233-8) Solutions to Assignment 3. x f(x)

DCDM BUSINESS SCHOOL NUMERICAL METHODS (COS 233-8) Solutions to Assignment 3. x f(x) DCDM BUSINESS SCHOOL NUMEICAL METHODS (COS -8) Solutons to Assgnment Queston Consder the followng dt: 5 f() 8 7 5 () Set up dfference tble through fourth dfferences. (b) Wht s the mnmum degree tht n nterpoltng

More information

Soft Set Theoretic Approach for Dimensionality Reduction 1

Soft Set Theoretic Approach for Dimensionality Reduction 1 Interntonl Journl of Dtbse Theory nd pplcton Vol No June 00 Soft Set Theoretc pproch for Dmensonlty Reducton Tutut Herwn Rozd Ghzl Mustf Mt Ders Deprtment of Mthemtcs Educton nversts hmd Dhln Yogykrt Indones

More information

Using Predictions in Online Optimization: Looking Forward with an Eye on the Past

Using Predictions in Online Optimization: Looking Forward with an Eye on the Past Usng Predctons n Onlne Optmzton: Lookng Forwrd wth n Eye on the Pst Nngjun Chen Jont work wth Joshu Comden, Zhenhu Lu, Anshul Gndh, nd Adm Wermn 1 Predctons re crucl for decson mkng 2 Predctons re crucl

More information

Quiz: Experimental Physics Lab-I

Quiz: Experimental Physics Lab-I Mxmum Mrks: 18 Totl tme llowed: 35 mn Quz: Expermentl Physcs Lb-I Nme: Roll no: Attempt ll questons. 1. In n experment, bll of mss 100 g s dropped from heght of 65 cm nto the snd contner, the mpct s clled

More information

Statistics 423 Midterm Examination Winter 2009

Statistics 423 Midterm Examination Winter 2009 Sttstcs 43 Mdterm Exmnton Wnter 009 Nme: e-ml: 1. Plese prnt your nme nd e-ml ddress n the bove spces.. Do not turn ths pge untl nstructed to do so. 3. Ths s closed book exmnton. You my hve your hnd clcultor

More information

We consider a finite-state, finite-action, infinite-horizon, discounted reward Markov decision process and

We consider a finite-state, finite-action, infinite-horizon, discounted reward Markov decision process and MANAGEMENT SCIENCE Vol. 53, No. 2, Februry 2007, pp. 308 322 ssn 0025-1909 essn 1526-5501 07 5302 0308 nforms do 10.1287/mnsc.1060.0614 2007 INFORMS Bs nd Vrnce Approxmton n Vlue Functon Estmtes She Mnnor

More information

Numerical Analysis: Trapezoidal and Simpson s Rule

Numerical Analysis: Trapezoidal and Simpson s Rule nd Simpson s Mthemticl question we re interested in numericlly nswering How to we evlute I = f (x) dx? Clculus tells us tht if F(x) is the ntiderivtive of function f (x) on the intervl [, b], then I =

More information

1 Online Learning and Regret Minimization

1 Online Learning and Regret Minimization 2.997 Decision-Mking in Lrge-Scle Systems My 10 MIT, Spring 2004 Hndout #29 Lecture Note 24 1 Online Lerning nd Regret Minimiztion In this lecture, we consider the problem of sequentil decision mking in

More information

GAUSS ELIMINATION. Consider the following system of algebraic linear equations

GAUSS ELIMINATION. Consider the following system of algebraic linear equations Numercl Anlyss for Engneers Germn Jordnn Unversty GAUSS ELIMINATION Consder the followng system of lgebrc lner equtons To solve the bove system usng clsscl methods, equton () s subtrcted from equton ()

More information

INTRODUCTION TO COMPLEX NUMBERS

INTRODUCTION TO COMPLEX NUMBERS INTRODUCTION TO COMPLEX NUMBERS The numers -4, -3, -, -1, 0, 1,, 3, 4 represent the negtve nd postve rel numers termed ntegers. As one frst lerns n mddle school they cn e thought of s unt dstnce spced

More information

Reinforcement Learning with a Gaussian Mixture Model

Reinforcement Learning with a Gaussian Mixture Model Renforcement Lernng wth Gussn Mxture Model Alejndro Agostn, Member, IEEE nd Enrc Cely Abstrct Recent pproches to Renforcement Lernng (RL) wth functon pproxmton nclude Neurl Ftted Q Iterton nd the use of

More information

Review of linear algebra. Nuno Vasconcelos UCSD

Review of linear algebra. Nuno Vasconcelos UCSD Revew of lner lgebr Nuno Vsconcelos UCSD Vector spces Defnton: vector spce s set H where ddton nd sclr multplcton re defned nd stsf: ) +( + ) (+ )+ 5) λ H 2) + + H 6) 3) H, + 7) λ(λ ) (λλ ) 4) H, - + 8)

More information

LOCAL FRACTIONAL LAPLACE SERIES EXPANSION METHOD FOR DIFFUSION EQUATION ARISING IN FRACTAL HEAT TRANSFER

LOCAL FRACTIONAL LAPLACE SERIES EXPANSION METHOD FOR DIFFUSION EQUATION ARISING IN FRACTAL HEAT TRANSFER Yn, S.-P.: Locl Frctonl Lplce Seres Expnson Method for Dffuson THERMAL SCIENCE, Yer 25, Vol. 9, Suppl., pp. S3-S35 S3 LOCAL FRACTIONAL LAPLACE SERIES EXPANSION METHOD FOR DIFFUSION EQUATION ARISING IN

More information

Lecture 4: Piecewise Cubic Interpolation

Lecture 4: Piecewise Cubic Interpolation Lecture notes on Vrtonl nd Approxmte Methods n Appled Mthemtcs - A Perce UBC Lecture 4: Pecewse Cubc Interpolton Compled 6 August 7 In ths lecture we consder pecewse cubc nterpolton n whch cubc polynoml

More information

The Schur-Cohn Algorithm

The Schur-Cohn Algorithm Modelng, Estmton nd Otml Flterng n Sgnl Processng Mohmed Njm Coyrght 8, ISTE Ltd. Aendx F The Schur-Cohn Algorthm In ths endx, our m s to resent the Schur-Cohn lgorthm [] whch s often used s crteron for

More information

Pyramid Algorithms for Barycentric Rational Interpolation

Pyramid Algorithms for Barycentric Rational Interpolation Pyrmd Algorthms for Brycentrc Rtonl Interpolton K Hormnn Scott Schefer Astrct We present new perspectve on the Floter Hormnn nterpolnt. Ths nterpolnt s rtonl of degree (n, d), reproduces polynomls of degree

More information

Dynamic Power Management in a Mobile Multimedia System with Guaranteed Quality-of-Service

Dynamic Power Management in a Mobile Multimedia System with Guaranteed Quality-of-Service Dynmc Power Mngement n Moble Multmed System wth Gurnteed Qulty-of-Servce Qnru Qu, Qng Wu, nd Mssoud Pedrm Dept. of Electrcl Engneerng-Systems Unversty of Southern Clforn Los Angeles CA 90089 Outlne! Introducton

More information

An Introduction to Support Vector Machines

An Introduction to Support Vector Machines An Introducton to Support Vector Mchnes Wht s good Decson Boundry? Consder two-clss, lnerly seprble clssfcton problem Clss How to fnd the lne (or hyperplne n n-dmensons, n>)? Any de? Clss Per Lug Mrtell

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

6 Roots of Equations: Open Methods

6 Roots of Equations: Open Methods HK Km Slghtly modfed 3//9, /8/6 Frstly wrtten t Mrch 5 6 Roots of Equtons: Open Methods Smple Fed-Pont Iterton Newton-Rphson Secnt Methods MATLAB Functon: fzero Polynomls Cse Study: Ppe Frcton Brcketng

More information

p-adic Egyptian Fractions

p-adic Egyptian Fractions p-adic Egyptin Frctions Contents 1 Introduction 1 2 Trditionl Egyptin Frctions nd Greedy Algorithm 2 3 Set-up 3 4 p-greedy Algorithm 5 5 p-egyptin Trditionl 10 6 Conclusion 1 Introduction An Egyptin frction

More information

A Family of Multivariate Abel Series Distributions. of Order k

A Family of Multivariate Abel Series Distributions. of Order k Appled Mthemtcl Scences, Vol. 2, 2008, no. 45, 2239-2246 A Fmly of Multvrte Abel Seres Dstrbutons of Order k Rupk Gupt & Kshore K. Ds 2 Fculty of Scence & Technology, The Icf Unversty, Agrtl, Trpur, Ind

More information

The Regulated and Riemann Integrals

The Regulated and Riemann Integrals Chpter 1 The Regulted nd Riemnn Integrls 1.1 Introduction We will consider severl different pproches to defining the definite integrl f(x) dx of function f(x). These definitions will ll ssign the sme vlue

More information

W. We shall do so one by one, starting with I 1, and we shall do it greedily, trying

W. We shall do so one by one, starting with I 1, and we shall do it greedily, trying Vitli covers 1 Definition. A Vitli cover of set E R is set V of closed intervls with positive length so tht, for every δ > 0 nd every x E, there is some I V with λ(i ) < δ nd x I. 2 Lemm (Vitli covering)

More information

Study of Trapezoidal Fuzzy Linear System of Equations S. M. Bargir 1, *, M. S. Bapat 2, J. D. Yadav 3 1

Study of Trapezoidal Fuzzy Linear System of Equations S. M. Bargir 1, *, M. S. Bapat 2, J. D. Yadav 3 1 mercn Interntonl Journl of Reserch n cence Technology Engneerng & Mthemtcs vlble onlne t http://wwwsrnet IN (Prnt: 38-349 IN (Onlne: 38-3580 IN (CD-ROM: 38-369 IJRTEM s refereed ndexed peer-revewed multdscplnry

More information

Introduction to Numerical Integration Part II

Introduction to Numerical Integration Part II Introducton to umercl Integrton Prt II CS 75/Mth 75 Brn T. Smth, UM, CS Dept. Sprng, 998 4/9/998 qud_ Intro to Gussn Qudrture s eore, the generl tretment chnges the ntegrton prolem to ndng the ntegrl w

More information

523 P a g e. is measured through p. should be slower for lesser values of p and faster for greater values of p. If we set p*

523 P a g e. is measured through p. should be slower for lesser values of p and faster for greater values of p. If we set p* R. Smpth Kumr, R. Kruthk, R. Rdhkrshnn / Interntonl Journl of Engneerng Reserch nd Applctons (IJERA) ISSN: 48-96 www.jer.com Vol., Issue 4, July-August 0, pp.5-58 Constructon Of Mxed Smplng Plns Indexed

More information

Riemann is the Mann! (But Lebesgue may besgue to differ.)

Riemann is the Mann! (But Lebesgue may besgue to differ.) Riemnn is the Mnn! (But Lebesgue my besgue to differ.) Leo Livshits My 2, 2008 1 For finite intervls in R We hve seen in clss tht every continuous function f : [, b] R hs the property tht for every ɛ >

More information

7.2 Volume. A cross section is the shape we get when cutting straight through an object.

7.2 Volume. A cross section is the shape we get when cutting straight through an object. 7. Volume Let s revew the volume of smple sold, cylnder frst. Cylnder s volume=se re heght. As llustrted n Fgure (). Fgure ( nd (c) re specl cylnders. Fgure () s rght crculr cylnder. Fgure (c) s ox. A

More information

Best Approximation. Chapter The General Case

Best Approximation. Chapter The General Case Chpter 4 Best Approximtion 4.1 The Generl Cse In the previous chpter, we hve seen how n interpolting polynomil cn be used s n pproximtion to given function. We now wnt to find the best pproximtion to given

More information

Research Article On the Upper Bounds of Eigenvalues for a Class of Systems of Ordinary Differential Equations with Higher Order

Research Article On the Upper Bounds of Eigenvalues for a Class of Systems of Ordinary Differential Equations with Higher Order Hndw Publshng Corporton Interntonl Journl of Dfferentl Equtons Volume 0, Artcle ID 7703, pges do:055/0/7703 Reserch Artcle On the Upper Bounds of Egenvlues for Clss of Systems of Ordnry Dfferentl Equtons

More information

CALIBRATION OF SMALL AREA ESTIMATES IN BUSINESS SURVEYS

CALIBRATION OF SMALL AREA ESTIMATES IN BUSINESS SURVEYS CALIBRATION OF SMALL AREA ESTIMATES IN BUSINESS SURVES Rodolphe Prm, Ntle Shlomo Southmpton Sttstcl Scences Reserch Insttute Unverst of Southmpton Unted Kngdom SAE, August 20 The BLUE-ETS Project s fnnced

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION. Indu Manickam, Andrew S. Lan, and Richard G.

CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION. Indu Manickam, Andrew S. Lan, and Richard G. CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION Indu Mnckm, Andrew S. Ln, nd Rchrd G. Brnuk Rce Unversty ABSTRACT Optmzng the selecton of lernng resources nd prctce

More information

CIS587 - Artificial Intelligence. Uncertainty CIS587 - AI. KB for medical diagnosis. Example.

CIS587 - Artificial Intelligence. Uncertainty CIS587 - AI. KB for medical diagnosis. Example. CIS587 - rtfcl Intellgence Uncertnty K for medcl dgnoss. Exmple. We wnt to uld K system for the dgnoss of pneumon. rolem descrpton: Dsese: pneumon tent symptoms fndngs, l tests: Fever, Cough, leness, WC

More information

THE EXISTENCE-UNIQUENESS THEOREM FOR FIRST-ORDER DIFFERENTIAL EQUATIONS.

THE EXISTENCE-UNIQUENESS THEOREM FOR FIRST-ORDER DIFFERENTIAL EQUATIONS. THE EXISTENCE-UNIQUENESS THEOREM FOR FIRST-ORDER DIFFERENTIAL EQUATIONS RADON ROSBOROUGH https://intuitiveexplntionscom/picrd-lindelof-theorem/ This document is proof of the existence-uniqueness theorem

More information

Chapter 0. What is the Lebesgue integral about?

Chapter 0. What is the Lebesgue integral about? Chpter 0. Wht is the Lebesgue integrl bout? The pln is to hve tutoril sheet ech week, most often on Fridy, (to be done during the clss) where you will try to get used to the ides introduced in the previous

More information

A Tri-Valued Belief Network Model for Information Retrieval

A Tri-Valued Belief Network Model for Information Retrieval December 200 A Tr-Vlued Belef Networ Model for Informton Retrevl Fernndo Ds-Neves Computer Scence Dept. Vrgn Polytechnc Insttute nd Stte Unversty Blcsburg, VA 24060. IR models t Combnng Evdence Grphcl

More information

Bi-level models for OD matrix estimation

Bi-level models for OD matrix estimation TNK084 Trffc Theory seres Vol.4, number. My 2008 B-level models for OD mtrx estmton Hn Zhng, Quyng Meng Abstrct- Ths pper ntroduces two types of O/D mtrx estmton model: ME2 nd Grdent. ME2 s mxmum-entropy

More information

5.7 Improper Integrals

5.7 Improper Integrals 458 pplictions of definite integrls 5.7 Improper Integrls In Section 5.4, we computed the work required to lift pylod of mss m from the surfce of moon of mss nd rdius R to height H bove the surfce of the

More information

M/G/1/GD/ / System. ! Pollaczek-Khinchin (PK) Equation. ! Steady-state probabilities. ! Finding L, W q, W. ! π 0 = 1 ρ

M/G/1/GD/ / System. ! Pollaczek-Khinchin (PK) Equation. ! Steady-state probabilities. ! Finding L, W q, W. ! π 0 = 1 ρ M/G//GD/ / System! Pollcze-Khnchn (PK) Equton L q 2 2 λ σ s 2( + ρ ρ! Stedy-stte probbltes! π 0 ρ! Fndng L, q, ) 2 2 M/M/R/GD/K/K System! Drw the trnston dgrm! Derve the stedy-stte probbltes:! Fnd L,L

More information

Sequences of Intuitionistic Fuzzy Soft G-Modules

Sequences of Intuitionistic Fuzzy Soft G-Modules Interntonl Mthemtcl Forum, Vol 13, 2018, no 12, 537-546 HIKARI Ltd, wwwm-hkrcom https://doorg/1012988/mf201881058 Sequences of Intutonstc Fuzzy Soft G-Modules Velyev Kemle nd Huseynov Afq Bku Stte Unversty,

More information

SCALED GRADIENT DESCENT LEARNING RATE Reinforcement learning with light-seeking robot

SCALED GRADIENT DESCENT LEARNING RATE Reinforcement learning with light-seeking robot SCALED GRADIET DESCET LEARIG RATE Renforcement lernng wth lght-seekng robot Kry Främlng Helsnk Unversty of Technology, P.O. Box 54, FI-5 HUT, Fnlnd. Eml: Kry.Frmlng@hut.f Keywords: Abstrct: Lner functon

More information

Advanced Calculus: MATH 410 Notes on Integrals and Integrability Professor David Levermore 17 October 2004

Advanced Calculus: MATH 410 Notes on Integrals and Integrability Professor David Levermore 17 October 2004 Advnced Clculus: MATH 410 Notes on Integrls nd Integrbility Professor Dvid Levermore 17 October 2004 1. Definite Integrls In this section we revisit the definite integrl tht you were introduced to when

More information

Lecture 36. Finite Element Methods

Lecture 36. Finite Element Methods CE 60: Numercl Methods Lecture 36 Fnte Element Methods Course Coordntor: Dr. Suresh A. Krth, Assocte Professor, Deprtment of Cvl Engneerng, IIT Guwht. In the lst clss, we dscussed on the ppromte methods

More information

Bellman Optimality Equation for V*

Bellman Optimality Equation for V* Bellmn Optimlity Eqution for V* The vlue of stte under n optiml policy must equl the expected return for the best ction from tht stte: V (s) mx Q (s,) A(s) mx A(s) mx A(s) Er t 1 V (s t 1 ) s t s, t s

More information

LECTURE 5: LOGICS OF COOPERATION (II) 1.1 Examples of ATL Formulae. 1.2 Extended Example: Train Controller. 1 Alternating-time Temporal Logic

LECTURE 5: LOGICS OF COOPERATION (II) 1.1 Examples of ATL Formulae. 1.2 Extended Example: Train Controller. 1 Alternating-time Temporal Logic ctr out-of -gte reuest % " F F wll be true untl wll lwys be true 2 gwb nd tb hve jont strtegy for ensurng tht eventully there s pece gwb tb pece mp hs no strtegy for ensurng tht the udence s lwys excted

More information

Math 426: Probability Final Exam Practice

Math 426: Probability Final Exam Practice Mth 46: Probbility Finl Exm Prctice. Computtionl problems 4. Let T k (n) denote the number of prtitions of the set {,..., n} into k nonempty subsets, where k n. Argue tht T k (n) kt k (n ) + T k (n ) by

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Torsion, Thermal Effects and Indeterminacy

Torsion, Thermal Effects and Indeterminacy ENDS Note Set 7 F007bn orson, herml Effects nd Indetermncy Deformton n orsonlly Loded Members Ax-symmetrc cross sectons subjected to xl moment or torque wll remn plne nd undstorted. At secton, nternl torque

More information

SUMMER KNOWHOW STUDY AND LEARNING CENTRE

SUMMER KNOWHOW STUDY AND LEARNING CENTRE SUMMER KNOWHOW STUDY AND LEARNING CENTRE Indices & Logrithms 2 Contents Indices.2 Frctionl Indices.4 Logrithms 6 Exponentil equtions. Simplifying Surds 13 Opertions on Surds..16 Scientific Nottion..18

More information

1B40 Practical Skills

1B40 Practical Skills B40 Prcticl Skills Comining uncertinties from severl quntities error propgtion We usully encounter situtions where the result of n experiment is given in terms of two (or more) quntities. We then need

More information

CHAPTER - 7. Firefly Algorithm based Strategic Bidding to Maximize Profit of IPPs in Competitive Electricity Market

CHAPTER - 7. Firefly Algorithm based Strategic Bidding to Maximize Profit of IPPs in Competitive Electricity Market CHAPTER - 7 Frefly Algorthm sed Strtegc Bddng to Mxmze Proft of IPPs n Compettve Electrcty Mrket 7. Introducton The renovton of electrc power systems plys mjor role on economc nd relle operton of power

More information

Administrivia CSE 190: Reinforcement Learning: An Introduction

Administrivia CSE 190: Reinforcement Learning: An Introduction Administrivi CSE 190: Reinforcement Lerning: An Introduction Any emil sent to me bout the course should hve CSE 190 in the subject line! Chpter 4: Dynmic Progrmming Acknowledgment: A good number of these

More information

Online Learning Algorithms for Stochastic Water-Filling

Online Learning Algorithms for Stochastic Water-Filling Onlne Lernng Algorthms for Stochstc Wter-Fllng Y G nd Bhskr Krshnmchr Mng Hseh Deprtment of Electrcl Engneerng Unversty of Southern Clforn Los Angeles, CA 90089, USA Eml: {yg, bkrshn}@usc.edu Abstrct Wter-fllng

More information

Two Activation Function Wavelet Network for the Identification of Functions with High Nonlinearity

Two Activation Function Wavelet Network for the Identification of Functions with High Nonlinearity Interntonl Journl of Engneerng & Computer Scence IJECS-IJENS Vol:1 No:04 81 Two Actvton Functon Wvelet Network for the Identfcton of Functons wth Hgh Nonlnerty Wsm Khld Abdulkder Abstrct-- The ntegrton

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

A New Markov Chain Based Acceptance Sampling Policy via the Minimum Angle Method

A New Markov Chain Based Acceptance Sampling Policy via the Minimum Angle Method Irnn Journl of Opertons Reserch Vol. 3, No., 202, pp. 04- A New Mrkov Chn Bsed Acceptnce Smplng Polcy v the Mnmum Angle Method M. S. Fllh Nezhd * We develop n optmzton model bsed on Mrkovn pproch to determne

More information

Multiple view geometry

Multiple view geometry EECS 442 Computer vson Multple vew geometry Perspectve Structure from Moton - Perspectve structure from moton prolem - mgutes - lgerc methods - Fctorzton methods - Bundle djustment - Self-clrton Redng:

More information

Simultaneous estimation of rewards and dynamics from noisy expert demonstrations

Simultaneous estimation of rewards and dynamics from noisy expert demonstrations Smultneous estmton of rewrds nd dynmcs from nosy expert demonstrtons Mchel Hermn,2, Tobs Gndele, Jo rg Wgner, Felx Schmtt, nd Wolfrm Burgrd2 - Robert Bosch GmbH - 70442 Stuttgrt - Germny 2- Unversty of

More information

NUMERICAL MODELLING OF A CILIUM USING AN INTEGRAL EQUATION

NUMERICAL MODELLING OF A CILIUM USING AN INTEGRAL EQUATION NUEICAL ODELLING OF A CILIU USING AN INTEGAL EQUATION IHAI EBICAN, DANIEL IOAN Key words: Cl, Numercl nlyss, Electromgnetc feld, gnetton. The pper presents fst nd ccurte method to model the mgnetc behvour

More information

Chapter 5 Supplemental Text Material R S T. ij i j ij ijk

Chapter 5 Supplemental Text Material R S T. ij i j ij ijk Chpter 5 Supplementl Text Mterl 5-. Expected Men Squres n the Two-fctor Fctorl Consder the two-fctor fxed effects model y = µ + τ + β + ( τβ) + ε k R S T =,,, =,,, k =,,, n gven s Equton (5-) n the textook.

More information

Attribute reduction theory and approach to concept lattice

Attribute reduction theory and approach to concept lattice Scence n Chn Ser F Informton Scences 2005 Vol48 No6 713 726 713 Attrbute reducton theory nd pproch to concept lttce ZHANG Wenxu 1, WEI Lng 1,2 & QI Jnun 3 1 Insttute for Informton nd System Scences, Fculty

More information

UNIFORM CONVERGENCE. Contents 1. Uniform Convergence 1 2. Properties of uniform convergence 3

UNIFORM CONVERGENCE. Contents 1. Uniform Convergence 1 2. Properties of uniform convergence 3 UNIFORM CONVERGENCE Contents 1. Uniform Convergence 1 2. Properties of uniform convergence 3 Suppose f n : Ω R or f n : Ω C is sequence of rel or complex functions, nd f n f s n in some sense. Furthermore,

More information

Jean Fernand Nguema LAMETA UFR Sciences Economiques Montpellier. Abstract

Jean Fernand Nguema LAMETA UFR Sciences Economiques Montpellier. Abstract Stochstc domnnce on optml portfolo wth one rsk less nd two rsky ssets Jen Fernnd Nguem LAMETA UFR Scences Economques Montpeller Abstrct The pper provdes restrctons on the nvestor's utlty functon whch re

More information

4.4 Areas, Integrals and Antiderivatives

4.4 Areas, Integrals and Antiderivatives . res, integrls nd ntiderivtives 333. Ares, Integrls nd Antiderivtives This section explores properties of functions defined s res nd exmines some connections mong res, integrls nd ntiderivtives. In order

More information

A Reinforcement Learning System with Chaotic Neural Networks-Based Adaptive Hierarchical Memory Structure for Autonomous Robots

A Reinforcement Learning System with Chaotic Neural Networks-Based Adaptive Hierarchical Memory Structure for Autonomous Robots Interntonl Conference on Control, Automton nd ystems 008 Oct. 4-7, 008 n COEX, eoul, Kore A Renforcement ernng ystem wth Chotc Neurl Networs-Bsed Adptve Herrchcl Memory tructure for Autonomous Robots Msno

More information

Math 8 Winter 2015 Applications of Integration

Math 8 Winter 2015 Applications of Integration Mth 8 Winter 205 Applictions of Integrtion Here re few importnt pplictions of integrtion. The pplictions you my see on n exm in this course include only the Net Chnge Theorem (which is relly just the Fundmentl

More information

Reinforcement learning II

Reinforcement learning II CS 1675 Introduction to Mchine Lerning Lecture 26 Reinforcement lerning II Milos Huskrecht milos@cs.pitt.edu 5329 Sennott Squre Reinforcement lerning Bsics: Input x Lerner Output Reinforcement r Critic

More information

Group-based active query selection. for rapid diagnosis in time-critical situations

Group-based active query selection. for rapid diagnosis in time-critical situations Group-bsed ctve query selecton for rpd dgnoss n tme-crtcl stutons *Gowthm Belll, Student Member, IEEE, Suresh K. Bhvnn, nd Clyton Scott, Member, IEEE Abstrct In pplctons such s ctve lernng or dsese/fult

More information

Vol. 5, No. 5 May 2014 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved.

Vol. 5, No. 5 May 2014 ISSN Journal of Emerging Trends in Computing and Information Sciences CIS Journal. All rights reserved. Vol. 5, No. 5 y 04 ISSN 079-8407 Jornl of Emergng Trends n Comptng nd Informton Scences 009-04 CIS Jornl. ll rghts reserved. http://www.csornl.org Notes on lt Soft trces D.Sngh, Onyeozl, I.., 3 lkl..j.,

More information

Online Stochastic Matching: New Algorithms with Better Bounds

Online Stochastic Matching: New Algorithms with Better Bounds Onlne Stochstc Mtchng: New Algorthms wth Better Bounds Ptrck Jllet Xn Lu My 202; revsed Jnury 203, June 203 Abstrct We consder vrnts of the onlne stochstc bprtte mtchng problem motvted by Internet dvertsng

More information

1 Probability Density Functions

1 Probability Density Functions Lis Yn CS 9 Continuous Distributions Lecture Notes #9 July 6, 28 Bsed on chpter by Chris Piech So fr, ll rndom vribles we hve seen hve been discrete. In ll the cses we hve seen in CS 9, this ment tht our

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

CISE 301: Numerical Methods Lecture 5, Topic 4 Least Squares, Curve Fitting

CISE 301: Numerical Methods Lecture 5, Topic 4 Least Squares, Curve Fitting CISE 3: umercl Methods Lecture 5 Topc 4 Lest Squres Curve Fttng Dr. Amr Khouh Term Red Chpter 7 of the tetoo c Khouh CISE3_Topc4_Lest Squre Motvton Gven set of epermentl dt 3 5. 5.9 6.3 The reltonshp etween

More information