Automated Recommendation Systems
|
|
- Hugo West
- 6 years ago
- Views:
Transcription
1 Automted Recommendtion Systems Collbortive Filtering Through Reinorcement Lerning Most Akhmizdeh Deprtment o MS&E, Stnord University Emil: mkhmi@stnord.edu Alexei Avkov Deprtment o Electricl Engineering, Stnord University Emil: linrv@stnord.edu Rez Tkpoui Deprtment o Electricl Engineering, Stnord University Emil: tkpoui@stnord.edu Abstrct Within this work we explore the topic o lrge scle, utomted recommendtion systems. We ocus on collbortive iltering pproches, wherein system suggests new products to users bsed on their viewing history s well s other known demogrphics. There re severl pproches to this in current literture, the simplest o which tret it s mtrix completion problem. We explore the setting rom reinorcement lerning perspective by pplying trditionl lgorithms or reinorcement lerning to the problem. I. PROBLEM FORMULATION Numerous online services such s Netlix, Amzon, Yelp, Pndor, online dvertisings, etc. provide utomted recommendtions to help users to nvigte through lrge collection o items. Every time user queries the system or new item, suggestion is mde on the bsis o the user s pst history nd when vilble their demogrphic proile. Two typicl wys o producing these recommendtions re collbortive iltering nd content-bsed iltering. There re two simultneous gols to be stisied: helping the user to explore the vilble items nd probing the user s preerences. One o the models tht cptures this setting well is the multirm bndit, n importnt model or decision mking under uncertinty. In this model set o rms with unknown rewrd proiles is given nd, t ech time slot, the decision mker must choose n rm to mximize his expected rewrd. Clerly, the decision t ech time slot should depend on previous observtions. Thus, there is trde-o between explortion, trying rms with more uncertin rewrd in order to gther more inormtion, nd exploittion, pulling rms with reltively high rewrd expecttions. For our purposes the rms hve very speciic structure nd this setting hs previously been reerred to s the linerbndits model see [3]. Here, it is ssumed tht the underlying mtrix o preerences which contins the rting user i gives to item j t entry i, j hs low-rnk structure. Hence, rtings mde by user i to item j cn be pproximted by scler product o two eture vectors i, b j R p, chrcterizing user nd item respectively. In other words our observtions, r ij cn be viewed s r ij T i b j + z ij where z ij represents the unexplined ctors. In the generl setting, both the user nd item eture vectors re treted s unknown, nd our recommendtion lgorithms must estimte them over time. However, some works like [1] mke simpliying ssumption tht the item eture vectors re known. We explore both settings, but ind more meningul results in the cse where the item eture vectors re known. In this cse, the item eture vectors cn be either constructed explicitly, or derived rom users eedbck using mtrix ctoriztion methods. With the item ltent vectors in hnd, we cn tret ech user independently nd throughout the explore-exploit trde-o, we cn try to estimte nd exploit the users ltent vectors. These eture vectors cn depend on users demogrphic inormtion nd their pst behvior in rting items. The gol o our system is to develop recommendtion policy, which suggests items to users. This policy will, t ech time slot, output recommendtion bsed on the previous observtions. This policy must properly djust or the exploreexploit trde-o, nd clssiclly there re two types o policies, which dier in the wy they perorm explortion: optimistic policies, e.g. upper conidence bound UCB, nd probbilistic policies, e.g. posterior smpling. UCB lgorithms hve been pplied to this problem in the pst, but posterior smpling is less common. Posterior smpling lso known s Thompson smpling ws introduced in 1933 nd oers signiicnt dvntges over UCB methods s shown in [], however until recently it hs not been very populr or esible. Our primry objective is to explore the esibility o collbortive iltering through posterior smpling. We nlyze its perormnce on rel world dt, speciiclly the reely vilble MovieLens dtsets, nd compre it to existing methods such s UCB nd the work done in [1] II. SYSTEM MODELS AND ALGORITHMS In this section we will introduce some nottion used throughout the rest o this work, s well s the lgorithms tht we seek to implement. A. Nottion We hve set o users, i 1,,..., m, with corresponding eture vectors i R p ; nd items, j 1,,..., n, with corresponding eture vectors b j R p. We reer to these eture vectors collectively s A R p m nd B R p n, thus the true rtings cn be cptured in the mtrix A T B. At ech time t Z +, user i t will enter the system nd
2 be recommended n item j t, ter which they will give it rting r t ccording to r t T i t b j t + z t Where z t cptures the unexplinble devition o the observtion rom our model. We reer to the viewing history t time t s the sequence H t { i τ, j τ, r τ} t, i.e. ll the viewings in the system beore time t. Thus on high-level, t time t our progrm seeks to use its knowledge o user i t to mke the best possible recommendtion. The job o recommendtion system is to deine unction µ H, which given user will output recommendtion or tht user. Unknown to the system, there is some optiml policy which t ech time t would output recommendtion j t. To mesure the perormnce o our system, we will compre the system s recommendtions to the best recommendtion. Speciiclly deine the regret o the system, t time t, to be Rt t T i b τ τ j E [r τ] Tht is, t ech time-step we increse our regret by how r the expected rting o our recommendtion diers rom the best possible rting. Ultimtely we seek to derive policy which chieves miniml regret. B. Posterior Smpling Algorithm 1 Posterior Smpling Strt with prior distribution on A, B, A, B or t 1,,... do observe rrivl o user i t smple Â, ˆB A, B H t compute nd output recommendtion j t where j t rg mx E j. observe the user s rting r t end or [â T i ˆb j ] The ide behind posterior smpling lgorithm is to orce optimism through probbilistic ction. Speciiclly t ech time step, t, we will mke recommendtion j t bsed on the probbility tht it is the best possible recommendtion, P j t j t. However, this probbility is inccessible, so insted the lgorithm smples model or the unknown eture vectors bsed on the probbility tht they re the true eture vectors given the viewing history, nd inds the optiml recommendtion should this be the true model. It cn be shown tht this smpling technique is equivlent to smpling recommendtion bsed on the probbility it is optiml, nd more detiled description o the lgorithm nd its motivtions cn be seen in []. Thus the lgorithm proceeds to keep trck o the distribution o model prmeters t ech time step, nd updtes them ccordingly. To implement this lgorithm ll tht remins is to choose prior on the model prmeters, nd compute their posterior distribution given viewing history. As in [3], nd other prior literture, we ssume i, b j N, I p /p i.i.d.. Furthermore we ssume tht unexplined devitions o the observtions re Gussin, i.e. z t N, σ z. Now we re redy to compute the posterior distribution. Using Byes rule observe or compctness we use to denote the distribution o the rgument: A, B H t H t A, B A B H t AB t z τ A,B AB t z τ dadb In the bove z τ r τ T b i τ j τ, nd the integrl o the denomintor is over the entire spce o R p m R p n. For the rest o this report, we consider the simpler cse where the vectors b j re given nd we tret ech user independently. This problem is extensively studied in literture, but s r s we cn tell hs never been solved or nlyzed through posterior smpling. We explore it more concretely below. For compctness we will consider only the eture vector o single user, R p, priori we ssume it comes rom N, I p /p s bove, nd we now consider the viewing history H t to be the history o the ctive user s opposed to ll users. We cn now compute the posterior distribution s ollows: H H t H H t H H t t z z τ R p t z z τ d r τ T b j τ t z R p t z r τ T b j τ d But observe in this simple cse computing the posterior is much simpler. The numertor is clerly Gussin, nd the denomintor is just normlizing term, thus we determine H t N µ t, Σ t. We cn ormulte recursive updte rule or the prmeters µ t, Σ t by mssging the numertor into n pproprite orm this is done in the ppendix. We ind the ollowing updte equtions or the posterior: Σ t Σ t µ t Σ t + bt b tt σ z Σ t µ t + rt b t These recursive updte equtions re convenient or implementtion, nd cn be used eiciently by storing Σ, however some intuition s to their opertion cn be seen by pplying
3 the mtrix inversion lemm. Through it, we ind: Σ t Σ t µ t µ t Σt b j tb T Σ t j t σ z + b T j t Σ t + rt Σ t Σ t + b T Σ t j t b j t b j t H t b j tµ t T b j t Thus, essentilly, t ech step the posterior men shits towrds or wy the eture vector o the recommended item. Similrly the covrince Σ thins out to select or single direction. The rest o our work revolves mostly round nlyzing the simpliied problem setting, however this simpliiction is extremely useul or the generl cse s well. Observe, A, B H t A B, H t B H t Thus we cn perorm posterior smpling in the generl cse by irst smpling item etures ˆB, ccording to B H t, nd then smpling A rom gussin distribution with men nd vrince determined by the previously derived updte equtions given the selected etures ˆB. Unortuntely the distribution o B H t is quite complicted; ter vectorizing the mtrix B into vector B R np we ind: B H t B / p exp k B T B + c T B In the bove p is polynomil unction in the entries o B, k is some sclr, nd c is vector in R np. Unortuntely, even in this orm, it is still uncler how to smple rom this distribution. C. A UCB Approch Algorithm UCB Strt with prior distribution on A, B, A, B, nd n optimism prmeter p, 1 or t 1,,... do Observe rrivl o user i t Compute the distribution on the rewrd o ech item For ll items, compute U j, the p-th percentile o the rewrd o item j Compute nd output recommendtion j t where j t rg mx j Observe the user s rting r t end or UCB is completely dierent pproch rom posterior smpling. At ech timestep the lgorithm computes n upper conidence bound on the rewrd o ech o the items. The lgorithm will then suggest the lgorithm with the highest UCB. For our purposes, we will use speciic percentile o the rewrd s the UCB o ech item. This is generlly hrd to do nd other literture uses vrious heuristics to determine U j UCB. In the generl problem setting, it is uncler how to implement UCB in ny meningul wy, however it is rther elegnt in the simpliied cse o given item eture vectors. In the simpliied cse, using the priors described in the previous section we observe tht the posterior o given is Gussin, thus the distribution on the rewrd o recommending item j is lso Gussin. We compute the men nd vrince s ollows: σ j b T j Σ b j + σ z µ j b T j µ Thus computing the p-th percentile o the rewrd cn be done, simply by inverting the cd o the norml distribution. D. Mixed Approches From evlution we observe tht U CB nd posterior smpling ech hve unique dvntges. Thus we propose vrious schemes tht llow you to chieve the vrious perormnce trde-os o both. First we propose n ɛ-greedy pproch nd second we propose two-phse pproch. These were both studied in the simpliied cse, but could potentilly be pplied to the generl setting s well. The ɛ-ucb lgorithm will lip weighted coin t ech timestep to decide wether to obtin recommendtion through posterior smpling or through UCB. Speciiclly the lgorithm will elect to perorm UCB ɛ percent o the time. The two phsed pproch will begin by lerning through posterior smpling until some time T, ter which it proceeds to output recommendtions through the UCB pproch. In the next section, we will thoroughly study the perormnce o ll o the lgorithms presented in this section. E. The Cse o No-repet Recommendtions Throughout this work we ssumed tht it is relevnt to recommend the sme item severl times. However, in some settings this is not very nturl. For instnce, i the system provides recommendtions or viewing movies, ll o the bove lgorithms would eventully chose to show the sme movie over nd over. Clerly this is not very useul, nd this cn be resolved in severl wys. We could lower the rewrd o successive viewings, but this dds complicted time dependence to our model. More simply we cn prohibit the lgorithm rom suggesting the sme item multiple times. In the cse o suggesting movies this is nturl since users would rrely view the sme production multiple times. III. IMPLEMENTATION AND EVALUATION In this section, we present our implementtion results or the orementioned lgorithms. For the purpose o numericl simultions, we used MATLAB. We hve crried out lgorithms both on synthetic dt nd reely vilble MovieLens dtset. For the purpose o synthetic dt, we generte rndom mtrix the sme size s MovieLens dt with rnk 3 by generting rndom Gussin eture mtrices nd multiplying them together. Ech o the entries o eture
4 mtrices comes rom N, 1/p. Then, we will tke item eture vectors s grnted nd try to estimte user eture mtrix by considering Gussin prior. Figure 1 shows the cumultive regret versus time, or posterior smpling nd UCB with our dierent prmeters. We cn see tht posterior smpling might work worse t irst by exploring too much, but it pys o lter when the better understnding o the rms comes to help lter. Cumultive regret Posterior Smpling UCB with. percentile UCB with.9 percentile UCB with.99 percentile UCB with.999 percentile Fig. 1. Cumultive regret o posterior smpling nd UCB lgorithms on synthetic dt Notice tht the regret observed by ech o these lgorithms is very good compred to the totl rewrd they obtin the cumultive rewrds t t8 is in the order o while the dierence in rewrds is in the order o 1. In order to show this, we hve plotted the cumultive rewrd versus time or ll o these lgorithms in Figure nd it cn be seen tht it is close to the optiml rewrd. Cumultive rewrd Posterior Smpling UCB with. percentile UCB with.9 percentile UCB with.99 percentile UCB with.999 percentile Optiml rewrd Fig.. Cumultive rewrd o posterior smpling nd UCB lgorithms on synthetic dt We lso tried similr simultions or dierent rnks o the underlying mtrix. Figure 3 shows the perormnce o these lgorithms when the rnk o the preerence mtrix is 1. It cn be seen tht posterior smpling outperorms UCB. One interesting observtion is tht unlike posterior smpling, UCB methods re very sensitive to the prmeters used in lgorithms in our cse the percentile prmeter, nd using inpproprite prmeter my result in non-zero symptotic regret. We observe tht the optiml tuning is highly sensitive to the dt, speciiclly to its rnk. As vrition o the introduced lgorithms, we hve crried out combintion o posterior smpling nd greedy Cumultive regret Posterior Smpling UCB with. percentile UCB with.9 percentile UCB with.99 percentile UCB with.999 percentile Fig. 3. Cumultive regret o posterior smpling nd UCB lgorithms or synthetic dt with rnk 1 lgorithms. Here, the greedy lgorithm chooses the rm tht mximized the expected instntneous rewrd nd cn be considered s UCB with percentile %. At ech time, the ɛ-greedy lgorithm mkes greedy decision with probbility ɛ, nd perorms n itertion o posterior smpling with probbility 1 ɛ. By looking t Figure 4, we cn see tht the perormnce o the greedy lgorithm improves drmticlly when it s combined with posterior smpling % o the time. This will result in even more computtionlly eicient methods while the regret still remins cceptbly low. Cumultive regret Posterior Smpling eps. eps.8 Greedy Fig. 4. Cumultive regret or hybrid pproch on synthetic dt We lso crried out the posterior smpling nd UCB lgorithms under the ssumption tht no item cn be recommended to user more thn once. Notice tht in this cse, we expect the regret to be decresing t some point, becuse the expected regret t time 168 is equl to zero. Figure shows the perormnce o these lgorithms in this cse. We hve implemented ll these methods or the MovieLens dtset nd got similr results. For exmple Figure 6 shows the cumultive regret or the posterior smpling nd UCB with dierent prmeters on MovieLens dtset. As seen in Figure 6, UCB lgorithm with prmeter.9 works better thn the other instnces o UCB, which urther shows tht there is no rule or inding the best prmeters or UCB lgorithms. IV. CONCLUSIONS All o the lgorithms we nlyzed perorm extremely well. The dierence in regret between them is negligible compred
5 Cumultive Regret Posterior Smpling UCB. UCB.9 UCB Fig.. Cumultive regret or the cse o no repetitions, on synthetic dt Cumultive regret Posterior Smpling 1 UCB with. percentile UCB with.9 percentile UCB with.99 percentile UCB with.999 percentile Fig. 6. Cumultive regret o posterior smpling nd UCB on MovieLens dtset to the totl rewrd collected. Thus we dvocte posterior smpling s the best generl purpose solution or severl resons. First, it is extremely eicient compre to the UCB style pproch. Second, it does not require ny tuning; while we observed tht UCB cn outperorm posterior smpling it is extremely relint on proper tuning which cn be hrd to determine in prctice. Furthermore, posterior smpling cn clerly be extended to the generl problem setting, wheres our stted UCB method is not. Lstly we note tht using previously mentioned hybrid pproches it is possible to chieve mny dierent eiciency/regret trde-os. V. FUTURE WORK It would be interesting to more closely nlyze the generl cse. Posterior smpling cn be implemented through Gibb s smpling, or the Metropolis-Hstings lgorithm. UCB s described in this pper would be much more diicult to implement, but we could try vrious heuristics nd other UCB style lgorithms. Alterntively this work could be continued in the prcticl direction by building rel-lie recommendtion system utilizing these lgorithms nd studying its perormnce. APPENDIX A. Derivtion o Posterior Updte Rules Agin consider the simpliied cse where we know the ltent eture vectors o the items. For compctness we will consider only the eture vector o single user, R p. Then: H t t z r τ T b j τ R p t z r τ T b j τ d t C 1 z r τ T b j τ C z r t T b j t H t r t T b j t C 3 exp σ z exp 1 µt T Σ t µ t Now it is cler tht the distribution remins Gussin. At this point simply compute the coeicients o the qudrtic nd liner terms to solve or the new men nd covrince. This yields Σ t Σ t µ t Σ t + bt b tt σ z Σ t µ t B. Woodury Mtrix Identity & Updte rules Recll the Woodury Mtrix Identity: + rt b t A + UCV A A U C + V A U V A For ese o nottion in this section we will reer to Σ t s Σ t, µ t s µ t, r t s r, nd lstly we reer to b t simply s b. Apply the lemm to the previously derived updte rules: Σ t Σ t + bbt σ z Σ t Σ tbb T Σ t σ z + b T Σ t b Now we cn plug this into the derivtion o µ t : µ t Σ t Σ t µ t + r b Σ t Σ tbb T Σ t + b T Σ t µ t + r Σ t b b µ t Σ tbµ T tb + b T Σ t b + r σ z Σ t b + b T Σ t bσ t b Σ t bb T Σ t b + b T Σ t b rσt Σ t bµ T t µ t + + b T b Σ t b REFERENCES [1] Ysh Deshpnde, Andre Montnri, Liner Bndits in High Dimension nd Recommendtion Systems. Avilble online: [] Dniel Russo, Benjmin Vn Roy, Lerning to Optimize Vi Posterior Smpling. Avilble online: [3] Pt Rusmevichientong, John N. Tsitsiklis, Linerly Prmeterized Bndits. Avilble online:
Multi-Armed Bandits: Non-adaptive and Adaptive Sampling
CSE 547/Stt 548: Mchine Lerning for Big Dt Lecture Multi-Armed Bndits: Non-dptive nd Adptive Smpling Instructor: Shm Kkde 1 The (stochstic) multi-rmed bndit problem The bsic prdigm is s follows: K Independent
More information1 Online Learning and Regret Minimization
2.997 Decision-Mking in Lrge-Scle Systems My 10 MIT, Spring 2004 Hndout #29 Lecture Note 24 1 Online Lerning nd Regret Minimiztion In this lecture, we consider the problem of sequentil decision mking in
More informationENGI 3424 Engineering Mathematics Five Tutorial Examples of Partial Fractions
ENGI 44 Engineering Mthemtics Five Tutoril Exmples o Prtil Frctions 1. Express x in prtil rctions: x 4 x 4 x 4 b x x x x Both denomintors re liner non-repeted ctors. The cover-up rule my be used: 4 4 4
More informationReinforcement Learning
Reinforcement Lerning Tom Mitchell, Mchine Lerning, chpter 13 Outline Introduction Comprison with inductive lerning Mrkov Decision Processes: the model Optiml policy: The tsk Q Lerning: Q function Algorithm
More informationReinforcement learning II
CS 1675 Introduction to Mchine Lerning Lecture 26 Reinforcement lerning II Milos Huskrecht milos@cs.pitt.edu 5329 Sennott Squre Reinforcement lerning Bsics: Input x Lerner Output Reinforcement r Critic
More informationReview of Calculus, cont d
Jim Lmbers MAT 460 Fll Semester 2009-10 Lecture 3 Notes These notes correspond to Section 1.1 in the text. Review of Clculus, cont d Riemnn Sums nd the Definite Integrl There re mny cses in which some
More informationNumerical integration
2 Numericl integrtion This is pge i Printer: Opque this 2. Introduction Numericl integrtion is problem tht is prt of mny problems in the economics nd econometrics literture. The orgniztion of this chpter
More information19 Optimal behavior: Game theory
Intro. to Artificil Intelligence: Dle Schuurmns, Relu Ptrscu 1 19 Optiml behvior: Gme theory Adversril stte dynmics hve to ccount for worst cse Compute policy π : S A tht mximizes minimum rewrd Let S (,
More informationp-adic Egyptian Fractions
p-adic Egyptin Frctions Contents 1 Introduction 1 2 Trditionl Egyptin Frctions nd Greedy Algorithm 2 3 Set-up 3 4 p-greedy Algorithm 5 5 p-egyptin Trditionl 10 6 Conclusion 1 Introduction An Egyptin frction
More informationBest Approximation. Chapter The General Case
Chpter 4 Best Approximtion 4.1 The Generl Cse In the previous chpter, we hve seen how n interpolting polynomil cn be used s n pproximtion to given function. We now wnt to find the best pproximtion to given
More informationCS 188 Introduction to Artificial Intelligence Fall 2018 Note 7
CS 188 Introduction to Artificil Intelligence Fll 2018 Note 7 These lecture notes re hevily bsed on notes originlly written by Nikhil Shrm. Decision Networks In the third note, we lerned bout gme trees
More informationSolution for Assignment 1 : Intro to Probability and Statistics, PAC learning
Solution for Assignment 1 : Intro to Probbility nd Sttistics, PAC lerning 10-701/15-781: Mchine Lerning (Fll 004) Due: Sept. 30th 004, Thursdy, Strt of clss Question 1. Bsic Probbility ( 18 pts) 1.1 (
More informationThe Regulated and Riemann Integrals
Chpter 1 The Regulted nd Riemnn Integrls 1.1 Introduction We will consider severl different pproches to defining the definite integrl f(x) dx of function f(x). These definitions will ll ssign the sme vlue
More informationCMDA 4604: Intermediate Topics in Mathematical Modeling Lecture 19: Interpolation and Quadrature
CMDA 4604: Intermedite Topics in Mthemticl Modeling Lecture 19: Interpoltion nd Qudrture In this lecture we mke brief diversion into the res of interpoltion nd qudrture. Given function f C[, b], we sy
More informationLecture 14: Quadrature
Lecture 14: Qudrture This lecture is concerned with the evlution of integrls fx)dx 1) over finite intervl [, b] The integrnd fx) is ssumed to be rel-vlues nd smooth The pproximtion of n integrl by numericl
More informationMonte Carlo method in solving numerical integration and differential equation
Monte Crlo method in solving numericl integrtion nd differentil eqution Ye Jin Chemistry Deprtment Duke University yj66@duke.edu Abstrct: Monte Crlo method is commonly used in rel physics problem. The
More informationGeneralized Fano and non-fano networks
Generlized Fno nd non-fno networks Nildri Ds nd Brijesh Kumr Ri Deprtment of Electronics nd Electricl Engineering Indin Institute of Technology Guwhti, Guwhti, Assm, Indi Emil: {d.nildri, bkri}@iitg.ernet.in
More informationCHAPTER 4a. ROOTS OF EQUATIONS
CHAPTER 4. ROOTS OF EQUATIONS A. J. Clrk School o Engineering Deprtment o Civil nd Environmentl Engineering by Dr. Ibrhim A. Asskk Spring 00 ENCE 03 - Computtion Methods in Civil Engineering II Deprtment
More informationPart I: Basic Concepts of Thermodynamics
Prt I: Bsic Concepts o Thermodynmics Lecture 4: Kinetic Theory o Gses Kinetic Theory or rel gses 4-1 Kinetic Theory or rel gses Recll tht or rel gses: (i The volume occupied by the molecules under ordinry
More informationHere we study square linear systems and properties of their coefficient matrices as they relate to the solution set of the linear system.
Section 24 Nonsingulr Liner Systems Here we study squre liner systems nd properties of their coefficient mtrices s they relte to the solution set of the liner system Let A be n n Then we know from previous
More informationODE: Existence and Uniqueness of a Solution
Mth 22 Fll 213 Jerry Kzdn ODE: Existence nd Uniqueness of Solution The Fundmentl Theorem of Clculus tells us how to solve the ordinry differentil eqution (ODE) du = f(t) dt with initil condition u() =
More information2D1431 Machine Learning Lab 3: Reinforcement Learning
2D1431 Mchine Lerning Lb 3: Reinforcement Lerning Frnk Hoffmnn modified by Örjn Ekeberg December 7, 2004 1 Introduction In this lb you will lern bout dynmic progrmming nd reinforcement lerning. It is ssumed
More informationContinuous Random Variables
STAT/MATH 395 A - PROBABILITY II UW Winter Qurter 217 Néhémy Lim Continuous Rndom Vribles Nottion. The indictor function of set S is rel-vlued function defined by : { 1 if x S 1 S (x) if x S Suppose tht
More informationOn Second Derivative-Free Zero Finding Methods
010 Americn Control Conerence Mrriott Wterront, Bltimore, MD, USA June 30-July 0, 010 FrC07.4 On Second Derivtive-Free Zero Finding Methods Mohmmed A. Hsn Deprtment o Electricl & Computer Engineering University
More informationLECTURE NOTE #12 PROF. ALAN YUILLE
LECTURE NOTE #12 PROF. ALAN YUILLE 1. Clustering, K-mens, nd EM Tsk: set of unlbeled dt D = {x 1,..., x n } Decompose into clsses w 1,..., w M where M is unknown. Lern clss models p(x w)) Discovery of
More informationMath Lecture 23
Mth 8 - Lecture 3 Dyln Zwick Fll 3 In our lst lecture we delt with solutions to the system: x = Ax where A is n n n mtrix with n distinct eigenvlues. As promised, tody we will del with the question of
More informationAdministrivia CSE 190: Reinforcement Learning: An Introduction
Administrivi CSE 190: Reinforcement Lerning: An Introduction Any emil sent to me bout the course should hve CSE 190 in the subject line! Chpter 4: Dynmic Progrmming Acknowledgment: A good number of these
More informationSection 11.5 Estimation of difference of two proportions
ection.5 Estimtion of difference of two proportions As seen in estimtion of difference of two mens for nonnorml popultion bsed on lrge smple sizes, one cn use CLT in the pproximtion of the distribution
More information1 nonlinear.mcd Find solution root to nonlinear algebraic equation f(x)=0. Instructor: Nam Sun Wang
nonlinermc Fin solution root to nonliner lgebric eqution ()= Instructor: Nm Sun Wng Bckgroun In science n engineering, we oten encounter lgebric equtions where we wnt to in root(s) tht stisies given eqution
More information5.7 Improper Integrals
458 pplictions of definite integrls 5.7 Improper Integrls In Section 5.4, we computed the work required to lift pylod of mss m from the surfce of moon of mss nd rdius R to height H bove the surfce of the
More informationNew data structures to reduce data size and search time
New dt structures to reduce dt size nd serch time Tsuneo Kuwbr Deprtment of Informtion Sciences, Fculty of Science, Kngw University, Hirtsuk-shi, Jpn FIT2018 1D-1, No2, pp1-4 Copyright (c)2018 by The Institute
More informationChapter 0. What is the Lebesgue integral about?
Chpter 0. Wht is the Lebesgue integrl bout? The pln is to hve tutoril sheet ech week, most often on Fridy, (to be done during the clss) where you will try to get used to the ides introduced in the previous
More informationData Assimilation. Alan O Neill Data Assimilation Research Centre University of Reading
Dt Assimiltion Aln O Neill Dt Assimiltion Reserch Centre University of Reding Contents Motivtion Univrite sclr dt ssimiltion Multivrite vector dt ssimiltion Optiml Interpoltion BLUE 3d-Vritionl Method
More informationMath 426: Probability Final Exam Practice
Mth 46: Probbility Finl Exm Prctice. Computtionl problems 4. Let T k (n) denote the number of prtitions of the set {,..., n} into k nonempty subsets, where k n. Argue tht T k (n) kt k (n ) + T k (n ) by
More informationUNIT 1 FUNCTIONS AND THEIR INVERSES Lesson 1.4: Logarithmic Functions as Inverses Instruction
Lesson : Logrithmic Functions s Inverses Prerequisite Skills This lesson requires the use of the following skills: determining the dependent nd independent vribles in n exponentil function bsed on dt from
More informationPhysics 116C Solution of inhomogeneous ordinary differential equations using Green s functions
Physics 6C Solution of inhomogeneous ordinry differentil equtions using Green s functions Peter Young November 5, 29 Homogeneous Equtions We hve studied, especilly in long HW problem, second order liner
More informationTests for the Ratio of Two Poisson Rates
Chpter 437 Tests for the Rtio of Two Poisson Rtes Introduction The Poisson probbility lw gives the probbility distribution of the number of events occurring in specified intervl of time or spce. The Poisson
More informationTHE EXISTENCE-UNIQUENESS THEOREM FOR FIRST-ORDER DIFFERENTIAL EQUATIONS.
THE EXISTENCE-UNIQUENESS THEOREM FOR FIRST-ORDER DIFFERENTIAL EQUATIONS RADON ROSBOROUGH https://intuitiveexplntionscom/picrd-lindelof-theorem/ This document is proof of the existence-uniqueness theorem
More informationMath 1B, lecture 4: Error bounds for numerical methods
Mth B, lecture 4: Error bounds for numericl methods Nthn Pflueger 4 September 0 Introduction The five numericl methods descried in the previous lecture ll operte by the sme principle: they pproximte the
More informationHow do we solve these things, especially when they get complicated? How do we know when a system has a solution, and when is it unique?
XII. LINEAR ALGEBRA: SOLVING SYSTEMS OF EQUATIONS Tody we re going to tlk bout solving systems of liner equtions. These re problems tht give couple of equtions with couple of unknowns, like: 6 2 3 7 4
More informationLecture 19: Continuous Least Squares Approximation
Lecture 19: Continuous Lest Squres Approximtion 33 Continuous lest squres pproximtion We begn 31 with the problem of pproximting some f C[, b] with polynomil p P n t the discrete points x, x 1,, x m for
More informationThe graphs of Rational Functions
Lecture 4 5A: The its of Rtionl Functions s x nd s x + The grphs of Rtionl Functions The grphs of rtionl functions hve severl differences compred to power functions. One of the differences is the behvior
More informationAcceptance Sampling by Attributes
Introduction Acceptnce Smpling by Attributes Acceptnce smpling is concerned with inspection nd decision mking regrding products. Three spects of smpling re importnt: o Involves rndom smpling of n entire
More informationWe partition C into n small arcs by forming a partition of [a, b] by picking s i as follows: a = s 0 < s 1 < < s n = b.
Mth 255 - Vector lculus II Notes 4.2 Pth nd Line Integrls We begin with discussion of pth integrls (the book clls them sclr line integrls). We will do this for function of two vribles, but these ides cn
More informationAdvanced Calculus: MATH 410 Notes on Integrals and Integrability Professor David Levermore 17 October 2004
Advnced Clculus: MATH 410 Notes on Integrls nd Integrbility Professor Dvid Levermore 17 October 2004 1. Definite Integrls In this section we revisit the definite integrl tht you were introduced to when
More informationMath& 152 Section Integration by Parts
Mth& 5 Section 7. - Integrtion by Prts Integrtion by prts is rule tht trnsforms the integrl of the product of two functions into other (idelly simpler) integrls. Recll from Clculus I tht given two differentible
More informationChapter 5 : Continuous Random Variables
STAT/MATH 395 A - PROBABILITY II UW Winter Qurter 216 Néhémy Lim Chpter 5 : Continuous Rndom Vribles Nottions. N {, 1, 2,...}, set of nturl numbers (i.e. ll nonnegtive integers); N {1, 2,...}, set of ll
More informationModule 6 Value Iteration. CS 886 Sequential Decision Making and Reinforcement Learning University of Waterloo
Module 6 Vlue Itertion CS 886 Sequentil Decision Mking nd Reinforcement Lerning University of Wterloo Mrkov Decision Process Definition Set of sttes: S Set of ctions (i.e., decisions): A Trnsition model:
More informationDiscrete Least-squares Approximations
Discrete Lest-squres Approximtions Given set of dt points (x, y ), (x, y ),, (x m, y m ), norml nd useful prctice in mny pplictions in sttistics, engineering nd other pplied sciences is to construct curve
More information8 Laplace s Method and Local Limit Theorems
8 Lplce s Method nd Locl Limit Theorems 8. Fourier Anlysis in Higher DImensions Most of the theorems of Fourier nlysis tht we hve proved hve nturl generliztions to higher dimensions, nd these cn be proved
More informationRiemann is the Mann! (But Lebesgue may besgue to differ.)
Riemnn is the Mnn! (But Lebesgue my besgue to differ.) Leo Livshits My 2, 2008 1 For finite intervls in R We hve seen in clss tht every continuous function f : [, b] R hs the property tht for every ɛ >
More informationTheoretical foundations of Gaussian quadrature
Theoreticl foundtions of Gussin qudrture 1 Inner product vector spce Definition 1. A vector spce (or liner spce) is set V = {u, v, w,...} in which the following two opertions re defined: (A) Addition of
More information7.2 The Definite Integral
7.2 The Definite Integrl the definite integrl In the previous section, it ws found tht if function f is continuous nd nonnegtive, then the re under the grph of f on [, b] is given by F (b) F (), where
More informationBayesian Networks: Approximate Inference
pproches to inference yesin Networks: pproximte Inference xct inference Vrillimintion Join tree lgorithm pproximte inference Simplify the structure of the network to mkxct inferencfficient (vritionl methods,
More informationQuadratic Forms. Quadratic Forms
Qudrtic Forms Recll the Simon & Blume excerpt from n erlier lecture which sid tht the min tsk of clculus is to pproximte nonliner functions with liner functions. It s ctully more ccurte to sy tht we pproximte
More informationGoals: Determine how to calculate the area described by a function. Define the definite integral. Explore the relationship between the definite
Unit #8 : The Integrl Gols: Determine how to clculte the re described by function. Define the definite integrl. Eplore the reltionship between the definite integrl nd re. Eplore wys to estimte the definite
More informationMIXED MODELS (Sections ) I) In the unrestricted model, interactions are treated as in the random effects model:
1 2 MIXED MODELS (Sections 17.7 17.8) Exmple: Suppose tht in the fiber breking strength exmple, the four mchines used were the only ones of interest, but the interest ws over wide rnge of opertors, nd
More informationMatrix Solution to Linear Equations and Markov Chains
Trding Systems nd Methods, Fifth Edition By Perry J. Kufmn Copyright 2005, 2013 by Perry J. Kufmn APPENDIX 2 Mtrix Solution to Liner Equtions nd Mrkov Chins DIRECT SOLUTION AND CONVERGENCE METHOD Before
More informationChapter 10: Symmetrical Components and Unbalanced Faults, Part II
Chpter : Symmetricl Components nd Unblnced Fults, Prt.4 Sequence Networks o Loded Genertor n the igure to the right is genertor supplying threephse lod with neutrl connected through impednce n to ground.
More informationNumerical Integration
Chpter 5 Numericl Integrtion Numericl integrtion is the study of how the numericl vlue of n integrl cn be found. Methods of function pproximtion discussed in Chpter??, i.e., function pproximtion vi the
More information3.4 Numerical integration
3.4. Numericl integrtion 63 3.4 Numericl integrtion In mny economic pplictions it is necessry to compute the definite integrl of relvlued function f with respect to "weight" function w over n intervl [,
More informationA recursive construction of efficiently decodable list-disjunct matrices
CSE 709: Compressed Sensing nd Group Testing. Prt I Lecturers: Hung Q. Ngo nd Atri Rudr SUNY t Bufflo, Fll 2011 Lst updte: October 13, 2011 A recursive construction of efficiently decodble list-disjunct
More informationLecture 3. In this lecture, we will discuss algorithms for solving systems of linear equations.
Lecture 3 3 Solving liner equtions In this lecture we will discuss lgorithms for solving systems of liner equtions Multiplictive identity Let us restrict ourselves to considering squre mtrices since one
More informationBest Approximation in the 2-norm
Jim Lmbers MAT 77 Fll Semester 1-11 Lecture 1 Notes These notes correspond to Sections 9. nd 9.3 in the text. Best Approximtion in the -norm Suppose tht we wish to obtin function f n (x) tht is liner combintion
More informationInformation synergy, part 3:
Informtion synergy prt : belief updting These notes describe belief updting for dynmic Kelly-Ross investments where initil conditions my mtter. This note diers from the first two notes on informtion synergy
More informationNon-Linear & Logistic Regression
Non-Liner & Logistic Regression If the sttistics re boring, then you've got the wrong numbers. Edwrd R. Tufte (Sttistics Professor, Yle University) Regression Anlyses When do we use these? PART 1: find
More informationECO 317 Economics of Uncertainty Fall Term 2007 Notes for lectures 4. Stochastic Dominance
Generl structure ECO 37 Economics of Uncertinty Fll Term 007 Notes for lectures 4. Stochstic Dominnce Here we suppose tht the consequences re welth mounts denoted by W, which cn tke on ny vlue between
More informationWeek 10: Line Integrals
Week 10: Line Integrls Introduction In this finl week we return to prmetrised curves nd consider integrtion long such curves. We lredy sw this in Week 2 when we integrted long curve to find its length.
More informationDiscrete Mathematics and Probability Theory Summer 2014 James Cook Note 17
CS 70 Discrete Mthemtics nd Proility Theory Summer 2014 Jmes Cook Note 17 I.I.D. Rndom Vriles Estimting the is of coin Question: We wnt to estimte the proportion p of Democrts in the US popultion, y tking
More information1 Linear Least Squares
Lest Squres Pge 1 1 Liner Lest Squres I will try to be consistent in nottion, with n being the number of dt points, nd m < n being the number of prmeters in model function. We re interested in solving
More informationLecture 3 Gaussian Probability Distribution
Introduction Lecture 3 Gussin Probbility Distribution Gussin probbility distribution is perhps the most used distribution in ll of science. lso clled bell shped curve or norml distribution Unlike the binomil
More information1. Gauss-Jacobi quadrature and Legendre polynomials. p(t)w(t)dt, p {p(x 0 ),...p(x n )} p(t)w(t)dt = w k p(x k ),
1. Guss-Jcobi qudrture nd Legendre polynomils Simpson s rule for evluting n integrl f(t)dt gives the correct nswer with error of bout O(n 4 ) (with constnt tht depends on f, in prticulr, it depends on
More informationThe steps of the hypothesis test
ttisticl Methods I (EXT 7005) Pge 78 Mosquito species Time of dy A B C Mid morning 0.0088 5.4900 5.5000 Mid Afternoon.3400 0.0300 0.8700 Dusk 0.600 5.400 3.000 The Chi squre test sttistic is the sum of
More informationDiscrete Mathematics and Probability Theory Spring 2013 Anant Sahai Lecture 17
EECS 70 Discrete Mthemtics nd Proility Theory Spring 2013 Annt Shi Lecture 17 I.I.D. Rndom Vriles Estimting the is of coin Question: We wnt to estimte the proportion p of Democrts in the US popultion,
More informationChapter 4 Contravariance, Covariance, and Spacetime Diagrams
Chpter 4 Contrvrince, Covrince, nd Spcetime Digrms 4. The Components of Vector in Skewed Coordintes We hve seen in Chpter 3; figure 3.9, tht in order to show inertil motion tht is consistent with the Lorentz
More informationMath 8 Winter 2015 Applications of Integration
Mth 8 Winter 205 Applictions of Integrtion Here re few importnt pplictions of integrtion. The pplictions you my see on n exm in this course include only the Net Chnge Theorem (which is relly just the Fundmentl
More information1.9 C 2 inner variations
46 CHAPTER 1. INDIRECT METHODS 1.9 C 2 inner vritions So fr, we hve restricted ttention to liner vritions. These re vritions of the form vx; ǫ = ux + ǫφx where φ is in some liner perturbtion clss P, for
More informationP 3 (x) = f(0) + f (0)x + f (0) 2. x 2 + f (0) . In the problem set, you are asked to show, in general, the n th order term is a n = f (n) (0)
1 Tylor polynomils In Section 3.5, we discussed how to pproximte function f(x) round point in terms of its first derivtive f (x) evluted t, tht is using the liner pproximtion f() + f ()(x ). We clled this
More informationAQA Further Pure 1. Complex Numbers. Section 1: Introduction to Complex Numbers. The number system
Complex Numbers Section 1: Introduction to Complex Numbers Notes nd Exmples These notes contin subsections on The number system Adding nd subtrcting complex numbers Multiplying complex numbers Complex
More information1B40 Practical Skills
B40 Prcticl Skills Comining uncertinties from severl quntities error propgtion We usully encounter situtions where the result of n experiment is given in terms of two (or more) quntities. We then need
More informationfractions Let s Learn to
5 simple lgebric frctions corne lens pupil retin Norml vision light focused on the retin concve lens Shortsightedness (myopi) light focused in front of the retin Corrected myopi light focused on the retin
More informationFrobenius numbers of generalized Fibonacci semigroups
Frobenius numbers of generlized Fiboncci semigroups Gretchen L. Mtthews 1 Deprtment of Mthemticl Sciences, Clemson University, Clemson, SC 29634-0975, USA gmtthe@clemson.edu Received:, Accepted:, Published:
More informationNew Expansion and Infinite Series
Interntionl Mthemticl Forum, Vol. 9, 204, no. 22, 06-073 HIKARI Ltd, www.m-hikri.com http://dx.doi.org/0.2988/imf.204.4502 New Expnsion nd Infinite Series Diyun Zhng College of Computer Nnjing University
More informationCS667 Lecture 6: Monte Carlo Integration 02/10/05
CS667 Lecture 6: Monte Crlo Integrtion 02/10/05 Venkt Krishnrj Lecturer: Steve Mrschner 1 Ide The min ide of Monte Crlo Integrtion is tht we cn estimte the vlue of n integrl by looking t lrge number of
More informationODE: Existence and Uniqueness of a Solution
Mth 22 Fll 213 Jerry Kzdn ODE: Existence nd Uniqueness of Solution The Fundmentl Theorem of Clculus tells us how to solve the ordinry dierentil eqution (ODE) du f(t) dt with initil condition u() : Just
More informationCredibility Hypothesis Testing of Fuzzy Triangular Distributions
666663 Journl of Uncertin Systems Vol.9, No., pp.6-74, 5 Online t: www.jus.org.uk Credibility Hypothesis Testing of Fuzzy Tringulr Distributions S. Smpth, B. Rmy Received April 3; Revised 4 April 4 Abstrct
More informationReversals of Signal-Posterior Monotonicity for Any Bounded Prior
Reversls of Signl-Posterior Monotonicity for Any Bounded Prior Christopher P. Chmbers Pul J. Hely Abstrct Pul Milgrom (The Bell Journl of Economics, 12(2): 380 391) showed tht if the strict monotone likelihood
More informationThe First Fundamental Theorem of Calculus. If f(x) is continuous on [a, b] and F (x) is any antiderivative. f(x) dx = F (b) F (a).
The Fundmentl Theorems of Clculus Mth 4, Section 0, Spring 009 We now know enough bout definite integrls to give precise formultions of the Fundmentl Theorems of Clculus. We will lso look t some bsic emples
More information( dg. ) 2 dt. + dt. dt j + dh. + dt. r(t) dt. Comparing this equation with the one listed above for the length of see that
Arc Length of Curves in Three Dimensionl Spce If the vector function r(t) f(t) i + g(t) j + h(t) k trces out the curve C s t vries, we cn mesure distnces long C using formul nerly identicl to one tht we
More informationInfinite Geometric Series
Infinite Geometric Series Finite Geometric Series ( finite SUM) Let 0 < r < 1, nd let n be positive integer. Consider the finite sum It turns out there is simple lgebric expression tht is equivlent to
More informationCzechoslovak Mathematical Journal, 55 (130) (2005), , Abbotsford. 1. Introduction
Czechoslovk Mthemticl Journl, 55 (130) (2005), 933 940 ESTIMATES OF THE REMAINDER IN TAYLOR S THEOREM USING THE HENSTOCK-KURZWEIL INTEGRAL, Abbotsford (Received Jnury 22, 2003) Abstrct. When rel-vlued
More informationEntropy and Ergodic Theory Notes 10: Large Deviations I
Entropy nd Ergodic Theory Notes 10: Lrge Devitions I 1 A chnge of convention This is our first lecture on pplictions of entropy in probbility theory. In probbility theory, the convention is tht ll logrithms
More informationMATH34032: Green s Functions, Integral Equations and the Calculus of Variations 1
MATH34032: Green s Functions, Integrl Equtions nd the Clculus of Vritions 1 Section 1 Function spces nd opertors Here we gives some brief detils nd definitions, prticulrly relting to opertors. For further
More informationLecture Note 9: Orthogonal Reduction
MATH : Computtionl Methods of Liner Algebr 1 The Row Echelon Form Lecture Note 9: Orthogonl Reduction Our trget is to solve the norml eution: Xinyi Zeng Deprtment of Mthemticl Sciences, UTEP A t Ax = A
More informationLecture 1: Introduction to integration theory and bounded variation
Lecture 1: Introduction to integrtion theory nd bounded vrition Wht is this course bout? Integrtion theory. The first question you might hve is why there is nything you need to lern bout integrtion. You
More informationMATH 144: Business Calculus Final Review
MATH 144: Business Clculus Finl Review 1 Skills 1. Clculte severl limits. 2. Find verticl nd horizontl symptotes for given rtionl function. 3. Clculte derivtive by definition. 4. Clculte severl derivtives
More informationStudent Activity 3: Single Factor ANOVA
MATH 40 Student Activity 3: Single Fctor ANOVA Some Bsic Concepts In designed experiment, two or more tretments, or combintions of tretments, is pplied to experimentl units The number of tretments, whether
More informationMapping the delta function and other Radon measures
Mpping the delt function nd other Rdon mesures Notes for Mth583A, Fll 2008 November 25, 2008 Rdon mesures Consider continuous function f on the rel line with sclr vlues. It is sid to hve bounded support
More informationNUMERICAL INTEGRATION
NUMERICAL INTEGRATION How do we evlute I = f (x) dx By the fundmentl theorem of clculus, if F (x) is n ntiderivtive of f (x), then I = f (x) dx = F (x) b = F (b) F () However, in prctice most integrls
More informationAn approximation to the arithmetic-geometric mean. G.J.O. Jameson, Math. Gazette 98 (2014), 85 95
An pproximtion to the rithmetic-geometric men G.J.O. Jmeson, Mth. Gzette 98 (4), 85 95 Given positive numbers > b, consider the itertion given by =, b = b nd n+ = ( n + b n ), b n+ = ( n b n ) /. At ech
More information