The Information Bottleneck Revisited or How to Choose a Good Distortion Measure
|
|
- Marian Garrett
- 5 years ago
- Views:
Transcription
1 Te Informaton Bottleneck Revsted or How to Coose a Good Dstorton Measure Peter Harremoës Centrum voor Wskunde en Informatca P.O , 090 GB Amsterdam Te Nederlands P.Harremoes@cw.nl Naftal Tsby Te Hebrew Unversty Jerusalem Israel tsby@cs.uj.ac.l Abstract It s well-known tat te nformaton bottleneck metod and rate dstorton teory are related. Here t s descrbed ow te nformaton bottleneck can be consdered as rate dstorton teory for a famly of probablty measures were nformaton dvergence s used as dstorton measure. It s sown tat te nformaton bottleneck metod as some propertes tat are not sared wt rate dstorton teory based on any oter dvergence measure. In ts sense te nformaton bottleneck metod s unque. I. INTRODUCTION For many problems n te teory of lossy compresson t s dfcult to specfy a dstorton measure. In some cases we want to compress a varable but wat we are really nterested n s not te value of but te value of a varable Y correlated or coupled wt. We are only nterested n retrevng te value of to te extent tat t gves nformaton about Y. Tus, eac value of gves a dstrbuton on Y. Te nformaton bottleneck metod was ntroduced n [] to solve ts problem. It as always been known tat te nformaton bottleneck metod s related to rate dstorton teory. In ts paper we sall explore te relaton n detal. Te nformaton bottleneck metod as found natural nterpretatons and a number of applcatons as descrbed n [], [3], [4], [5], [6], [7]. Te results n tese papers do not rule out te possblty tat smlar results could ave been obtaned by oter means (oter dstorton measure). Our approac wll be va rate dstorton teory and as we sall see we are lead drectly to te nformaton bottleneck va a number of assumptons or smplcatons tat efcently rule out all oter dstorton or dvergence measures tan nformaton dvergence. Let A and B be alpabets and let and Y be random varables wt values n A and B For smplcty we sall assume tat A and B are nte but most of te results n ts paper olds for nnte sets as well. Te jont dstrbuton of and Y s gven and gven by a dstrbuton on A and a Markov kernel A! M + (B) ; were M + (B) denotes te set of probablty measures on B As reconstructon alpabet we use ^A M + (B)! Y As dstorton measure d A ^A! R we use d(x; ^x) D ( (x) ; ^x) () were D denotes some dvergence measure on M+ (B) Our goal s to mnmze bot te rate I ; ^ and te dstorton E d ; ^ over all jont dstrbuton of ; ^ A ^A wt prescrbed margnal dstrbuton of Te trade-off between rate and dstorton s gven by te rate dstorton curve. To nd te pont on te rate dstorton curve wt slope one sould mnmze I ; ^ Te most mportant dvergence measure s nformaton dvergence (or Kullback-Lebler nformaton or relatve entropy) dened by D (P kq) p log p q yb We know from Sanov's Teorem [8] tat te dfculty n dstngusng a dstrbuton P from a dstrbuton Q by a statstcal test s gven by te nformaton dvergence. Ts suggests to use nformaton dvergence as dstorton measure. We sall now formalze tese deas. Informaton dvergence belong to a class of dvergence measures called Csszár f-dvergences dened by D f (P; Q) p f B were f s denotes a convex functon satsfyng f () 0 [9], [0]. For f (x) x log x we get nformaton dvergence. For f (x) (x ) we get -dvergence. For f (x) x we get Hellnger dvergence. for f (x) jx j we get varatonal dstance. Informaton dvergence also belongs to te class of Bregman dvergences []. For a nte output alpabet a Bregman dvergence on M + (B) s dened by B f (P; Q) f (P ) (f (Q) + (P Q) rf (Q)) were f M + (B)! R s some convex functon. Informaton dvergence s obtaned wen q f (P ) B p log p p
2 Rate dstorton teory wt Bregman dvergences s well studed n []. Sometmes so-called Burbea-Rao dvergences are used. Te Burbea-Rao dvergence between P and Q s dened by BR f (P; Q) B f (p ) + f (q ) p + q f for some convex functon f [3]. An mportant class of dvergence measures are te separable dvergences ntroduced n [4]. Tese are dvergences dened by D (P; Q) (p ; q ) B for some functon [0; ]! R We note tat Csszár f-dvergences and Burbea-Rao dvergences a separable. A separable Bregman dvergence s gven by B f (P; Q) B g (p ) (g (q ) + (p q ) g 0 (q )) for some convex functon g [0; ]! R A dvergence measure s sad to be reexve f D (P; Q) 0 wt equalty for P Q For dstrbutons P and Q close to eac oter all tese dvergence measures are approxmately proportonal [5] f te functons used to dene tem are sufcently smoot. II. ADDITIVITY Rate dstorton teory s most nterestng wen you ave a rate dstorton teorem. In order to get a rate dstorton teorem one as to consder sequences nstead of sngle events and one as to extend te denton of dstorton from a sequence of nputs x n n A n to a sequence of reconstructon ponts ^x n by d (x n ; ^x n ) n n d (x ; ^x ) ; see [8]. Te factor n s just a matter of normalzaton but for our problem t s essental tat our dvergence measure s addtve,.e. D (P P ; Q Q ) D (P ; Q ) + D (P ; Q ) () As te dvergence on te left and sde lves on a dfferent space tan te dvergences on te rgt and sde Equaton could be used to dene te dvergence on a product space and n ts sense Equaton cannot be used to caracterze dvergence measures sutable for rate dstorton teory. If we requre tat te dvergence measures n te left and te rgt and sde of Equaton belong to specal classes of dvergence measures one can obtan non-trval caracterzatons. Teorem A separable reexve dvergence measure tat s addtve s a lnear combnaton og nformaton dvergence and reversed nformaton dvergence. Proof Let te Dvergence measure D be based on ]0; [! R Take P (p ) B ; Q (p ) B ; P (s; B s) and Q (; ) Ten (p ; q ) + s; + s; B B sp ; q + ( s) p ; q By takng te second dervatve at bot sdes we get 00 s; + 00 s; 00 sp ; q p + 00 For s we get 00 ; B ( s) p ; q 00 p ; q p p Ts equaton sould old for all probablty vectors P and Q wc mples tat (x; y)! 00 (x; y) x s lnear n x and y Tus tere exsts contants c ; c and suc tat 00 (x; y) x c x + c y avng solutons of te form (x; y) c x log x y + c y log y + f (y) x + g (y) x for some functons f; g Te rst two terms gves nformaton dvergence and reversed nformaton dvergence so we just ave to ceck addtvty of a dvergence based on ~ (x; y) f (y) x + g (y) Takng te second dervatve wt respect to y leads to (f 00 (y) x + g 00 (y)) y c 4 x + c 5 y n te same way as above. Ts sould old for all x leadng to f 00 (y) c 4 y and g 00 (y) c 5 y Te solutons are f (y) c 4 log y + c 6 y + c 7 g (y) c 5 y log y + c 8 y + c 9 for some constants c 6 ; c 7 ; c 8 and c 9 Tus ~ (x; y) ( c 4 log y + c 6 y + c 7 ) x + c 5 y log y + c 8 y + c 9 Te terms tat are lnear n x or y may be replaced by constants wtout cangng te dvergence so we may assume tat ~ (x; y) c 4 x log y + c 5 y log y + c 7 xy + c 0 One easly cecks tat te rst two terms satses addtvty and te second ones do not except for c 7 c 0 0 For x y we ave 0 ~ (x; x) (c 5 c 4 ) x log x wc mples tat c 5 c 4 and c 7 c 0 Tus ~ (x; y) c 4 (y log y x log y) and postvty s only obtaned for c 0
3 III. SUFFICIENCY Te number of parameters n our rate dstorton problem can often be reduced by replacng te orgnal varables wt sufcent varables. Ts dea s actually used already n formulatng te model. In prncple any model can always be extended by ncludng less relevant or even rrelevant varables. Normally one would leave out tese less relevant varables at an early stage, but as we sall see n te nformaton bottle neck metod one can also get rd of rrelevant varables wtn te model. Frst we sall consder sufcency on te nput sde. Assume tat ( ; ) and tat s ndependent of Y gven. Equvalently,!! Y s assumed to be a Markov can. As Y only depend on va we would lke to leave out from te analyss. Tus we sould compare te bottleneck problem! Y wt te bottleneck problem! Y and sow tat tey ave te same rate dstorton functon. Obvously any jont dstrbuton on ^; ; Y gves a jont dstrbuton on ^; ; Y wt te same mean dstorton and a smaller (or equal) rate. Let a jont dstrbuton on ^; ; Y be gven were ^ and Y are ndependent gven Te jont dstrbuton on ( ; ; Y ) denes a Markov kernel from to Now consder te jont dstrbuton on ^; ; ; Y were ; ^ and Y are ndependent gven For ts jont dstrbuton temean dstorton s equal to te mean dstorton of ^; ; Y and te rate equals I ^; I ^; I ^; + I ^; j We note tat ts sufcency result on te nput sde olds for any dstorton measure. Next we sall consder sufcency on te output sde. Assume tat Y (Y ; Y ) and tat s ndependent of Y gven Y. Equvalently,! Y! Y s assumed to be a Markov can. As Y only depend on va Y we would lke to leave out Y from te analyss. Tus we sould compare te bottleneck problem! Y wt te bottleneck problem! Y and sow tat tey ave te same rate dstorton functon. We ave D P Y ( j ) kp Y 0 P Y ( j ) kp Y +D P Y ( j Y ; ) kp Y j Y ; ^ A Terefore E D P Y ( j ) kp Y E D P Y ( j ) kp Y (3) Note tat Equaton (3) olds for any f-dvergence. We sall sow tat t essentally only olds for f-dvergences. It s easy to nd examples of pars of general dvergence measures tat fulll Equaton (3) for some specc jont dstrbuton on (; Y ) but t s natural to requre tat te dstorton measure on M+ (B) sould not depend on te jont dstrbuton on (; Y ) Teorem A separable dvergence measure tat satses te sufcency condton as te form D (P; Q) p f q (4) B Remark 3 Te functon f n Equaton 4 s not necessarly convex. If f s convex and f () 0 te dvergence s a Csszár f-dvergence. Proof Te proof of ts result was essentally gven as part of te proof of Teorem n [6]. Ter teorem states tat a separable dvergence satsfyng a data processng nequalty must be a Csszár f-dvergence. Under some condtons ts result can even be extended to Bregman dvergences tat are not separable. q IV. FINITENESS Assume tat te number of elements n A s n Let P M+ (A) denote te margnal dstrbuton of Consder dstrbutons on AM+ (B) ;.e. a jont dstrbuton on ; ^ Te jont dstrbuton can be speced by te dstrbuton Q of ^ and te condtonal dstrbuton of gven ^ gven by a Markov kernel E M+ (B)! M+ (A) Te set of dstrbutons on M+ (B) s an nnte smplex. Consder te convex set C of dstrbuton Q suc tat Q and E determnes jont dstrbuton of ; ^ as P as te margnal dstrbuton of Te condton tat te margnal dstrbuton of as P as prescrbed dstrbuton gves n lnearly ndependent condtons. Terefore te extreme ponts of C are mxtures of at most n ponts. We ave I ; ^ H () H + E E d ; ^ H () + E E d ; ^ H Note tat E d ; ^ H s a functon of ^;.e. a functon M + (B)! R Tus te mnmum of I ; ^ s attaned for an extreme pont of C,.e. te mnmum s attaned for a dstrbuton on ^A wt support on at most n ponts. In te
4 nformaton bottleneck lterature one wll normally nd te result tat te support as at most n + ponts. Note tat no partcular propertes of te dstorton measure d ave been used. Let b A denote a set wt n elements. From now on we sall dentfy a couplng between B and M + (A) wt a couplng between B and b A togeter wt a map b A! M + (B) We know tat ts s no restrcton for calculatng te rate dstorton functon. V. THE BOTTLENECK EQUATIONS Let a jont dstrbuton on ; ^ A A b be gven. Ts gves Markov kernel A b! M+ (A) It can be composed wt to gve te map A b! M+ (B) dened by (^x) xa (x j ^x) (x) We sall use tat nformaton dvergence satses te socalled compensaton equalty rst recognzed n [7]. Let (s ; s ; ; s k ) denote a probablty vector and let Q and P ; P ; ; P k denote probablty measures on te same set. Ten s j D (P j kq) s j D P j kp + D P kq were P P j s jp j In partcular s j D (P j kq) Ts leads to E d(; (^x)) ^x E s j D P j kp (6) E D ( () k (^x)) ^x D ( () k (^x)) ^x Terefore E d(; ( ^)) E D () k ^ E d(; ( ^)) We also ave tat I ; ^ I ; ^ so nstead of mnmzng over all possble maps and all possble jont dstrbutons on A A b we just ave to mnmze over jont dstrbutons and put ^!! Y For a jont dstrbuton of ; ^ we ave tat (^x) s te dstrbuton of Y gven ^ ^x f ^ and Y are ndependent gven ;.e. ^!! Y form a Markov can. Now I(; Y ) I ^; Y + I ; Y I ^; Y + E D () k ^ I ^; Y + E d ; ^ Tus E d(; ( ^)) I(; Y ) I ^; Y We note tat I (; Y ) s a constant and get te followng teorem. Teorem 4 Wen nformaton dvergence s used as dstorton measure nf I ; ^ I(; Y ) + nf I ; ^ I ^; Y Te soluton satses te so-called bottleneck equatons,.e. te mnmum s attaned for reconstructon ponts satsfyng and jont dstrbuton satses te Kun-Tucker condtons for rate dstorton wt tese reconstructons ponts. Te last term s essentally te one tat sall be mnmzed nte nformaton bottleneck were one wants to mnmze I ; ^ and at te same tme maxmze I ^; Y We ave seen tat nequalty (6) s essentally n dervng Teorem 4. Teorem 5 If d M+ (A) M+ (A)! R s a Csszár f-dvergenc for some dfferentable functon f and s j d (P j ; Q) s j d P j ; P for all mxtures P P k s jp j and all Q ten d s proportonal to nformaton dvergence. Proof Assume tat d s te Csszár f-dvergence gven by d (P; Q) D f (P kq) P () f Q () Q () B We sall wrte f as f (x) xg x so tat D f (P kq) Q () g P () P () B We ave Q () s j g P j () P j () B s j B P () g P j () P j () Te probablty vector (Q () ; Q () ; ) satses P Q () We ntroduce a Lagrange multpler and Q () s j g P j () Q () () P j () Dene (x) g 0 B x Ten Pj () s j P () Q () s j g 0 P j ()
5 olds for all because P s a statonary pont. In partcular t sould old f tere exsts an 0 suc tat P j ( 0 ) does not depend on j In tat case Pj ( 0 ) s j s P ( 0 j () () ) Tus for any B we ave Pj () s j () P () Tat mples tat (x) () + (x ) for some constant Ten Z Z g (x) g 0 (x) dx () + x dx ( () ) x + log x + c for come constant c R Ten f (x) x ( () ) x + log x + c () + x log x + cx Te condton f () 0 mples () c Hence D f (P; Q) B B f (x) x log x + c (x ) P () P () P () log Q () Q () + c Q () P () log P () D (P kq) Q () Q () In [] t was sown tat a dvergence measure satsfyng (6) must be a Bregman dvergence. Ts leads us to te followng corollary. Corollary 6 A dvergence measure tat s bot a f- dvergence and a Bregman dvergence must be proportonal to nformaton dvergence. In [6, Tm. 4] te ntersecton of te set of Bregman and f-dvergences s caracterzed by an equaton and a concavty condton and nformaton dvergence s gven as an example of an element n te ntersecton. Corollary 6 mples tat tere are essentally no oter elements n te ntersecton. VI. CONCLUSION In ts paper varous dstorton measures ave been consdered and ter propertes wt respect to rate dstorton teory ave been studed. Some of te results are summarzed n te followng table. Property Addtvty Sufcency on nput sde Sufcency on output sde Data processng nequalty Fnteness Bottleneck equaton Class of dvergences Inf. dv. and reversed All non-convex "f-dv." Csszár f-dv. All Bregman dv. We see tat f one wants to ave all te propertes fullled te dvergence s equal to or proportonal to nformaton dvergence. A dvergence tat s proportonal to nformaton dvergence s essentally nformaton dvergence were nformaton dvergence s measured by a dfferent unt (for nstance bts nstead of nats). Tus t s normally desrable to use nformaton dvergence and tat leads to te nformaton bottleneck metod. Acknowledgement 7 Te autors want to tank Peter Grünwald and Tm van Erven for useful dscussons and comments. REFERENCES [] N. Tsby, F. Perera, and W. Balek, Te nformaton bottleneck metod, n Proceedngs of te 37-t Annual Allerton Conference on Communcaton, Control and Computng, pp , 999. [] N. Slonm and N. Tsby, Document clusterng usng word clusters va te nformaton bottleneck metod, n SIGIR '00 Proceedngs of te 3rd annual nternatonal ACM SIGIR conference on Researc and development n nformaton retreval, (New York, NY, USA), pp. 08 5, ACM Press, 000. [3] G. Ceck, A. Globerson, N. Tsby, and Y. Wess, Informaton bottleneck for Gaussan varables, n Advances n Neural Informaton Processng Systems 6 (S. Trun, L. Saul, and B. Scölkopf, eds.), Cambrdge, MA MIT Press, 004. [4] C. R. Salz and J. P. Crutceld, Informaton bottlenecks, causal states, and statstcal relevance bases How to represent relevant nformaton n memoryless transducton, tec. rep., Santa Fe Insttute, 00. [5] N. Slonm and Y. Wess, Maxmum lkelood and te nformaton bottleneck, n Advances n Neural Informaton Processng Systems 5, Vancouver, Canada, 00 (S. Becker, S. Trun, and K. Obermayer, eds.), pp , 003. [6] G. Eldan and N. Fredman, Te nformaton bottleneck EM algortm, n Proceedngs of te 9t Annual Conference on Uncertanty n Artcal Intellgence (UAI-03), (San Francsco, CA), pp , Morgan Kaufmann, 003. [7] D. Gondek and T. Hofmann, Condtonal nformaton bottleneck clusterng, n Worksop on Clusterng Large Data Sets, IEEE Internatonal Conference on Data Mnng, Melbourne, USA, November 9-, 003., 003. [8] T. Cover and J. A. Tomas, Elements of Informaton Teory. Wley, 99. [9] I. Csszár and P. Selds, Informaton Teory and Statstcs A Tutoral. Foundatons and Trends n Communcatons and Informaton Teory, Now Publsers Inc., 004. [0] F. Lese and I. Vajda, On dvergence and nformatons n statstcs and nformaton teory, IEEE Tranns. Inform. Teory, vol. 5, pp , Oct [] L. M. Bregman, Te relaxaton metod of ndng te common pont of convex sets and ts applcaton to te soluton of problems n convex programmng, USSR Comput. Mat. and Mat. Pys., vol. 7, pp. 00 7, 967. Translated from Russan. [] A. Banerjee, S. Merugu, I. S. Dllon, and J. Gos, Clusterng wt Bregman dvergences, Journal of Macne Learnng Researc, vol. 6, pp , 005. [3] J. Burbea and C. R. Rao, On te convexty of some dvergence measures based on entropy functons, IEEE Trans. Inform. Teory, vol. 8, pp , 98. [4] P. D. Grünwald and A. P. Dawd, Game teory, maxmum entropy, mnmum dscrepancy, and robust Bayesan decson teory, Annals of Matematcal Statstcs, vol. 3, no. 4, pp , 004. [5] M. C. Pardo and I. Vajda, On asymptotc propertes of nformatonteoretc dvergences., IEEE Trans. Inform. Teory, vol. 49, no. 7, pp , 003. [6] M. C. Pardo and I. Vajda, About dstances of dscrete dstrbutons satsfyng te data processng teorem of nformaton teory, IEEE Trans. Inform. Teory, vol. 43, pp , 997. [7] F. Topsøe, An nformaton teoretcal dentty and a problem nvolvng capacty, Studa Scentarum Matematcarum Hungarca, vol., pp. 9 9, 967.
The Information Bottleneck Revisited or How to Choose a Good Distortion Measure
The Information Bottleneck Revisited or How to Choose a Good Distortion Measure Peter Harremoës Centrum voor Wiskunde en Informatica PO 94079, 1090 GB Amsterdam The Nederlands PHarremoes@cwinl Naftali
More informationStanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7
Stanford Unversty CS54: Computatonal Complexty Notes 7 Luca Trevsan January 9, 014 Notes for Lecture 7 1 Approxmate Countng wt an N oracle We complete te proof of te followng result: Teorem 1 For every
More informationOn Pfaff s solution of the Pfaff problem
Zur Pfaff scen Lösung des Pfaff scen Probles Mat. Ann. 7 (880) 53-530. On Pfaff s soluton of te Pfaff proble By A. MAYER n Lepzg Translated by D. H. Delpenc Te way tat Pfaff adopted for te ntegraton of
More informationThe Finite Element Method: A Short Introduction
Te Fnte Element Metod: A Sort ntroducton Wat s FEM? Te Fnte Element Metod (FEM) ntroduced by engneers n late 50 s and 60 s s a numercal tecnque for solvng problems wc are descrbed by Ordnary Dfferental
More information3.1 ML and Empirical Distribution
67577 Intro. to Machne Learnng Fall semester, 2008/9 Lecture 3: Maxmum Lkelhood/ Maxmum Entropy Dualty Lecturer: Amnon Shashua Scrbe: Amnon Shashua 1 In the prevous lecture we defned the prncple of Maxmum
More informationCOMP4630: λ-calculus
COMP4630: λ-calculus 4. Standardsaton Mcael Norrs Mcael.Norrs@ncta.com.au Canberra Researc Lab., NICTA Semester 2, 2015 Last Tme Confluence Te property tat dvergent evaluatons can rejon one anoter Proof
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationSolution for singularly perturbed problems via cubic spline in tension
ISSN 76-769 England UK Journal of Informaton and Computng Scence Vol. No. 06 pp.6-69 Soluton for sngularly perturbed problems va cubc splne n tenson K. Aruna A. S. V. Rav Kant Flud Dynamcs Dvson Scool
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationOn a nonlinear compactness lemma in L p (0, T ; B).
On a nonlnear compactness lemma n L p (, T ; B). Emmanuel Matre Laboratore de Matématques et Applcatons Unversté de Haute-Alsace 4, rue des Frères Lumère 6893 Mulouse E.Matre@ua.fr 3t February 22 Abstract
More informationMultivariate Ratio Estimator of the Population Total under Stratified Random Sampling
Open Journal of Statstcs, 0,, 300-304 ttp://dx.do.org/0.436/ojs.0.3036 Publsed Onlne July 0 (ttp://www.scrp.org/journal/ojs) Multvarate Rato Estmator of te Populaton Total under Stratfed Random Samplng
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationProblem Set 4: Sketch of Solutions
Problem Set 4: Sketc of Solutons Informaton Economcs (Ec 55) George Georgads Due n class or by e-mal to quel@bu.edu at :30, Monday, December 8 Problem. Screenng A monopolst can produce a good n dfferent
More informationLecture 17 : Stochastic Processes II
: Stochastc Processes II 1 Contnuous-tme stochastc process So far we have studed dscrete-tme stochastc processes. We studed the concept of Makov chans and martngales, tme seres analyss, and regresson analyss
More informationTR/95 February Splines G. H. BEHFOROOZ* & N. PAPAMICHAEL
TR/9 February 980 End Condtons for Interpolatory Quntc Splnes by G. H. BEHFOROOZ* & N. PAPAMICHAEL *Present address: Dept of Matematcs Unversty of Tabrz Tabrz Iran. W9609 A B S T R A C T Accurate end condtons
More informationOutline. Bayesian Networks: Maximum Likelihood Estimation and Tree Structure Learning. Our Model and Data. Outline
Outlne Bayesan Networks: Maxmum Lkelhood Estmaton and Tree Structure Learnng Huzhen Yu janey.yu@cs.helsnk.f Dept. Computer Scence, Unv. of Helsnk Probablstc Models, Sprng, 200 Notces: I corrected a number
More informationImage classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?
Image classfcaton Gven te bag-of-features representatons of mages from dfferent classes ow do we learn a model for dstngusng tem? Classfers Learn a decson rule assgnng bag-offeatures representatons of
More informationHidden Markov Models & The Multivariate Gaussian (10/26/04)
CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationStat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors
Stat60: Bayesan Modelng and Inference Lecture Date: February, 00 Reference Prors Lecturer: Mchael I. Jordan Scrbe: Steven Troxler and Wayne Lee In ths lecture, we assume that θ R; n hgher-dmensons, reference
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationThe Order Relation and Trace Inequalities for. Hermitian Operators
Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence
More information1 Introducton Nonlnearty crtera of Boolean functons Cryptograpc transformatons sould be nonlnear to be secure aganst varous attacks. For example, te s
KUIS{94{000 Nonlnearty crtera of Boolean functons HIROSE Souc IKEDA Katsuo Tel +81 75 753 5387 Fax +81 75 751 048 E-mal frose, kedag@kus.kyoto-u.ac.jp July 14, 1994 1 Introducton Nonlnearty crtera of Boolean
More informationThroughput Capacities and Optimal Resource Allocation in Multiaccess Fading Channels
Trougput Capactes and Optmal esource Allocaton n ultaccess Fadng Cannels Hao Zou arc 7, 003 Unversty of Notre Dame Abstract oble wreless envronment would ntroduce specal penomena suc as multpat fadng wc
More informationMAXIMUM A POSTERIORI TRANSDUCTION
MAXIMUM A POSTERIORI TRANSDUCTION LI-WEI WANG, JU-FU FENG School of Mathematcal Scences, Peng Unversty, Bejng, 0087, Chna Center for Informaton Scences, Peng Unversty, Bejng, 0087, Chna E-MIAL: {wanglw,
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationNot-for-Publication Appendix to Optimal Asymptotic Least Aquares Estimation in a Singular Set-up
Not-for-Publcaton Aendx to Otmal Asymtotc Least Aquares Estmaton n a Sngular Set-u Antono Dez de los Ros Bank of Canada dezbankofcanada.ca December 214 A Proof of Proostons A.1 Proof of Prooston 1 Ts roof
More information, rst we solve te PDE's L ad L ad n g g (x) = ; = ; ; ; n () (x) = () Ten, we nd te uncton (x), te lnearzng eedbac and coordnates transormaton are gve
Freedom n Coordnates Transormaton or Exact Lnearzaton and ts Applcaton to Transent Beavor Improvement Kenj Fujmoto and Tosaru Suge Dvson o Appled Systems Scence, Kyoto Unversty, Uj, Kyoto, Japan suge@robotuassyoto-uacjp
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More information2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We revew the crteron of statstcally optmal tranng data (Fukumzu et al., 1994). We
Advances n Neural Informaton Processng Systems 8 Actve Learnng n Multlayer Perceptrons Kenj Fukumzu Informaton and Communcaton R&D Center, Rcoh Co., Ltd. 3-2-3, Shn-yokohama, Yokohama, 222 Japan E-mal:
More informationMATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)
1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationThe finite element method explicit scheme for a solution of one problem of surface and ground water combined movement
IOP Conference Seres: Materals Scence and Engneerng PAPER OPEN ACCESS e fnte element metod explct sceme for a soluton of one problem of surface and ground water combned movement o cte ts artcle: L L Glazyrna
More informationIranian Journal of Mathematical Chemistry, Vol. 5, No.2, November 2014, pp
Iranan Journal of Matematcal Cemstry, Vol. 5, No.2, November 204, pp. 85 90 IJMC Altan dervatves of a grap I. GUTMAN (COMMUNICATED BY ALI REZA ASHRAFI) Faculty of Scence, Unversty of Kragujevac, P. O.
More informationEstimation: Part 2. Chapter GREG estimation
Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the
More informationChapter 7 Channel Capacity and Coding
Wreless Informaton Transmsson System Lab. Chapter 7 Channel Capacty and Codng Insttute of Communcatons Engneerng atonal Sun Yat-sen Unversty Contents 7. Channel models and channel capacty 7.. Channel models
More informationA Discrete Approach to Continuous Second-Order Boundary Value Problems via Monotone Iterative Techniques
Internatonal Journal of Dfference Equatons ISSN 0973-6069, Volume 12, Number 1, pp. 145 160 2017) ttp://campus.mst.edu/jde A Dscrete Approac to Contnuous Second-Order Boundary Value Problems va Monotone
More informationDirect Methods for Solving Macromolecular Structures Ed. by S. Fortier Kluwer Academic Publishes, The Netherlands, 1998, pp
Drect Metods for Solvng Macromolecular Structures Ed. by S. Forter Kluwer Academc Publses, Te Neterlands, 998, pp. 79-85. SAYRE EQUATION, TANGENT FORMULA AND SAYTAN FAN HAI-FU Insttute of Pyscs, Cnese
More informationThermodynamics and statistical mechanics in materials modelling II
Course MP3 Lecture 8/11/006 (JAE) Course MP3 Lecture 8/11/006 Thermodynamcs and statstcal mechancs n materals modellng II A bref résumé of the physcal concepts used n materals modellng Dr James Ellott.1
More informationNatural Language Processing and Information Retrieval
Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support
More informationSolving Singularly Perturbed Differential Difference Equations via Fitted Method
Avalable at ttp://pvamu.edu/aam Appl. Appl. Mat. ISSN: 193-9466 Vol. 8, Issue 1 (June 013), pp. 318-33 Applcatons and Appled Matematcs: An Internatonal Journal (AAM) Solvng Sngularly Perturbed Dfferental
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationChapter 7 Channel Capacity and Coding
Chapter 7 Channel Capacty and Codng Contents 7. Channel models and channel capacty 7.. Channel models Bnary symmetrc channel Dscrete memoryless channels Dscrete-nput, contnuous-output channel Waveform
More informationEM and Structure Learning
EM and Structure Learnng Le Song Machne Learnng II: Advanced Topcs CSE 8803ML, Sprng 2012 Partally observed graphcal models Mxture Models N(μ 1, Σ 1 ) Z X N N(μ 2, Σ 2 ) 2 Gaussan mxture model Consder
More information2 S. S. DRAGOMIR, N. S. BARNETT, AND I. S. GOMM Theorem. Let V :(d d)! R be a twce derentable varogram havng the second dervatve V :(d d)! R whch s bo
J. KSIAM Vol.4, No., -7, 2 FURTHER BOUNDS FOR THE ESTIMATION ERROR VARIANCE OF A CONTINUOUS STREAM WITH STATIONARY VARIOGRAM S. S. DRAGOMIR, N. S. BARNETT, AND I. S. GOMM Abstract. In ths paper we establsh
More informationAPPENDIX A Some Linear Algebra
APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,
More informationx i1 =1 for all i (the constant ).
Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by
More informationISSN On error probability exponents of many hypotheses optimal testing illustrations
Journal Afrka Statstka Vol. 6, 2011, pages 307 315. DOI: http://dx.do.org/10.4314/afst.v61.1 Journal Afrka Statstka ISS 0825-0305 On error probablty exponents of many hypotheses optmal testng llustratons
More informationLossy Compression. Compromise accuracy of reconstruction for increased compression.
Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost
More information5 The Laplace Equation in a convex polygon
5 Te Laplace Equaton n a convex polygon Te most mportant ellptc PDEs are te Laplace, te modfed Helmoltz and te Helmoltz equatons. Te Laplace equaton s u xx + u yy =. (5.) Te real and magnary parts of an
More informationbounds, but Mao [{4] only dscussed te mean square (te case of p = ) almost sure exponental stablty. Due to te new tecnques developed n ts paper, te re
Asymptotc Propertes of Neutral Stocastc Derental Delay Equatons Xuerong Mao Department of Statstcs Modellng Scence Unversty of Stratclyde Glasgow G XH, Scotl, U.K. Abstract: Ts paper dscusses asymptotc
More informationThe Second Anti-Mathima on Game Theory
The Second Ant-Mathma on Game Theory Ath. Kehagas December 1 2006 1 Introducton In ths note we wll examne the noton of game equlbrum for three types of games 1. 2-player 2-acton zero-sum games 2. 2-player
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EQUATION
Advanced Mathematcal Models & Applcatons Vol.3, No.3, 2018, pp.215-222 ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EUATION
More informationNumerical Simulation of One-Dimensional Wave Equation by Non-Polynomial Quintic Spline
IOSR Journal of Matematcs (IOSR-JM) e-issn: 78-578, p-issn: 319-765X. Volume 14, Issue 6 Ver. I (Nov - Dec 018), PP 6-30 www.osrournals.org Numercal Smulaton of One-Dmensonal Wave Equaton by Non-Polynomal
More informationMarginal Models for categorical data.
Margnal Models for categorcal data Applcaton to condtonal ndependence and graphcal models Wcher Bergsma 1 Marcel Croon 2 Jacques Hagenaars 2 Tamas Rudas 3 1 London School of Economcs and Poltcal Scence
More informationn α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0
MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector
More informationAdaptive Kernel Estimation of the Conditional Quantiles
Internatonal Journal of Statstcs and Probablty; Vol. 5, No. ; 206 ISSN 927-7032 E-ISSN 927-7040 Publsed by Canadan Center of Scence and Educaton Adaptve Kernel Estmaton of te Condtonal Quantles Rad B.
More informationA Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach
A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationIntroduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:
CE304, Sprng 2004 Lecture 4 Introducton to Vapor/Lqud Equlbrum, part 2 Raoult s Law: The smplest model that allows us do VLE calculatons s obtaned when we assume that the vapor phase s an deal gas, and
More information338 A^VÇÚO 1n ò Lke n Mancn (211), we make te followng assumpton to control te beavour of small jumps. Assumpton 1.1 L s symmetrc α-stable, were α (,
A^VÇÚO 1n ò 1oÏ 215c8 Cnese Journal of Appled Probablty and Statstcs Vol.31 No.4 Aug. 215 Te Speed of Convergence of te Tresold Verson of Bpower Varaton for Semmartngales Xao Xaoyong Yn Hongwe (Department
More informationInformation Geometry of Gibbs Sampler
Informaton Geometry of Gbbs Sampler Kazuya Takabatake Neuroscence Research Insttute AIST Central 2, Umezono 1-1-1, Tsukuba JAPAN 305-8568 k.takabatake@ast.go.jp Abstract: - Ths paper shows some nformaton
More informationCENTROID (AĞIRLIK MERKEZİ )
CENTOD (ĞLK MEKEZİ ) centrod s a geometrcal concept arsng from parallel forces. Tus, onl parallel forces possess a centrod. Centrod s tougt of as te pont were te wole wegt of a pscal od or sstem of partcles
More informationShuai Dong. Isaac Newton. Gottfried Leibniz
Computatonal pyscs Sua Dong Isaac Newton Gottred Lebnz Numercal calculus poston dervatve ntegral v velocty dervatve ntegral a acceleraton Numercal calculus Numercal derentaton Numercal ntegraton Roots
More informationPHYS 705: Classical Mechanics. Calculus of Variations II
1 PHYS 705: Classcal Mechancs Calculus of Varatons II 2 Calculus of Varatons: Generalzaton (no constrant yet) Suppose now that F depends on several dependent varables : We need to fnd such that has a statonary
More informationCS : Algorithms and Uncertainty Lecture 14 Date: October 17, 2016
CS 294-128: Algorthms and Uncertanty Lecture 14 Date: October 17, 2016 Instructor: Nkhl Bansal Scrbe: Antares Chen 1 Introducton In ths lecture, we revew results regardng follow the regularzed leader (FTRL.
More informationxp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ
CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and
More informationEGR 544 Communication Theory
EGR 544 Communcaton Theory. Informaton Sources Z. Alyazcoglu Electrcal and Computer Engneerng Department Cal Poly Pomona Introducton Informaton Source x n Informaton sources Analog sources Dscrete sources
More information8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS
SECTION 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS 493 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS All the vector spaces you have studed thus far n the text are real vector spaces because the scalars
More informationThe Karush-Kuhn-Tucker. Nuno Vasconcelos ECE Department, UCSD
e Karus-Kun-ucker condtons and dualt Nuno Vasconcelos ECE Department, UCSD Optmzaton goal: nd mamum or mnmum o a uncton Denton: gven unctons, g, 1,...,k and, 1,...m dened on some doman Ω R n mn w, w Ω
More informationPerron Vectors of an Irreducible Nonnegative Interval Matrix
Perron Vectors of an Irreducble Nonnegatve Interval Matrx Jr Rohn August 4 2005 Abstract As s well known an rreducble nonnegatve matrx possesses a unquely determned Perron vector. As the man result of
More informationMultigrid Methods and Applications in CFD
Multgrd Metods and Applcatons n CFD Mcael Wurst 0 May 009 Contents Introducton Typcal desgn of CFD solvers 3 Basc metods and ter propertes for solvng lnear systems of equatons 4 Geometrc Multgrd 3 5 Algebrac
More informationProbabilistic & Unsupervised Learning
Probablstc & Unsupervsed Learnng Convex Algorthms n Approxmate Inference Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computatonal Neuroscence Unt Unversty College London Term 1, Autumn 2008 Convexty A convex
More informationLecture 3: Probability Distributions
Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the
More informationLecture 3: Shannon s Theorem
CSE 533: Error-Correctng Codes (Autumn 006 Lecture 3: Shannon s Theorem October 9, 006 Lecturer: Venkatesan Guruswam Scrbe: Wdad Machmouch 1 Communcaton Model The communcaton model we are usng conssts
More informationCENTROID (AĞIRLIK MERKEZİ )
CENTOD (ĞLK MEKEZİ ) centrod s a geometrcal concept arsng from parallel forces. Tus, onl parallel forces possess a centrod. Centrod s tougt of as te pont were te wole wegt of a pscal od or sstem of partcles
More informationECE559VV Project Report
ECE559VV Project Report (Supplementary Notes Loc Xuan Bu I. MAX SUM-RATE SCHEDULING: THE UPLINK CASE We have seen (n the presentaton that, for downlnk (broadcast channels, the strategy maxmzng the sum-rate
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.65/15.070J Fall 013 Lecture 1 10/1/013 Martngale Concentraton Inequaltes and Applcatons Content. 1. Exponental concentraton for martngales wth bounded ncrements.
More informationCHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE
CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE Analytcal soluton s usually not possble when exctaton vares arbtrarly wth tme or f the system s nonlnear. Such problems can be solved by numercal tmesteppng
More informationCase A. P k = Ni ( 2L i k 1 ) + (# big cells) 10d 2 P k.
THE CELLULAR METHOD In ths lecture, we ntroduce the cellular method as an approach to ncdence geometry theorems lke the Szemeréd-Trotter theorem. The method was ntroduced n the paper Combnatoral complexty
More informationLOGIT ANALYSIS. A.K. VASISHT Indian Agricultural Statistics Research Institute, Library Avenue, New Delhi
LOGIT ANALYSIS A.K. VASISHT Indan Agrcultural Statstcs Research Insttute, Lbrary Avenue, New Delh-0 02 amtvassht@asr.res.n. Introducton In dummy regresson varable models, t s assumed mplctly that the dependent
More informationHidden Markov Models
Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,
More informationSnce h( q^; q) = hq ~ and h( p^ ; p) = hp, one can wrte ~ h hq hp = hq ~hp ~ (7) the uncertanty relaton for an arbtrary state. The states that mnmze t
8.5: Many-body phenomena n condensed matter and atomc physcs Last moded: September, 003 Lecture. Squeezed States In ths lecture we shall contnue the dscusson of coherent states, focusng on ther propertes
More informationLecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.
prnceton u. sp 02 cos 598B: algorthms and complexty Lecture 20: Lft and Project, SDP Dualty Lecturer: Sanjeev Arora Scrbe:Yury Makarychev Today we wll study the Lft and Project method. Then we wll prove
More informationMath 324 Advanced Financial Mathematics Spring 2008 Final Exam Solutions May 2, 2008
Mat 324 Advanced Fnancal Matematcs Sprng 28 Fnal Exam Solutons May 2, 28 Ts s an open book take-ome exam. You may work wt textbooks and notes but do not consult any oter person. Sow all of your work and
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationTHE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY. William A. Pearlman. References: S. Arimoto - IEEE Trans. Inform. Thy., Jan.
THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY Wllam A. Pearlman 2002 References: S. Armoto - IEEE Trans. Inform. Thy., Jan. 1972 R. Blahut - IEEE Trans. Inform. Thy., July 1972 Recall
More informationLinear discriminants. Nuno Vasconcelos ECE Department, UCSD
Lnear dscrmnants Nuno Vasconcelos ECE Department UCSD Classfcaton a classfcaton problem as to tpes of varables e.g. X - vector of observatons features n te orld Y - state class of te orld X R 2 fever blood
More informationSolutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.
Solutons HW #2 Dual of general LP. Fnd the dual functon of the LP mnmze subject to c T x Gx h Ax = b. Gve the dual problem, and make the mplct equalty constrants explct. Soluton. 1. The Lagrangan s L(x,
More informationOn the Correlation between Boolean Functions of Sequences of Random Variables
On the Correlaton between Boolean Functons of Sequences of Random Varables Farhad Shran Chaharsoogh Electrcal Engneerng and Computer Scence Unversty of Mchgan Ann Arbor, Mchgan, 48105 Emal: fshran@umch.edu
More informationModelli Clamfim Equazione del Calore Lezione ottobre 2014
CLAMFIM Bologna Modell 1 @ Clamfm Equazone del Calore Lezone 17 15 ottobre 2014 professor Danele Rtell danele.rtell@unbo.t 1/24? Convoluton The convoluton of two functons g(t) and f(t) s the functon (g
More informationOn mutual information estimation for mixed-pair random variables
On mutual nformaton estmaton for mxed-par random varables November 3, 218 Aleksandr Beknazaryan, Xn Dang and Haln Sang 1 Department of Mathematcs, The Unversty of Msssspp, Unversty, MS 38677, USA. E-mal:
More informationLecture 4: September 12
36-755: Advanced Statstcal Theory Fall 016 Lecture 4: September 1 Lecturer: Alessandro Rnaldo Scrbe: Xao Hu Ta Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer: These notes have not been
More informationCanonical transformations
Canoncal transformatons November 23, 2014 Recall that we have defned a symplectc transformaton to be any lnear transformaton M A B leavng the symplectc form nvarant, Ω AB M A CM B DΩ CD Coordnate transformatons,
More informationAssortment Optimization under MNL
Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.
More informationNUMERICAL DIFFERENTIATION
NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the
More informationDuality and Auxiliary Functions for Bregman Distances
Dualty and Auxlary Functons for Bregman Dstances Stephen Della Petra, Vncent Della Petra, and John Lafferty October 8, 2001 CMU-CS-01-109 School of Computer Scence Carnege Mellon Unversty Pttsburgh, PA
More information