On mutual information estimation for mixed-pair random variables

Size: px
Start display at page:

Download "On mutual information estimation for mixed-pair random variables"

Transcription

1 On mutual nformaton estmaton for mxed-par random varables November 3, 218 Aleksandr Beknazaryan, Xn Dang and Haln Sang 1 Department of Mathematcs, The Unversty of Msssspp, Unversty, MS 38677, USA. E-mal: abeknaza@olemss.edu, xdang@olemss.edu, sang@olemss.edu Abstract We study the mutual nformaton estmaton for mxed-par random varables. One random varable s dscrete and the other one s contnuous. We develop a kernel method to estmate the mutual nformaton between the two random varables. The estmates enjoy a central lmt theorem under some regular condtons on the dstrbutons. The theoretcal results are demonstrated by smulaton study. Keywords: central lmt theorem, entropy, kernel estmaton, mxed-par, mutual nformaton. MSC 21 subject classfcaton: 62G5, 62G2 1 Introducton The entropy of a dscrete random varable X wth countable support {x 1, x 2,...} and p = P(X = x ) s defned to be H(X) = p log p, and the (dfferental) entropy of a contnuous random varable Y wth probablty densty functon f(y) s defned as H(Y ) = f(y) log f(y)dy. If d 2, H(X) or H(Y ) s also called the jont entropy of the components n X or Y. Entropy s a measure of dstrbuton uncertanty and naturally t has applcaton n the felds of nformaton theory, statstcal classfcaton, pattern recognton and so on. Let P X, P Y be probablty measures on some arbtrary measure spaces X and Y respectvely. Let P XY be the jont probablty measure on the space X Y. If P XY s absolutely contnuous dp wth respect to the product measure P X P Y, let XY d(p X P Y ) be the Radon-Nkodym dervatve. Then the general defnton of the mutual nformaton (e.g., [3]) s gven by dp XY I(X, Y ) = dp XY log d(p X P Y ). (1) 1 Correspondng author. X Y 1

2 If two random varables X and Y are ether both dscrete or both contnuous then the mutual nformaton of X and Y can be expressed n terms of entropes as I(X, Y ) = H(X) + H(Y ) H(X, Y ). (2) However, n practce and applcaton, we often need to work on a mxture of contnuous and dscrete random varables. There are several ways for the mxture. 1). One random varable X s dscrete and the other random varable Y s contnuous; 2). A random varable Z has both dscrete and contnuous components,.e., Z = X wth probablty p and Z = Y wth probablty 1 p, where < p < 1, X s a dscrete random varable and Y s a contnuous random varable; 3). a random vector wth each dmenson component beng dscrete, contnuous or mxture as n 2). In [11], the authors extend the defnton of the jont entropy for the frst case mxture,.e., for the par of random varables, where the frst random varable s dscrete and the second one s contnuous. Our goal s to study the mutual nformaton for that case and provde the estmaton of the mutual nformaton from a gven..d. sample {X, Y } N =1. In [3], the authors appled the k-nearest neghbor method to estmate the Radon-Nkodym dervatve and, therefore, to estmate the mutual nformaton for all three mxed cases. In the lterature, f the random varables X and Y are ether both dscrete or both contnuous, the estmaton of mutual nformaton s usually performed by the estmaton of the three entropes n (2). The estmaton of a dfferental entropy has been well studed. An ncomplete lst of the related research ncludes the nearest-neghbor estmator [7], [12], [1]; the kernel estmator [1], [6], [4], [5] and the orthogonal projecton estmator [8], [9]. Basharn [2] studed the plug-n entropy estmator for the fnte value dscrete case and obtaned the mean, the varance and the central lmt theorem of ths estmator. Vu, Yu and Kass [13] studed the coverage-adjusted entropy estmator wth unobserved values for the nfnte value dscrete case. 2 Man results Consder a random vector Z = (X, Y ). We call Z a mxed-par f X R s a dscrete random varable wth countable support X = {x 1, x 2,...} whle Y s a contnuous random varable. Observe that Z = (X, Y ) nduces measures {µ 1, µ 2, } that are absolutely contnuous wth respect to the Lebesgue measure, where µ (A) = P(X = x, Y A), for every Borel set A n. There exsts a non-negatve functon g(x, y) wth h(x) := g(x, y)dy be the probablty mass functon on X and f(y) := g (y) be the margnal densty functon of Y. Here, g (y) = g(x, y), N. In partcular, denote p = h(x ), N. We have that f (y) = 1 p g (y) s the probablty densty functon of Y condtoned on X = x. In [11], the authors gave the followng regulaton of mxed-par and then defned the jont entropy of a mxed-par. Defnton 2.1 (Good mxed-par). A mxed-par random varables Z = (X, Y ) s called good f the followng condton s satsfed: g(x, y) log g(x, y) dxdy = g (y) log g (y) dy <. X Essentally, we have a good mxed-par random varables when restrcted to any of the X values, the condtonal dfferental entropy of Y s well-defned. 2

3 Defnton 2.2 (Entropy of a mxed-par). The entropy of a good mxed-par random varable s defned by H(Z) = g(x, y) log g(x, y)dxdy = g (y) log g (y)dy. X As g (y) = p f (y) then we have that H(Z) = g (y) log g (y)dy = p f (y) log p f (y)dy = p log p f (y)dy p f (y) log f (y)dy (3) = p log p p f (y) log f (y)dy = H(X) + p H(Y X = x ). We take the conventon log = and log / =. From the general formula of the mutual nformaton (1), we get that g(x, y)dxdy I(X, Y ) = g(x, y) log X h(x)f(y)dxdy dxdy = g (y) log g (y) p f(y) dy = g (y) log g (y)dy g (y) log p dy g (y) log f(y)dy = p f (y) log[p f (y)]dy (4) p log p f (y)dy f(y) log f(y)dy = p log p f (y)dy + p f (y) log f (y)dy p log p f(y) log f(y)dy = H(Z) + H(X) + H(Y ) = H(Y ) p H(Y X = x ) := H(Y ) I. Let (X, Y ), (X 1, Y 1 ),..., (X N, Y N ) be a random sample drawn from a mxed dstrbuton wth dscrete component havng support {, 1,, m}, and let p = P(X = ), m wth < p < 1, p = 1. Also suppose that the contnuous component has pdf f(y). Denote ˆp = N I(X k = )/N, m, and let Ī = ˆp [ N ˆp ] 1 N I(X k = ) log f (Y k ) N = N 1 I(X k = ) log f (Y k ) (5) 3

4 and H(Y ) = N 1 N log f(y k ) (6) be the estmators of I = p H(Y X = ), m, and H(Y ) respectvely, where f (y) s the probablty densty functon of Y condtoned on X =, m. Denote a = (1, 1,, 1). Let Σ be the covarance matrx of (log f(y ), I(X = ) log f (Y ),, I(X = m) log f m (Y )). Theorem 2.1 a Σa > f and only f X and Y are dependent. For the estmator Ī(X, Y ) = H Ī (7) = of I(X, Y ) we have that N( Ī(X, Y ) I(X, Y )) N(, a Σa) (8) gven that X and Y are dependent. Furthermore, the varance a Σa can be calculated by a Σa = var ( log f(y ) ) p E [log f (Y )] 2 = = p [E log f (Y ) log f(y ) E log f (Y )E log f(y )] = <j m p p j [E log f (Y )][E j log f j (Y )], where E s the condtonal expectaton of Y gven X =, m. p 2 ( E [log f (Y )] ) 2 (9) Proof. Frst of all, a Σa snce Σ s the varance covarance matrx. If a Σa = then ( ) var log f(y ) I(X = ) log f (Y ) = a Σa = and log f(y ) m = I(X = ) log f (Y ) C for some constant C. But log f(y ) = I(X = ) log f (Y ) = = = I(X = ) log f(y ) f (Y ). Hence log f(y ) f (Y ) C. Then f (y) = cf(y) for some constant c > and for all m. But f(y) = m = p f (y) = cf(y) m = p = cf(y). Hence, c 1 and f (y) = f(y) for all m. Then X and Y are ndependent. On the other hand, f X and Y are ndependent, then f (y) = f(y) for all m. Therefore, log f(y ) m = I(X = ) log f (Y ) = and a Σa =. Hence, a Σa = f and only f X and Y are ndependent. Notce that the vector ( H(Y ), Ī,, Īm) s the sample mean of a sequence of..d. random vectors {(log f(y k ), I(X k = ) log f (Y k ),, I(X k = m) log f m (Y k )) } N wth mean (H(Y ), I,, I m ). Then, by central lmt theorem, we have H H Ī N. I. N(, Σ), Ī m I m 4

5 and, gven a Σa >, we have (8). By the formula for varance decomposton, we have var ( I(X = ) log f (Y ) ) = E { var[i(x = ) log f (Y ) X] } + var { E[I(X = ) log f (Y ) X] } = E { I(X = )var[log f (Y ) X] } + var { I(X = )E[log f (Y ) X] } = E { I(X = ) var j (log f j (Y ))I(X = j) } j= + var { I(X = ) E j (log f j (Y ))I(X = j) } j= = var [log f (Y )]E { I(X = ) } + ( E [log f (Y )] ) 2 var { I(X = ) } = p var [log f (Y )] + (p p 2 ) ( E [log f (Y )] ) 2 = p E [log f (Y )] 2 p 2 ( E [log f (Y )] ) 2, (1) m. Here var s the condtonal varance of Y when X =, m. By smlar calculaton, ( ) Cov I(X = ) log f (Y ), I(X = j) log f j (Y ) (11) = p p j [E log f (Y )][E j log f j (Y )], for all < j m, and ( ) Cov I(X = ) log f (Y ), log f(y ) = p [E log f (Y ) log f(y ) E log f (Y )E log f(y )]. (12) Thus, the covarance matrx Σ of (log f(y ), I(X = ) log f (Y ),, I(X = m) log f m (Y )) and therefore a Σa can be calculated by the above calculaton (1)-(12). We then have (9). We consder the case when the random varables X and Y are dependent. Note that n ths case a Σa > and we have (8). However, Ī(X, Y ) s not a practcal estmator snce the densty functons nvolved are not known. Now let K( ) be a kernel functon n and let h be the bandwdth. Then ˆf k (y) = { } 1 (N ˆp 1)h d I(X j = )K{(y Y j )/h} j k are the leave-one-out estmators of the functons f, m, and Î = N 1 N are estmators of I = p H(Y X = ), m. Also Ĥ = N 1 I(X k = ) log ˆf k (Y k ) (13) N log ˆf k (Y k ) (14) 5

6 s an estmator of H(Y ), where ˆf k (y) = { (N 1)h d } 1 j k K{(y Y j )/h} = = { (N 1)h d } 1 j k[ = N ˆp 1 N 1 ˆf k (y). I(X k = )]K{(y Y j )/h} = (15) Theorem 2.2 Assume that the tals of f,, f m are decreasng lke x α,, x αm, respectvely, as x. Also assume that the kernel functon has approprately heavy tals as n [4]. If h = o(n 1/8 ) and α, α m are all greater than 7/3 n the case d = 1, greater than 6 n the case d = 2 and greater than 15 n the case d = 3, then for the estmator m Î(X, Y ) = Ĥ Î, (16) = we have N( Î(X, Y ) I(X, Y )) N(, a Σa). (17) Proof. Under the condtons n the theorem, applyng the formula (3.1) or (3.2) from [5], we have Ĥ = H + o(n 1/2 ), Î = Ī + o(n 1/2 ),, Î m = Īm + o(n 1/2 ). Together wth Theorem 2.1, we have (17). We may take the probablty densty functon of Student-t dstrbuton wth proper degree of freedom nstead of the normal densty functon as the kernel functon. On the other hand, f X and Y are ndependent then I(X, Y ) = Ī(X, Y ) = and we have that Î(X, Y ) = o(n 1/2 ). 3 Smulaton study In ths secton we conduct a smulaton study wth m = 1,.e., the random varable X takes two possble values and 1, to confrm the man results stated n (17) for the kernel mutual nformaton estmaton of good mxed-pars. Frst we study some one dmensonal examples. Let t(ν, µ, σ) be the Student t dstrbuton wth degree of freedom ν, locaton parameter µ and scale parameter σ and let pareto(x m, α) be the Pareto dstrbuton wth densty functon f(x) = αx α mx (α+1) I(x x m ). We study the mxture for the followng four cases: 1). t(3,, 1) and t(12,, 1); 2). t(3,, 1) and t(3, 2, 1); 3). t(3,, 1) and t(3,, 3); 4). pareto(1, 2) and pareto(1, 1). For each case, p =.3 for the frst dstrbuton and p 1 =.7 for the second dstrbuton. The second row of Table 1 lsts the mathematca calculaton of the mutual nformaton (MI) as stated n (4) for each case. The thrd row of Table 1 gves the average of 4 estmates based on formula (16). For each estmate, we use the probablty densty functon of the Student t dstrbuton wth degree of freedom 3,.e. t(3,, 1), as the kernel functon. We also have smulaton study wth kernel functons satsfyng the condtons n the man results and obtaned smlar results. We take h = N 1/5 as the bandwdth for the frst three cases and h = N 1/5 /24 for the last case. The data sze for each estmate s N = 5, n each case. The Pareto dstrbutons pareto(1, 2) and pareto(1, 1) have very dense area on the rght of 1. Ths s the reason that we take a relatvely small bandwdth for ths case. To apply the kernel method n estmaton, one should 6

7 select an optmal bandwdth based on some crtera, for example, to mnmze the mean squared error. It s nterestng to nvestgate the bandwdth selecton problem from both theoretcal and applcaton vewponts. However, t seems that the study n ths drecton s very dffcult. We leave t as an open queston for future study. It s clear that the average of the estmates matches the true value of mutual nformaton. We apply mathematca to calculate the covarance matrx Σ of (log f(y ), I(X = ) log f (Y ), I(X = 1) log f 1 (Y )) and, therefore, the value of a Σa for each case by formulae (1)-(12) or (9). The values of a Σa are , , and respectvely for the four cases. The fourth row of Table 1 lsts the values of (a Σa/N) 1/2 whch serves as the asymptotc approxmaton of the standard devaton of the estmator Î(X, Y ) n the central lmt theorem (17). The last row gves the sample standard devaton from M = 4 estmates. These two values also have good match. mxture t(3,, 1) t(3,, 1) t(3,, 1) pareto(1, 2) t(12,, 1) t(3, 2, 1) t(3,, 3) pareto(1, 1) MI mean of estmates (a Σa/N) 1/ sample sd Table 1: True value of the mutual nformaton and the mean value of the estmates Fgure 1: The hstograms wth kernel densty fts of M = 4 estmates. Top left: t(3,, 1) and t(12,, 1). Top rght: t(3,, 1) and t(3, 2, 1). Bottom left: t(3,, 1) and t(3,, 3). Bottom rght: pareto(1, 2) and pareto(1, 1). 7

8 Quantles of Input Sample Quantles of Input Sample Quantles of Input Sample Quantles of Input Sample Standard Normal Quantles Standard Normal Quantles Standard Normal Quantles Standard Normal Quantles Fgure 2: The Q-Q plots of M = 4 estmates. Top left: t(3,, 1) and t(12,, 1). Top rght: t(3,, 1) and t(3, 2, 1). Bottom left: t(3,, 1) and t(3,, 3). Bottom rght: pareto(1, 2) and pareto(1, 1). Fgure 1 and 2 show the hstograms wth kernel densty fts and normal Q-Q plots of 4 estmates for each case. It s clear that the values of Î(X, Y ) follow a normal dstrbuton. We study two examples n the two dmensonal case. Let t ν (µ, Σ ) be the two dmensonal Student t dstrbuton wth degree of freedom ν, mean µ and shape matrx Σ. We study the mxture n two cases: 1). t 5 (, I) and t 25 (, I); 2). t 5 (, I) and t 5 (, 3I). Here I s the dentty matrx. For each case, p =.3 for the frst dstrbuton and p 1 =.7 for the second dstrbuton. Table 2 summarzes 2 estmates of the mutual nformaton wth h = N 1/5 and sample sze N = 5, for each estmate. We take t 3 (, I) as the kernel functon. Same as the one dmensonal case, we apply mathematca to calculate the true value of MI and (a Σa/N) 1/2 whch s gven n formula (9). Fgure 3 shows the hstograms wth kernel densty fts and normal Q-Q plots of 2 estmates for each example. It s clear that the values of Î(X, Y ) also follow a normal dstrbuton n the two dmensonal case. In summary, the smulaton study confrms the central lmt theorem as stated n (17). 8

9 Quantles of Input Sample Quantles of Input Sample mxture t 5 (, I) t 5 (, I) t 25 (, I) t 5 (, 3I) MI mean of estmates (a Σa/N) 1/ sample sd Table 2: True value of the mutual nformaton and the mean value of the estmates Standard Normal Quantles Standard Normal Quantles Fgure 3: The hstograms and Q-Q plots of M = 2 estmates. Left: t 5 (, I) and t 25 (, I). Rght: t 5 (, I) and t 5 (, 3I). Acknowledgement The authors thank the edtor and the referees for carefully readng the manuscrpt and for the suggestons that mproved the presentaton. Ths research s supported by the College of Lberal Arts Faculty Grants for Research and Creatve Achevement at the Unversty of Msssspp. The research of Haln Sang s also supported by the Smons Foundaton Grant References [1] Ahmad, I. A. and Ln, P. E A nonparametrc estmaton of the entropy for absolutely contnuous dstrbutons. IEEE Trans. Informaton Theory. 22, [2] Basharn, G. P On a statstcal estmate for the entropy of a sequence of ndependent random varables. Theory of Probablty and Its Applcatons. 4,

10 [3] Gao, W., Kannan, S., Oh, S. and Vswanath, P Estmatng mutual nformaton for dscrete-contnuous mxtures. Advances n Neural Informaton Processng Systems [4] Hall, P On Kullback-Lebler Loss and Densty Estmaton. Ann. Statst. 15, no. 4, [5] Hall, P. and Morton, S On the estmaton of entropy. Ann. Inst. Statst. Math. 45, [6] Joe, H On the estmaton of entropy and other functonals of a multvarate densty. Ann. Inst. Statst. Math. 41, [7] Kozachenko, L. F. and Leonenko, N. N Sample estmate of entropy of a random vector. Problems of Informaton Transmsson, 23, [8] Laurent, B Effcent estmaton of ntegral functonals of a densty. Ann. Statst. 24, [9] Laurent, B Estmaton of ntegral functonals of a densty and ts dervatves. Bernoull 3, [1] Leonenko, N., Pronzato, L. and Savan, V. 28. A class of Rény nformaton estmators for multdmensonal denstes. Ann. Statst. 36, Correctons, Ann. Statst. 38 (21), [11] Nar, C., Prabhakar, B. and Shah, D. On entropy for mxtures of dscrete and contnuous varables. arxv:cs/6775 [12] Tsybakov, A. B. and van der Meulen, E. C Root-n consstent estmators of entropy for denstes wth unbounded support. Scand. J. Statst., 23, [13] Vu, V. Q., Yu, B. and Kass, R. E. 27. Coverage-adjusted entropy estmaton. Statst. Med., 26,

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

EGR 544 Communication Theory

EGR 544 Communication Theory EGR 544 Communcaton Theory. Informaton Sources Z. Alyazcoglu Electrcal and Computer Engneerng Department Cal Poly Pomona Introducton Informaton Source x n Informaton sources Analog sources Dscrete sources

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Multiple Choice. Choose the one that best completes the statement or answers the question.

Multiple Choice. Choose the one that best completes the statement or answers the question. ECON 56 Homework Multple Choce Choose the one that best completes the statement or answers the queston ) The probablty of an event A or B (Pr(A or B)) to occur equals a Pr(A) Pr(B) b Pr(A) + Pr(B) f A

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Google PageRank with Stochastic Matrix

Google PageRank with Stochastic Matrix Google PageRank wth Stochastc Matrx Md. Sharq, Puranjt Sanyal, Samk Mtra (M.Sc. Applcatons of Mathematcs) Dscrete Tme Markov Chan Let S be a countable set (usually S s a subset of Z or Z d or R or R d

More information

Probability Theory (revisited)

Probability Theory (revisited) Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted

More information

Another converse of Jensen s inequality

Another converse of Jensen s inequality Another converse of Jensen s nequalty Slavko Smc Abstract. We gve the best possble global bounds for a form of dscrete Jensen s nequalty. By some examples ts frutfulness s shown. 1. Introducton Throughout

More information

Estimation: Part 2. Chapter GREG estimation

Estimation: Part 2. Chapter GREG estimation Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the

More information

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES SVANTE JANSON Abstract. We gve explct bounds for the tal probabltes for sums of ndependent geometrc or exponental varables, possbly wth dfferent

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

The Order Relation and Trace Inequalities for. Hermitian Operators

The Order Relation and Trace Inequalities for. Hermitian Operators Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

Goodness of fit and Wilks theorem

Goodness of fit and Wilks theorem DRAFT 0.0 Glen Cowan 3 June, 2013 Goodness of ft and Wlks theorem Suppose we model data y wth a lkelhood L(µ) that depends on a set of N parameters µ = (µ 1,...,µ N ). Defne the statstc t µ ln L(µ) L(ˆµ),

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models

More information

Modelli Clamfim Equazione del Calore Lezione ottobre 2014

Modelli Clamfim Equazione del Calore Lezione ottobre 2014 CLAMFIM Bologna Modell 1 @ Clamfm Equazone del Calore Lezone 17 15 ottobre 2014 professor Danele Rtell danele.rtell@unbo.t 1/24? Convoluton The convoluton of two functons g(t) and f(t) s the functon (g

More information

Appendix B. Criterion of Riemann-Stieltjes Integrability

Appendix B. Criterion of Riemann-Stieltjes Integrability Appendx B. Crteron of Remann-Steltes Integrablty Ths note s complementary to [R, Ch. 6] and [T, Sec. 3.5]. The man result of ths note s Theorem B.3, whch provdes the necessary and suffcent condtons for

More information

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X

More information

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth

More information

Quantum and Classical Information Theory with Disentropy

Quantum and Classical Information Theory with Disentropy Quantum and Classcal Informaton Theory wth Dsentropy R V Ramos rubensramos@ufcbr Lab of Quantum Informaton Technology, Department of Telenformatc Engneerng Federal Unversty of Ceara - DETI/UFC, CP 6007

More information

Singular Value Decomposition: Theory and Applications

Singular Value Decomposition: Theory and Applications Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real

More information

ISQS 6348 Final Open notes, no books. Points out of 100 in parentheses. Y 1 ε 2

ISQS 6348 Final Open notes, no books. Points out of 100 in parentheses. Y 1 ε 2 ISQS 6348 Fnal Open notes, no books. Ponts out of 100 n parentheses. 1. The followng path dagram s gven: ε 1 Y 1 ε F Y 1.A. (10) Wrte down the usual model and assumptons that are mpled by ths dagram. Soluton:

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

A Note on Bound for Jensen-Shannon Divergence by Jeffreys

A Note on Bound for Jensen-Shannon Divergence by Jeffreys OPEN ACCESS Conference Proceedngs Paper Entropy www.scforum.net/conference/ecea- A Note on Bound for Jensen-Shannon Dvergence by Jeffreys Takuya Yamano, * Department of Mathematcs and Physcs, Faculty of

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Mamum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models for

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors Stat60: Bayesan Modelng and Inference Lecture Date: February, 00 Reference Prors Lecturer: Mchael I. Jordan Scrbe: Steven Troxler and Wayne Lee In ths lecture, we assume that θ R; n hgher-dmensons, reference

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

Maximizing the number of nonnegative subsets

Maximizing the number of nonnegative subsets Maxmzng the number of nonnegatve subsets Noga Alon Hao Huang December 1, 213 Abstract Gven a set of n real numbers, f the sum of elements of every subset of sze larger than k s negatve, what s the maxmum

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

Random Partitions of Samples

Random Partitions of Samples Random Parttons of Samples Klaus Th. Hess Insttut für Mathematsche Stochastk Technsche Unverstät Dresden Abstract In the present paper we construct a decomposton of a sample nto a fnte number of subsamples

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.65/15.070J Fall 013 Lecture 1 10/1/013 Martngale Concentraton Inequaltes and Applcatons Content. 1. Exponental concentraton for martngales wth bounded ncrements.

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

Non-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT

Non-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT Malaysan Journal of Mathematcal Scences 8(S): 37-44 (2014) Specal Issue: Internatonal Conference on Mathematcal Scences and Statstcs 2013 (ICMSS2013) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES Journal

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

More metrics on cartesian products

More metrics on cartesian products More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of

More information

MATH 5707 HOMEWORK 4 SOLUTIONS 2. 2 i 2p i E(X i ) + E(Xi 2 ) ä i=1. i=1

MATH 5707 HOMEWORK 4 SOLUTIONS 2. 2 i 2p i E(X i ) + E(Xi 2 ) ä i=1. i=1 MATH 5707 HOMEWORK 4 SOLUTIONS CİHAN BAHRAN 1. Let v 1,..., v n R m, all lengths v are not larger than 1. Let p 1,..., p n [0, 1] be arbtrary and set w = p 1 v 1 + + p n v n. Then there exst ε 1,..., ε

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 where (x 1,..., x N ) X N, N s called the populaton sze, f(x) f (x) for at least one {1, 2,..., N}, and those dfferent from f(x) are called the tral dstrbutons n terms of mportance samplng. Dfferent ways

More information

Computing MLE Bias Empirically

Computing MLE Bias Empirically Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.

More information

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD Matrx Approxmaton va Samplng, Subspace Embeddng Lecturer: Anup Rao Scrbe: Rashth Sharma, Peng Zhang 0/01/016 1 Solvng Lnear Systems Usng SVD Two applcatons of SVD have been covered so far. Today we loo

More information

Expected Value and Variance

Expected Value and Variance MATH 38 Expected Value and Varance Dr. Neal, WKU We now shall dscuss how to fnd the average and standard devaton of a random varable X. Expected Value Defnton. The expected value (or average value, or

More information

Lecture 17 : Stochastic Processes II

Lecture 17 : Stochastic Processes II : Stochastc Processes II 1 Contnuous-tme stochastc process So far we have studed dscrete-tme stochastc processes. We studed the concept of Makov chans and martngales, tme seres analyss, and regresson analyss

More information

Inductance Calculation for Conductors of Arbitrary Shape

Inductance Calculation for Conductors of Arbitrary Shape CRYO/02/028 Aprl 5, 2002 Inductance Calculaton for Conductors of Arbtrary Shape L. Bottura Dstrbuton: Internal Summary In ths note we descrbe a method for the numercal calculaton of nductances among conductors

More information

Chapter 7 Channel Capacity and Coding

Chapter 7 Channel Capacity and Coding Wreless Informaton Transmsson System Lab. Chapter 7 Channel Capacty and Codng Insttute of Communcatons Engneerng atonal Sun Yat-sen Unversty Contents 7. Channel models and channel capacty 7.. Channel models

More information

Complete Convergence for Weighted Sums of Weakly Negative Dependent of Random Variables

Complete Convergence for Weighted Sums of Weakly Negative Dependent of Random Variables Flomat 30:12 (2016), 3177 3186 DOI 10.2298/FIL1612177N Publshed by Faculty of Scences and Mathematcs, Unversty of Nš, Serba Avalable at: http://www.pmf.n.ac.rs/flomat Complete Convergence for Weghted Sums

More information

On quasiperfect numbers

On quasiperfect numbers Notes on Number Theory and Dscrete Mathematcs Prnt ISSN 1310 5132, Onlne ISSN 2367 8275 Vol. 23, 2017, No. 3, 73 78 On quasperfect numbers V. Sva Rama Prasad 1 and C. Suntha 2 1 Nalla Malla Reddy Engneerng

More information

The lower and upper bounds on Perron root of nonnegative irreducible matrices

The lower and upper bounds on Perron root of nonnegative irreducible matrices Journal of Computatonal Appled Mathematcs 217 (2008) 259 267 wwwelsevercom/locate/cam The lower upper bounds on Perron root of nonnegatve rreducble matrces Guang-Xn Huang a,, Feng Yn b,keguo a a College

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

Asymptotics of the Solution of a Boundary Value. Problem for One-Characteristic Differential. Equation Degenerating into a Parabolic Equation

Asymptotics of the Solution of a Boundary Value. Problem for One-Characteristic Differential. Equation Degenerating into a Parabolic Equation Nonl. Analyss and Dfferental Equatons, ol., 4, no., 5 - HIKARI Ltd, www.m-har.com http://dx.do.org/.988/nade.4.456 Asymptotcs of the Soluton of a Boundary alue Problem for One-Characterstc Dfferental Equaton

More information

Convexity preserving interpolation by splines of arbitrary degree

Convexity preserving interpolation by splines of arbitrary degree Computer Scence Journal of Moldova, vol.18, no.1(52), 2010 Convexty preservng nterpolaton by splnes of arbtrary degree Igor Verlan Abstract In the present paper an algorthm of C 2 nterpolaton of dscrete

More information

Excess Error, Approximation Error, and Estimation Error

Excess Error, Approximation Error, and Estimation Error E0 370 Statstcal Learnng Theory Lecture 10 Sep 15, 011 Excess Error, Approxaton Error, and Estaton Error Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton So far, we have consdered the fnte saple

More information

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Solutions to exam in SF1811 Optimization, Jan 14, 2015 Solutons to exam n SF8 Optmzaton, Jan 4, 25 3 3 O------O -4 \ / \ / The network: \/ where all lnks go from left to rght. /\ / \ / \ 6 O------O -5 2 4.(a) Let x = ( x 3, x 4, x 23, x 24 ) T, where the varable

More information

Interval Estimation of Stress-Strength Reliability for a General Exponential Form Distribution with Different Unknown Parameters

Interval Estimation of Stress-Strength Reliability for a General Exponential Form Distribution with Different Unknown Parameters Internatonal Journal of Statstcs and Probablty; Vol. 6, No. 6; November 17 ISSN 197-73 E-ISSN 197-74 Publshed by Canadan Center of Scence and Educaton Interval Estmaton of Stress-Strength Relablty for

More information

Introduction to Random Variables

Introduction to Random Variables Introducton to Random Varables Defnton of random varable Defnton of random varable Dscrete and contnuous random varable Probablty functon Dstrbuton functon Densty functon Sometmes, t s not enough to descrbe

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics )

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics ) Ismor Fscher, 8//008 Stat 54 / -8.3 Summary Statstcs Measures of Center and Spread Dstrbuton of dscrete contnuous POPULATION Random Varable, numercal True center =??? True spread =???? parameters ( populaton

More information

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1 On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Error Probability for M Signals

Error Probability for M Signals Chapter 3 rror Probablty for M Sgnals In ths chapter we dscuss the error probablty n decdng whch of M sgnals was transmtted over an arbtrary channel. We assume the sgnals are represented by a set of orthonormal

More information

CS 468 Lecture 16: Isometry Invariance and Spectral Techniques

CS 468 Lecture 16: Isometry Invariance and Spectral Techniques CS 468 Lecture 16: Isometry Invarance and Spectral Technques Justn Solomon Scrbe: Evan Gawlk Introducton. In geometry processng, t s often desrable to characterze the shape of an object n a manner that

More information

3.1 ML and Empirical Distribution

3.1 ML and Empirical Distribution 67577 Intro. to Machne Learnng Fall semester, 2008/9 Lecture 3: Maxmum Lkelhood/ Maxmum Entropy Dualty Lecturer: Amnon Shashua Scrbe: Amnon Shashua 1 In the prevous lecture we defned the prncple of Maxmum

More information

Complete subgraphs in multipartite graphs

Complete subgraphs in multipartite graphs Complete subgraphs n multpartte graphs FLORIAN PFENDER Unverstät Rostock, Insttut für Mathematk D-18057 Rostock, Germany Floran.Pfender@un-rostock.de Abstract Turán s Theorem states that every graph G

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x

More information

Adaptive Kernel Estimation of the Conditional Quantiles

Adaptive Kernel Estimation of the Conditional Quantiles Internatonal Journal of Statstcs and Probablty; Vol. 5, No. ; 206 ISSN 927-7032 E-ISSN 927-7040 Publsed by Canadan Center of Scence and Educaton Adaptve Kernel Estmaton of te Condtonal Quantles Rad B.

More information

NP-Completeness : Proofs

NP-Completeness : Proofs NP-Completeness : Proofs Proof Methods A method to show a decson problem Π NP-complete s as follows. (1) Show Π NP. (2) Choose an NP-complete problem Π. (3) Show Π Π. A method to show an optmzaton problem

More information

Lecture 4: September 12

Lecture 4: September 12 36-755: Advanced Statstcal Theory Fall 016 Lecture 4: September 1 Lecturer: Alessandro Rnaldo Scrbe: Xao Hu Ta Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer: These notes have not been

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Neryškioji dichotominių testo klausimų ir socialinių rodiklių diferencijavimo savybių klasifikacija

Neryškioji dichotominių testo klausimų ir socialinių rodiklių diferencijavimo savybių klasifikacija Neryškoj dchotomnų testo klausmų r socalnų rodklų dferencjavmo savybų klasfkacja Aleksandras KRYLOVAS, Natalja KOSAREVA, Julja KARALIŪNAITĖ Technologcal and Economc Development of Economy Receved 9 May

More information

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013 ISSN: 2277-375 Constructon of Trend Free Run Orders for Orthogonal rrays Usng Codes bstract: Sometmes when the expermental runs are carred out n a tme order sequence, the response can depend on the run

More information

Small Area Interval Estimation

Small Area Interval Estimation .. Small Area Interval Estmaton Partha Lahr Jont Program n Survey Methodology Unversty of Maryland, College Park (Based on jont work wth Masayo Yoshmor, Former JPSM Vstng PhD Student and Research Fellow

More information

Testing for seasonal unit roots in heterogeneous panels

Testing for seasonal unit roots in heterogeneous panels Testng for seasonal unt roots n heterogeneous panels Jesus Otero * Facultad de Economía Unversdad del Rosaro, Colomba Jeremy Smth Department of Economcs Unversty of arwck Monca Gulett Aston Busness School

More information

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

Notes on Frequency Estimation in Data Streams

Notes on Frequency Estimation in Data Streams Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Statistics and Probability Theory in Civil, Surveying and Environmental Engineering

Statistics and Probability Theory in Civil, Surveying and Environmental Engineering Statstcs and Probablty Theory n Cvl, Surveyng and Envronmental Engneerng Pro. Dr. Mchael Havbro Faber ETH Zurch, Swtzerland Contents o Todays Lecture Overvew o Uncertanty Modelng Random Varables - propertes

More information

Learning Theory: Lecture Notes

Learning Theory: Lecture Notes Learnng Theory: Lecture Notes Lecturer: Kamalka Chaudhur Scrbe: Qush Wang October 27, 2012 1 The Agnostc PAC Model Recall that one of the constrants of the PAC model s that the data dstrbuton has to be

More information

Supplementary Notes for Chapter 9 Mixture Thermodynamics

Supplementary Notes for Chapter 9 Mixture Thermodynamics Supplementary Notes for Chapter 9 Mxture Thermodynamcs Key ponts Nne major topcs of Chapter 9 are revewed below: 1. Notaton and operatonal equatons for mxtures 2. PVTN EOSs for mxtures 3. General effects

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

A quantum-statistical-mechanical extension of Gaussian mixture model

A quantum-statistical-mechanical extension of Gaussian mixture model A quantum-statstcal-mechancal extenson of Gaussan mxture model Kazuyuk Tanaka, and Koj Tsuda 2 Graduate School of Informaton Scences, Tohoku Unversty, 6-3-09 Aramak-aza-aoba, Aoba-ku, Senda 980-8579, Japan

More information

CSCE 790S Background Results

CSCE 790S Background Results CSCE 790S Background Results Stephen A. Fenner September 8, 011 Abstract These results are background to the course CSCE 790S/CSCE 790B, Quantum Computaton and Informaton (Sprng 007 and Fall 011). Each

More information

Parameters Estimation of the Modified Weibull Distribution Based on Type I Censored Samples

Parameters Estimation of the Modified Weibull Distribution Based on Type I Censored Samples Appled Mathematcal Scences, Vol. 5, 011, no. 59, 899-917 Parameters Estmaton of the Modfed Webull Dstrbuton Based on Type I Censored Samples Soufane Gasm École Supereure des Scences et Technques de Tuns

More information

Convergence of random processes

Convergence of random processes DS-GA 12 Lecture notes 6 Fall 216 Convergence of random processes 1 Introducton In these notes we study convergence of dscrete random processes. Ths allows to characterze phenomena such as the law of large

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information