Bias Reduction when Data are Rounded

Size: px
Start display at page:

Download "Bias Reduction when Data are Rounded"

Transcription

1 Bias Reduction when Data are Rounded Christopher S. Withers & Saralees Nadarajah First version: 10 December 2013 Research Report No. 12, 2013, Probability and Statistics Group School of Mathematics, The University of Manchester

2 Bias reduction when data are rounded by Christopher S. Withers Applied Mathematics Group Industrial Research Limited Lower Hutt, NEW ZEALAND Saralees Nadarajah School of Mathematics University of Manchester Manchester M13 9PL, UK Abstract: Analytical bias reduction methods are developed for univariate rounded data for the first time. Extensions are given to rounding of multivariate data, and to smooth functionals of several distributions. As a by product, we give for the first time the relation between rounded and unrounded multivariate cumulants. Keywords: Bias reduction; Cumulants; Rounding. 1 Introduction There are many reasons and many applied areas for rounding of measurements. Some recent examples are: Measurements of continuous random variables are inevitably rounded to certain digits before or after the decimal point according to the precision of the measuring instruments or the recording/storage mechanism. For example, scientific measurements are often recorded to the nearest millimetres. The Consumer Price Index is commonly rounded to a single decimal place (Li and Bai, 2011); In general, measuring CD can be subjective, and such measurements are only estimates because observations are rounded to the nearest centimeter (Hoh et al., 2012); All observations are rounded to the nearest whole knot when recorded at the meteorological station, except that wind speeds below 1 kt are recorded as zero (Thorarinsdottir and Johnson, 2012); In practice, it is not uncommon that data are rounded. In the US, birth weights show ounces, for example 5 lb 3 oz, while in Europe, for example in Ukraine, intervals could be in 50 g or nearly 1.8 oz (Wang and Wertelecki, 2012). However, we are aware of no work in the literature considering bias reduction for rounded data. Bias reduction is a common method for providing accurate point as well as interval estimators. The aim of this paper to derive analytical bias reduction methods for univariate rounded data, multivariate rounded data and for multiple samples of rounded data. Analytical bias reduction methods have several advantages over other nonparametric methods for bias reduction like bootstrapping and jackknifing. Two main advantages are: (i) they take only O(n) calculations, while bootstrap estimators take O (n p ) calculations, p denotes the order of bias reduction and n denotes the sample size; (ii) they offer the advantage of analytic clarity as opposed to blind number crunching for a set of parameter 1

3 choices that may miss some features. Withers and Nadarajah (2008, 2012a) have developed analytic methods for nonparametric bias reduction that remove the need for computationally intensive methods like the bootstrap and the jackknife. They define an estimator as having pth order if its bias has magnitude n p 0 as n 0, n 0 is the sample size or the minimum sample size if the estimator is a function of more than one sample. For general p, Withers and Nadarajah (2008) have provided analytic pth order nonparametric estimators that require only O(N) calculations, N is the total sample size. These estimators are given in terms of the von Mises derivatives of the functional being estimated, evaluated at the empirical distribution function. For p 4, Withers and Nadarajah (2012a) have given explicit estimators in terms of the first 2p 2 von Mises derivatives of the functional evaluated at the empirical distribution functions. These may be used to obtain unbiased estimators (UEs) when these exist and are of known form in terms of the sample sizes. The analytic nonparametric bias reduction methods of Withers and Nadarajah (2008, 2012a) are for unrounded data. Here, we extend these methods for rounded data. We have written programs in the R statistical software (R Development Core Team, 2013) to implement the general methods of this paper for rounded data: Theorems 2.1 to 2.5 for bias reduction for univariate rounded data; Theorem 3.1 for bias reduction for multivariate rounded data. The electronic versions of these programs can be obtained from the corresponding author, mbbsssn2@manchester.ac.uk We have chosen the R statistical software since it is freely available and since it can be implemented in any platform. These programs are believed to be the first programs in R for any analytical bias reduction. The programs can be used to yield ready and fast analytic bias reduction. The approach will be as accurate and far less time consuming than using bootstrapping, jackknifing and other computer intensive methods. Let X 1,..., X n be a random sample of real univariate observations with empirical distribution function F n from a distribution function F (x) with finite derivatives, mean µ, and density f(x). The problem we consider is to find an estimator of low bias for any smooth functional T (F ) given a rounded sample from F. By Withers and Nadarajah (2008, 2012a), for p 1, an estimator of T (F ) is S np (F n ), In particular, p 1 S np (F, T ) = S np (F ) = S i (F )/[n 1] i, [n 1] i = (n 1) (n i). (1.1) i=0 S 0 (F ) = T (F ), S 1 (F ) = T [2]/2, S 2 (F ) = T [3]/3 + T [ 2 2] /8, (1.2) S 3 (F ) = T [4]/4 + 3T [ 2 2] /8 3T [32]/6 T [ 2 3] /48, S 4 (F ) = T [5]/5 2T [32]/3 3T [ 2 3] /16 + T [42]/8 + T [ 3 2] /18 +T [ 32 2] /24 + T [ 2 4] /384, T [r] = T F (x r ) df (x), T [rs] = T F (x r, y s ) df (x)df (y), T F (x 1,..., x r ) is the rth functional derivative of T (F ), T F ( x 2 ) = T F (x, x), T F ( x 3, y 2) = T F (x, x, x, y, y), T [ x 2] = T [xx], T [ x 3] = T [xxx], T [ xy 2] = T [xyy], and so on. S np (F n ) 2

4 is an UE if T (F ) is a polynomial of order p in F such as the product of moments or cumulants of total order p. Tchouproff (1918), Fisher (1929) and James (1958) gave UEs of this form for κ p and µ p. We can expand (1.1) in the form S np (F ) = T np (F ) + O ( n p), (1.3) p 1 T np (F ) = n i T i (F ), p 1. So, for p 1, T np (F n ) also estimates T (F ) with bias O (n p ). i=0 Now suppose that only the rounded values of the sample are recorded, say X h1,..., X hn, X hj is X j rounded to the nearest integral multiple of a given constant h > 0. Let F nh be their empirical distribution function, and let F h (x) be the distribution function of X h = X h1. Theorem 2.4 gives our main result for univariate data. We show that for any smooth functional T (F ), we have expansions of the form T (F h ) = T (F ) + T rh (F )/r! = r=1 T (F ) = T h (F h ), T h (F ) = T (F ) + H k T k (F ) = T h(f ) say, (1.4) Tr h (F )/r! = r=1 H k T k (F ), (1.5) H = h 2 /4. (For example, if T (F ) is a smooth function of (µ, µ 3 ), then T (F h ) = T (F ), that is, the expansions (1.4) and (1.5) for T (F h ) and T (F ) only have one term.) So, Snp h ( Fnh, T h) is an estimator of T (F ) based on rounded data with bias O (n p ). So, E [ H k Snp h ( Fnh, T k ) ] [ ( = E Snp h F nh, T h)] = T (F ) + O ( n p). This and its multivariate and multisample extensions in Sections 3 and 4, are the key results of this paper. If the series for T h is truncated at I terms, an additional bias O ( ( H I) is introduced. We can make this dimensionless by writing O H I) for O ( H I), H = H/µ 2. If T (F ) is a polynomial in moments or cumulants, so that it is a polynomial in F of total degree p say, then these expansions stop at h p, as we now illustrate. Define the pth non-central and central moments of X and X h by m p = m p (F ) = E [X p ], m ph = m p (F h ) = E [ X p h], µ p = µ p (F ) = E [(X µ) p ], µ ph = µ p (F h ), and their pth cumulants by κ p = κ p (F ), κ ph = κ p (F h ), p 1. Then, κ 1h = κ 1 = µ, κ ph = κ p + b ph, (1.6) 3

5 b ph = B p h p /p for p 2, and B p is the pth Bernoulli number, defined by t/ ( e t 1 ) = t p B p /p! = t/2 + B t /()!. p=0 (This does not contradict (1.4) or (1.5) since B p = 0 for p = 3, 5, 7,.... For B p to p = 14, see Table 23.1 on page 809 of Abramowitz and Stegun (1964).) So, κ ph = κ p for p odd. Kendall (1938) gave a proof for bounded variables, and noted proofs of earlier authors. We give a proof in Example 2.2 below, and an extension to multivariate data in Example 3.1. UEs for cumulants and their products easily follow: see Examples 2.2, 3.1. Sheppard (1899, 1907) showed that m ph = [p/2] ( ) p H k m p /( + 1) (1.7) for p 1, [p/2] is the integral part of p/2. This is a special case of (1.4). For example, m 1h = m 1, m 2h = m 2 + H/3, m 3h = m 3 + Hm 1, m 4h = m 4 + 2Hm 2 + H 2 /5, m 5h = m Hm 3 /3 + H 2 m 1, m 6h = m 6 + 5Hm 4 + 3H 2 m 2 + H 3 /7, m 7h = m 7 + 7Hm 5 + 7H 2 m 3 + H 3 m 1, m 8h = m Hm 6 /3 + 14H 2 m 4 + 8H 3 m 2 + H 4 /9, m 9h = m Hm H 2 m 5 /5 + 12H 3 m 3 + H 4. Kendall (1938) gave regularity conditions for (1.7) to hold, but only for bounded random variables. Corollary 2.1 extends (1.7) to show that for any smooth function g, g h (x) = E [g (X h )] = E [g h (X)], H k g (x)/( + 1)!. Equation (37) of Sheppard (1907) inverted (1.7) to give an UE of m p equivalent to that given in Example 2.1 below: say, p 1, m p = [p/2] ( ) p a k H k m p,h = m h p (F h ) (1.8) ( ) a k = 2 2 B = 2 B (1/2) (1.9) for k 1. This is a special case of (1.5). For example, m 1 = m 1h, m 2 = m 2h H/3, m 3 = m 3h Hm 1h, m 4 = m 4h 2Hm 2h + 7H 2 /15, m 5 = m 5h 10Hm 3h /3 + 7H 2 m 1h /3, m 6 = m 6h 5Hm 4h + 7H 2 m 2h 31H 3 /21, m 7 = m 7h 7Hm 5h + 49H 2 m 3h /3 31H 3 m 1h /3, m 8 = m 8h 28Hm 6h /3 + 98H 2 m 4h /3 124H 3 m 2h / H 4 /15, m 9 = m 9h 12Hm 7h + 294H 2 m 5h /5 124H 3 m 3h + 381H 4 m 1h /5. 4

6 So, for 1 p n, an UE of m p is m h p = m h p (F nh ), m h p(f ) = [p/2] ( ) p a k H k m p. Theorem 2.2 extends (1.8) to show that for any smooth function g, [ ] E [g(x)] = E g h (X h ), g h (x) = H k A k g (x), A k = a k /()!. (1.10) Example 2.2 gives an alternative proof of (1.6). Theorem 2.3 gives an UE of E [g(x)] for any smooth enough function g, based on the rounded sample. Theorem 2.4 extends this to nonlinear functionals of F. Equation (25) of Sheppard (1907) gave a result equivalent to µ 2 = µ 2h H/3, µ 3 = µ 3h, µ 4 = µ 4h 2Hµ 2h + 7H 2 /15, µ 5 = µ 5h 10Hµ 3h /3. Example 2.3 proves that (1.7) and (1.8) hold with {m i } replaced by {µ i }, that is, for p 1, µ ph = [p/2] ( ) p H k µ p /( + 1), µ p = [p/2] ( ) p a k H k µ p,h = µ h p (F h ). (1.11) These are also special cases of (1.4) and (1.5). For example, since µ 1 = 0 this gives µ 6 = µ 6h 5Hµ 4h + 7H 2 µ 2h 31H 3 /21, µ 7 = µ 7h 7Hµ 5h + 49H 2 µ 3h /3, µ 8 = µ 8h 28Hµ 6h /3 + 98H 2 µ 4h /3 124H 3 µ 2h / H 4 /15, µ 9 = µ 9h 12Hµ 7h + 294H 2 µ 5h /5 124H 3 µ 3h. The versions of the four expansions (1.7), (1.8), (1.11) given by the authors cited, are more cumbersome as they are in powers of h, not H. By (1.11), for 1 p n, an UE of µ p is µ h p = µ h pn (F nh ), (1.12) µ h pn(f ) = [p/2] ( ) p a k H k µ p,n (F ), and µ pn (F n ) is the UE of µ p given by James (1958) for p 6 and by Withers and Nadarajah (2008) for p 12 by a different method. 5

7 Section 3 generalizes Section 2 to s-dimensional rounding: we suppose that for 1 j s the jth component of the data is rounded to the nearest multiple of some constant h j > 0. The multivariate relations between moments analogous to (1.8) and (1.11) are given in the appendix, using the method of Example 3.2. Section 4 generalizes Section 3 to multisample data. Removing the effect of rounding on parametric estimators. The results given here are nonparametric, but apply immediately to bias reduction of estimators of the form θ = T (F n ). Related results. Withers and Nadarajah (2011) have given UEs for moments for the linear regression model when observations are rounded. Withers and Nadarajah (2012b) have shown how to adjust Cornish-Fisher expansions and confidence intervals for the effect of roundoff. Bias reduction for ) parametric problems has also been well studied. Withers (1987) gave an estimator t nk ( θ of bias n k for any smooth function t(θ) : R p R, when F is determined by θ. Notation. Set [r] k = r!/(r k)! = r(r 1) (r k + 1), D = d/dx, g k (x) = D k g(x) for g : R C, i = / x i, g k1,...,k r (x) = k 1 1 kr r g (x 1,..., x r ) for g : R r C. We use the notation g to denote both the univariate and multivariate versions. The distinction should be clear from the context. 2 Proof of (1.4) and (1.5) In this section, we give an UE of E [g (X 1,..., X r )] and then prove (1.4) and (1.5). Let g(x) : R C be any function with finite kth derivative g k (x), k 1. Set θ = E [g(x)], θ h = E [g (X h )] = g(x)df h (x), θ h = g(x)df nh (x) = n 1 n g (X hj ), (2.1) ] so that E [ θh = θ h. Corollary 2.1 below shows that θ h estimates θ with bias θ h θ = H k E [g (X)/( + 1)!] = O(H) (2.2) k=1 as H 0. Since F is continuous, θ h of (2.1) has mean θ h given by E [g (X h )] = g df h = g(mh) {F (mh + h/2) F (mh h/2)}. (2.3) m= We first expand this as a Taylor series in h, and then prove (2.2). Set G(x) = x 1 sinh x = x /( + 1)! 6

8 for x R. By equation ( ) in Gradshteyn and Ryzhik (2000) or equation ( ) in Abramowitz and Stegun (1964), for x < π, for A k of (1.10). Set Theorem 2.1 Given g and F, G(x) 1 = x/ sinh x = g h (x) = G(hD/2)g(x) = A k x (2.4) H k g (x)/( + 1)!, (2.5) g h (x) = G(hD/2) 1 g(x) = E [g (X h )] = H k A k g (x). (2.6) h k L k (g) when this converges, 2 k (e k ( ) + e k ( )) /k!, k = 1, 3, 5,..., L k (g) = 2 k e k+1 (x)dx/(k + 1)! + d k/2 (g), k = 0, 2, 4,..., (2.7) e k (x) = g(x)f k (x), d 0 (g) = 0, k 1 d k (g) = [B 2l / ( 2l)!] α k l (e 2l+1 ) 2 2l /(2l + 1)!, k 1, l=0 α k (v) = v 1 ( ) v 1 ( ). So, e k (x) 0 as x for k 1, [ ] E [g (X h )] = H k e +1 (x)dx/( + 1)! + d k (g) (2.8) when this converges. Proof: Since F (x + h/2) F (x h/2) = 2 (h/2) 2l+1 F 2l+1 (x)/(2l + 1)!, l=0 we have the right hand side of (2.3) equal to g(mh) h 2l+1 F 2l+1 (mh)2 2l /(2l + 1)! = h 2l+1 2 2l d (h, e 2l+1 ) /(2l + 1)!, (2.9) m= l=0 l=0 7

9 d(h, v) = d(h, v) = m= m=0 v(mh) = v(0) + d(h, v) + d (h, v), v(mh) = h 1 v(x)dx + {v(0) + v( )} /2 + h 2j 1 B 2j α j (v)/(2j)!, v(x) = v( x), α j (v) = v 2j 1 ( ) v 2j 1 (0) 0 by the Euler-MacLaurin formula, equation ( ), page 806 of Abramowitz and Stegun (1964) with a = 0, m =. So, d(h, v) = h 1 v + [v( ) + v( )] /2 + h 2j 1 B 2j β j (v)/(2j)!, (2.10) β j (v) = α j (v) + α j (v) = v 2j 1 ( ) 2v 2j 1 (0) + v 2j 1 ( ). Now apply (2.10) to (2.9) and collect terms to obtain E [g (X h )] as being equal to the right hand side of (2.8) plus O h, O h = [C h ( ) + C h ( )] g(x), C h (x) = (h/2) 2l+1 F 2l+1 (x)/(2l + 1)! = sinh(hd/2) F (x) l=0 = (F (x + h/2) F (x h/2)) /2 0 as x ±, by Taylor s expansion, so that O h = 0. Corollary 2.1 Suppose that k 1 h B +1 [e 2l+1 (x) + e 2l+1 ( x)] D 2(k l) 1 [e 2l+1 (x) + e 2l+1 ( x)] /( 2l)! 0 k=1 l=0 (2.11) as x, and g i (x)f j (x) 0 (2.12) as x ± for i, j 1. Then, θ h of (2.1) is given in terms of g h, f h of (2.5) by E [g (X h )] = g(x)f h (x)dx = f(x)g h (x)dx = E [g h (X)] = H k E [g (X)/( + 1)!]. (2.13) 8

10 Proof: Under these conditions, (2.7) reduces to L k (g) = I(k even)2 k e k+1 (x)dx/(k + 1)!, so that E [g (X h )] = g(x) H k f (x)dx/( + 1)! = g(x)f h (x)dx if absolutely integrable. Also, by integrating by parts, if g i (x)f k i (x) 0 as x ± for 1 i k 1, then g(x)f k (x)dx = g 1 (x)f k 1 (x)dx = ( 1) i g i (x)f k i (x)dx = ( 1) k g k (x)f(x)dx. So, by (2.12), gf h = g(x)g(hd/2)f(x)dx = f(x)g(hd/2)g(x)dx = fg h. The proof is complete. The problem of finding an UE of gdf is essentially that of inverting this series. The solution is given in terms of g h (x) of (2.6) and A k of (1.10) by Theorem 2.2 If the derivatives of F and g tend to zero at ± as in (2.11) and (2.12), then for g h of (2.6), g h (X h ) is an UE of E [g(x)]: [ ] E [g(x)] = E g h (X h ) = H k A k E [g (X h )]. (2.14) Proof: Replace g(x) in (2.13) by g h (x). The result follows from (2.4), (2.6). So, Theorem 2.3 If the derivatives of F and g tend to zero at ± as in (2.11) and (2.12), then an UE of E [g(x)] = gdf is T h (F nh ), [ ] T h (F ) = E g h (X) = g h df for g h of (2.6). So, for K 1, T Kh (F nh ) = K 1 H k A k n 1 n g (X hi ) estimates θ = ( gdf with bias O H K) h 2K, h = h/σ, σ = µ 1/2 2. 9

11 Example 2.1 Note that (1.7) and (1.8) are just (2.13) and (2.14) with g(x) = x p. Also E [ m h p (X h ) ] = m p, m h p(x) = G(hD/2) 1 x p = [p/2] A k H k D x p = [p/2] since D x p = [p] x p. Taking the mean of m h p(x) gives (1.8). ( ) p a k H k x p (2.15) That m p can be recovered from {m ph } even though there are infinitely many F such that Prob((j 1)h/2 < X jh/2) takes a given value Prob(X h = jh), is a reminder that the moments need not determine a distribution. Example 2.2 Take g(x) = e xt, t C. Then, g h (x) = G(ht/2)g(x), g h (x) = G(ht/2) 1 g(x), E [ e X ht ] /E [ e Xt] = G(ht/2), ln E [ e X ht ] ln E [ e Xt] = ln G(ht/2). But by equation ( ), page 53 of Gradshteyn and Ryzhik (2000), ln G(t/2) = t 2p k 2p /(2p)!, (2.16) k 2p = B 2p /(2p), so that p=1 κ 2p 1,h = κ 2p 1, κ 2p,h = κ 2p + h 2p B 2p /(2p), proving Kendall s (1.6). So, for p 2, an UE of κ p based on rounded data is κ h pn = k pn (F nh ) b ph for b ph of (1.6), k pn (F n ), Fisher s k p statistic, an UE of κ p for unrounded data, see Section 12.6 of Stuart and Ord (1987). So, for T (F ) a smooth function of the odd cumulants, T (F h ) = T (F ). Similarly, since an UE of κ p κ q is κ p κ q = (κ p b ph ) (κ q b qh ), κ h pqn = k pqn (F nh ) 2 b ph k qn (F nh ) + b ph b qh, pq k pqn (F n ), Fisher s UE (polykay) of κ p κ q for unrounded data, see Section of Stuart and Ord (1987). UEs of higher products of cumulants can be dealt with similarly. Equation (2.13) fails if E [g (X)] is unbounded for some k 0. Equation (2.14) fails if E [g (X h )] is unbounded for some k 0, for example, if g (0) is unbounded for some k 0, as is the case for g(x) = x 1/2 or ln(x), and Prob( X < h/2) > 0. So, (2.14) fails for these two cases of g if X has support (0, ), although it will work for g(x) = (x + x 0 ) 1/2 or ln (x + x 0 ) if x 0 h/2. So far we have been dealing with the functional E [g(x)] = gdf, using g h, g h of (2.5) and (2.6). This is linear in F. Applying (2.13) and (2.14) r times gives 10

12 Corollary 2.2 Given a smooth function g (x 1,..., x r ) : R R C, and independent X 1,..., X r from a distribution function F on R, rounded as above to X 1h,..., X rh with distribution function F h on R, So, g h (x 1,..., x r ) = g h (x 1,..., x r ) = E [g (X 1h,..., X rh )] = E [g h (X 1,..., X r )], (2.17) [ ] E g h (X 1h,..., X rh ) = E [g (X 1,..., X r )], (2.18) g h (x 1,..., x r ) = G h ( /2) g (x 1,..., x r ), g h (x 1,..., x r ) = G h ( /2) 1 g (x 1,..., x r ), r G h ( /2) = G h ( i /2). k 1,...,k r=1 k 1,...,k r=1 H k 1+ +k 2 r g 1,..., r (x 1,..., x r ) / By (2.18), for 1 r n, an UE of E [g (X 1,..., X r )] is N 1 r r ( i + 1)!, H k 1+ +k 2 r A k1 A kr g 1,..., r (x 1,..., x r ). N r g h (Y 1h,..., Y rh ) summed over all N r = ( n r) subsamples Y1h,..., Y rh of X 1h,..., X rh. The next three examples give UEs of central moments. As noted, James (1958) and Example 6.3 of Withers and Nadarajah (2008) gave an UE for µ r of the form µ rn (F n ) for r 6 and r 12, respectively. Example 2.3 An UE of µ p. Write µ p = E [(X µ) p ] = p i=0 ( ) p ( µ) i m p i = i p i=0 ( ) p ( 1) i J i+1, (2.19) i J i+1 = E [g i+1 (X 1,..., X i+1 )], g i+1 (x 1,..., x i+1 ) = x 1 x i x p i i+1. So, an UE of µ p is given by replacing J i+1 in (2.19) by for m h p(x) of Example 2.1. That is, J i+1 (X 1h,..., X i+1,h ) = X 1h X ih m h p i (X h,i+1 ) µ p = p i=0 ( p [ ] )( 1) i µ ih i E m h p i (X h ), 11

13 E [ m h p (X h ) ] is equal to the right hand side of (1.8). So, M pkh = i=0 µ p = [p/2] a k H k M pkh, p ( )( ) ( ) ( µ h ) i p i p m p i,h = µ p,h, i proving (1.11). So, an UE of µ p is given by (1.12). The next two examples illustrate this for p = 2, 3. Example 2.4 An UE of µ 2 using unrounded data is µ 2n (F n ), µ 2n (F ) = ( 1 n 1) 1 µ2 (F ). So, an UE of µ 2 = µ 2h H/3 using rounded data is µ 2n (F nh ) H/3. Example 2.5 An UE of µ 3 using unrounded data is µ 3n (F n ), ( ) n µ 3n (F ) = N3 1 n3 µ 3 (F nh ), N 3 =. (2.20) 3 So, an UE of µ 3 = µ 3h using rounded data is µ 3n (F nh ). Other polynomials in F of degree p say, can be estimated similarly by writing them as the sum of U-statistics: p T (F ) = g i (x 1,..., x i ) df (x 1 ) df (x i ). In Section 3, we prove Theorem 2.4 Given a distribution function F (x) on R, let T (F ) be a smooth functional with rth functional derivative T F (x 1,..., x r ), r 1, as defined in Withers (1983). Then, (I): say, T (F h ) = T (F ) + T rh (F )/r! = T h (F ) (2.21) r=1 T rh (F ) = E [T F (X 1h,..., X rh )] = E [T F h (X 1,..., X r )], T F h (x 1,..., x r ) = G h ( /2)T F (x 1,..., x r ) = H k T F (x 1,..., x r ), T F (x 1,..., x r ) = T F 1,..., r (x 1,..., x r ) / k 1 + +k r=k 12 r ( i + 1)!

14 when (2.21) converges. Also by (2.17), for k r 1, Also (II): T rh (F ) = T h (F ) = H k T kr (F ) = O ( H r ), T kr (F ) = E [ T F (X 1,..., X r ) ], k=r T h (F ) = T (F ) + H k T k (F ), T 0(F ) = T (F ), T k (F ) = k r=1 T (F ) = T h (F h ), T kr (F )/r!. (2.22) Tr h (F )/r!, (2.23) r=1 [ ] Tr h (F ) = E TF h (X 1,..., X r ), T h F (x 1,..., x r ) = G h ( /2) 1 T F (x 1,..., x r ) = T F (x 1,..., x r ) = k 1 + +k r=k when (2.23) converges. So, for k r 1, T h r (F ) = T h (F ) = T 0 (F ) = T (F ), T k (F ) = For example, k=r k r=1 T F 1,..., r (x 1,..., x r ) H k T F (x 1,..., x r ), r A ki H k E [ T F (X 1,..., X r ) ] = O ( H r ), H k T k (F ), T kr (F )/r!, T kr (F ) = E [ T F (X 1,..., X r ) ]. (2.24) T k1 (F ) = E [T F (X 1 ) /( + 1)!], T k1 (F ) = E [T F (X 1 ) A k ], [ ] k 1 T k2 (F ) = 2 E [T F 1, 2 (X 1, X 2 )] / ( i + 1)! k 1 =1 k 1 T k2 (F ) = k 1 =1 [ E [T F 1, 2 (X 1, X 2 )] 13 ] 2 A ki k 2 =k k 1 k 2 =k k 1,

15 since E [T F 0,2 (X 1, X 2 )] = 0 for k 2 1. For T (F ) a polynomial of degree p in F, its rth derivatives are zero for r > p so that T rh (F ) = T h r (F ) = 0, and the expansions (2.21) and (2.23) have only p terms. Example 2.6 Suppose that T (F ) = m p. Then, T F (x 1 ) = x p 1 m p, T k (F ) = T k1 (F ), T k ( + 1)!T k (F ) = T k1 (F ), (F ) = T k (F )/A k = E [T F (X 1 )] /( + 1)! = [p] m p since T F (x 1,..., x r ) = 0 implies T kr (F ) = T kr (F ) = 0 for k r 2. This proves (1.7) and (1.8) again. We now use this theorem to give another proof that the relations (1.7) and (1.8) between m p and m ph also hold between µ p and µ ph, so that an UE of µ p in terms of the rounded sample, is just a linear combination of the UEs of µ p in terms of the unrounded sample, with F and the unrounded sample replaced by F h and the rounded sample. To prove this, we need the following result, given in Example 5.6 of Withers and Nadarajah (2012a). Theorem 2.5 For p, r in {1, 2...}, the rth functional derivative of T (F ) = µ p (F ) is { r ( ) } r T 1,...,r = ( 1) r [p] r µ p r [p] r 1 h p r i µ p r+1 h 1 i h i. (2.25) Each term in (2.25) is linear in h i except for the term with h p r+1 i. Example 2.7 Suppose that T (F ) = µ p. Then, for r 2, k 1,..., k r 1, T F 1,..., r (x 1,..., x r ) = E [T F 1,..., r (X 1,..., X r )] = 0, T rh (F ) = T h r (F ) = 0. Also for k 1 and A k of (1.10), implies T F (x 1 ) = [p] h p 1 T k (F ) = T k1 (F ), T k (F ) = T k1 (F ), ( + 1)!T k (F ) = T k (F )/A k = E [T F (X)] = [p] µ p. So, T 1h (F ) = { ( ) } p µ p H k /( + 1) : 1 k p/2 = µ ph µ p, T1 h (F ) = {( ) } p a k H k µ p : 1 k p/2 = µ h p µ p, and (1.11) holds. So, an UE of µ p in terms of the rounded sample is given by (1.12). 14

16 A method for obtaining UEs using unrounded data, for products of moments, was given in Withers and Nadarajah (2012a). Example 2.8 By page 255 of Withers and Nadarajah (2012a), an UE of µ 2 µ 3 is µ 23n (F n ), µ 23n (F ) = { µ 2 µ 3 + ( µ 5 2µ 2 µ 3 ) n 1 + (µ 5 + 5µ 2 µ 3 ) n 2} / 4 (1 i/n). So, using rounded data, an UE of µ 2 µ 3 = (µ 2h H/3) µ 3h is µ h 23n (F nh), for µ 3n (F ) of (2.20). µ h 23n(F ) = µ 23n (F ) (H/3)µ 3n (F ) Theorem 2.5 can be used to obtain estimators of low bias for any smooth function of moments. (In contrast, such a method does not exist for non-polynomial functions of cumulants since the derivatives of the general cumulant are not known.) We now give two examples when T (F ) need not be a polynomial in F. Example 2.9 Estimating a function of the mean. Suppose that T (F ) = a(µ) for some smooth function a. Since m 1h = m 1, we just need to substitute F nh for F n in the estimators S np (F n ), T np (F n ) of bias O (n p ) given by (1.1) and (1.3). From Example 6.1 of Withers and Nadarajah (2008), S i (F ) of (1.1) is given by S i (F ) = 2i r=i+1 S 12 = µ 2 /2, S 23 = µ 3 /3, S 24 = µ 2 2/8, S ir a r (µ), S 34 = µ 4 /4 + 3µ 2 2/8, S 35 = 3µ 3 µ 2 /8, S 36 = µ 3 2/48, S 45 = µ 5 /5 2µ 3 µ 2 /3, S 46 = 3µ 3 2/16 + µ 4 µ 2 /8 µ 2 3/18, S 47 = µ 3 µ 2 2/24, S 48 = µ 4 2/24, S 56 = µ 6 /6 + 5µ 4 µ 2 /8 + 5µ 2 3/18, S 57 = µ 5 µ 2 /10 µ 4 µ 3 /12, S 58 = 3µ 4 2/64 µ 4 µ 2 2/32 µ 2 3µ 3 /36, S 59 = µ 3 µ 3 2/144, S 5,10 = µ 5 2/3840. So, S np (F n ) and S np (F nh ) are estimators based on unrounded and rounded observations for a(µ) with bias O (n p ) (and are UEs if a(µ) is a polynomial of degree less than or equal to 2p). For example, an estimator of bias O ( n 3) based on rounded observations is S n3 (F nh ), S n3 (F ) = a(µ) (µ 2 /2) a 2 (µ)/(n 1) + [ (µ 3 /3) a 3 (µ) + ( µ 2 2/8 ) a 4 (µ) ] /(n 1) 2. For a(µ) = µ q, substitute a i (µ) = [q] i µ q i. 15

17 Example 2.10 Suppose that T (F ) = a (µ 2 ) for some smooth function a. So, T (F ) = T h (F h ), T h (F ) = a (µ 2 H/3). For T (F ) = a (µ 2 ), by Example 5.12 of Withers and Nadarajah (2012a), S i (F ), i = 1, 2, 3 needed for (1.1) are given by S 11 = µ 2, S 12 = ( µ 4 µ 2 2) /2, 2i S i (F ) = S ir a i (µ 2 ), r=1 S 21 = 0, S 22 = 2µ 4 + 7µ 2 2/2, S 23 = µ 6 /3 µ 2 3 3µ 4 µ 2 /2 + 7µ 3 2/6, S 24 = ( µ 4 µ 2 2) 2 /8, S 31 = 0, S 32 = 3µ 4 + 9µ 2 2/2, S 33 = 3µ 6 27µ 4 µ 2 /2 7µ µ 3 2, S 34 = µ 8 /4 + 4µ 6 µ 2 /3 + 2µ 5 µ µ 2 4/8 6µ 4 µ 2 2 7µ 2 3µ µ 4 2/24, S 35 = ( µ 4 µ 2 2) ( 4µ6 + 15µ 4 µ 2 + 3µ µ 3 2) /24, S36 = ( µ 4 µ 2 2) 3 /48. So, for 1 p 4, an estimator of a (µ 2 ) with bias O (n p ) based on unrounded observations is S np (F n, a) = S np (F n ). But a (µ 2 ) = a h (µ 2 ), a h (µ 2 ) = a (µ 2h H/3). So, for 1 ( p 4, an estimator of a (µ 2 ) with bias O (n p ) based on rounded observations is S np Fnh, a h). For example, an estimator of a (µ 2 ) with bias O ( ( n 2) based on rounded observations is S n2 Fnh, a h), S n2 (F, a h) = a (µ 2 H/3) = [ µ 2 a 1 (µ 2 H/3) ( µ 4 µ 2 2) a 2 (µ 2 H/3) /2 ] /(n 1). For a (µ 2 ) = µ q 2, substitute a 1 (µ 2 ) = [q] i µ q i 2. For the standard deviation, σ = µ 1/2 q = 1/2. Example 2.11 By Example 5.16 of Withers and Nadarajah (2012a), an estimator of T (F ) = µ/σ of bias O (n p ) based on unrounded observations is given by S np (F n ) of (1.1) with S 1 (F ) = β 3 /2 β (3β 4 + 1) /8, S 2 (F ) = (48β 5 15β 4 β 3 23β 3 ) /64 +(µ/σ) ( 80β β β β 2 3) /128, β r = µ r µ r/2 2. Now replace each µ p by the right hand side of (1.11) and F n by F nh to obtain an estimator of bias O (n p ), p = 1, 2, 3 based on rounded observations. For example, writing 2, µ h = µ (F nh ), µ rh = µ r (F nh ), (2.26) an estimator of µ/σ with bias O ( n 1) based on rounded observations is µ h ( µ 2h H/3) 1/2. Example 2.12 By the appendix and Examples 5.22, 5.23 of Withers and Nadarajah (2012a), UEs of µ 11, the covariance of X 1 and X 2, and of µ 111 are µ 11 (F nh ) n/(n 1) and µ 111 (F nh ) n 2 /[n 1] 2. 16

18 Example 2.13 The correlation of X1 and X 2 is say, ρ = T (F ) = U 1 (U 2 U 3 ) 1/2 = g(u) U 1 = µ 11 = µ 11h, U 2 = µ 20 = µ 20h H 1 /3, U 3 = µ 02 = µ 02h H 2 /3. So, an estimator of correlation with bias O ( n 1) using rounded observations is ρ = µ 11h ( µ 20h H 1 /3) 1/2 ( µ 02h H 2 /3) 1/2. An estimator of bias O ( n 2) using unrounded observations is given by S n2 (F n ) of (1.1), by equation (A.8) of Withers and Nadarajah (2012a), T ( 1 2) = g ij U ij + g i U i ( 1 2 ), U ij = U ij (1, 1) = U if (x)u jf (x)df (x), g i, g ij are the first and second partial derivatives of g at µ 2, and the summation of i, j over their range 1, 2, 3 is implicit. (Withers and Nadarajah (2012a) have used a different notation for multivariate moments.) Now substitute g 1 = U 1/2 2 U 1/2 3, g 2 = ( 1/2)U 1 U 3/2 2 U 1/2 3, g 3 = ( 1/2)U 1 U 1/2 2 U 3/2 3, g 11 = 0, g 12 = ( 1/2)U 1 U 3/2 2 U 1/2 3, g 13 = ( 1/2)U 1 U 1/2 2 U 3/2 3, g 22 = (3/4)U 1 U 5/2 2 U 1/2 3, g 23 = (1/4)U 1 U 3/2 2 U 1/2 3, g 33 = (3/4)U 1 U 1/2 2 U 5/2 3, U 1 ( 1 2 ) = 2µ 11, U 2 ( 1 2 ) = 2µ 20, U 3 ( 1 2 ) = 2µ 02, U 11 = µ 22 µ 2 11, U 12 = µ 31 µ 11 µ 20, U 13 = µ 13 µ 11 µ 02, U 22 = µ 40 µ 2 20, U 23 = µ 22 µ 20 µ 02, U 33 = µ 04 µ to obtain T ( 1 2) for S 1 (F ) of (1.2) and an estimator S n2 (F nh ) of bias O ( n 2) using S n2 (F ) of (1.1). Example 2.14 Example 6.6 of Withers and Nadarajah (2008) gives T [2], T [3], T [ 2 2] needed for S i (F ) of (1.1) to estimate the skewness, T (F ) = µ 3 /µ 3/2 2, to O (n p ) for p = 1, 2, 3 using unrounded data. They are given in terms of the standardized moments, β r = µ r µ r/2 2. For rounded data, just replace each µ p by the right hand side of (1.11) and F n by F nh, as in Example 2.9. For example, in terms of µ h, µ rh of (2.26), an estimator of skewness with bias O ( n 1) using rounded observations is µ 3h ( µ 2h H/3) 3/2. Similarly, the bias reduced estimators of return times for unrounded data given in Examples of Withers and Nadarajah (2012a), can be adapted to rounded data. Example 2.15 By (1.6), for r 2, an UE of κ r is k r (F nh ) B r h r /r, k r = k r (F n ) is Fisher s k-statistic: see, for example, Stuart and Ord (1987). 17

19 3 Multivariate rounding In this section, we suppose that X is an s-vector with distribution function F on R s, and that g : R s C. Let h = (h 1,..., h s ) be given positive constants and suppose that X h is X with its ith component rounded to the nearest multiple of h i for 1 i s. Similarly, let X 1h,..., X nh be a random sample X 1,..., X n on R s rounded in this way, and let F nh and F n be their empirical distribution functions. What is E [g (X h )]? What is an UE of E [g(x)]? For x R s and 1 j s, set v j = var [(X 1 ) j ], for A kj = / x, j = / x j, k = G h (x) = s G (h j x j ), g h (x) = G h ( /2)g(x) = H k = s H k j s k j j, g (x) = g(x), H k g (x)/( + 1)!, s j, k! = k j!, H j = h 2 j/4, H 0 = max s H j/v j, (3.1) g h (x) = G h ( /2) 1 g(x) = A k H k g (x), A k = s A kj (3.2) of (1.9). So, H 0 is dimensionless. Applying (2.13) and (2.14) s times gives Corollary 3.1 Given a smooth function g : R s C, E [g (X h )] = E [g h (X)], [ ] E g h (X h ) = E [g(x)]. Applying this r times to a function of r arguments gives Corollary 3.2 Given a smooth function g (x 1,..., x r ) : R s R s C, and independent X 1,..., X r from a distribution function F on R s, rounded as above to X 1h,..., X rh with distribution function F h on R s, E [g (X 1h,..., X rh )] = E [g h (X 1,..., X r )], [ ] E g h (X 1h,..., X rh ) = E [g (X 1,..., X r )], (3.3) g h (x 1,..., x r ) = G h ( /2) g (x 1,..., x r ), g h (x 1,..., x r ) = G h ( /2) 1 g (x 1,..., x r ), r s G h ( /2) = G h ( i /2), G h ( i /2) = G (h j ij /2), ij = / x ij. 18

20 We now extend Example 2.2 to F on R s. Set N = {0, 1, 2,...}. For X R s and p = (p 1,..., p s ) N s, the pth cumulant of X is defined by ln E [e ] t X = κ p t p /p!, p 1 t p /p! = s t p j j /p j! for t C s. Example 3.1 Take g(x) = e x t, t C s. Then, g h (x) = G h (t/2)g(x), g h (x) = G h (t/2) 1 g(x), [ E e X t] [ ] h /E e X t = G h (t/2), [ ln E e X t] [ ] h ln E e X t = ln G h (t/2) = s ln G (h j t j /2) = p=1 k 2p (2p)! s (h j t j ) 2p by (2.16). So, κ ph = κ p for p = (p 1,..., p s ) except if p 1 = = p s = 2r for some positive integer r; in that case κ ph = κ p + (h 1 h s ) 2r b 2r,s for r 1, b 2r,s = (2r!) s 1 B 2r /(2r). For example, b 2r,2 = (2r 1)!B 2r, b 2r,3 = (2r 1)!(2r)!B 2r. So, if k pn (F n ) is Fisher s polykay, the UE of the multivariate cumulant κ p for unrounded data, then an UE of κ p for rounded data is k np (F n ) except if p 1 = = p s = 2r for some r 1; in that case an UE of κ p is κ h pn = k pn (F nh ) b rs (h 1 h s ) 2r for the even marginals of p, which can be obtained from Example 2.2. Wold (1934), equation (5), page 254 considered the case s = 2. Theorem 3.1 Given a distribution function F (x) on R s, Theorem 2.4 remains true if 0, k, k i are interpreted as s-vectors, H k, k!, A k are interpreted as in (3.1), (3.2), the expressions in (2.22), (2.24) for T k (F ), T k (F ) are replaced by T k (F ) = T kr (F )/r!, T k (F ) = T kr (F )/r!, 1 r k 1 r k k = s k j, and O ( H r ) is replaced by O ( H r 0) for H0 of (3.1). 19

21 Proof: By Withers (1983), T (G) = T (F ) + r=1 T F (x 1,..., x r ) dg (x 1 ) dg (x r ) /r!. Putting G = F h gives (2.21). Reversing F, F h and applying (3.3) gives (2.23). That E [T F h (X 1,..., X r )] = O ( H r 0) follows from its expansion in powers of H by noting that H k 1+ +k r = O ( H r ) 0 if k 1 1,..., k r 1. Furthermore, E [T F (X 1, x 2,..., x r )] = 0 implies E [ T F 0,k2,...,k r (X 1, x 2,..., x r ) ] = 0 which in turn implies E [ T F k1,k 2,...,k r (X 1, X 2,..., X r ) ] = 0 if k 1 = 0 or k 2 = 0 or or k r = 0. The proof is complete. We need to distinguish between unrounded sample values, X 1,..., X n, and the jth component of X = X 1, say X j, 1 j s. The multivariate moments are s m p = E X p j j [ ] s ( ) pj, µ j = E Xj, µ p = E Xj µ j. (3.4) Example 3.2 Take g(x) = x p = s xp j j, T (F ) = E [g(x)] = m p. Then, g h (x) = s G (h j j /2) 1 x p j j is the product of terms of the form (2.15). Taking the mean of m h p(x) gives m p = ( ) p a k H k m p, (3.5) ( ) p = s ( pj j ). This multiplication is easily done by writing (1.8) as m p = P ph (S)m h, P ph (S) = [p/2] ( ) p a k H k S, S i m h = m ih. Its multivariate form is s m p = P pj,h j (S j ) m h, 20

22 s S i j j m h = m i1,...,i sh. For example, for m 2, m 4 are given by (1.8) as So, So, setting (rs) = m rsh, P 2H (S) = S 2 H/3, P 4H (S) = S 4 2HS 2 + 7H 2 /15. P 2H2 (S 2 ) = S 2 2 H 2 /3, P 4H1 (S 1 ) = S 4 1 2H 1 S H 2 1 /15. m 42 = P 4H1 (S 1 ) P 2H2 (S 2 ) m h = (42) 2H 1 (22) + 7H 2 1 (02)/15 (H 2 /3) ( (40) 2H 1 (20) + 7H 2 1 /15 ), (3.6) and an UE estimator of m 42 based on rounded data is the right hand side of (3.6) with m rs (F h ) = m rsh = (rs) replaced by m rs (F nh ). In this way, we obtain the multivariate non-central moments of the appendix, and so UEs for them, by replacing F by F nh. As in the univariate case, (3.5) also holds with moments replaced by central moments: µ p = ( ) p a k H k µ p,h, (3.7) as illustrated in the appendix. An UE for µ p follows by replacing µ p in the right hand side of (3.7) by an UE based on F nh. For our next example, we replace the notation for the central moments of (3.4) by µ [i 1,..., i a ] = E [(X µ) i1 (X µ) ia ]. For example, by Appendix A, 4 µ[1234] = E (X µ) j = µ 1111 = µ 1111h, µ[1112] = E [ (X µ) 3 1(X µ) 2 ] = µ31 = µ 31h H 1 µ 11h. Example 3.3 Estimating a function of a multivariate mean. Suppose that T (F ) = a(µ) for some smooth function a with partial derivatives a i1,...,i r (µ) for i 1,..., i r {1, 2,..., s}. As in Example 2.9, we just need to substitute F nh for F n in the estimators S np (F n ), T np (F n ) of bias O (n p ) given by (1.1) and (1.3). For a, b,... 1, define T [a] = a i1,...,i a (µ)µ [i 1,..., i a ], T [ab] = a i1,...,i a,j 1,...,j b (µ)µ [i 1,..., i a ] µ [j 1,..., j b ], 21

23 and so on, and set T [ a 2] = T [aa], and so on, with the tensor summation convention that repeated pairs of suffixes i 1,..., i a, j 1,..., j b,... are implicitly summed over their range {1, 2,..., s}. For example, T [ 2 3] = a i1,i 2,j 1,j 2,k 1,k 2 (µ)µ [i 1, i 2 ] µ [j 1, j 2 ] µ [k 1, k 2 ] s = a i1,...,i 6 (µ)µ [i 1, i 2 ] µ [i 3, i 4 ] µ [i 5, i 6 ]. i 1,...,i 6 =1 So, for p 1, an estimator of a(µ) of bias O (n p ) is S np (F nh ) for S np (F ) of (1.1). S i (F ) has the form S i (F ) = 2i r=i+1 a j1,...,j r (µ)s i,j1,...,j r. If a(µ) is a polynomial of degree p i in µ i for i = 1,..., s, then S np (F nh ) is an UE of a(µ) if p p s p. 4 The multisample case Given k random samples with empirical distribution functions F n = (F n1 1,..., F nk k) from distribution functions F = (F 1,..., F k ) on R s 1,..., R s k and a smooth functional T (F ), equation (3.10) of Withers and Nadarajah (2008) gives an estimator of T (F ) with bias O (n p ) of the form p 1 S np (F n ) = Si n (F n ) (4.1) for p 7, n = min (n 1,..., n k ) is the minimum sample size and S n 0 (F ) = T (F ), S n 1 (F ) = (1/2) k T ( i 2) / (n i 1), T ( i 2) = ( ) i, i T F df i (x), x, x ( T i,j F x,y) is the second partial derivative of T (F ) with respect to Fi, F j as defined on page 194 of Withers and Nadarajah (2008), and so on: see page 199 of Withers and Nadarajah (2008). This was done using the expansion (2.21) with (F h, F ) replaced by (F, G) for G = (G 1,..., G k ) arbitrary distribution functions on R s 1,..., R s k, replacing E [T G (X 1,..., X r )] by an UE, (available for r n), then replacing G by F n. We illustrate how to adapt these results to rounded data with some examples. Let the rounding for the ith sample of dimension s i have corresponding rounding vector h i R s i. Let F ini h i be the empirical distribution function of the ith sample after rounding. Set F nh = (F 1n1 h 1,..., F knk h k ), the set of empirical distribution functions. Let µ i be the mean of the ith distribution. Example 4.1 Suppose that T (F ) = k U i (F i ). Then, an estimator based on rounded data of bias O (n p ) is k U h i inp (F in i h i ). 22

24 For example, if α i are t s i constants, then an UE of k α iµ i based on rounded data is k α i µ ihi. For another example, suppose that s i 1 and that α i are given scalars. Denote the variance of F i by µ 2i. Then, an UE of k α iµ 2i based on rounded data is for H i given by (3.1). k α i ( µ ihi H i /3) Example 4.2 Suppose that k = 2, s i 1 and T (F ) = µ 1 µ 2 / (µ 1 + µ 2 ). Then, S n i (F ) needed for (4.1) are given for 1 i 3 in Example 6.2.2, page 213 of Withers and Nadarajah (2008). Since µ i = µ ihi, an estimator of T (F ) based on rounded data is S np (F nh ) of (4.1). Example 4.3 For k univariate samples with T (F ) = α µ/β µ, an estimator of bias O (n p ) based on unrounded data( was ) given in Example 5.5, page 247 of Withers and Nadarajah (2012a) of the form T np F of (1.3). For example, T 1 (F ) = ( β µ ) 2 k a=1 (n/n a ) β 2 aµ 2 [a] (α a /β a T (F )). Since µ i = µ ihi, an estimator of T (F ) based on rounded data is S np (F nh ) of (4.1). Example 4.4 For two univariate samples with T (F ) = µ 1 µ 2 / (µ 1 + µ 2 ), S n4 (F ) of (4.1) is given by Example 6.2.2, page 213 of Withers and Nadarajah (2008). So, S n4 (F nh ) estimates T (F ) with bias O ( n 4). Appendix A: Multivariate moments up to order seven By pages 252 and 253 of Wold (1934), m 11 = m 11h, m 21 = m 21h H 1 m 01h /3, 2 m 31 = m 31h H 1 m 11h, m 22 = m 22h H1 m 02h /3 + H 1 H 2 /9, m 41 = m 41h 2H 1 m 21h + 7H1 2 m 01h /15, m 32 = m 32h (H 1 /3) m 12h H 2 (m 30h (H 1 /3) m 10h ), m 51 = m 51h 10H 1 m 31h + 7H1 2 m 11h /3, m 42 = m 42h 2H 1 m 22h + 7H1 2 m 02h /15 (H 2 /3) ( m 40h 2H 1 m 20h + 7H1 2 /15 ), m 33 = m 33 H 1 m 13h H 2 (m 31h H 1 m 11h ). 23

25 Also, by the method explained in Example 3.2, m 61 = m 61h 5H 1 m 41h + 7H 2 1 m 21h 31H 3 1 m 01h /21, m 52 = m 52h (10H 1 /3) m 32h + ( 7H 2 1 /3 ) m 12h (H 2 /3) [ m 50h (10H 1 /3) m 30h + ( 7H 2 1 /3 ) m 10h ], m 43 = m 43 H 2 m 41h 2H 1 (m 23h H 2 m 21h ) + ( 7H 2 1 /15 ) (m 03h H 2 m 01h ), m 111 = m 111h, m 211 = m 211h H 1 m 011h /3, m 311 = m 311h H 1 m 111h /3, m 221 = m 221h H 1 m 021h /3 H 2 m 201h /3 + H 1 H 2 m 001h /9, m 411 = m 411h 2H 1 m 211h /3 + 7H 2 1 m 011h /15, m 321 = m 321h H 1 (m 121h H 2 m 101h /3) H 2 m 301h /3, 3 3 m 222 = m 2212 H1 m 022h /3 + H1 H 2 m 002h /9 H 1 H 2 H 3 /27, m 511 = m 511h 10H 1 m 311h /3 + 7H 2 1 m 111h /3, m 421 = m 421h 2H 1 m 221h + 7H 2 1 m 021h /15 (H 2 /3) ( m 401h /3 2H 1 m 201h + 7H 2 1 m 001h /15 ), m 331 = m 331h H 1 m 131h H 2 m 311h + H 1 H 2 m 111h, m 1111 = m 1111h, m 2111 = m 2111h H 1 m 0111h /3, m 2211 = m 2211h H 1 m 0111h /3 H 2 m 2011h /3 + H 1 H 2 m 001h /9, m 3111 = m 3111h H 1 m 1111h /3, m 2221 = m 2221h H 1 m 0221h /3 H 2 m 2021h /3 + H 1 H 2 m 0011h /9, m 4111 = m 4111h 2H 1 m 2111h /3 + 7H 2 1 m 0111h /15, m 3211 = m 3211h H 1 m 1211h /3 H 2 (m 3011h /3 H 1 m 1011h ) /3, m 21 4 = m 21 4 h H 1 m 01 4 h/3, m 31 4 = m 31 4 h H 1 m 1 5 h, m = m h H 1 m h/3 H 2 m h/3 + H 1 H 2 m h/9, m 21 5 = m 21 5 h H 1 m 01 5 h/3. 24

26 As before, we can replace moments by central moments, giving µ 11 = µ 11h, µ 21 = µ 21h, 2 µ 31 = µ 31h H 1 µ 11h, µ 22 = µ 22h H1 µ 02h /3 + H 1 H 2 /9, µ 41 = µ 41h 2H 1 µ 21h, µ 32 = µ 32h H 1 µ 12h H 2 µ 30h, µ 51 = µ 51h 10H 1 µ 31h + 7H1 2 µ 11h /3, µ 42 = µ 42h 2H 1 µ 22h + 7H1 2 µ 02h /15 (H 2 /3) ( µ 40h 2H 1 µ 20h + 7H1 2 /15 ), µ 33 = µ 33 H 1 µ 13h H 2 (µ 31h H 1 µ 11h ), µ 61 = µ 61h 5H 1 µ 41h + 7H1 2 µ 21h, µ 52 = µ 52h (10H 1 /3) µ 32h + ( 7H1 2 /3 ) µ 12h (H 2 /3) [µ 50h (10H 1 /3) µ 30h ], µ 43 = µ 43 H 2 µ 41h 2H 1 (µ 23h H 2 µ 21h ) + 7H1 2 µ 03h /15, µ 111 = µ 111h, µ 211 = µ 211h H 1 µ 011h /3, µ 311 = µ 311h H 1 µ 111h /3, µ 221 = µ 221h H 1 µ 021h /3 H 2 µ 201h /3, µ 411 = µ 411h 2H 1 µ 211h /3 + 7H1 2 µ 011h /15, µ 321 = µ 321h H 1 (µ 121h H 2 µ 101h /3) H 2 µ 301h /3, 3 3 µ 222 = µ 2212 H1 µ 022h /3 + H1 H 2 µ 002h /9 H 1 H 2 H 3 /27, µ 511 = µ 511h 10H 1 µ 311h /3 + 7H1 2 µ 111h /3, µ 421 = µ 421h 2H 1 µ 221h + 7H1 2 µ 021h /15 H 2 µ 401h /9 + 2H 1 µ 201h /3, µ 331 = µ 331h H 1 µ 131h H 2 µ 311h + H 1 H 2 µ 111h, µ 1111 = µ 1111h, µ 2111 = µ 2111h H 1 µ 0111h /3, µ 3111 = µ 3111h H 1 µ 1111h /3, µ 2211 = µ 2211h H 1 µ 0111h /3 H 2 µ 2011h /3, µ 4111 = µ 4111h 2H 1 µ 2111h /3 + 7H1 2 µ 0111h /15, µ 3211 = µ 3211h H 1 µ 1211h /3 H 2 (µ 3011h /3 H 1 µ 1011h ) /3, µ 2221 = µ 2221h H 1 µ 0221h /3 H 2 µ 2021h /3 + H 1 H 2 µ 0011h /9, µ 21 4 = µ 21 4 h H 1 µ 01 4 h/3, µ 31 4 = µ 31 4 h H 1 µ 1 5 h, µ = µ h H 1 µ h/3 H 2 µ h/3 + H 1 H 2 µ h/9, µ 21 5 = µ 21 5 h H 1 µ 01 5 h/3. References [1] Abramowitz, M. and Stegun, I. A. (1964). Handbook of Mathematical Functions. United States Department of Commerce, National Bureau of Standards, Applied Mathematics Series, volume 55. [2] Fisher, R. A. (1929). Tests of significance in harmonic analysis. Proceedings of the Royal Society, A, 125, [3] Gradshteyn, I. S. and Ryzhik, I. M. (2000). Tables of Integrals, Series and Products, sixth edition. Academic Press, New York. 25

27 [4] Hoh, J. -K., Cha, K. -J., Park, M. -I., Lee, M. -L. T. and Park, Y. -S. (2012). Estimating time to full uterine cervical dilation using genetic algorithm. Kaohsiung Journal of Medical Sciences, 28, [5] James, G. S. (1958). On moments and cumulants of systems of statistics. Sankhyā, 20, [6] Kendall, M. G. (1938). Conditions under which Sheppard s corrections are valid. Journal of the Royal Statistical Society, 101, [7] Li, W. and Bai, Z. D. (2011). Analysis of accumulated rounding errors in autoregressive processes. Journal of Time Series Analysis, 32, [8] R Development Core Team (2013). R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing. Vienna, Austria. [9] Sheppard, W. F. (1899). Central-difference formulae. Proceedings of the London Mathematical Society, 31, [10] Sheppard, W. F. (1907). The calculation of moments of a frequency distribution. Biometrika, 5, [11] Stuart, A. and Ord, K. (1987). Kendall s Advanced Theory of Statistics, volume 1, fifth edition. Griffin, London. [12] Tchouproff, A. A. (1918). On the mathematical expectation of the moments of frequency distribution. Biometrika, 12, [13] Thorarinsdottir, T. L. and Johnson, M. S. (2012). Probabilistic wind gust forecasting using nonhomogeneous Gaussian regression. Monthly Weather Review, [14] Wang, A. and Wertelecki, W. (2012). Density estimation for data with rounding errors. Computational Statistics and Data Analysis, doi: /j.csda [15] Withers, C. S. (1983). Expansions for the distribution and quantiles of a regular functional of the empirical distribution with applications to nonparametric confidence intervals. Annals of Statistics, 11, [16] Withers, C. S. (1987). Bias reduction by Taylor series. Communications in Statistics Theory and Methods, 16, [17] Withers, C. S. and Nadarajah, S. (2008). Analytic bias reduction for k-sample functionals. Sankhyā, A, 70, [18] Withers, C. S. and Nadarajah, S. (2011). Unbiased estimates for linear regression with roundoff error. Probability and Mathematical Statistics, 31, [19] Withers, C. S. and Nadarajah, S. (2012a). Nonparametric estimates of low bias. REV- STAT, 10, [20] Withers, C. S. and Nadarajah, S. (2012b). Adjusting Cornish-Fisher expansions and confidence intervals for the effect of roundoff. Statistics, 46, [21] Wold, H. (1934). Sheppard s correction formula in several variables. Skandinavisk Aktuarietdskrift, 17,

EXPANSIONS FOR QUANTILES AND MULTIVARIATE MOMENTS OF EXTREMES FOR HEAVY TAILED DISTRIBUTIONS

EXPANSIONS FOR QUANTILES AND MULTIVARIATE MOMENTS OF EXTREMES FOR HEAVY TAILED DISTRIBUTIONS REVSTAT Statistical Journal Volume 15, Number 1, January 2017, 25 43 EXPANSIONS FOR QUANTILES AND MULTIVARIATE MOMENTS OF EXTREMES FOR HEAVY TAILED DISTRIBUTIONS Authors: Christopher Withers Industrial

More information

Solutions to the recurrence relation u n+1 = v n+1 + u n v n in terms of Bell polynomials

Solutions to the recurrence relation u n+1 = v n+1 + u n v n in terms of Bell polynomials Volume 31, N. 2, pp. 245 258, 2012 Copyright 2012 SBMAC ISSN 0101-8205 / ISSN 1807-0302 (Online) www.scielo.br/cam Solutions to the recurrence relation u n+1 = v n+1 + u n v n in terms of Bell polynomials

More information

EXPANSIONS FOR FUNCTIONS OF DETERMINANTS OF POWER SERIES

EXPANSIONS FOR FUNCTIONS OF DETERMINANTS OF POWER SERIES CANADIAN APPLIED MATHEMATICS QUARTERLY Volume 18, Number 1, Spring 010 EXPANSIONS FOR FUNCTIONS OF DETERMINANTS OF POWER SERIES CHRISTOPHER S. WITHERS AND SARALEES NADARAJAH ABSTRACT. Let Aɛ) be an analytic

More information

Negentropy as a Function of Cumulants

Negentropy as a Function of Cumulants Negentropy as a Function of Cumulants Christopher S. Withers & Saralees Nadarajah First version: 19 December 2011 Research Report No. 15, 2011, Probability and Statistics Group School of Mathematics, The

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

Simulating Uniform- and Triangular- Based Double Power Method Distributions

Simulating Uniform- and Triangular- Based Double Power Method Distributions Journal of Statistical and Econometric Methods, vol.6, no.1, 2017, 1-44 ISSN: 1792-6602 (print), 1792-6939 (online) Scienpress Ltd, 2017 Simulating Uniform- and Triangular- Based Double Power Method Distributions

More information

From Histograms to Multivariate Polynomial Histograms and Shape Estimation. Assoc Prof Inge Koch

From Histograms to Multivariate Polynomial Histograms and Shape Estimation. Assoc Prof Inge Koch From Histograms to Multivariate Polynomial Histograms and Shape Estimation Assoc Prof Inge Koch Statistics, School of Mathematical Sciences University of Adelaide Inge Koch (UNSW, Adelaide) Poly Histograms

More information

High-Order Corrected Trapezoidal Quadrature Rules for Functions with a Logarithmic Singularity in 2-D

High-Order Corrected Trapezoidal Quadrature Rules for Functions with a Logarithmic Singularity in 2-D High-Order Corrected Trapezoidal Quadrature Rules for Functions with a Logarithmic Singularity in 2-D Juan C. Aguilar 1 Yu Chen 2 April 24, 2002 Abstract In this report we construct correction coefficients

More information

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department

More information

Explicit polynomial expansions of regular real functions by means of even order Bernoulli polynomials and boundary values

Explicit polynomial expansions of regular real functions by means of even order Bernoulli polynomials and boundary values Journal of Computational and Applied Mathematics 176 (5 77 9 www.elsevier.com/locate/cam Explicit polynomial expansions of regular real functions by means of even order Bernoulli polynomials and boundary

More information

PARAMETER ESTIMATION FOR THE LOG-LOGISTIC DISTRIBUTION BASED ON ORDER STATISTICS

PARAMETER ESTIMATION FOR THE LOG-LOGISTIC DISTRIBUTION BASED ON ORDER STATISTICS PARAMETER ESTIMATION FOR THE LOG-LOGISTIC DISTRIBUTION BASED ON ORDER STATISTICS Authors: Mohammad Ahsanullah Department of Management Sciences, Rider University, New Jersey, USA ahsan@rider.edu) Ayman

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Multivariate Normal-Laplace Distribution and Processes

Multivariate Normal-Laplace Distribution and Processes CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by

More information

Approximating the Conway-Maxwell-Poisson normalizing constant

Approximating the Conway-Maxwell-Poisson normalizing constant Filomat 30:4 016, 953 960 DOI 10.98/FIL1604953S Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Approximating the Conway-Maxwell-Poisson

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

Econ Slides from Lecture 10

Econ Slides from Lecture 10 Econ 205 Sobel Econ 205 - Slides from Lecture 10 Joel Sobel September 2, 2010 Example Find the tangent plane to {x x 1 x 2 x 2 3 = 6} R3 at x = (2, 5, 2). If you let f (x) = x 1 x 2 x3 2, then this is

More information

LEGENDRE POLYNOMIALS AND APPLICATIONS. We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates.

LEGENDRE POLYNOMIALS AND APPLICATIONS. We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates. LEGENDRE POLYNOMIALS AND APPLICATIONS We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates.. Legendre equation: series solutions The Legendre equation is

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

A nonparametric test for trend based on initial ranks

A nonparametric test for trend based on initial ranks Journal of Statistical Computation and Simulation Vol. 76, No. 9, September 006, 89 837 A nonparametric test for trend based on initial ranks GLENN HOFMANN* and N. BALAKRISHNAN HSBC, USA ID Analytics,

More information

HIGHER ORDER CUMULANTS OF RANDOM VECTORS, DIFFERENTIAL OPERATORS, AND APPLICATIONS TO STATISTICAL INFERENCE AND TIME SERIES

HIGHER ORDER CUMULANTS OF RANDOM VECTORS, DIFFERENTIAL OPERATORS, AND APPLICATIONS TO STATISTICAL INFERENCE AND TIME SERIES HIGHER ORDER CUMULANTS OF RANDOM VECTORS DIFFERENTIAL OPERATORS AND APPLICATIONS TO STATISTICAL INFERENCE AND TIME SERIES S RAO JAMMALAMADAKA T SUBBA RAO AND GYÖRGY TERDIK Abstract This paper provides

More information

Irr. Statistical Methods in Experimental Physics. 2nd Edition. Frederick James. World Scientific. CERN, Switzerland

Irr. Statistical Methods in Experimental Physics. 2nd Edition. Frederick James. World Scientific. CERN, Switzerland Frederick James CERN, Switzerland Statistical Methods in Experimental Physics 2nd Edition r i Irr 1- r ri Ibn World Scientific NEW JERSEY LONDON SINGAPORE BEIJING SHANGHAI HONG KONG TAIPEI CHENNAI CONTENTS

More information

YALE UNIVERSITY DEPARTMENT OF COMPUTER SCIENCE

YALE UNIVERSITY DEPARTMENT OF COMPUTER SCIENCE fa it) On the Jacobi Polynomials PA ' Gregory Matviyenko Research Report YALETJ/DCS/RR-1076 June 13, 1995 DTI,* QUALITY DEFECTED 3 YALE UNIVERSITY DEPARTMENT OF COMPUTER SCIENCE & k.-. *,«.

More information

3. Probability and Statistics

3. Probability and Statistics FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important

More information

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES 1. Order statistics Let X 1,...,X n be n real-valued observations. One can always arrangetheminordertogettheorder statisticsx (1) X (2) X (n). SinceX (k)

More information

1 Solutions to selected problems

1 Solutions to selected problems Solutions to selected problems Section., #a,c,d. a. p x = n for i = n : 0 p x = xp x + i end b. z = x, y = x for i = : n y = y + x i z = zy end c. y = (t x ), p t = a for i = : n y = y(t x i ) p t = p

More information

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity International Mathematics and Mathematical Sciences Volume 2011, Article ID 249564, 7 pages doi:10.1155/2011/249564 Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity J. Martin van

More information

Chapter 11 - Sequences and Series

Chapter 11 - Sequences and Series Calculus and Analytic Geometry II Chapter - Sequences and Series. Sequences Definition. A sequence is a list of numbers written in a definite order, We call a n the general term of the sequence. {a, a

More information

K-ANTITHETIC VARIATES IN MONTE CARLO SIMULATION ISSN k-antithetic Variates in Monte Carlo Simulation Abdelaziz Nasroallah, pp.

K-ANTITHETIC VARIATES IN MONTE CARLO SIMULATION ISSN k-antithetic Variates in Monte Carlo Simulation Abdelaziz Nasroallah, pp. K-ANTITHETIC VARIATES IN MONTE CARLO SIMULATION ABDELAZIZ NASROALLAH Abstract. Standard Monte Carlo simulation needs prohibitive time to achieve reasonable estimations. for untractable integrals (i.e.

More information

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Ann Inst Stat Math (0) 64:359 37 DOI 0.007/s0463-00-036-3 Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Paul Vos Qiang Wu Received: 3 June 009 / Revised:

More information

3 Applications of partial differentiation

3 Applications of partial differentiation Advanced Calculus Chapter 3 Applications of partial differentiation 37 3 Applications of partial differentiation 3.1 Stationary points Higher derivatives Let U R 2 and f : U R. The partial derivatives

More information

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 120 minutes.

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 120 minutes. Closed book and notes. 10 minutes. Two summary tables from the concise notes are attached: Discrete distributions and continuous distributions. Eight Pages. Score _ Final Exam, Fall 1999 Cover Sheet, Page

More information

Nonparametric Regression. Changliang Zou

Nonparametric Regression. Changliang Zou Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 Chapter 11 Taylor Series Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 First-Order Approximation We want to approximate function f by some simple function. Best possible approximation

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

RESEARCH REPORT. Estimation of sample spacing in stochastic processes. Anders Rønn-Nielsen, Jon Sporring and Eva B.

RESEARCH REPORT. Estimation of sample spacing in stochastic processes.   Anders Rønn-Nielsen, Jon Sporring and Eva B. CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING www.csgb.dk RESEARCH REPORT 6 Anders Rønn-Nielsen, Jon Sporring and Eva B. Vedel Jensen Estimation of sample spacing in stochastic processes No. 7,

More information

Statistics 3657 : Moment Approximations

Statistics 3657 : Moment Approximations Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the

More information

On the positivity of linear weights in WENO approximations. Abstract

On the positivity of linear weights in WENO approximations. Abstract On the positivity of linear weights in WENO approximations Yuanyuan Liu, Chi-Wang Shu and Mengping Zhang 3 Abstract High order accurate weighted essentially non-oscillatory (WENO) schemes have been used

More information

Week 1 Quantitative Analysis of Financial Markets Distributions A

Week 1 Quantitative Analysis of Financial Markets Distributions A Week 1 Quantitative Analysis of Financial Markets Distributions A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October

More information

Time Series Analysis. Asymptotic Results for Spatial ARMA Models

Time Series Analysis. Asymptotic Results for Spatial ARMA Models Communications in Statistics Theory Methods, 35: 67 688, 2006 Copyright Taylor & Francis Group, LLC ISSN: 036-0926 print/532-45x online DOI: 0.080/036092050049893 Time Series Analysis Asymptotic Results

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Knowable moments and K-climacogram

Knowable moments and K-climacogram Lecture notes on stochastics Part 3 Rome, 3-4 May 208 Knowable moments and K-climacogram Demetris Koutsoyiannis Department of Water Resources and Environmental Engineering School of Civil Engineering National

More information

Engg. Math. II (Unit-IV) Numerical Analysis

Engg. Math. II (Unit-IV) Numerical Analysis Dr. Satish Shukla of 33 Engg. Math. II (Unit-IV) Numerical Analysis Syllabus. Interpolation and Curve Fitting: Introduction to Interpolation; Calculus of Finite Differences; Finite Difference and Divided

More information

The System of Linear Equations. Direct Methods. Xiaozhou Li.

The System of Linear Equations. Direct Methods. Xiaozhou Li. 1/16 The Direct Methods xiaozhouli@uestc.edu.cn http://xiaozhouli.com School of Mathematical Sciences University of Electronic Science and Technology of China Chengdu, China Does the LU factorization always

More information

Asymptotic statistics using the Functional Delta Method

Asymptotic statistics using the Functional Delta Method Quantiles, Order Statistics and L-Statsitics TU Kaiserslautern 15. Februar 2015 Motivation Functional The delta method introduced in chapter 3 is an useful technique to turn the weak convergence of random

More information

Bootstrap Approach to Comparison of Alternative Methods of Parameter Estimation of a Simultaneous Equation Model

Bootstrap Approach to Comparison of Alternative Methods of Parameter Estimation of a Simultaneous Equation Model Bootstrap Approach to Comparison of Alternative Methods of Parameter Estimation of a Simultaneous Equation Model Olubusoye, O. E., J. O. Olaomi, and O. O. Odetunde Abstract A bootstrap simulation approach

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Introduction to Regression

Introduction to Regression Introduction to Regression Chad M. Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/100 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression

More information

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs Martin J. Wolfsegger Department of Biostatistics, Baxter AG, Vienna, Austria Thomas Jaki Department of Statistics, University of South Carolina,

More information

Formulas for probability theory and linear models SF2941

Formulas for probability theory and linear models SF2941 Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms

More information

RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA

RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA Discussiones Mathematicae General Algebra and Applications 23 (2003 ) 125 137 RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA Seok-Zun Song and Kyung-Tae Kang Department of Mathematics,

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

Lecture 4: Numerical solution of ordinary differential equations

Lecture 4: Numerical solution of ordinary differential equations Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor

More information

EXPLICIT EXPRESSIONS FOR MOMENTS OF χ 2 ORDER STATISTICS

EXPLICIT EXPRESSIONS FOR MOMENTS OF χ 2 ORDER STATISTICS Bulletin of the Institute of Mathematics Academia Sinica (New Series) Vol. 3 (28), No. 3, pp. 433-444 EXPLICIT EXPRESSIONS FOR MOMENTS OF χ 2 ORDER STATISTICS BY SARALEES NADARAJAH Abstract Explicit closed

More information

Asymptotic Variance Formulas, Gamma Functions, and Order Statistics

Asymptotic Variance Formulas, Gamma Functions, and Order Statistics Statistical Models and Methods for Lifetime Data, Second Edition by Jerald F. Lawless Copyright 2003 John Wiley & Sons, Inc. APPENDIX Β Asymptotic Variance Formulas, Gamma Functions, and Order Statistics

More information

Introduction to Regression

Introduction to Regression Introduction to Regression p. 1/97 Introduction to Regression Chad Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/97 Acknowledgement Larry Wasserman, All of Nonparametric

More information

JUST THE MATHS UNIT NUMBER DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) A.J.Hobson

JUST THE MATHS UNIT NUMBER DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) A.J.Hobson JUST THE MATHS UNIT NUMBER.5 DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) by A.J.Hobson.5. Maclaurin s series.5. Standard series.5.3 Taylor s series.5.4 Exercises.5.5 Answers to exercises

More information

Section x7 +

Section x7 + Difference Equations to Differential Equations Section 5. Polynomial Approximations In Chapter 3 we discussed the problem of finding the affine function which best approximates a given function about some

More information

4. CONTINUOUS RANDOM VARIABLES

4. CONTINUOUS RANDOM VARIABLES IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We

More information

Matrix Multiplication

Matrix Multiplication 228 hapter Three Maps etween Spaces IV2 Matrix Multiplication After representing addition and scalar multiplication of linear maps in the prior subsection, the natural next operation to consider is function

More information

Evaluating Singular and Nearly Singular Integrals Numerically

Evaluating Singular and Nearly Singular Integrals Numerically Evaluating Singular and early Singular Integrals umerically Matthew J. Atwood Duke University Mentor: J. Thomas Beale This work is supported by SF VIGRE grant number DMS-998332. Abstract We look for a

More information

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances Advances in Decision Sciences Volume 211, Article ID 74858, 8 pages doi:1.1155/211/74858 Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances David Allingham 1 andj.c.w.rayner

More information

Monte Carlo Study on the Successive Difference Replication Method for Non-Linear Statistics

Monte Carlo Study on the Successive Difference Replication Method for Non-Linear Statistics Monte Carlo Study on the Successive Difference Replication Method for Non-Linear Statistics Amang S. Sukasih, Mathematica Policy Research, Inc. Donsig Jang, Mathematica Policy Research, Inc. Amang S. Sukasih,

More information

arxiv:math/ v1 [math.pr] 9 Sep 2003

arxiv:math/ v1 [math.pr] 9 Sep 2003 arxiv:math/0309164v1 [math.pr] 9 Sep 003 A NEW TEST FOR THE MULTIVARIATE TWO-SAMPLE PROBLEM BASED ON THE CONCEPT OF MINIMUM ENERGY G. Zech and B. Aslan University of Siegen, Germany August 8, 018 Abstract

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition) NONLINEAR PROGRAMMING (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Nonlinear Programming g Linear programming has a fundamental role in OR. In linear programming all its functions

More information

Chapter 2: Resampling Maarten Jansen

Chapter 2: Resampling Maarten Jansen Chapter 2: Resampling Maarten Jansen Randomization tests Randomized experiment random assignment of sample subjects to groups Example: medical experiment with control group n 1 subjects for true medicine,

More information

Applied Math for Engineers

Applied Math for Engineers Applied Math for Engineers Ming Zhong Lecture 15 March 28, 2018 Ming Zhong (JHU) AMS Spring 2018 1 / 28 Recap Table of Contents 1 Recap 2 Numerical ODEs: Single Step Methods 3 Multistep Methods 4 Method

More information

Goodness-of-fit tests for the cure rate in a mixture cure model

Goodness-of-fit tests for the cure rate in a mixture cure model Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,

More information

On corrections of classical multivariate tests for high-dimensional data

On corrections of classical multivariate tests for high-dimensional data On corrections of classical multivariate tests for high-dimensional data Jian-feng Yao with Zhidong Bai, Dandan Jiang, Shurong Zheng Overview Introduction High-dimensional data and new challenge in statistics

More information

LIST OF FORMULAS FOR STK1100 AND STK1110

LIST OF FORMULAS FOR STK1100 AND STK1110 LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function

More information

Linear Algebra: Lecture notes from Kolman and Hill 9th edition.

Linear Algebra: Lecture notes from Kolman and Hill 9th edition. Linear Algebra: Lecture notes from Kolman and Hill 9th edition Taylan Şengül March 20, 2019 Please let me know of any mistakes in these notes Contents Week 1 1 11 Systems of Linear Equations 1 12 Matrices

More information

Chapter 10. U-statistics Statistical Functionals and V-Statistics

Chapter 10. U-statistics Statistical Functionals and V-Statistics Chapter 10 U-statistics When one is willing to assume the existence of a simple random sample X 1,..., X n, U- statistics generalize common notions of unbiased estimation such as the sample mean and the

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

Introduction to Regression

Introduction to Regression Introduction to Regression Chad M. Schafer May 20, 2015 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures Cross Validation Local Polynomial Regression

More information

The Distributions of Sums, Products and Ratios of Inverted Bivariate Beta Distribution 1

The Distributions of Sums, Products and Ratios of Inverted Bivariate Beta Distribution 1 Applied Mathematical Sciences, Vol. 2, 28, no. 48, 2377-2391 The Distributions of Sums, Products and Ratios of Inverted Bivariate Beta Distribution 1 A. S. Al-Ruzaiza and Awad El-Gohary 2 Department of

More information

36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1

36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1 36. Multisample U-statistics jointly distributed U-statistics Lehmann 6.1 In this topic, we generalize the idea of U-statistics in two different directions. First, we consider single U-statistics for situations

More information

THIRD-MOMENT STANDARDIZATION FOR STRUCTURAL RELIABILITY ANALYSIS

THIRD-MOMENT STANDARDIZATION FOR STRUCTURAL RELIABILITY ANALYSIS THIRD-MOMENT STANDARDIZATION FOR STRUCTURAL RELIABILITY ANALYSIS By Yan-Gang Zhao and Tetsuro Ono ABSTRACT: First- and second-order reliability methods are generally considered to be among the most useful

More information

H NT Z N RT L 0 4 n f lt r h v d lt n r n, h p l," "Fl d nd fl d " ( n l d n l tr l t nt r t t n t nt t nt n fr n nl, th t l n r tr t nt. r d n f d rd n t th nd r nt r d t n th t th n r lth h v b n f

More information

Introduction to the Numerical Solution of IVP for ODE

Introduction to the Numerical Solution of IVP for ODE Introduction to the Numerical Solution of IVP for ODE 45 Introduction to the Numerical Solution of IVP for ODE Consider the IVP: DE x = f(t, x), IC x(a) = x a. For simplicity, we will assume here that

More information

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction Estimation Tasks Short Course on Image Quality Matthew A. Kupinski Introduction Section 13.3 in B&M Keep in mind the similarities between estimation and classification Image-quality is a statistical concept

More information

DEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE

DEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE Estimating the error distribution in nonparametric multiple regression with applications to model testing Natalie Neumeyer & Ingrid Van Keilegom Preprint No. 2008-01 July 2008 DEPARTMENT MATHEMATIK ARBEITSBEREICH

More information

EE4601 Communication Systems

EE4601 Communication Systems EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two

More information

Estimation of cumulative distribution function with spline functions

Estimation of cumulative distribution function with spline functions INTERNATIONAL JOURNAL OF ECONOMICS AND STATISTICS Volume 5, 017 Estimation of cumulative distribution function with functions Akhlitdin Nizamitdinov, Aladdin Shamilov Abstract The estimation of the cumulative

More information

Advanced Machine Learning & Perception

Advanced Machine Learning & Perception Advanced Machine Learning & Perception Instructor: Tony Jebara Topic 6 Standard Kernels Unusual Input Spaces for Kernels String Kernels Probabilistic Kernels Fisher Kernels Probability Product Kernels

More information

Algorithms: Background

Algorithms: Background Algorithms: Background Amotz Bar-Noy CUNY Amotz Bar-Noy (CUNY) Algorithms: Background 1 / 66 What is a Proof? Definition I: The cogency of evidence that compels acceptance by the mind of a truth or a fact.

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

Optimal global rates of convergence for interpolation problems with random design

Optimal global rates of convergence for interpolation problems with random design Optimal global rates of convergence for interpolation problems with random design Michael Kohler 1 and Adam Krzyżak 2, 1 Fachbereich Mathematik, Technische Universität Darmstadt, Schlossgartenstr. 7, 64289

More information

Power Series Solutions to the Legendre Equation

Power Series Solutions to the Legendre Equation Department of Mathematics IIT Guwahati The Legendre equation The equation (1 x 2 )y 2xy + α(α + 1)y = 0, (1) where α is any real constant, is called Legendre s equation. When α Z +, the equation has polynomial

More information

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1. MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:

More information

A correlation coefficient for circular data

A correlation coefficient for circular data BiomelriL-a (1983). 70. 2, pp. 327-32 327 Prinltd in Great Britain A correlation coefficient for circular data BY N. I. FISHER CSIRO Division of Mathematics and Statistics, Lindfield, N.S.W., Australia

More information

Finite Population Sampling and Inference

Finite Population Sampling and Inference Finite Population Sampling and Inference A Prediction Approach RICHARD VALLIANT ALAN H. DORFMAN RICHARD M. ROYALL A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane

More information

Structural Reliability

Structural Reliability Structural Reliability Thuong Van DANG May 28, 2018 1 / 41 2 / 41 Introduction to Structural Reliability Concept of Limit State and Reliability Review of Probability Theory First Order Second Moment Method

More information

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html

More information

Networks: Lectures 9 & 10 Random graphs

Networks: Lectures 9 & 10 Random graphs Networks: Lectures 9 & 10 Random graphs Heather A Harrington Mathematical Institute University of Oxford HT 2017 What you re in for Week 1: Introduction and basic concepts Week 2: Small worlds Week 3:

More information

Introduction to Regression

Introduction to Regression Introduction to Regression David E Jones (slides mostly by Chad M Schafer) June 1, 2016 1 / 102 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures

More information

A nonparametric two-sample wald test of equality of variances

A nonparametric two-sample wald test of equality of variances University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 211 A nonparametric two-sample wald test of equality of variances David

More information

1 Solutions to selected problems

1 Solutions to selected problems Solutions to selected problems Section., #a,c,d. a. p x = n for i = n : 0 p x = xp x + i end b. z = x, y = x for i = : n y = y + x i z = zy end c. y = (t x ), p t = a for i = : n y = y(t x i ) p t = p

More information