University of Toronto

Size: px
Start display at page:

Download "University of Toronto"

Transcription

1 A Limit Result for the Prior Predictive by Michael Evans Department of Statistics University of Toronto and Gun Ho Jang Department of Statistics University of Toronto Technical Report No April 15, 2010 TECHNICAL REPORT SERIES University of Toronto Department of Statistics

2 A Limit Result for the Prior Predictive Michael Evans and Gun Ho Jang Department of Statistics University of Toronto Abstract We establish results concerning the convergence of the prior predictive distribution. An application is made to the problem of checking for prior-data con ict. Keywords: minimal su ciency, prior predictive, convergence, prior-data con ict 1 Introduction Suppose we have a model given by a collection of probability measures fp : 2 g where P (A) = R A f (x) (dx); i.e., each P is absolutely continuous with respect to a support measure on the sample space X ; with the density denoted by f : With this formulation a prior leads to a prior predictive probability measure on X given by M(A) = R P (A) (d) = R A m(x) (dx); where m(x) = R f (x) (d): If T : X! T is a minimal su cient statistic for fp : 2 g; then it is well known that the posterior distribution for is the same whether we observe x or T (x) and so we denote the posterior by ( j T ). Furthermore, the conditional distribution of x given T is independent of and we denote this conditional measure by P ( j T ): The joint distribution P can then be factored as P = M ( j x) = P ( j T ) M T ( j T ) where M T is the marginal prior predictive distribution of T: If f T denotes the marginal density of T; with respect to a support measure T on T ; then m T (t) = R f T (t) (d) denotes the 1

3 density of M T with respect to T : If denotes the density of ; with respect to a support measure on ; then we can write m T (t) = R f T (t)() (d): Our concern here is with the behavior of M T as the amount of data grows. A simple example illustrates the asymptotic behavior of this distribution that we might expect to hold in more general situations. Example 1 (Location normal). Suppose that x = (x 1 ; : : : ; x n ) is a sample from a N(; 1) distribution where 2 R 1 is unknown. Then a minimal su cient statistic is given by (x) = x and (x) converges almost surely to the true value as n! 1: Suppose we put a N( 0 ; 0) 2 prior on : The prior predictive distribution M Tn is then easily obtained from x = + z where z N(0; 1=n) independent of N( 0 ; 2 0); namely, M Tn is the N( 0 ; =n) distribution. We see immediately that M Tn converges in distribution to the N( 0 ; 2 0) distribution as n! 1: Furthermore, m Tn (t) converges almost surely to (2) 1=2 1 0 expf (t 0 ) 2 =2 2 0g as n! 1; uniformly for t in a compact set. Simple computations show that these results do not depend on using a normal prior, namely, if we use prior measure with continuous density ; then M Tn converges in distribution to as n! 1; and m Tn (t) converges almost surely to as n! 1; uniformly for t in a compact set. So in Example 1 we can think of m T (T (x)) as a consistent estimator of the prior evaluated at the true value of the parameter. The signi cance of this is that the value of m T (T (x)) gives an indication of whether or not the prior has been poorly chosen, in the sense that the true value of may lie in a region where little prior probability has been assigned. Of course, we cannot tell this from the value m T (T (x)) itself but need to calibrate this on some scale. In Evans and Moshonov (2006, 2007) the P-value M T (m T (t) m T (T (x))); (1) and some variations of this, was proposed for checking for prior-data con ict. Note that this P-value is a modi cation of a P-value proposed by Box (1980) for general model checking in Bayesian contexts. In Example 1, when using the normal prior, we see that 2

4 (1) converges to 2(1 (j 0 j= 0 )) as n! 1; where is the true value of : So (1) is a consistent estimator of the P-value which measures whether the true value of the parameter lies in the tails of the prior. In Section 2 we prove that (1) converges to (() ( )); where is the true value of ; in fairly general circumstances. The P-value (() ( )) will be small whenever the true value lies in a region of low prior probability and so we have an instance of prior-data con ict. As such, (1) is seen to be an appropriate measure of prior-data con ict. A criticism of (1) is that, in the case of continuous models at least, the P-value is not invariant under smooth transformations. In particular, suppose that W : T! W is 1-1 and smooth. Let J W (t) be the reciprocal of the Jacobian determinant of W evaluated at t: Then, if instead of T we use W (T ) as the minimal su cient statistic, the P-value is M W (m W (w) m W (W (T (x)))) = M T (m T (t)j W (t) m T (T (x))j W (T (x))) and this will not equal (1) unless J W (t) is constant. In Evans and Jang (2010) the general problem of computing P-values, based on the density of a discrepancy statistic, to assess whether or not the data came from a single xed distribution, was considered. An invariant P-value was proposed. For (1) this entails using instead the P-value M T (m T (t) m T (T (x))); (2) where m T (t) = m T (t)e(j 1 T (X) j T (X) = t); J T (x) = j det(dt (x) dt 0 (x))j 1=2 and dt denotes the di erential of T: The factor E(J 1 T (X) j T (X) = t) corrects for volume distortions due to the transformation T and is independent of because T is minimal su cient. Note that m T (t) is the density of M T with respect to the support measure (E(J 1 T (X) j T (X) = t)) 1 (dt): In Example 1 we have that J T (x) is constant and so (1) equals (2). While the P-value (2) will generally di er from (1), it is often the case that the e ect of E(J 1 T (X) j T (X) = t) is negligible. We establish a convergence result for (2) in Section 3. While there are numerous discussions concerning asymptotics for a posterior analysis, for example, Walker (1969), Heyde and Johnstone (1979), and Chen (1985), there seem 3

5 to be almost no discussions concerned with convergence issues associated with the prior predictive distribution. Such results also have implications for methods that choose the prior based on the prior predictive. This paper addresses some of these problems. 2 Convergence of the Basic P-value In the Appendix we provide the proof of the following result. Theorem 1. Suppose is an open subset of a Euclidean space and assume (i)! a.s. P for every ; (ii) m Tn (t)! (t) uniformly on compact subsets of ; (iii) is continuous and the prior distribution of () has no atoms, then M Tn (m Tn (t) m Tn ( (x n )))! (() ( )) a.s. P where is the true value of : Note that Theorem 1 implicitly assumes that the sampling model for is continuous. We will subsequently discuss how to handle the discrete case. To apply this result we need to establish (ii). We discuss several examples. Example 2 (Scale-Gamma). Let x = (x 1 ; : : : ; x n ) be a sample from a Gamma( 0 ; ) distribution where the scale parameter > 0 is unknown. Then the statistic (x) = (n 0 ) 1 P n i=1 x i Gamma(n 0 ; =(n 0 )) is minimal su cient and it converges almost surely to the true value of : We prove the following result in the Appendix. Lemma 2. If (x) Gamma(n 0 ; =(n 0 )) and the prior on is continuous, then (ii) of Theorem 1 holds. So if, in addition the prior distribution on (0; 1) has no atoms, then Theorem 1 applies and we have the convergence of (1). Certainly these conditions apply to the commonly used priors on a scale parameter. The following example uses Example 2 in a problem of considerable importance for statistical practice. 4

6 Example 3 (Normal linear regression). We consider rst the situation where we have a sample x = (x 1 ; : : : ; x n ) from a N(; 2 ) distribution with 2 R 1 and > 0 unknown. Then (x) = (T 1n (x); T 2n (x)) = (x; s 2 ) is a minimal su cient statistic and (x)! (; 2 ) as n! 1: We prove the following result in the Appendix. Lemma 3. If (x) = (x; s) where x N(; 2 =n) independent of s 2 Gamma((n 1)=2; 2 2 =(n 1)) and the prior on (; 2 ) is continuous, then (ii) of Theorem 1 holds. The prior is commonly prescribed by rst stating a prior 2 for 2 and then a conditional prior 1 ( j 2 ) for : As discussed in Evans and Moshonov (2006, 2007), it then makes sense to check 2 rst and, if 2 passes, then check 1 : With this approach we can learn more about where the prior is de cient, if indeed there is a problem: Following that development, the check for 2 is based on T 2n (x) = s 2 via the P-value M T2n (m T2n (t 2 ) m T2n (T 2n (x))): Now T 2n (x) Gamma((n 1)=2; 2 2 =(n 1)): So, when 2 satis es the conditions of Theorem 1, Example 2 applies with 0 = 1=2 and M T2n (m T2n (t 2 ) m T2n (T 2n (x)))! 2 ( 2 ( 2 ) 2 ()): 2 To check 1 the relevant P- value to use is M T1n (m T1n (t 1 j T 2n (x) = s 2 ) m T1n (T 1n (x) j T 2n (x) = s 2 ) j T 2n (x) = s 2 ): (3) Now consider m T1n (t 1 j T 2n = t 2 ) = m Tn (t 1 ; t 2 )=m T2n (t 2 ): Since m Tn (t 1 ; t 2 )! (t 1 ; t 2 ) and m T2n (t 2 )! 2 (t 2 ) uniformly on compact sets, then we have that m T1n (t 1 j T 2n = t 2 )! (t 1 ; t 2 )= 2 (t 2 ) = 1 (t 1 j t 2 ) uniformly on compact sets. Furthermore, the measures M T1n ( j T 2n = s 2 ) converge in distribution to 1 (t 1 j ); 2 m T1n (t 1 j T 2n = s 2 ) converges almost surely N( ; ) 2 to 1 (t 1 j ); 2 and m T1n (x j T 2n = s 2 ) converges almost surely N( ; ) 2 to 1 ( j ): 2 This implies the convergence almost surely N( ; ) 2 of (3) to 1 ( 1 ( j ) 2 1 ( j ) 2 j ): 2 For a normal linear regression model y n = X n + e; where X n 2 R nk and e N n (0; 2 I); we have that (y) = (T 1n (y); T 2n (y)) = (b; s 2 ) where b = (XnX 0 n ) 1 Xny 0 and s 2 = (n k) 1 jjy Xbjj 2 : Under suitable conditions on the X n matrices we have that! (; 2 ) almost surely: The convergence results for this situation then proceed 5

7 just as in as in the location-scale case. The following result, with the proof provided in the Appendix, is sometimes useful in establishing condition (ii) in Theorem 1. Theorem 4. Suppose is an open subset of R k ; and for any > 0; K compact there exist c 1, c 2 > 0 and N > 0 such that f ;Tn (t) c 1 e c2n whenever t 2 K; jjt jj > ; and n N: (4) The following are equivalent: (a) for any prior with continuous density ; then m Tn (t)! (t) uniformly for t 2 K (b) for any compact K and > 0; then Z f ;Tn (t)i(jjt jj < ) d! 1 as n! 1; uniformly for t 2 K: (5) We now consider an example where the distribution of is on a discrete subset of R 1 : In such a case we can t expect condition (ii) of Theorem 1 to hold at values of t where (t) > 0 but m Tn (t) = 0 for all n: Suppose, however, that has a lattice distribution with lattice step equal to h: Then for kh < t (k +1)h we de ne m cont (t) = m Tn ((k + 1)h)=h and 0 otherwise and treat m cont (t) as a density with respect to length measure. Since (x) is always on the lattice we see immediately that M cont (m cont (t) m cont ( (x))) = M Tn (m Tn (t) m Tn ( (x))): We can then apply Theorem 1 to m cont and this proves the convergence of (1). Note that m cont (t) = R f ;T cont n (t) (d) where f;t cont n (t) = f ;Tn ((k + 1)h)=h when kh < t (k + 1)h: Example 4 (Binomial). Suppose that x = (x 1 ; : : : ; x n ) is a sample from a Bernoulli() distribution where 2 (0; 1) is unknown. Then (x) = x is minimal su cient and converges to : For t 2 f0; 1=n; : : : ; 1g then f ;Tn (t) = P ( = t) = n nt nt (1 ) n(1 t) : In this case x has a discrete distribution on the lattice with step size equal to 1=n: In the Appendix we prove the following result. 6

8 Lemma 5. If n (x) Binomial(n; ) and is a continuous on (0; 1), then (ii) of Theorem 1 holds for m cont : Therefore, m cont (t) converges to (t) uniformly on each compact set and (1) converges provided the prior satis es the conditions of Theorem 1. One interesting case where does not satisfy the conditions of Theorem 1 arises when Uniform(0; 1) as the prior distribution of () has all of its mass at 1. In this case, however, we have that m cont (t) n=(n+1)! 1 uniformly for all t 2 (0; 1) and moreover M Tn (m Tn (t) m Tn ( (x))) = 1 = (() ( )) and so the convergence result is obvious. 3 Convergence of the Invariant P-value We now consider the convergence of (2). As noted, this P-value is invariant under smooth transformations and will agree with (1) whenever T is linear or the sampling model for T is discrete. This applies in Examples 1, 2, and 4 but not in Example 3. Example 5 (Normal linear regression). Consider the location-scale case. Clearly T 1n (x) = x is linear, and so the P-value for checking 1 agrees with the invariant version. But T 2n (x) = s 2 is nonlinear and so the P-value (1) for checking 2 is not the same as the invariant version. In this case dt 2n (x) = (2=(n 1))(x 1 x; : : : ; x n x) giving J T2n (x) = j det(dt 2n (x) dt 0 2n(x))j 1=2 = ( p n 1=2)s 1 and so E(J 1 T 2n (X) j (X) = (x; s 2 )) = (2= p n 1)s: Therefore, the invariant P-value is equal to M T2n (m T2n (t 2 )t 1=2 2 m T2n (T 2n (x))(t 2n (x)) 1=2 ) and this converges almost surely to 2 ( 2 ( 2 ) 2 ( 2 ) ) (see Theorem 6). The proof of the following result is virtually identical to that of Theorem 1. Theorem 6. Suppose is an open subset of a Euclidean space and assume (i)! a.s. P for every ; 7

9 (ii) w n (t) = E(J 1 (X) j (X) = t) is continuous and a n w n (t)! w(t) for some sequence a n ; (iii) a n m Tn (t)w n (t)! (t)w(t) uniformly on compact subsets of ; (iv) is continuous and the prior distribution of ()w() has no atoms, then M Tn (m (t) m ( (x n )))! (()w() ( )w( )) a.s. P where is the true value of : Note that ()w() is the density of with respect to the support measure (w()) 1 (d): Also note that when is k-dimensional and p n( ) is asymptotically normal, then, in many cases, we can take a n = n k=2 : The developments in this paper have required that the minimal su cient statistic be a consistent estimator of : The existence of such a minimal su cient statistic is guaranteed for exponential models. Suppose, however, that we reparameterize via the 1-1, smooth function ; namely, = (): Then we must replace by ( ) for the convergence results to hold as stated. If is nonlinear, however, then (1) will typically depend on whether we use or ( ); namely, it will implicitly depend on the parameterization. Using (2) this dependence is avoided and the P-value is independent of the choice of the minimal su cient statistic or equivalently the parameterization. The use of (2) seems more appropriate than (1) for this reason, although there is typically very little di erence in the P-values obtained. 4 Conclusions We have established convergence results for various prior predictive P-values that show directly that these are appropriate for checking for prior-data con ict, namely, assessing if the true value of the parameter is in the tails of the prior. Essentially these results are restricted to situations where a version of the minimal su cient statistic is a consistent estimate of the model parameter and this means our results apply in the context of exponential models. Similar results can undoubtedly be established in other contexts, in 8

10 particular for group models, and these are currently being developed. More generally convergence results for the prior predictive have implications for empirical Bayes methods. For example, suppose we have a family of priors f : 2 Ag with corresponding prior predictives m ;Tn for a minimal su cient statistic : Then convergence of m ;Tn ( (x)) to ( ) has the implication that maximizing m ;Tn ( (x)) over to select the prior is essentially nding the prior that has maximal value at the true value of the parameter. One might argue that it makes more sense to maximize M ;Tn (m ;Tn (t) m ;Tn ( (x))) over as then, based on the convergence of this P- value to ( () ( )); we are nding the prior for which the true value is least surprising. The implications of this are currently being investigated. Appendix Proof of Theorem 1 Let > 0: Then there exists > 0 such that j(t) ( 0 )j < =2 whenever t 2 B ( 0 ) and there exists N 1 such that for all n > N 1 ; (x n ) 2 B ( 0 ): Also there exists N 2 such that for all n > N 2 and for all t 2 B ( 0 ); then jm Tn (t) (t)j < =2: So, if n > maxfn 1 ; N 2 g then M Tn (m Tn (t) ( 0 ) ) M Tn (m Tn (t) m Tn ( (x n ))) M Tn (m Tn (t) ( 0 ) + ): Now we prove that M Tn (m Tn (t) ( 0 ) )! (() ( 0 ) ): Let 0 > 0: Let C be compact such that 0 2 C; (@C) = 0; and (C) 1 0 =2: By (i) and Slutsky s Theorem M Tn converges in law to. Therefore, M Tn (C)! (C) and so there exists N 3 such that for all n > N 3 ; M Tn (C) > 1 0 : Therefore, for all n > N 3 ; M Tn (m Tn (t) ( 0 ) ) 0 M Tn (m Tn (t) ( 0 ) ; C) M Tn (m Tn (t) ( 0 ) ) and we can make the LHS and RHS as close as we like by choosing 0 small. Let 00 > 0: There exists N 4 such that for all n > N 4 then jm Tn (t) (t)j < 00 for all t 2 C: When n > maxfn 3 ; N 4 g then, M Tn ((t) ( 0 ) 00 ; C) M Tn (m Tn (t) ( 0 ) ; C) M Tn ((t) ( 0 ) + 00 ; C) and the LHS converges to ((t) ( 0 ) 00 ; C) while the RHS converges to ((t) ( 0 ) + 00 ; C): By choosing 00 small we can make these quantities as close to ((t) ( 0 ) ; C) as we 9

11 like. This proves that M Tn (m Tn (t) ( 0 ) ; C)! ((t) ( 0 ) ; C) and this establishes that M Tn (m Tn (t) ( 0 ) )! ((t) ( 0 ) ): A similar argument shows that M Tn (m Tn (t) ( 0 )+)! ((t) ( 0 )+) and this completes the proof. Proof of Lemma 2 Suppose K is a compact set in (0; 1): Then, there are 0 < a < b < 1 such that K [a; b]. Fix > 0 satisfying < minfa=3; 1g. We prove (4) and (5), then apply Theorem 4. For t >, t= t=(t+) 1 =(b+). Also for t <, t= t=(t ) 1+=(a ). Hence, jt= 1j > = =(b+) for jt j > and t 2 K. Note (n 0 ) p 2(n0 ) n0 1=2 e n0. Since ue u+1 has peak 1 at u = 1, 1 = sup u:ju 1j> ue u+1 < 1. Also there exists N 1 > 1 such that n 1=2 n0=2 1 1 for all n N 1. Let u = t=. Then for jt j >, we get ju 1j > and f ;Tn (t) = [(n 0 ) n0 e n0 = (n 0 )t](ue u+1 ) n0 a 1 (n 0 =2) 1=2 n0 1 a 1 ( 0 =2) 1=2 e n2 1 0 log(1= 1). Hence, (4) holds. For (5), let I 0 = R f ;Tn(t)I(jt j < ) d R 1 0 [(n 0) n0 = (n 0 )]u n0 2 e n0u du = n 0 =(n 0 1). Also for t 2 K, I(jt j < ) I(jt= 1j < 1 ) I(jt= 1j < 1 n 1=2 log n) where 1 = =(b + ). Then, I 0 (n Z 0) n0 e n0 u n0 2 e n0(u 1) I(ju 1j < 1 ) du: (n 0 ) For ju 1j < 1 n 1=2 log n, a lower bound of the logarithm of the integrand is given by log(u n0 2 e n0(u 1) ) = n 0 (u 1) + (n 0 2) log(1 (1 u)) n 0 (n 1) (n 0 2)(1 u + (1 u) 2 =2 + j1 uj 3 ) n 0 (1 u) 2 =2 (2 + 0 (log n) 2 )j1 uj: The change of variable v = p n 0 (u I 0 (n 0) n0 1=2 Z e n0 (n 0 ) By Stirling s formula (n 0 ) n0 1) gives e v2 =2 (2+ 0(log n) 2 )jvj= p n 0 I(jvj < 1 1=2 0 log n) dv: 1=2 e n0 = (n 0 )! (2) 1=2. The integral converges to R e v 2 =2 dv by the Lebesgue dominated convergence theorem. Hence, I 0! 1 as n! 1. Thus (5) holds and, by Theorem 4, Theorem 1 (ii) holds. Proof of Lemma 3 We prove (4) and (5). The density f ; 2(x; s 2 ) of = (X; S 2 ) is given by ((n 1)=2) (n 1)=2 (n 1)=2 e h s 2 (2=n) 1=2 ((n 1)=2)(s 2 ) 3=2 n 1 2 exp n ( s2 2 1) (x ) 2 2 i n=2 : 10

12 Let I 1 be the rst part and I 2 be the part inside the brackets. To prove (4), x a compact set K. Let a = inffs 2 : (x; s 2 ) 2 Kg > 0 and b = supfs 2 : (x; s 2 ) 2 Kg < 1. Also we consider 0 < < min(a=3; 1). Then, Stirling s formula gives I 1 =n! 1=[(2s 2 ) 3=2 ] 1=[(2a) 3=2 ]. For (x; s 2 ) 2 K and jj(x; s 2 ) (; 2 )jj >, we have jx j > =2 or js 2 2 j > =2. Then there exists > 0 such that js 2 = 2 1j >. Since ve (v 1)(n 1)=n is unimodal having peak e 1=n n=(n 1) at v = n=(n 1), an upper bound of I 2 is obtained at s 2 = 2 = 1 + or s 2 = 2 = 1 provided by n > 1 + 1=. So I 2 max((1 ) exp(n=(n 1)); (1 + ) exp( n=(n 1))) < 1. If js 2 = 2 1j, then jx j > =2. Thus (x ) 2 = 2 = (x ) 2 (s 2 = 2 )=s 2 (=2) 2 (1 )=b. So I 2 e 1=n (n=(n 1)) exp( 2 (1 )=4b) < 1 when n > (1 + exp( 2 (1 )=4b)) 1. Hence (4) holds. The integration range jj(x; s 2 ) (; 2 )jj < contains jx j < =2 and js 2 2 j < =2. Again, this region contains jx j= < 2 = =( + 2b). Then, Z Z I 3 = f ; 2(x; s 2 )I(jj(x; s 2 ) (; 2 )jj < ) d d 2 Z Z (2 2 =n) 1=2 exp( (n=2 2 )(x ) 2 )I(jx j= < 2 ) ((n 1)=2) (n 1)=2 (n 1)=2 1 s e ((n 1)=2)( 2 ) (n 1)=2 (n 1)s 2 =2 2 I(js 2 2 j < =2) d d 2 Using v = p n(x )=, this integral can be separated into two parts, let I 4 and I 5 be the two integrals. Then, I 4 = ( p n 2 ) ( p n 2 )! 1 as n! 1; and I 5! 1 as n! 1 by Lemma 2 for 0 = 1=2. So I 3! 1 and (5) holds. Finally, Theorem 1 (ii) holds by Theorem 4. Proof of Theorem 4 Suppose (a) holds. If > 0 and K is compact, then R f ;Tn (t)()i(jjt jj ) d R c1 e c2n I(jjt jj ) (d) c 1 e c2n! 0 uniformly in t 2 K: So if (a) holds, then j R f ;Tn (t)()i(jjt jj < ) d (t)j! 0 uniformly for t 2 K: The set K = f : jj 0 jj for some 0 2 Kg is also compact when is small enough. Note that, for given ; the convergence in (5) follows whenever this convergence holds for a smaller value of : Let be a continuous density that is constant and positive on K ; then (b) follows. 11

13 Now suppose (b) holds, has a continuous density, K is a compact subset of and > 0. Then K is also compact for is small enough. Since is uniformly continuous on K ; there exists 0 > 0 such that j( 1 ) ( 2 )j < =4 whenever 1 ; 2 2 K and jj 1 2 jj < 0. From (b), there exists L 1 > 0 such that j R f ;Tn (t)i(jjt jj < ) d 1j < =(4 sup t2k (t)) for all n L 1 and t 2 K where = min( 0 ; ): Also, there exist c 1, c 2 > 0 and L 2 > 0 such that f ;Tn (t) c 1 e c2n whenever t 2 K; jjt jj ; n L 2 : Therefore, there exists L 3 such that R f ;Tn (t)()i(jjt jj ) d c 1 e c2n =4 for all n L 3. Finally, for n L = max(l 1 ; L 2 ; L 3 ) and t 2 K; we have that jm Tn (t) (t)j R f ;Tn (t)()i(jjt jj ) d + R f ;Tn (t)j() (t)ji(jjt jj < ) d + (t)j R f ;Tn (t)i(jjt jj < ) d 1j =4 + (=4)(1 + =4) + =4 < and we see that (a) holds. Proof of Lemma 5 Since K is compact, there is a > 0 such that 0 < a t 1 a < 1 for all t 2 K. For t 2 K; Stirling s formula implies that log nt n = 2 1 log 2nt(1 t) n(t log t + (1 t) log(1 t)) + r(n; t) where r(n; t) < (12n) 1 (12nt + 1) 1 (12n(1 t) + 1) 1 < 1. So we have that f ;Tn (t) = (2t(1 t)=n) 1=2 exp(r(n; t)) exp(ng(; t)) where g(; t) = t log(=t) + (1 t) log((1 )=(1 t)): Let 0 < < a=2: Note g(; t) has maximum value 0 at = t. Therefore, since g(; t) is continuous when jt j > ; we have that b = sup t;:jt j> g(; t) > 0: Also there is N 1 so that n 1=2 e bn=2 1 for n N 1. So, when t 2 K, jt j > and n N 1, f ;Tn (t) (2a(1 a)) 1=2 e exp( nb=2). Hence (4) holds. Let I 0 = R 1 0 f cont ; (t)i(jt j < ) d. Then, I 0 R 1 0 f cont ; (t) d = n n nt R 1 0 nt (1 ) n(1 t) d = n=(n + 1). Thus, using (4), we get I 0 = n n + 1 Z 1 0 f cont ; (t)i(jt j > ) d n n + 1 c 1 e c2n! 1: Hence, (5) holds. So Theorem 1 (ii) holds by Theorem 4. 12

14 References Box, G. E. P., Sampling and Bayes inference in scienti c modelling and robustness. J. Roy. Statist. Soc. Ser. A 143 (4), , with discussion. Chen, C. F., On asymptotic normality of limiting density functions with Bayesian implications. J. Roy. Statist. Soc. Ser. B 47 (3), Evans, M., Jang, G. H., Invariant P -values for model checking. Ann. Statist. 38 (1), Evans, M., Moshonov, H., Checking for prior-data con ict. Bayesian Anal. 1 (4), Evans, M., Moshonov, H., Checking for prior-data con ict with hierarchically speci ed priors. In: Upadhyay, A., Singh, U., Dey, D. (Eds.), Bayesian Statistics and its Applications. Anamaya Publishers, New Delhi, pp Heyde, C. C., Johnstone, I. M., On asymptotic posterior normality for stochastic processes. J. Roy. Statist. Soc. Ser. B 41 (2), Walker, A. M., On the asymptotic behaviour of posterior distributions. J. Roy. Statist. Soc. Ser. B 31,

Measuring Statistical Evidence Using Relative Belief

Measuring Statistical Evidence Using Relative Belief Measuring Statistical Evidence Using Relative Belief Michael Evans Department of Statistics University of Toronto Abstract: A fundamental concern of a theory of statistical inference is how one should

More information

3 Random Samples from Normal Distributions

3 Random Samples from Normal Distributions 3 Random Samples from Normal Distributions Statistical theory for random samples drawn from normal distributions is very important, partly because a great deal is known about its various associated distributions

More information

Stochastic Processes

Stochastic Processes Stochastic Processes A very simple introduction Péter Medvegyev 2009, January Medvegyev (CEU) Stochastic Processes 2009, January 1 / 54 Summary from measure theory De nition (X, A) is a measurable space

More information

Appendix for "O shoring in a Ricardian World"

Appendix for O shoring in a Ricardian World Appendix for "O shoring in a Ricardian World" This Appendix presents the proofs of Propositions - 6 and the derivations of the results in Section IV. Proof of Proposition We want to show that Tm L m T

More information

MC3: Econometric Theory and Methods. Course Notes 4

MC3: Econometric Theory and Methods. Course Notes 4 University College London Department of Economics M.Sc. in Economics MC3: Econometric Theory and Methods Course Notes 4 Notes on maximum likelihood methods Andrew Chesher 25/0/2005 Course Notes 4, Andrew

More information

Desire-as-belief revisited

Desire-as-belief revisited Desire-as-belief revisited Richard Bradley and Christian List June 30, 2008 1 Introduction On Hume s account of motivation, beliefs and desires are very di erent kinds of propositional attitudes. Beliefs

More information

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t ) LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data

More information

Supplemental Material 1 for On Optimal Inference in the Linear IV Model

Supplemental Material 1 for On Optimal Inference in the Linear IV Model Supplemental Material 1 for On Optimal Inference in the Linear IV Model Donald W. K. Andrews Cowles Foundation for Research in Economics Yale University Vadim Marmer Vancouver School of Economics University

More information

ON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS. Abstract. We introduce a wide class of asymmetric loss functions and show how to obtain

ON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS. Abstract. We introduce a wide class of asymmetric loss functions and show how to obtain ON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS FUNCTIONS Michael Baron Received: Abstract We introduce a wide class of asymmetric loss functions and show how to obtain asymmetric-type optimal decision

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

Linearized methods for ordinary di erential equations

Linearized methods for ordinary di erential equations Applied Mathematics and Computation 104 (1999) 109±19 www.elsevier.nl/locate/amc Linearized methods for ordinary di erential equations J.I. Ramos 1 Departamento de Lenguajes y Ciencias de la Computacion,

More information

Checking for Prior-Data Conflict

Checking for Prior-Data Conflict Bayesian Analysis (2006) 1, Number 4, pp. 893 914 Checking for Prior-Data Conflict Michael Evans and Hadas Moshonov Abstract. Inference proceeds from ingredients chosen by the analyst and data. To validate

More information

Chapter 1. GMM: Basic Concepts

Chapter 1. GMM: Basic Concepts Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating

More information

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.) Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori

More information

MA 8101 Stokastiske metoder i systemteori

MA 8101 Stokastiske metoder i systemteori MA 811 Stokastiske metoder i systemteori AUTUMN TRM 3 Suggested solution with some extra comments The exam had a list of useful formulae attached. This list has been added here as well. 1 Problem In this

More information

A NONPARAMETRIC TEST FOR SERIAL INDEPENDENCE OF REGRESSION ERRORS

A NONPARAMETRIC TEST FOR SERIAL INDEPENDENCE OF REGRESSION ERRORS A NONPARAMETRIC TEST FOR SERIAL INDEPENDENCE OF REGRESSION ERRORS Miguel A. Delgado and Juan Mora WP-AD 99-28 Correspondence to J. Mora: University of Alicante. Departamento de Fundamentos del Análisis

More information

E cient Simulation and Conditional Functional Limit Theorems for Ruinous Heavy-tailed Random Walks

E cient Simulation and Conditional Functional Limit Theorems for Ruinous Heavy-tailed Random Walks E cient Simulation and Conditional Functional Limit Theorems for Ruinous Heavy-tailed Random Walks Jose Blanchet and Jingchen Liu y June 14, 21 Abstract The contribution of this paper is to introduce change

More information

Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models

Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models Ryan Greenaway-McGrevy y Bureau of Economic Analysis Chirok Han Korea University February 7, 202 Donggyu Sul University

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Applications of Subsampling, Hybrid, and Size-Correction Methods

Applications of Subsampling, Hybrid, and Size-Correction Methods Applications of Subsampling, Hybrid, and Size-Correction Methods Donald W. K. Andrews Cowles Foundation for Research in Economics Yale University Patrik Guggenberger Department of Economics UCLA November

More information

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 Instructions: Answer all four (4) questions. Point totals for each question are given in parenthesis; there are 00 points possible. Within

More information

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43 Panel Data March 2, 212 () Applied Economoetrics: Topic March 2, 212 1 / 43 Overview Many economic applications involve panel data. Panel data has both cross-sectional and time series aspects. Regression

More information

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University

More information

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor Aguirregabiria

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor Aguirregabiria ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor guirregabiria SOLUTION TO FINL EXM Monday, pril 14, 2014. From 9:00am-12:00pm (3 hours) INSTRUCTIONS:

More information

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,

More information

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1815 COWLES FOUNDATION FOR RESEARCH IN ECONOMICS

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

n E(X t T n = lim X s Tn = X s

n E(X t T n = lim X s Tn = X s Stochastic Calculus Example sheet - Lent 15 Michael Tehranchi Problem 1. Let X be a local martingale. Prove that X is a uniformly integrable martingale if and only X is of class D. Solution 1. If If direction:

More information

13 Endogeneity and Nonparametric IV

13 Endogeneity and Nonparametric IV 13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,

More information

ECON0702: Mathematical Methods in Economics

ECON0702: Mathematical Methods in Economics ECON0702: Mathematical Methods in Economics Yulei Luo SEF of HKU January 14, 2009 Luo, Y. (SEF of HKU) MME January 14, 2009 1 / 44 Comparative Statics and The Concept of Derivative Comparative Statics

More information

c 2007 Je rey A. Miron

c 2007 Je rey A. Miron Review of Calculus Tools. c 007 Je rey A. Miron Outline 1. Derivatives. Optimization 3. Partial Derivatives. Optimization again 5. Optimization subject to constraints 1 Derivatives The basic tool we need

More information

Scattering for the NLS equation

Scattering for the NLS equation Scattering for the NLS equation joint work with Thierry Cazenave (UPMC) Ivan Naumkin Université Nice Sophia Antipolis February 2, 2017 Introduction. Consider the nonlinear Schrödinger equation with the

More information

Testing for Regime Switching: A Comment

Testing for Regime Switching: A Comment Testing for Regime Switching: A Comment Andrew V. Carter Department of Statistics University of California, Santa Barbara Douglas G. Steigerwald Department of Economics University of California Santa Barbara

More information

Likelihood inference for a nonstationary fractional autoregressive model

Likelihood inference for a nonstationary fractional autoregressive model Likelihood inference for a nonstationary fractional autoregressive model Søren Johansen y University of Copenhagen and CREATES and Morten Ørregaard Nielsen z Cornell University and CREATES Preliminary

More information

Time is discrete and indexed by t =0; 1;:::;T,whereT<1. An individual is interested in maximizing an objective function given by. tu(x t ;a t ); (0.

Time is discrete and indexed by t =0; 1;:::;T,whereT<1. An individual is interested in maximizing an objective function given by. tu(x t ;a t ); (0. Chapter 0 Discrete Time Dynamic Programming 0.1 The Finite Horizon Case Time is discrete and indexed by t =0; 1;:::;T,whereT

More information

Optimal Target Criteria for Stabilization Policy

Optimal Target Criteria for Stabilization Policy Optimal Target Criteria for Stabilization Policy Marc P. Giannoni Columbia University y Michael Woodford Columbia University z February 7, Abstract This paper considers a general class of nonlinear rational-expectations

More information

Exact unconditional tests for a 2 2 matched-pairs design

Exact unconditional tests for a 2 2 matched-pairs design Statistical Methods in Medical Research 2003; 12: 91^108 Exact unconditional tests for a 2 2 matched-pairs design RL Berger Statistics Department, North Carolina State University, Raleigh, NC, USA and

More information

GMM estimation of spatial panels

GMM estimation of spatial panels MRA Munich ersonal ReEc Archive GMM estimation of spatial panels Francesco Moscone and Elisa Tosetti Brunel University 7. April 009 Online at http://mpra.ub.uni-muenchen.de/637/ MRA aper No. 637, posted

More information

Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models

Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models Fall 22 Contents Introduction 2. An illustrative example........................... 2.2 Discussion...................................

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

h=1 exp (X : J h=1 Even the direction of the e ect is not determined by jk. A simpler interpretation of j is given by the odds-ratio

h=1 exp (X : J h=1 Even the direction of the e ect is not determined by jk. A simpler interpretation of j is given by the odds-ratio Multivariate Response Models The response variable is unordered and takes more than two values. The term unordered refers to the fact that response 3 is not more favored than response 2. One choice from

More information

Bayesian Modeling of Conditional Distributions

Bayesian Modeling of Conditional Distributions Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings

More information

Volume 30, Issue 3. Monotone comparative statics with separable objective functions. Christian Ewerhart University of Zurich

Volume 30, Issue 3. Monotone comparative statics with separable objective functions. Christian Ewerhart University of Zurich Volume 30, Issue 3 Monotone comparative statics with separable objective functions Christian Ewerhart University of Zurich Abstract The Milgrom-Shannon single crossing property is essential for monotone

More information

Vector fields Lecture 2

Vector fields Lecture 2 Vector fields Lecture 2 Let U be an open subset of R n and v a vector field on U. We ll say that v is complete if, for every p U, there exists an integral curve, γ : R U with γ(0) = p, i.e., for every

More information

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean

More information

The Degree of the Splitting Field of a Random Polynomial over a Finite Field

The Degree of the Splitting Field of a Random Polynomial over a Finite Field The Degree of the Splitting Field of a Random Polynomial over a Finite Field John D. Dixon and Daniel Panario School of Mathematics and Statistics Carleton University, Ottawa, Canada fjdixon,danielg@math.carleton.ca

More information

8 Periodic Linear Di erential Equations - Floquet Theory

8 Periodic Linear Di erential Equations - Floquet Theory 8 Periodic Linear Di erential Equations - Floquet Theory The general theory of time varying linear di erential equations _x(t) = A(t)x(t) is still amazingly incomplete. Only for certain classes of functions

More information

LOGARITHMIC MULTIFRACTAL SPECTRUM OF STABLE. Department of Mathematics, National Taiwan University. Taipei, TAIWAN. and. S.

LOGARITHMIC MULTIFRACTAL SPECTRUM OF STABLE. Department of Mathematics, National Taiwan University. Taipei, TAIWAN. and. S. LOGARITHMIC MULTIFRACTAL SPECTRUM OF STABLE OCCUPATION MEASURE Narn{Rueih SHIEH Department of Mathematics, National Taiwan University Taipei, TAIWAN and S. James TAYLOR 2 School of Mathematics, University

More information

A Remark on Complete Convergence for Arrays of Rowwise Negatively Associated Random Variables

A Remark on Complete Convergence for Arrays of Rowwise Negatively Associated Random Variables Proceedings of The 3rd Sino-International Symposium October, 2006 on Probability, Statistics, and Quantitative Management ICAQM/CDMS Taipei, Taiwan, ROC June 10, 2006 pp. 9-18 A Remark on Complete Convergence

More information

Stochastic Processes (Master degree in Engineering) Franco Flandoli

Stochastic Processes (Master degree in Engineering) Franco Flandoli Stochastic Processes (Master degree in Engineering) Franco Flandoli Contents Preface v Chapter. Preliminaries of Probability. Transformation of densities. About covariance matrices 3 3. Gaussian vectors

More information

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University Instructions: Answer all four (4) questions. Be sure to show your work or provide su cient justi cation for

More information

ECON2285: Mathematical Economics

ECON2285: Mathematical Economics ECON2285: Mathematical Economics Yulei Luo Economics, HKU September 17, 2018 Luo, Y. (Economics, HKU) ME September 17, 2018 1 / 46 Static Optimization and Extreme Values In this topic, we will study goal

More information

Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels.

Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels. Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels. Pedro Albarran y Raquel Carrasco z Jesus M. Carro x June 2014 Preliminary and Incomplete Abstract This paper presents and evaluates

More information

Local disaggregation of demand and excess demand functions: a new question

Local disaggregation of demand and excess demand functions: a new question Local disaggregation of demand and excess demand functions: a new question Pierre-Andre Chiappori Ivar Ekeland y Martin Browning z January 1999 Abstract The literature on the characterization of aggregate

More information

Real Analysis: Homework # 12 Fall Professor: Sinan Gunturk Fall Term 2008

Real Analysis: Homework # 12 Fall Professor: Sinan Gunturk Fall Term 2008 Eduardo Corona eal Analysis: Homework # 2 Fall 2008 Professor: Sinan Gunturk Fall Term 2008 #3 (p.298) Let X be the set of rational numbers and A the algebra of nite unions of intervals of the form (a;

More information

Parametric Inference on Strong Dependence

Parametric Inference on Strong Dependence Parametric Inference on Strong Dependence Peter M. Robinson London School of Economics Based on joint work with Javier Hualde: Javier Hualde and Peter M. Robinson: Gaussian Pseudo-Maximum Likelihood Estimation

More information

EconS Advanced Microeconomics II Handout on Mechanism Design

EconS Advanced Microeconomics II Handout on Mechanism Design EconS 503 - Advanced Microeconomics II Handout on Mechanism Design 1. Public Good Provision Imagine that you and your colleagues want to buy a co ee machine for your o ce. Suppose that some of you may

More information

Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation

Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation Xu Cheng y Department of Economics Yale University First Draft: August, 27 This Version: December 28 Abstract In this paper,

More information

Week 2 Spring Lecture 3. The Canonical normal means estimation problem (cont.).! (X) = X+ 1 X X, + (X) = X+ 1

Week 2 Spring Lecture 3. The Canonical normal means estimation problem (cont.).! (X) = X+ 1 X X, + (X) = X+ 1 Week 2 Spring 2009 Lecture 3. The Canonical normal means estimation problem (cont.). Shrink toward a common mean. Theorem. Let X N ; 2 I n. Let 0 < C 2 (n 3) (hence n 4). De ne!! (X) = X+ 1 Then C 2 X

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Notes on Time Series Modeling

Notes on Time Series Modeling Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

On the synchronization of a class of electronic circuits that exhibit chaos

On the synchronization of a class of electronic circuits that exhibit chaos Chaos, Solitons and Fractals 13 2002) 1515±1521 www.elsevier.com/locate/chaos On the synchronization of a class of electronic circuits that exhibit chaos Er-Wei Bai a, *, Karl E. Lonngren a, J.C. Sprott

More information

LECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT

LECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes

More information

Statistical inference on Lévy processes

Statistical inference on Lévy processes Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline

More information

Serial Correlation Robust LM Type Tests for a Shift in Trend

Serial Correlation Robust LM Type Tests for a Shift in Trend Serial Correlation Robust LM Type Tests for a Shift in Trend Jingjing Yang Department of Economics, The College of Wooster Timothy J. Vogelsang Department of Economics, Michigan State University March

More information

Rank Estimation of Partially Linear Index Models

Rank Estimation of Partially Linear Index Models Rank Estimation of Partially Linear Index Models Jason Abrevaya University of Texas at Austin Youngki Shin University of Western Ontario October 2008 Preliminary Do not distribute Abstract We consider

More information

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012 SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 Revised March 2012 COWLES FOUNDATION DISCUSSION PAPER NO. 1815R COWLES FOUNDATION FOR

More information

Introduction to Linear Algebra. Tyrone L. Vincent

Introduction to Linear Algebra. Tyrone L. Vincent Introduction to Linear Algebra Tyrone L. Vincent Engineering Division, Colorado School of Mines, Golden, CO E-mail address: tvincent@mines.edu URL: http://egweb.mines.edu/~tvincent Contents Chapter. Revew

More information

Stochastic solutions of nonlinear pde s: McKean versus superprocesses

Stochastic solutions of nonlinear pde s: McKean versus superprocesses Stochastic solutions of nonlinear pde s: McKean versus superprocesses R. Vilela Mendes CMAF - Complexo Interdisciplinar, Universidade de Lisboa (Av. Gama Pinto 2, 1649-3, Lisbon) Instituto de Plasmas e

More information

Dichotomy Of Poincare Maps And Boundedness Of Some Cauchy Sequences

Dichotomy Of Poincare Maps And Boundedness Of Some Cauchy Sequences Applied Mathematics E-Notes, 2(202), 4-22 c ISSN 607-250 Available free at mirror sites of http://www.math.nthu.edu.tw/amen/ Dichotomy Of Poincare Maps And Boundedness Of Some Cauchy Sequences Abar Zada

More information

EDRP lecture 7. Poisson process. Pawe J. Szab owski

EDRP lecture 7. Poisson process. Pawe J. Szab owski EDRP lecture 7. Poisson process. Pawe J. Szab owski 2007 Counting process Random process fn t ; t 0g is called a counting process, if N t is equal total number of events that have happened up to moment

More information

Microeconometrics: Clustering. Ethan Kaplan

Microeconometrics: Clustering. Ethan Kaplan Microeconometrics: Clustering Ethan Kaplan Gauss Markov ssumptions OLS is minimum variance unbiased (MVUE) if Linear Model: Y i = X i + i E ( i jx i ) = V ( i jx i ) = 2 < cov i ; j = Normally distributed

More information

Approximately Most Powerful Tests for Moment Inequalities

Approximately Most Powerful Tests for Moment Inequalities Approximately Most Powerful Tests for Moment Inequalities Richard C. Chiburis Department of Economics, Princeton University September 26, 2008 Abstract The existing literature on testing moment inequalities

More information

MAXIMUM LIKELIHOOD ESTIMATION AND UNIFORM INFERENCE WITH SPORADIC IDENTIFICATION FAILURE. Donald W. K. Andrews and Xu Cheng.

MAXIMUM LIKELIHOOD ESTIMATION AND UNIFORM INFERENCE WITH SPORADIC IDENTIFICATION FAILURE. Donald W. K. Andrews and Xu Cheng. MAXIMUM LIKELIHOOD ESTIMATION AND UNIFORM INFERENCE WITH SPORADIC IDENTIFICATION FAILURE By Donald W. K. Andrews and Xu Cheng October COWLES FOUNDATION DISCUSSION PAPER NO. 8 COWLES FOUNDATION FOR RESEARCH

More information

Stochastic Processes

Stochastic Processes Introduction and Techniques Lecture 4 in Financial Mathematics UiO-STK4510 Autumn 2015 Teacher: S. Ortiz-Latorre Stochastic Processes 1 Stochastic Processes De nition 1 Let (E; E) be a measurable space

More information

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Winter 2017 TOPIC 3: MULTINOMIAL CHOICE MODELS

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Winter 2017 TOPIC 3: MULTINOMIAL CHOICE MODELS ECONOMETRICS II (ECO 2401) Victor Aguirregabiria Winter 2017 TOPIC 3: MULTINOMIAL CHOICE MODELS 1. Introduction 2. Nonparametric model 3. Random Utility Models - De nition; - Common Speci cation and Normalizations;

More information

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()

More information

Introduction Wavelet shrinage methods have been very successful in nonparametric regression. But so far most of the wavelet regression methods have be

Introduction Wavelet shrinage methods have been very successful in nonparametric regression. But so far most of the wavelet regression methods have be Wavelet Estimation For Samples With Random Uniform Design T. Tony Cai Department of Statistics, Purdue University Lawrence D. Brown Department of Statistics, University of Pennsylvania Abstract We show

More information

Economics 620, Lecture 18: Nonlinear Models

Economics 620, Lecture 18: Nonlinear Models Economics 620, Lecture 18: Nonlinear Models Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 18: Nonlinear Models 1 / 18 The basic point is that smooth nonlinear

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 11 The Prior Distribution Definition Suppose that one has a statistical model with parameter

More information

Estimation under Ambiguity (Very preliminary)

Estimation under Ambiguity (Very preliminary) Estimation under Ambiguity (Very preliminary) Ra aella Giacomini, Toru Kitagawa, y and Harald Uhlig z Abstract To perform a Bayesian analysis for a set-identi ed model, two distinct approaches exist; the

More information

A formal statistical test for the number of factors in. the approximate factor models

A formal statistical test for the number of factors in. the approximate factor models A formal statistical test for the number of factors in the approximate factor models Alexei Onatski Economics Department, Columbia University September 26, 2006 Abstract In this paper we study i.i.d. sequences

More information

Measuring robustness

Measuring robustness Measuring robustness 1 Introduction While in the classical approach to statistics one aims at estimates which have desirable properties at an exactly speci ed model, the aim of robust methods is loosely

More information

Oscillatory Mixed Di erential Systems

Oscillatory Mixed Di erential Systems Oscillatory Mixed Di erential Systems by José M. Ferreira Instituto Superior Técnico Department of Mathematics Av. Rovisco Pais 49- Lisboa, Portugal e-mail: jferr@math.ist.utl.pt Sra Pinelas Universidade

More information

Central Limit Theorem for Non-stationary Markov Chains

Central Limit Theorem for Non-stationary Markov Chains Central Limit Theorem for Non-stationary Markov Chains Magda Peligrad University of Cincinnati April 2011 (Institute) April 2011 1 / 30 Plan of talk Markov processes with Nonhomogeneous transition probabilities

More information

1 Review of di erential calculus

1 Review of di erential calculus Review of di erential calculus This chapter presents the main elements of di erential calculus needed in probability theory. Often, students taking a course on probability theory have problems with concepts

More information

Not Only What But also When A Theory of Dynamic Voluntary Disclosure

Not Only What But also When A Theory of Dynamic Voluntary Disclosure Not Only What But also When A Theory of Dynamic Voluntary Disclosure PRELIMINARY AND INCOMPLETE Ilan Guttman, Ilan Kremer and Andrzej Skrzypacz Stanford Graduate School of Business November 2011 1 Introduction

More information

4. Duality and Sensitivity

4. Duality and Sensitivity 4. Duality and Sensitivity For every instance of an LP, there is an associated LP known as the dual problem. The original problem is known as the primal problem. There are two de nitions of the dual pair

More information

TitleCoupled Nosé-Hoover Equations of Mo.

TitleCoupled Nosé-Hoover Equations of Mo. TitleCoupled Nosé-Hoover Equations of Mo Author(s) Fukuda, Ikuo; Moritsugu, Kei Citation Issue 15-8-18 Date Text Version author UL http://hdl.handle.net/119/56 DOI ights Osaka University Supplemental Material:

More information

Nonparametric Trending Regression with Cross-Sectional Dependence

Nonparametric Trending Regression with Cross-Sectional Dependence Nonparametric Trending Regression with Cross-Sectional Dependence Peter M. Robinson London School of Economics January 5, 00 Abstract Panel data, whose series length T is large but whose cross-section

More information

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank David Glickenstein November 3, 4 Representing graphs as matrices It will sometimes be useful to represent graphs

More information

Lecture Notes - Dynamic Moral Hazard

Lecture Notes - Dynamic Moral Hazard Lecture Notes - Dynamic Moral Hazard Simon Board and Moritz Meyer-ter-Vehn October 27, 2011 1 Marginal Cost of Providing Utility is Martingale (Rogerson 85) 1.1 Setup Two periods, no discounting Actions

More information

EXTREMAL POLYNOMIALS ON DISCRETE SETS

EXTREMAL POLYNOMIALS ON DISCRETE SETS EXTREMAL POLYNOMIALS ON DISCRETE SETS A. B. J. KUIJLAARS and W. VAN ASSCHE [Received 14 April 1997] 1. Introduction It is well known that orthonormal polynomials p n on the real line can be studied from

More information

INEQUALITIES OF LIPSCHITZ TYPE FOR POWER SERIES OF OPERATORS IN HILBERT SPACES

INEQUALITIES OF LIPSCHITZ TYPE FOR POWER SERIES OF OPERATORS IN HILBERT SPACES INEQUALITIES OF LIPSCHITZ TYPE FOR POWER SERIES OF OPERATORS IN HILBERT SPACES S.S. DRAGOMIR ; Abstract. Let (z) := P anzn be a power series with complex coe - cients convergent on the open disk D (; R)

More information

Checking for Prior-Data Conßict. by Michael Evans Department of Statistics University of Toronto. and

Checking for Prior-Data Conßict. by Michael Evans Department of Statistics University of Toronto. and Checking for Prior-Data Conßict by Michael Evans Department of Statistics University of Toronto and Hadas Moshonov Department of Statistics University of Toronto Technical Report No. 413, February 24,

More information

Ch 4. Linear Models for Classification

Ch 4. Linear Models for Classification Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,

More information

An Analysis of the Difference of Code Lengths Between Two-Step Codes Based on MDL Principle and Bayes Codes

An Analysis of the Difference of Code Lengths Between Two-Step Codes Based on MDL Principle and Bayes Codes IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 47, NO. 3, MARCH 2001 927 An Analysis of the Difference of Code Lengths Between Two-Step Codes Based on MDL Principle Bayes Codes Masayuki Goto, Member, IEEE,

More information

Lecture Notes Part 7: Systems of Equations

Lecture Notes Part 7: Systems of Equations 17.874 Lecture Notes Part 7: Systems of Equations 7. Systems of Equations Many important social science problems are more structured than a single relationship or function. Markets, game theoretic models,

More information