University of Toronto
|
|
- Roderick Griffin
- 6 years ago
- Views:
Transcription
1 A Limit Result for the Prior Predictive by Michael Evans Department of Statistics University of Toronto and Gun Ho Jang Department of Statistics University of Toronto Technical Report No April 15, 2010 TECHNICAL REPORT SERIES University of Toronto Department of Statistics
2 A Limit Result for the Prior Predictive Michael Evans and Gun Ho Jang Department of Statistics University of Toronto Abstract We establish results concerning the convergence of the prior predictive distribution. An application is made to the problem of checking for prior-data con ict. Keywords: minimal su ciency, prior predictive, convergence, prior-data con ict 1 Introduction Suppose we have a model given by a collection of probability measures fp : 2 g where P (A) = R A f (x) (dx); i.e., each P is absolutely continuous with respect to a support measure on the sample space X ; with the density denoted by f : With this formulation a prior leads to a prior predictive probability measure on X given by M(A) = R P (A) (d) = R A m(x) (dx); where m(x) = R f (x) (d): If T : X! T is a minimal su cient statistic for fp : 2 g; then it is well known that the posterior distribution for is the same whether we observe x or T (x) and so we denote the posterior by ( j T ). Furthermore, the conditional distribution of x given T is independent of and we denote this conditional measure by P ( j T ): The joint distribution P can then be factored as P = M ( j x) = P ( j T ) M T ( j T ) where M T is the marginal prior predictive distribution of T: If f T denotes the marginal density of T; with respect to a support measure T on T ; then m T (t) = R f T (t) (d) denotes the 1
3 density of M T with respect to T : If denotes the density of ; with respect to a support measure on ; then we can write m T (t) = R f T (t)() (d): Our concern here is with the behavior of M T as the amount of data grows. A simple example illustrates the asymptotic behavior of this distribution that we might expect to hold in more general situations. Example 1 (Location normal). Suppose that x = (x 1 ; : : : ; x n ) is a sample from a N(; 1) distribution where 2 R 1 is unknown. Then a minimal su cient statistic is given by (x) = x and (x) converges almost surely to the true value as n! 1: Suppose we put a N( 0 ; 0) 2 prior on : The prior predictive distribution M Tn is then easily obtained from x = + z where z N(0; 1=n) independent of N( 0 ; 2 0); namely, M Tn is the N( 0 ; =n) distribution. We see immediately that M Tn converges in distribution to the N( 0 ; 2 0) distribution as n! 1: Furthermore, m Tn (t) converges almost surely to (2) 1=2 1 0 expf (t 0 ) 2 =2 2 0g as n! 1; uniformly for t in a compact set. Simple computations show that these results do not depend on using a normal prior, namely, if we use prior measure with continuous density ; then M Tn converges in distribution to as n! 1; and m Tn (t) converges almost surely to as n! 1; uniformly for t in a compact set. So in Example 1 we can think of m T (T (x)) as a consistent estimator of the prior evaluated at the true value of the parameter. The signi cance of this is that the value of m T (T (x)) gives an indication of whether or not the prior has been poorly chosen, in the sense that the true value of may lie in a region where little prior probability has been assigned. Of course, we cannot tell this from the value m T (T (x)) itself but need to calibrate this on some scale. In Evans and Moshonov (2006, 2007) the P-value M T (m T (t) m T (T (x))); (1) and some variations of this, was proposed for checking for prior-data con ict. Note that this P-value is a modi cation of a P-value proposed by Box (1980) for general model checking in Bayesian contexts. In Example 1, when using the normal prior, we see that 2
4 (1) converges to 2(1 (j 0 j= 0 )) as n! 1; where is the true value of : So (1) is a consistent estimator of the P-value which measures whether the true value of the parameter lies in the tails of the prior. In Section 2 we prove that (1) converges to (() ( )); where is the true value of ; in fairly general circumstances. The P-value (() ( )) will be small whenever the true value lies in a region of low prior probability and so we have an instance of prior-data con ict. As such, (1) is seen to be an appropriate measure of prior-data con ict. A criticism of (1) is that, in the case of continuous models at least, the P-value is not invariant under smooth transformations. In particular, suppose that W : T! W is 1-1 and smooth. Let J W (t) be the reciprocal of the Jacobian determinant of W evaluated at t: Then, if instead of T we use W (T ) as the minimal su cient statistic, the P-value is M W (m W (w) m W (W (T (x)))) = M T (m T (t)j W (t) m T (T (x))j W (T (x))) and this will not equal (1) unless J W (t) is constant. In Evans and Jang (2010) the general problem of computing P-values, based on the density of a discrepancy statistic, to assess whether or not the data came from a single xed distribution, was considered. An invariant P-value was proposed. For (1) this entails using instead the P-value M T (m T (t) m T (T (x))); (2) where m T (t) = m T (t)e(j 1 T (X) j T (X) = t); J T (x) = j det(dt (x) dt 0 (x))j 1=2 and dt denotes the di erential of T: The factor E(J 1 T (X) j T (X) = t) corrects for volume distortions due to the transformation T and is independent of because T is minimal su cient. Note that m T (t) is the density of M T with respect to the support measure (E(J 1 T (X) j T (X) = t)) 1 (dt): In Example 1 we have that J T (x) is constant and so (1) equals (2). While the P-value (2) will generally di er from (1), it is often the case that the e ect of E(J 1 T (X) j T (X) = t) is negligible. We establish a convergence result for (2) in Section 3. While there are numerous discussions concerning asymptotics for a posterior analysis, for example, Walker (1969), Heyde and Johnstone (1979), and Chen (1985), there seem 3
5 to be almost no discussions concerned with convergence issues associated with the prior predictive distribution. Such results also have implications for methods that choose the prior based on the prior predictive. This paper addresses some of these problems. 2 Convergence of the Basic P-value In the Appendix we provide the proof of the following result. Theorem 1. Suppose is an open subset of a Euclidean space and assume (i)! a.s. P for every ; (ii) m Tn (t)! (t) uniformly on compact subsets of ; (iii) is continuous and the prior distribution of () has no atoms, then M Tn (m Tn (t) m Tn ( (x n )))! (() ( )) a.s. P where is the true value of : Note that Theorem 1 implicitly assumes that the sampling model for is continuous. We will subsequently discuss how to handle the discrete case. To apply this result we need to establish (ii). We discuss several examples. Example 2 (Scale-Gamma). Let x = (x 1 ; : : : ; x n ) be a sample from a Gamma( 0 ; ) distribution where the scale parameter > 0 is unknown. Then the statistic (x) = (n 0 ) 1 P n i=1 x i Gamma(n 0 ; =(n 0 )) is minimal su cient and it converges almost surely to the true value of : We prove the following result in the Appendix. Lemma 2. If (x) Gamma(n 0 ; =(n 0 )) and the prior on is continuous, then (ii) of Theorem 1 holds. So if, in addition the prior distribution on (0; 1) has no atoms, then Theorem 1 applies and we have the convergence of (1). Certainly these conditions apply to the commonly used priors on a scale parameter. The following example uses Example 2 in a problem of considerable importance for statistical practice. 4
6 Example 3 (Normal linear regression). We consider rst the situation where we have a sample x = (x 1 ; : : : ; x n ) from a N(; 2 ) distribution with 2 R 1 and > 0 unknown. Then (x) = (T 1n (x); T 2n (x)) = (x; s 2 ) is a minimal su cient statistic and (x)! (; 2 ) as n! 1: We prove the following result in the Appendix. Lemma 3. If (x) = (x; s) where x N(; 2 =n) independent of s 2 Gamma((n 1)=2; 2 2 =(n 1)) and the prior on (; 2 ) is continuous, then (ii) of Theorem 1 holds. The prior is commonly prescribed by rst stating a prior 2 for 2 and then a conditional prior 1 ( j 2 ) for : As discussed in Evans and Moshonov (2006, 2007), it then makes sense to check 2 rst and, if 2 passes, then check 1 : With this approach we can learn more about where the prior is de cient, if indeed there is a problem: Following that development, the check for 2 is based on T 2n (x) = s 2 via the P-value M T2n (m T2n (t 2 ) m T2n (T 2n (x))): Now T 2n (x) Gamma((n 1)=2; 2 2 =(n 1)): So, when 2 satis es the conditions of Theorem 1, Example 2 applies with 0 = 1=2 and M T2n (m T2n (t 2 ) m T2n (T 2n (x)))! 2 ( 2 ( 2 ) 2 ()): 2 To check 1 the relevant P- value to use is M T1n (m T1n (t 1 j T 2n (x) = s 2 ) m T1n (T 1n (x) j T 2n (x) = s 2 ) j T 2n (x) = s 2 ): (3) Now consider m T1n (t 1 j T 2n = t 2 ) = m Tn (t 1 ; t 2 )=m T2n (t 2 ): Since m Tn (t 1 ; t 2 )! (t 1 ; t 2 ) and m T2n (t 2 )! 2 (t 2 ) uniformly on compact sets, then we have that m T1n (t 1 j T 2n = t 2 )! (t 1 ; t 2 )= 2 (t 2 ) = 1 (t 1 j t 2 ) uniformly on compact sets. Furthermore, the measures M T1n ( j T 2n = s 2 ) converge in distribution to 1 (t 1 j ); 2 m T1n (t 1 j T 2n = s 2 ) converges almost surely N( ; ) 2 to 1 (t 1 j ); 2 and m T1n (x j T 2n = s 2 ) converges almost surely N( ; ) 2 to 1 ( j ): 2 This implies the convergence almost surely N( ; ) 2 of (3) to 1 ( 1 ( j ) 2 1 ( j ) 2 j ): 2 For a normal linear regression model y n = X n + e; where X n 2 R nk and e N n (0; 2 I); we have that (y) = (T 1n (y); T 2n (y)) = (b; s 2 ) where b = (XnX 0 n ) 1 Xny 0 and s 2 = (n k) 1 jjy Xbjj 2 : Under suitable conditions on the X n matrices we have that! (; 2 ) almost surely: The convergence results for this situation then proceed 5
7 just as in as in the location-scale case. The following result, with the proof provided in the Appendix, is sometimes useful in establishing condition (ii) in Theorem 1. Theorem 4. Suppose is an open subset of R k ; and for any > 0; K compact there exist c 1, c 2 > 0 and N > 0 such that f ;Tn (t) c 1 e c2n whenever t 2 K; jjt jj > ; and n N: (4) The following are equivalent: (a) for any prior with continuous density ; then m Tn (t)! (t) uniformly for t 2 K (b) for any compact K and > 0; then Z f ;Tn (t)i(jjt jj < ) d! 1 as n! 1; uniformly for t 2 K: (5) We now consider an example where the distribution of is on a discrete subset of R 1 : In such a case we can t expect condition (ii) of Theorem 1 to hold at values of t where (t) > 0 but m Tn (t) = 0 for all n: Suppose, however, that has a lattice distribution with lattice step equal to h: Then for kh < t (k +1)h we de ne m cont (t) = m Tn ((k + 1)h)=h and 0 otherwise and treat m cont (t) as a density with respect to length measure. Since (x) is always on the lattice we see immediately that M cont (m cont (t) m cont ( (x))) = M Tn (m Tn (t) m Tn ( (x))): We can then apply Theorem 1 to m cont and this proves the convergence of (1). Note that m cont (t) = R f ;T cont n (t) (d) where f;t cont n (t) = f ;Tn ((k + 1)h)=h when kh < t (k + 1)h: Example 4 (Binomial). Suppose that x = (x 1 ; : : : ; x n ) is a sample from a Bernoulli() distribution where 2 (0; 1) is unknown. Then (x) = x is minimal su cient and converges to : For t 2 f0; 1=n; : : : ; 1g then f ;Tn (t) = P ( = t) = n nt nt (1 ) n(1 t) : In this case x has a discrete distribution on the lattice with step size equal to 1=n: In the Appendix we prove the following result. 6
8 Lemma 5. If n (x) Binomial(n; ) and is a continuous on (0; 1), then (ii) of Theorem 1 holds for m cont : Therefore, m cont (t) converges to (t) uniformly on each compact set and (1) converges provided the prior satis es the conditions of Theorem 1. One interesting case where does not satisfy the conditions of Theorem 1 arises when Uniform(0; 1) as the prior distribution of () has all of its mass at 1. In this case, however, we have that m cont (t) n=(n+1)! 1 uniformly for all t 2 (0; 1) and moreover M Tn (m Tn (t) m Tn ( (x))) = 1 = (() ( )) and so the convergence result is obvious. 3 Convergence of the Invariant P-value We now consider the convergence of (2). As noted, this P-value is invariant under smooth transformations and will agree with (1) whenever T is linear or the sampling model for T is discrete. This applies in Examples 1, 2, and 4 but not in Example 3. Example 5 (Normal linear regression). Consider the location-scale case. Clearly T 1n (x) = x is linear, and so the P-value for checking 1 agrees with the invariant version. But T 2n (x) = s 2 is nonlinear and so the P-value (1) for checking 2 is not the same as the invariant version. In this case dt 2n (x) = (2=(n 1))(x 1 x; : : : ; x n x) giving J T2n (x) = j det(dt 2n (x) dt 0 2n(x))j 1=2 = ( p n 1=2)s 1 and so E(J 1 T 2n (X) j (X) = (x; s 2 )) = (2= p n 1)s: Therefore, the invariant P-value is equal to M T2n (m T2n (t 2 )t 1=2 2 m T2n (T 2n (x))(t 2n (x)) 1=2 ) and this converges almost surely to 2 ( 2 ( 2 ) 2 ( 2 ) ) (see Theorem 6). The proof of the following result is virtually identical to that of Theorem 1. Theorem 6. Suppose is an open subset of a Euclidean space and assume (i)! a.s. P for every ; 7
9 (ii) w n (t) = E(J 1 (X) j (X) = t) is continuous and a n w n (t)! w(t) for some sequence a n ; (iii) a n m Tn (t)w n (t)! (t)w(t) uniformly on compact subsets of ; (iv) is continuous and the prior distribution of ()w() has no atoms, then M Tn (m (t) m ( (x n )))! (()w() ( )w( )) a.s. P where is the true value of : Note that ()w() is the density of with respect to the support measure (w()) 1 (d): Also note that when is k-dimensional and p n( ) is asymptotically normal, then, in many cases, we can take a n = n k=2 : The developments in this paper have required that the minimal su cient statistic be a consistent estimator of : The existence of such a minimal su cient statistic is guaranteed for exponential models. Suppose, however, that we reparameterize via the 1-1, smooth function ; namely, = (): Then we must replace by ( ) for the convergence results to hold as stated. If is nonlinear, however, then (1) will typically depend on whether we use or ( ); namely, it will implicitly depend on the parameterization. Using (2) this dependence is avoided and the P-value is independent of the choice of the minimal su cient statistic or equivalently the parameterization. The use of (2) seems more appropriate than (1) for this reason, although there is typically very little di erence in the P-values obtained. 4 Conclusions We have established convergence results for various prior predictive P-values that show directly that these are appropriate for checking for prior-data con ict, namely, assessing if the true value of the parameter is in the tails of the prior. Essentially these results are restricted to situations where a version of the minimal su cient statistic is a consistent estimate of the model parameter and this means our results apply in the context of exponential models. Similar results can undoubtedly be established in other contexts, in 8
10 particular for group models, and these are currently being developed. More generally convergence results for the prior predictive have implications for empirical Bayes methods. For example, suppose we have a family of priors f : 2 Ag with corresponding prior predictives m ;Tn for a minimal su cient statistic : Then convergence of m ;Tn ( (x)) to ( ) has the implication that maximizing m ;Tn ( (x)) over to select the prior is essentially nding the prior that has maximal value at the true value of the parameter. One might argue that it makes more sense to maximize M ;Tn (m ;Tn (t) m ;Tn ( (x))) over as then, based on the convergence of this P- value to ( () ( )); we are nding the prior for which the true value is least surprising. The implications of this are currently being investigated. Appendix Proof of Theorem 1 Let > 0: Then there exists > 0 such that j(t) ( 0 )j < =2 whenever t 2 B ( 0 ) and there exists N 1 such that for all n > N 1 ; (x n ) 2 B ( 0 ): Also there exists N 2 such that for all n > N 2 and for all t 2 B ( 0 ); then jm Tn (t) (t)j < =2: So, if n > maxfn 1 ; N 2 g then M Tn (m Tn (t) ( 0 ) ) M Tn (m Tn (t) m Tn ( (x n ))) M Tn (m Tn (t) ( 0 ) + ): Now we prove that M Tn (m Tn (t) ( 0 ) )! (() ( 0 ) ): Let 0 > 0: Let C be compact such that 0 2 C; (@C) = 0; and (C) 1 0 =2: By (i) and Slutsky s Theorem M Tn converges in law to. Therefore, M Tn (C)! (C) and so there exists N 3 such that for all n > N 3 ; M Tn (C) > 1 0 : Therefore, for all n > N 3 ; M Tn (m Tn (t) ( 0 ) ) 0 M Tn (m Tn (t) ( 0 ) ; C) M Tn (m Tn (t) ( 0 ) ) and we can make the LHS and RHS as close as we like by choosing 0 small. Let 00 > 0: There exists N 4 such that for all n > N 4 then jm Tn (t) (t)j < 00 for all t 2 C: When n > maxfn 3 ; N 4 g then, M Tn ((t) ( 0 ) 00 ; C) M Tn (m Tn (t) ( 0 ) ; C) M Tn ((t) ( 0 ) + 00 ; C) and the LHS converges to ((t) ( 0 ) 00 ; C) while the RHS converges to ((t) ( 0 ) + 00 ; C): By choosing 00 small we can make these quantities as close to ((t) ( 0 ) ; C) as we 9
11 like. This proves that M Tn (m Tn (t) ( 0 ) ; C)! ((t) ( 0 ) ; C) and this establishes that M Tn (m Tn (t) ( 0 ) )! ((t) ( 0 ) ): A similar argument shows that M Tn (m Tn (t) ( 0 )+)! ((t) ( 0 )+) and this completes the proof. Proof of Lemma 2 Suppose K is a compact set in (0; 1): Then, there are 0 < a < b < 1 such that K [a; b]. Fix > 0 satisfying < minfa=3; 1g. We prove (4) and (5), then apply Theorem 4. For t >, t= t=(t+) 1 =(b+). Also for t <, t= t=(t ) 1+=(a ). Hence, jt= 1j > = =(b+) for jt j > and t 2 K. Note (n 0 ) p 2(n0 ) n0 1=2 e n0. Since ue u+1 has peak 1 at u = 1, 1 = sup u:ju 1j> ue u+1 < 1. Also there exists N 1 > 1 such that n 1=2 n0=2 1 1 for all n N 1. Let u = t=. Then for jt j >, we get ju 1j > and f ;Tn (t) = [(n 0 ) n0 e n0 = (n 0 )t](ue u+1 ) n0 a 1 (n 0 =2) 1=2 n0 1 a 1 ( 0 =2) 1=2 e n2 1 0 log(1= 1). Hence, (4) holds. For (5), let I 0 = R f ;Tn(t)I(jt j < ) d R 1 0 [(n 0) n0 = (n 0 )]u n0 2 e n0u du = n 0 =(n 0 1). Also for t 2 K, I(jt j < ) I(jt= 1j < 1 ) I(jt= 1j < 1 n 1=2 log n) where 1 = =(b + ). Then, I 0 (n Z 0) n0 e n0 u n0 2 e n0(u 1) I(ju 1j < 1 ) du: (n 0 ) For ju 1j < 1 n 1=2 log n, a lower bound of the logarithm of the integrand is given by log(u n0 2 e n0(u 1) ) = n 0 (u 1) + (n 0 2) log(1 (1 u)) n 0 (n 1) (n 0 2)(1 u + (1 u) 2 =2 + j1 uj 3 ) n 0 (1 u) 2 =2 (2 + 0 (log n) 2 )j1 uj: The change of variable v = p n 0 (u I 0 (n 0) n0 1=2 Z e n0 (n 0 ) By Stirling s formula (n 0 ) n0 1) gives e v2 =2 (2+ 0(log n) 2 )jvj= p n 0 I(jvj < 1 1=2 0 log n) dv: 1=2 e n0 = (n 0 )! (2) 1=2. The integral converges to R e v 2 =2 dv by the Lebesgue dominated convergence theorem. Hence, I 0! 1 as n! 1. Thus (5) holds and, by Theorem 4, Theorem 1 (ii) holds. Proof of Lemma 3 We prove (4) and (5). The density f ; 2(x; s 2 ) of = (X; S 2 ) is given by ((n 1)=2) (n 1)=2 (n 1)=2 e h s 2 (2=n) 1=2 ((n 1)=2)(s 2 ) 3=2 n 1 2 exp n ( s2 2 1) (x ) 2 2 i n=2 : 10
12 Let I 1 be the rst part and I 2 be the part inside the brackets. To prove (4), x a compact set K. Let a = inffs 2 : (x; s 2 ) 2 Kg > 0 and b = supfs 2 : (x; s 2 ) 2 Kg < 1. Also we consider 0 < < min(a=3; 1). Then, Stirling s formula gives I 1 =n! 1=[(2s 2 ) 3=2 ] 1=[(2a) 3=2 ]. For (x; s 2 ) 2 K and jj(x; s 2 ) (; 2 )jj >, we have jx j > =2 or js 2 2 j > =2. Then there exists > 0 such that js 2 = 2 1j >. Since ve (v 1)(n 1)=n is unimodal having peak e 1=n n=(n 1) at v = n=(n 1), an upper bound of I 2 is obtained at s 2 = 2 = 1 + or s 2 = 2 = 1 provided by n > 1 + 1=. So I 2 max((1 ) exp(n=(n 1)); (1 + ) exp( n=(n 1))) < 1. If js 2 = 2 1j, then jx j > =2. Thus (x ) 2 = 2 = (x ) 2 (s 2 = 2 )=s 2 (=2) 2 (1 )=b. So I 2 e 1=n (n=(n 1)) exp( 2 (1 )=4b) < 1 when n > (1 + exp( 2 (1 )=4b)) 1. Hence (4) holds. The integration range jj(x; s 2 ) (; 2 )jj < contains jx j < =2 and js 2 2 j < =2. Again, this region contains jx j= < 2 = =( + 2b). Then, Z Z I 3 = f ; 2(x; s 2 )I(jj(x; s 2 ) (; 2 )jj < ) d d 2 Z Z (2 2 =n) 1=2 exp( (n=2 2 )(x ) 2 )I(jx j= < 2 ) ((n 1)=2) (n 1)=2 (n 1)=2 1 s e ((n 1)=2)( 2 ) (n 1)=2 (n 1)s 2 =2 2 I(js 2 2 j < =2) d d 2 Using v = p n(x )=, this integral can be separated into two parts, let I 4 and I 5 be the two integrals. Then, I 4 = ( p n 2 ) ( p n 2 )! 1 as n! 1; and I 5! 1 as n! 1 by Lemma 2 for 0 = 1=2. So I 3! 1 and (5) holds. Finally, Theorem 1 (ii) holds by Theorem 4. Proof of Theorem 4 Suppose (a) holds. If > 0 and K is compact, then R f ;Tn (t)()i(jjt jj ) d R c1 e c2n I(jjt jj ) (d) c 1 e c2n! 0 uniformly in t 2 K: So if (a) holds, then j R f ;Tn (t)()i(jjt jj < ) d (t)j! 0 uniformly for t 2 K: The set K = f : jj 0 jj for some 0 2 Kg is also compact when is small enough. Note that, for given ; the convergence in (5) follows whenever this convergence holds for a smaller value of : Let be a continuous density that is constant and positive on K ; then (b) follows. 11
13 Now suppose (b) holds, has a continuous density, K is a compact subset of and > 0. Then K is also compact for is small enough. Since is uniformly continuous on K ; there exists 0 > 0 such that j( 1 ) ( 2 )j < =4 whenever 1 ; 2 2 K and jj 1 2 jj < 0. From (b), there exists L 1 > 0 such that j R f ;Tn (t)i(jjt jj < ) d 1j < =(4 sup t2k (t)) for all n L 1 and t 2 K where = min( 0 ; ): Also, there exist c 1, c 2 > 0 and L 2 > 0 such that f ;Tn (t) c 1 e c2n whenever t 2 K; jjt jj ; n L 2 : Therefore, there exists L 3 such that R f ;Tn (t)()i(jjt jj ) d c 1 e c2n =4 for all n L 3. Finally, for n L = max(l 1 ; L 2 ; L 3 ) and t 2 K; we have that jm Tn (t) (t)j R f ;Tn (t)()i(jjt jj ) d + R f ;Tn (t)j() (t)ji(jjt jj < ) d + (t)j R f ;Tn (t)i(jjt jj < ) d 1j =4 + (=4)(1 + =4) + =4 < and we see that (a) holds. Proof of Lemma 5 Since K is compact, there is a > 0 such that 0 < a t 1 a < 1 for all t 2 K. For t 2 K; Stirling s formula implies that log nt n = 2 1 log 2nt(1 t) n(t log t + (1 t) log(1 t)) + r(n; t) where r(n; t) < (12n) 1 (12nt + 1) 1 (12n(1 t) + 1) 1 < 1. So we have that f ;Tn (t) = (2t(1 t)=n) 1=2 exp(r(n; t)) exp(ng(; t)) where g(; t) = t log(=t) + (1 t) log((1 )=(1 t)): Let 0 < < a=2: Note g(; t) has maximum value 0 at = t. Therefore, since g(; t) is continuous when jt j > ; we have that b = sup t;:jt j> g(; t) > 0: Also there is N 1 so that n 1=2 e bn=2 1 for n N 1. So, when t 2 K, jt j > and n N 1, f ;Tn (t) (2a(1 a)) 1=2 e exp( nb=2). Hence (4) holds. Let I 0 = R 1 0 f cont ; (t)i(jt j < ) d. Then, I 0 R 1 0 f cont ; (t) d = n n nt R 1 0 nt (1 ) n(1 t) d = n=(n + 1). Thus, using (4), we get I 0 = n n + 1 Z 1 0 f cont ; (t)i(jt j > ) d n n + 1 c 1 e c2n! 1: Hence, (5) holds. So Theorem 1 (ii) holds by Theorem 4. 12
14 References Box, G. E. P., Sampling and Bayes inference in scienti c modelling and robustness. J. Roy. Statist. Soc. Ser. A 143 (4), , with discussion. Chen, C. F., On asymptotic normality of limiting density functions with Bayesian implications. J. Roy. Statist. Soc. Ser. B 47 (3), Evans, M., Jang, G. H., Invariant P -values for model checking. Ann. Statist. 38 (1), Evans, M., Moshonov, H., Checking for prior-data con ict. Bayesian Anal. 1 (4), Evans, M., Moshonov, H., Checking for prior-data con ict with hierarchically speci ed priors. In: Upadhyay, A., Singh, U., Dey, D. (Eds.), Bayesian Statistics and its Applications. Anamaya Publishers, New Delhi, pp Heyde, C. C., Johnstone, I. M., On asymptotic posterior normality for stochastic processes. J. Roy. Statist. Soc. Ser. B 41 (2), Walker, A. M., On the asymptotic behaviour of posterior distributions. J. Roy. Statist. Soc. Ser. B 31,
Measuring Statistical Evidence Using Relative Belief
Measuring Statistical Evidence Using Relative Belief Michael Evans Department of Statistics University of Toronto Abstract: A fundamental concern of a theory of statistical inference is how one should
More information3 Random Samples from Normal Distributions
3 Random Samples from Normal Distributions Statistical theory for random samples drawn from normal distributions is very important, partly because a great deal is known about its various associated distributions
More informationStochastic Processes
Stochastic Processes A very simple introduction Péter Medvegyev 2009, January Medvegyev (CEU) Stochastic Processes 2009, January 1 / 54 Summary from measure theory De nition (X, A) is a measurable space
More informationAppendix for "O shoring in a Ricardian World"
Appendix for "O shoring in a Ricardian World" This Appendix presents the proofs of Propositions - 6 and the derivations of the results in Section IV. Proof of Proposition We want to show that Tm L m T
More informationMC3: Econometric Theory and Methods. Course Notes 4
University College London Department of Economics M.Sc. in Economics MC3: Econometric Theory and Methods Course Notes 4 Notes on maximum likelihood methods Andrew Chesher 25/0/2005 Course Notes 4, Andrew
More informationDesire-as-belief revisited
Desire-as-belief revisited Richard Bradley and Christian List June 30, 2008 1 Introduction On Hume s account of motivation, beliefs and desires are very di erent kinds of propositional attitudes. Beliefs
More informationLecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )
LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data
More informationSupplemental Material 1 for On Optimal Inference in the Linear IV Model
Supplemental Material 1 for On Optimal Inference in the Linear IV Model Donald W. K. Andrews Cowles Foundation for Research in Economics Yale University Vadim Marmer Vancouver School of Economics University
More informationON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS. Abstract. We introduce a wide class of asymmetric loss functions and show how to obtain
ON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS FUNCTIONS Michael Baron Received: Abstract We introduce a wide class of asymmetric loss functions and show how to obtain asymmetric-type optimal decision
More informationDensity estimators for the convolution of discrete and continuous random variables
Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract
More informationLinearized methods for ordinary di erential equations
Applied Mathematics and Computation 104 (1999) 109±19 www.elsevier.nl/locate/amc Linearized methods for ordinary di erential equations J.I. Ramos 1 Departamento de Lenguajes y Ciencias de la Computacion,
More informationChecking for Prior-Data Conflict
Bayesian Analysis (2006) 1, Number 4, pp. 893 914 Checking for Prior-Data Conflict Michael Evans and Hadas Moshonov Abstract. Inference proceeds from ingredients chosen by the analyst and data. To validate
More informationChapter 1. GMM: Basic Concepts
Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating
More informationComputer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)
Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori
More informationMA 8101 Stokastiske metoder i systemteori
MA 811 Stokastiske metoder i systemteori AUTUMN TRM 3 Suggested solution with some extra comments The exam had a list of useful formulae attached. This list has been added here as well. 1 Problem In this
More informationA NONPARAMETRIC TEST FOR SERIAL INDEPENDENCE OF REGRESSION ERRORS
A NONPARAMETRIC TEST FOR SERIAL INDEPENDENCE OF REGRESSION ERRORS Miguel A. Delgado and Juan Mora WP-AD 99-28 Correspondence to J. Mora: University of Alicante. Departamento de Fundamentos del Análisis
More informationE cient Simulation and Conditional Functional Limit Theorems for Ruinous Heavy-tailed Random Walks
E cient Simulation and Conditional Functional Limit Theorems for Ruinous Heavy-tailed Random Walks Jose Blanchet and Jingchen Liu y June 14, 21 Abstract The contribution of this paper is to introduce change
More informationEstimating the Number of Common Factors in Serially Dependent Approximate Factor Models
Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models Ryan Greenaway-McGrevy y Bureau of Economic Analysis Chirok Han Korea University February 7, 202 Donggyu Sul University
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationApplications of Subsampling, Hybrid, and Size-Correction Methods
Applications of Subsampling, Hybrid, and Size-Correction Methods Donald W. K. Andrews Cowles Foundation for Research in Economics Yale University Patrik Guggenberger Department of Economics UCLA November
More informationECONOMETRICS FIELD EXAM Michigan State University May 9, 2008
ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 Instructions: Answer all four (4) questions. Point totals for each question are given in parenthesis; there are 00 points possible. Within
More informationPanel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43
Panel Data March 2, 212 () Applied Economoetrics: Topic March 2, 212 1 / 43 Overview Many economic applications involve panel data. Panel data has both cross-sectional and time series aspects. Regression
More informationNonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity
Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University
More informationECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor Aguirregabiria
ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor guirregabiria SOLUTION TO FINL EXM Monday, pril 14, 2014. From 9:00am-12:00pm (3 hours) INSTRUCTIONS:
More informationGMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails
GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,
More informationSIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011
SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1815 COWLES FOUNDATION FOR RESEARCH IN ECONOMICS
More information5 Introduction to the Theory of Order Statistics and Rank Statistics
5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order
More informationn E(X t T n = lim X s Tn = X s
Stochastic Calculus Example sheet - Lent 15 Michael Tehranchi Problem 1. Let X be a local martingale. Prove that X is a uniformly integrable martingale if and only X is of class D. Solution 1. If If direction:
More information13 Endogeneity and Nonparametric IV
13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,
More informationECON0702: Mathematical Methods in Economics
ECON0702: Mathematical Methods in Economics Yulei Luo SEF of HKU January 14, 2009 Luo, Y. (SEF of HKU) MME January 14, 2009 1 / 44 Comparative Statics and The Concept of Derivative Comparative Statics
More informationc 2007 Je rey A. Miron
Review of Calculus Tools. c 007 Je rey A. Miron Outline 1. Derivatives. Optimization 3. Partial Derivatives. Optimization again 5. Optimization subject to constraints 1 Derivatives The basic tool we need
More informationScattering for the NLS equation
Scattering for the NLS equation joint work with Thierry Cazenave (UPMC) Ivan Naumkin Université Nice Sophia Antipolis February 2, 2017 Introduction. Consider the nonlinear Schrödinger equation with the
More informationTesting for Regime Switching: A Comment
Testing for Regime Switching: A Comment Andrew V. Carter Department of Statistics University of California, Santa Barbara Douglas G. Steigerwald Department of Economics University of California Santa Barbara
More informationLikelihood inference for a nonstationary fractional autoregressive model
Likelihood inference for a nonstationary fractional autoregressive model Søren Johansen y University of Copenhagen and CREATES and Morten Ørregaard Nielsen z Cornell University and CREATES Preliminary
More informationTime is discrete and indexed by t =0; 1;:::;T,whereT<1. An individual is interested in maximizing an objective function given by. tu(x t ;a t ); (0.
Chapter 0 Discrete Time Dynamic Programming 0.1 The Finite Horizon Case Time is discrete and indexed by t =0; 1;:::;T,whereT
More informationOptimal Target Criteria for Stabilization Policy
Optimal Target Criteria for Stabilization Policy Marc P. Giannoni Columbia University y Michael Woodford Columbia University z February 7, Abstract This paper considers a general class of nonlinear rational-expectations
More informationExact unconditional tests for a 2 2 matched-pairs design
Statistical Methods in Medical Research 2003; 12: 91^108 Exact unconditional tests for a 2 2 matched-pairs design RL Berger Statistics Department, North Carolina State University, Raleigh, NC, USA and
More informationGMM estimation of spatial panels
MRA Munich ersonal ReEc Archive GMM estimation of spatial panels Francesco Moscone and Elisa Tosetti Brunel University 7. April 009 Online at http://mpra.ub.uni-muenchen.de/637/ MRA aper No. 637, posted
More informationChapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models
Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models Fall 22 Contents Introduction 2. An illustrative example........................... 2.2 Discussion...................................
More informationTime Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley
Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the
More informationh=1 exp (X : J h=1 Even the direction of the e ect is not determined by jk. A simpler interpretation of j is given by the odds-ratio
Multivariate Response Models The response variable is unordered and takes more than two values. The term unordered refers to the fact that response 3 is not more favored than response 2. One choice from
More informationBayesian Modeling of Conditional Distributions
Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings
More informationVolume 30, Issue 3. Monotone comparative statics with separable objective functions. Christian Ewerhart University of Zurich
Volume 30, Issue 3 Monotone comparative statics with separable objective functions Christian Ewerhart University of Zurich Abstract The Milgrom-Shannon single crossing property is essential for monotone
More informationVector fields Lecture 2
Vector fields Lecture 2 Let U be an open subset of R n and v a vector field on U. We ll say that v is complete if, for every p U, there exists an integral curve, γ : R U with γ(0) = p, i.e., for every
More informationMathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector
On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean
More informationThe Degree of the Splitting Field of a Random Polynomial over a Finite Field
The Degree of the Splitting Field of a Random Polynomial over a Finite Field John D. Dixon and Daniel Panario School of Mathematics and Statistics Carleton University, Ottawa, Canada fjdixon,danielg@math.carleton.ca
More information8 Periodic Linear Di erential Equations - Floquet Theory
8 Periodic Linear Di erential Equations - Floquet Theory The general theory of time varying linear di erential equations _x(t) = A(t)x(t) is still amazingly incomplete. Only for certain classes of functions
More informationLOGARITHMIC MULTIFRACTAL SPECTRUM OF STABLE. Department of Mathematics, National Taiwan University. Taipei, TAIWAN. and. S.
LOGARITHMIC MULTIFRACTAL SPECTRUM OF STABLE OCCUPATION MEASURE Narn{Rueih SHIEH Department of Mathematics, National Taiwan University Taipei, TAIWAN and S. James TAYLOR 2 School of Mathematics, University
More informationA Remark on Complete Convergence for Arrays of Rowwise Negatively Associated Random Variables
Proceedings of The 3rd Sino-International Symposium October, 2006 on Probability, Statistics, and Quantitative Management ICAQM/CDMS Taipei, Taiwan, ROC June 10, 2006 pp. 9-18 A Remark on Complete Convergence
More informationStochastic Processes (Master degree in Engineering) Franco Flandoli
Stochastic Processes (Master degree in Engineering) Franco Flandoli Contents Preface v Chapter. Preliminaries of Probability. Transformation of densities. About covariance matrices 3 3. Gaussian vectors
More informationECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University
ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University Instructions: Answer all four (4) questions. Be sure to show your work or provide su cient justi cation for
More informationECON2285: Mathematical Economics
ECON2285: Mathematical Economics Yulei Luo Economics, HKU September 17, 2018 Luo, Y. (Economics, HKU) ME September 17, 2018 1 / 46 Static Optimization and Extreme Values In this topic, we will study goal
More informationEstimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels.
Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels. Pedro Albarran y Raquel Carrasco z Jesus M. Carro x June 2014 Preliminary and Incomplete Abstract This paper presents and evaluates
More informationLocal disaggregation of demand and excess demand functions: a new question
Local disaggregation of demand and excess demand functions: a new question Pierre-Andre Chiappori Ivar Ekeland y Martin Browning z January 1999 Abstract The literature on the characterization of aggregate
More informationReal Analysis: Homework # 12 Fall Professor: Sinan Gunturk Fall Term 2008
Eduardo Corona eal Analysis: Homework # 2 Fall 2008 Professor: Sinan Gunturk Fall Term 2008 #3 (p.298) Let X be the set of rational numbers and A the algebra of nite unions of intervals of the form (a;
More informationParametric Inference on Strong Dependence
Parametric Inference on Strong Dependence Peter M. Robinson London School of Economics Based on joint work with Javier Hualde: Javier Hualde and Peter M. Robinson: Gaussian Pseudo-Maximum Likelihood Estimation
More informationEconS Advanced Microeconomics II Handout on Mechanism Design
EconS 503 - Advanced Microeconomics II Handout on Mechanism Design 1. Public Good Provision Imagine that you and your colleagues want to buy a co ee machine for your o ce. Suppose that some of you may
More informationRobust Con dence Intervals in Nonlinear Regression under Weak Identi cation
Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation Xu Cheng y Department of Economics Yale University First Draft: August, 27 This Version: December 28 Abstract In this paper,
More informationWeek 2 Spring Lecture 3. The Canonical normal means estimation problem (cont.).! (X) = X+ 1 X X, + (X) = X+ 1
Week 2 Spring 2009 Lecture 3. The Canonical normal means estimation problem (cont.). Shrink toward a common mean. Theorem. Let X N ; 2 I n. Let 0 < C 2 (n 3) (hence n 4). De ne!! (X) = X+ 1 Then C 2 X
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationOn the synchronization of a class of electronic circuits that exhibit chaos
Chaos, Solitons and Fractals 13 2002) 1515±1521 www.elsevier.com/locate/chaos On the synchronization of a class of electronic circuits that exhibit chaos Er-Wei Bai a, *, Karl E. Lonngren a, J.C. Sprott
More informationLECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT
MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes
More informationStatistical inference on Lévy processes
Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline
More informationSerial Correlation Robust LM Type Tests for a Shift in Trend
Serial Correlation Robust LM Type Tests for a Shift in Trend Jingjing Yang Department of Economics, The College of Wooster Timothy J. Vogelsang Department of Economics, Michigan State University March
More informationRank Estimation of Partially Linear Index Models
Rank Estimation of Partially Linear Index Models Jason Abrevaya University of Texas at Austin Youngki Shin University of Western Ontario October 2008 Preliminary Do not distribute Abstract We consider
More informationSIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012
SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 Revised March 2012 COWLES FOUNDATION DISCUSSION PAPER NO. 1815R COWLES FOUNDATION FOR
More informationIntroduction to Linear Algebra. Tyrone L. Vincent
Introduction to Linear Algebra Tyrone L. Vincent Engineering Division, Colorado School of Mines, Golden, CO E-mail address: tvincent@mines.edu URL: http://egweb.mines.edu/~tvincent Contents Chapter. Revew
More informationStochastic solutions of nonlinear pde s: McKean versus superprocesses
Stochastic solutions of nonlinear pde s: McKean versus superprocesses R. Vilela Mendes CMAF - Complexo Interdisciplinar, Universidade de Lisboa (Av. Gama Pinto 2, 1649-3, Lisbon) Instituto de Plasmas e
More informationDichotomy Of Poincare Maps And Boundedness Of Some Cauchy Sequences
Applied Mathematics E-Notes, 2(202), 4-22 c ISSN 607-250 Available free at mirror sites of http://www.math.nthu.edu.tw/amen/ Dichotomy Of Poincare Maps And Boundedness Of Some Cauchy Sequences Abar Zada
More informationEDRP lecture 7. Poisson process. Pawe J. Szab owski
EDRP lecture 7. Poisson process. Pawe J. Szab owski 2007 Counting process Random process fn t ; t 0g is called a counting process, if N t is equal total number of events that have happened up to moment
More informationMicroeconometrics: Clustering. Ethan Kaplan
Microeconometrics: Clustering Ethan Kaplan Gauss Markov ssumptions OLS is minimum variance unbiased (MVUE) if Linear Model: Y i = X i + i E ( i jx i ) = V ( i jx i ) = 2 < cov i ; j = Normally distributed
More informationApproximately Most Powerful Tests for Moment Inequalities
Approximately Most Powerful Tests for Moment Inequalities Richard C. Chiburis Department of Economics, Princeton University September 26, 2008 Abstract The existing literature on testing moment inequalities
More informationMAXIMUM LIKELIHOOD ESTIMATION AND UNIFORM INFERENCE WITH SPORADIC IDENTIFICATION FAILURE. Donald W. K. Andrews and Xu Cheng.
MAXIMUM LIKELIHOOD ESTIMATION AND UNIFORM INFERENCE WITH SPORADIC IDENTIFICATION FAILURE By Donald W. K. Andrews and Xu Cheng October COWLES FOUNDATION DISCUSSION PAPER NO. 8 COWLES FOUNDATION FOR RESEARCH
More informationStochastic Processes
Introduction and Techniques Lecture 4 in Financial Mathematics UiO-STK4510 Autumn 2015 Teacher: S. Ortiz-Latorre Stochastic Processes 1 Stochastic Processes De nition 1 Let (E; E) be a measurable space
More informationECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Winter 2017 TOPIC 3: MULTINOMIAL CHOICE MODELS
ECONOMETRICS II (ECO 2401) Victor Aguirregabiria Winter 2017 TOPIC 3: MULTINOMIAL CHOICE MODELS 1. Introduction 2. Nonparametric model 3. Random Utility Models - De nition; - Common Speci cation and Normalizations;
More informationCS 540: Machine Learning Lecture 2: Review of Probability & Statistics
CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()
More informationIntroduction Wavelet shrinage methods have been very successful in nonparametric regression. But so far most of the wavelet regression methods have be
Wavelet Estimation For Samples With Random Uniform Design T. Tony Cai Department of Statistics, Purdue University Lawrence D. Brown Department of Statistics, University of Pennsylvania Abstract We show
More informationEconomics 620, Lecture 18: Nonlinear Models
Economics 620, Lecture 18: Nonlinear Models Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 18: Nonlinear Models 1 / 18 The basic point is that smooth nonlinear
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables
More informationClassical and Bayesian inference
Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 11 The Prior Distribution Definition Suppose that one has a statistical model with parameter
More informationEstimation under Ambiguity (Very preliminary)
Estimation under Ambiguity (Very preliminary) Ra aella Giacomini, Toru Kitagawa, y and Harald Uhlig z Abstract To perform a Bayesian analysis for a set-identi ed model, two distinct approaches exist; the
More informationA formal statistical test for the number of factors in. the approximate factor models
A formal statistical test for the number of factors in the approximate factor models Alexei Onatski Economics Department, Columbia University September 26, 2006 Abstract In this paper we study i.i.d. sequences
More informationMeasuring robustness
Measuring robustness 1 Introduction While in the classical approach to statistics one aims at estimates which have desirable properties at an exactly speci ed model, the aim of robust methods is loosely
More informationOscillatory Mixed Di erential Systems
Oscillatory Mixed Di erential Systems by José M. Ferreira Instituto Superior Técnico Department of Mathematics Av. Rovisco Pais 49- Lisboa, Portugal e-mail: jferr@math.ist.utl.pt Sra Pinelas Universidade
More informationCentral Limit Theorem for Non-stationary Markov Chains
Central Limit Theorem for Non-stationary Markov Chains Magda Peligrad University of Cincinnati April 2011 (Institute) April 2011 1 / 30 Plan of talk Markov processes with Nonhomogeneous transition probabilities
More information1 Review of di erential calculus
Review of di erential calculus This chapter presents the main elements of di erential calculus needed in probability theory. Often, students taking a course on probability theory have problems with concepts
More informationNot Only What But also When A Theory of Dynamic Voluntary Disclosure
Not Only What But also When A Theory of Dynamic Voluntary Disclosure PRELIMINARY AND INCOMPLETE Ilan Guttman, Ilan Kremer and Andrzej Skrzypacz Stanford Graduate School of Business November 2011 1 Introduction
More information4. Duality and Sensitivity
4. Duality and Sensitivity For every instance of an LP, there is an associated LP known as the dual problem. The original problem is known as the primal problem. There are two de nitions of the dual pair
More informationTitleCoupled Nosé-Hoover Equations of Mo.
TitleCoupled Nosé-Hoover Equations of Mo Author(s) Fukuda, Ikuo; Moritsugu, Kei Citation Issue 15-8-18 Date Text Version author UL http://hdl.handle.net/119/56 DOI ights Osaka University Supplemental Material:
More informationNonparametric Trending Regression with Cross-Sectional Dependence
Nonparametric Trending Regression with Cross-Sectional Dependence Peter M. Robinson London School of Economics January 5, 00 Abstract Panel data, whose series length T is large but whose cross-section
More informationMath 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank
Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank David Glickenstein November 3, 4 Representing graphs as matrices It will sometimes be useful to represent graphs
More informationLecture Notes - Dynamic Moral Hazard
Lecture Notes - Dynamic Moral Hazard Simon Board and Moritz Meyer-ter-Vehn October 27, 2011 1 Marginal Cost of Providing Utility is Martingale (Rogerson 85) 1.1 Setup Two periods, no discounting Actions
More informationEXTREMAL POLYNOMIALS ON DISCRETE SETS
EXTREMAL POLYNOMIALS ON DISCRETE SETS A. B. J. KUIJLAARS and W. VAN ASSCHE [Received 14 April 1997] 1. Introduction It is well known that orthonormal polynomials p n on the real line can be studied from
More informationINEQUALITIES OF LIPSCHITZ TYPE FOR POWER SERIES OF OPERATORS IN HILBERT SPACES
INEQUALITIES OF LIPSCHITZ TYPE FOR POWER SERIES OF OPERATORS IN HILBERT SPACES S.S. DRAGOMIR ; Abstract. Let (z) := P anzn be a power series with complex coe - cients convergent on the open disk D (; R)
More informationChecking for Prior-Data Conßict. by Michael Evans Department of Statistics University of Toronto. and
Checking for Prior-Data Conßict by Michael Evans Department of Statistics University of Toronto and Hadas Moshonov Department of Statistics University of Toronto Technical Report No. 413, February 24,
More informationCh 4. Linear Models for Classification
Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,
More informationAn Analysis of the Difference of Code Lengths Between Two-Step Codes Based on MDL Principle and Bayes Codes
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 47, NO. 3, MARCH 2001 927 An Analysis of the Difference of Code Lengths Between Two-Step Codes Based on MDL Principle Bayes Codes Masayuki Goto, Member, IEEE,
More informationLecture Notes Part 7: Systems of Equations
17.874 Lecture Notes Part 7: Systems of Equations 7. Systems of Equations Many important social science problems are more structured than a single relationship or function. Markets, game theoretic models,
More information