(a). Bumps (b). Wavelet Coefficients

Size: px
Start display at page:

Download "(a). Bumps (b). Wavelet Coefficients"

Transcription

1 Incorporating Information on Neighboring Coecients into Wavelet Estimation T. Tony Cai Bernard W. Silverman Department of Statistics Department of Mathematics Purdue University University of Bristol West Lafayette, IN Bristol BS8 1TW U.S.A. U.K. Abstract In standard wavelet methods, the empirical wavelet coecients are thresholded term by term, on the basis of their individual magnitudes. Information on other coecients has no inuence on the treatment of particular coecients. We propose a wavelet shrinkage method that incorporates information on neighboring coecients into the decision making. The coecients are considered in overlapping blocks; the treatment of coecients in the middle of each block depends on the data in the whole block. The asymptotic and numerical performances of two particular versions of the estimator are investigated. We show that, asymptotically, one version of the estimator achieves the exact optimal rates of convergence over a range of Besov classes for global estimation, and attains adaptive minimax rate for estimating functions at a point. In numerical comparisons with various methods, both versions of the estimator perform excellently. Keywords: Adaptivity; Besov Space; Block Thresholding; James-Stein Estimator; Local Adaptivity; Nonparametric Regression; Wavelets; White Noise Model. 1

2 1 Introduction Consider the nonparametric regression model y i = f(t i )+z i (1) where t i = i=n for i = 1; 2;:::n; is the noise level, and the z i are i.i.d. N(0; 1). The function f() is an unknown function of interest. Wavelet methods have demonstrated success in nonparametric function estimation in terms of spatial adaptivity, computational eciency and asymptotic optimality. Standard wavelet methods achieve adaptivity through term-by-term thresholding of the empirical wavelet coecients. To obtain the wavelet coecients of the function estimate, each individual empirical wavelet coecient y is compared with a predetermined threshold, and is processed taking account solely of its own magnitude. Other coecients have no inuence on the estimate. Examples of shrinkage functions applied to individual coecients include the hard thresholding function h (y) = y I(jyj > ) and the soft thresholding function (y) s = sgn(y) (jyj,) +. For example, Donoho and Johnstone's (1995a) VisuShrink estimates the true wavelet coecients by soft thresholding with the universal threshold = (2 log n) 1=2. Hall et al. (1999) and Cai (1996, 1999a and 1999b) studied local block thresholding rules for wavelet function estimation. These threshold the empirical wavelet coecients in groups rather than individually, making simultaneous decisions to retain or to discard all the coecients within a block. The aim is to increase estimation accuracy by utilizing information about neighboring wavelet coecients. Estimators obtained by block thresholding enjoy a higher degree of spatial adaptivity than the standard term-by-term thresholding methods. The multiwavelet threshold estimators considered by Downie and Silverman (1998) also utilize block thresholding ideas. In the present paper, we propose a wavelet shrinkage method that incorporates into the thresholding decision information about neighboring coecients outside the block of current interest. The basic motivation is that if neighboring coecients contain some signal, then it is likely that the coecients of current direct interest also do, and so a lower threshold should be used. Two particular cases are considered. One method, which we call NeighBlock, estimates wavelet coecients simultaneously in groups, with the aim of gaining the advantages of the block thresholding method. In the other approach, called NeighCoe coecients are estimated individually, but with reference to their neighbors. Both methods are specied completely, with explicit denition of both the block size and the threshold level. After Section 2.1 in which basic notation and denitions are reviewed, the two estimators are dened in Section 2.2. We then investigate the two estimators both theoretically and by simulation. We show in Section 3 that the NeighBlock estimator enjoys a high degree of adaptivity and spatial adaptivity. Specically, we prove that the estimator simultaneously attains the exact optimal rate of convergence over a wide interval of the Besov classes with p 2 without prior knowledge of the smoothness of the underlying functions. Over the Besov classes with p < 2, the estimator simultaneously achieves the optimal convergence rate within a logarithmic factor. For estimating functions at a point, the estimator also 2

3 attains the local adaptive minimax rate. The theoretical properties of NeighCoe are discussed in Section 4. The estimator is within a logarithmic factor of being minimax over a range of Besov classes, and shares the pointwise optimality properties of NeighBlock. Technical details and proofs are given in Section 6. In Section 5, a simulation study of the two estimators is reported, together with a comparison on a data set collected in an anesthesiology study. Both estimators have excellent performance relative to conventional wavelet shrinkage methods. Our nite-sample comparison of the two estimators seems to rank the estimators in the opposite order to the theoretical discussion. If anything, the NeighCoe method is slightly superior. The estimators are appealing visually as well as quantitatively. The reconstructions jump where the target function jump; the reconstruction is smooth where the target function is smooth. They do not contain the spurious ne-scale structure contained in some wavelet estimators, but adapt well to subtle changes in the underlying functions. In comparison to the estimator of Hall et al. (1999), the NeighBlock estimator also enjoys a number of advantages. First, their estimator is not fully specied and no criterion is given for choosing the parameters objectively in nite sample cases. So the users need to select the block size and threshold level empirically. The NeighBlock estimator, on the other hand, is completely specied and easy to implement. Asymptotically, NeighBlock achieves the optimal local adaptivity while the estimator of Hall et al. does not. The web site Cai and Silverman (1999) contains SPlus scripts implementing both our estimators. It also describes additional simulation results not included in this paper. 2 The estimation method 2.1 Notation and conventions We shall assume that we are working within an orthonormal wavelet basis generated by dilation and translation of a compactly supported scaling function and a mother wavelet. We call a wavelet r-regular if has r vanishing moments and r continuous derivatives. For simplicity in exposition, we work with periodized wavelet bases on [0; 1], letting p jk(t) = jk (t, l); jk(t) p = jk (t, l); for t 2 [0; 1] l2z l2z where jk (t) =2 j=2 (2 j t, k); jk (t) =2 j=2 (2 j t, k): The collection f p j 0 k; k =1;:::;2 j 0 ; p jk; j j 0 0;k =1; :::; 2 j g is then an orthonormal basis of L 2 [0; 1], provided the primary resolution level j 0 is large enough to ensure that the support of the scaling functions and wavelets at level j 0 is not the whole of [0; 1]. The superscript \p" will be suppressed from the notation for convenience. An orthonormal wavelet basis has an associated exact orthogonal Discrete Wavelet Transform (DWT) that is norm-preserving and transforms sampled data into the wavelet coecient domain in O(n) steps. We use the standard device of transforming the problem 3

4 in the function domain into a problem, in the sequence domain, of estimating the wavelet coecients. See Daubechies (1992) and Strang (1992) for further details about the wavelets and the discrete wavelet transform. Wavelets are known for their excellent compression and localization properties. In very many cases of interest, information about a function is essentially contained in relatively small number of large coecients. Figure 1 displays the wavelet coecients of the wellknown test function Bumps (Donoho & Johnstone (1994)). It shows that large detail coecients come as groups; they cluster around the areas where the function changes signicantly. (a). Bumps (b). Wavelet Coefficients d1 d2 d3 d4 d5 d6 s Figure 1: Wavelet coecients. This example illustrates the motivation for our methods a coecient is more likely to contain information about a signal if neighboring coecients do also. Therefore when the observations are contaminated with noise, estimation accuracy might be improved by incorporating information on neighboring coecients. Indeed, as we shall see, our estimators show signicant numerical improvement over the conventional term-by-term thresholding estimators. Suppose we observe the data Y = fy i g as in (1). We shall assume that the noise level is known. Let ~ = W Y be the discrete wavelet transform of Y. Then ~ is an n-vector with elements ~ j0 k (k = 1; ; 2 j 0 ), which are the gross structure wavelet terms at the lowest resolution level, and ~ jk (j = 1; ;J, 1;k = 1; ; 2 j ), which are ne structure wavelet terms. Since the DWT is an orthogonal transform, the coecients are independently normally distributed with variance 2. For any particular estimation procedure based on the wavelet coecients, use the notation ^ for the estimate of the DWT of the values of f at the sample points. Up to the error involved in approximating f at the nest level by a wavelet series, the mean integrated square error of the estimation satises Ek ^f, fk 2 2 = n,1 Ek ^, k 2 : We therefore measure quality of recovery in terms of the mean square error in wavelet coecient space. 4

5 2.2 The NeighBlock and NeighCoe procedures We now dene the estimates studied in this paper. following stages: The NeighBlock estimator has the 1. Transform the data into the wavelet domain via the discrete wavelet transform: ~ = W Y. 2. At each resolution level j, group the empirical wavelet coecients into disjoint blocks b j i of length L 0 = [(log n)=2]. Each block b j i is extended by an amount L 1 = max(1; [L 0 =2]) in each direction to form overlapping larger blocks B j i of length L = L 0 +2L Within each block b j i, estimate the coecients simultaneously via a shrinkage rule ^ j;k = j i ~ j;k ; for all (j; k) 2 b j i : The shrinkage factor j i is chosen with reference to the coecients in the larger block B j i : j i =(1, L 2 =S 2 ji) + (2) where S 2 ji = (j;k)2b j i ~ 2 j;k (3) and = 4:5053. We can envision B j i as a sliding window which moves L 0 positions each time and, for each given window, only the half of the coecients in the center of the window are estimated. The choice of the block length L 0 and the threshold are discussed below. 4. Obtain the estimate of the function via the inverse discrete wavelet transform of the denoised wavelet coecients. The procedure is simple and easy to implement, at a computational cost of O(n). The NeighBlock procedure aims to combine the advantages previously found for block thresholding methods with those obtained by using information about neighboring coecients. In this case L log n, and the threshold is chosen by a James-Stein procedure; see Remark 2 below. Remark 1: If L 0 is not a power of two, then one or both of the b j i at the boundary is shortened to ensure all the b j i are nonoverlapping. In the periodic case, the corresponding B j i are kept of length L with b j i at the center. If periodic boundary conditions are not being used, then the b j i at the boundary are only extended in one direction to form B j i, again of length L. Remark 2: In the NeighBlock procedure, the thresholding constant is set to = 4:505:::, which is the solution of the equation, log =3. The value is derived from an oracle inequality introduced in Cai (1999). The estimator then enjoys superior numerical performance and asymptotic optimality, as our subsequent discussion shows. 5

6 Remark 3: The block length L 0 = [(log n)=2] depends on the sample size n only and the thresholding constant is an absolute constant. Therefore, in contrast to some other block thresholding methods, the various parameters are fully specied. Remark 4: The estimator can be modied by averaging over every possible position of the block centers. The resulting estimator sometimes has numerical advantages, at the cost of higher computational complexity. The NeighCoe procedure follows the same four steps as the NeighBlock estimator, but with L 0 = L 1 =1,L = 3, and =(2=3) log n. The eect is that each individual coecient is shrunk by an amount that may also depend on its immediate neighbors. See Section 4 for further details. Remark 5: NeighCoe uses lower threshold level than VisuShrink of Donoho and Johnstone (1995a). In NeighCoe, an coecient is estimated by zero only when the sum of squares of the empirical coecient and its immediate neighbors is less than (2 log n) 2, or the average of the squares is less than (2=3 log n) 2. We will investigate the asymptotic and numerical properties of the NeighBlock and NeighCoe estimators. Let us rst look at one example. The top-left panel of Figure 2 displays a noisy Doppler signal. We use NeighBlock and NeighCoe as well as three conventional methods, VisuShrink, SureShrink, and TI de-noising, to recover the true signal. The reconstructions are shown in Figure 2. Doppler + Noise NeighBlock VisuShrink NeighCoeff TI de-noising SureShrink Figure 2: The noisy Doppler signal and the reconstructions. The dotted line is the true signal. 6

7 All methods except SureShrink recover the smooth low frequency part reasonably well. SureShrink contains a signicant amount of local oscillations and is visually unpleasant. Both NeighBlock and NeighCoe automatically adapt to the changing frequency of the underlying signal. Both estimate the smooth and low frequency part accurately; at the same time, they also captures the more rapidly oscillating area between t = 0:1 and t = 0:4. In contrast, both VisuShrink and TI de-noising signicantly over-smooth in this region. SureShrink does better than VisuShrink and TI de-noising in recovering the high frequency part, but it contains a signicant amount of local oscillations and is visually unpleasant. None of the estimators do a particularly good job in the region t < 0:1 of very high frequency oscillation, partly because of the low sampling rate relative to the rate of oscillation; however, in contrast to any of the other estimators, both NeighBlock and NeighCoe do recover a little of the signal even in this region. Quantitatively, NeighBlock and NeighCoe are almost identical and both are signicantly better than the other methods. In this particular example, the ratios of the mean squared error of NeighBlock and NeighCoe to those of VisuShrink, SureShrink, and TI de-noising are 0.35, 0.72, and 0.45 respectively. See Section 5 for further numerical results. Inspection of wavelet coecients shows that NeighBlock NeighCoe, VisuShrink, and SureShrink use 33, 28, 15, and 61 detail coecients in the reconstruction, respectively. SureShrink retains many detail coecients in the low frequency area and as a result, the reconstruction contains spurious oscillations. VisuShrink keeps only 15 detail coecients and the reconstruction is over-smoothed. The additional smoothing inherent in the TIdenoising method has also led to over-smoothing. 3 Optimality of the NeighBlock procedure 3.1 Global properties As is traditional in the wavelet literature, we investigate the adaptivity of the NeighBlock procedure across Besov classes. Besov spaces are a very rich class of function spaces. They contain many traditional smoothness spaces such as Holder and Sobolev Spaces. We shall show that NeighBlock enjoys excellent adaptivity across a wide range of Besov classes. Full details of Besov spaces are given, for example, in DeVore and Popov (1988). For a given square-integrable function f on [0; 1], dene the scaling function and wavelet coecients jk = hf; jk i; jk = hf; jk i: The function f can be expanded into a wavelet series: f(x) = 2 j 0 k=1 j0 k j0 k(x)+ 1 2 j j=j 0 k=1 jk jk (x) (4) Let be the vector of the scaling function coecients, and for each j let j be the vector of the wavelet coecients at level j. Suppose > 0; 0 < p 1 and 0 < q 1. Then, roughly speaking, the Besov function norm of index (; p; q) quanties the size in an L p sense of the derivative of f 7

8 of order, with q giving a ner gradation; for a precise denition see DeVore and Popov (1988). Dene s = +1=2,1=p. For a given r-regular mother wavelet with r>, the Besov sequence norm of the wavelet coecients of a function f is then dened by where kk p + jj b s p;q ; jj q b s p;q = 1 j=j 0 2 jsq k j k q p: (5) It is an important fact (Meyer 1992) that the Besov function norm kfk B p;q is equivalent to the sequence norm of the wavelet coecients of f. The Besov class B p;q(m) is dened to be the set of all functions whose Besov norm is less than M. Denote the minimax risk over a function class F by R(F;n) = inf ^f n sup Ek ^f n, fk 2 2 f2f where ^f n are estimators based on n observed data points. Donoho and Johnstone (1998) showed that the minimax risk over a Besov class B p;q(m) is given by R(B p;q(m);n) n,2=(1+2) ; n!1 If attention is restricted to linear estimates, the corresponding minimax rate of convergence is n,0, with 0 +(1=p,, 1=p) = +1=2+(1=p,, 1=p) ; where p, = max(p; 2): (6) Therefore the traditional linear methods such as kernel, spline and orthogonal series estimates are suboptimal for estimation over the Besov bodies with p<2. We show in the following theorem that the NeighBlock method attains the exact optimal convergence rate over a wide range of the Besov scales. We denote by C a generic constant that may vary from place to place. Theorem 1 Suppose the wavelet is r-regular. Then the NeighBlock estimator satises sup f2b p;q (M ) Ek ^f n, fk 2 Cn,2=(1+2) for all M 2 (0; 1);2 (0;r);q 2 [1; 1] and p 2 [2; 1]. Thus, the NeighBlock estimator, without knowing the degree or amount of smoothness of the underlying function, attains the true optimal convergence rate that one could achieve knowing the regularity. The next theorem addresses the case of p<2, and shows that the NeighBlock method achieves advantages over linear methods even at the level of rates. Theorem 2 Assume that the wavelet is r-regular. Then the NeighBlock estimator is simultaneously within a logarithmic factor of minimax for p<2: sup f2b p;q (M ) Ek ^f n, fk 2 Cn,2=(1+2) (log n) (2,p)=p(1+2) for all M 2 (0; 1);2 [1=p; r);q 2 [1; 1] and p 2 [1; 2). The proofs of these theorems are given in Section 6. 8 (7) (8)

9 3.2 Local adaptation We now study the optimality of the NeighBlock procedure for estimating functions at a point. It is well known that for global estimation, it is possible to achieves complete adaptation without any cost in terms of convergence rate across a range of function classes. That is, one can achieve the same rate of convergence when the degree of smoothness is unknown as one could do if the degree of smoothness is known. But for estimation at a point, one must asymptotically pay a price (albeit only in log n terms) for not knowing the smoothness of the underlying function. Denote the minimax risk for estimating functions at a point t 0 over a function class F by R(F;n;t 0 ) = inf ^f n sup E( ^f n (t 0 ), f(t 0 )) 2 F Consider the Holder class (M). The optimal rate of convergence for estimating f(t 0 ) with known is n, where =2=(1+2). Brown and Low (1996b) and Lepski (1990) showed that even when is known to be one of two values, adaptation causes the rate of convergence to be worse by at least a logarithmic factor. They showed that the best one can do is (log n=n) when the smoothness parameter is unknown. We call (log n=n) the local adaptive minimax rate over the Holder class (M). The following theorem shows that NeighBlock achieves the local adaptive minimax rate over a wide range of Holder classes. Theorem 3 Suppose the wavelet is r-regular and has r vanishing moments with r. Let t 0 2 (0; 1) be xed. Then the NeighBlock estimator ^f n(t 0 ) of f(t 0 ) satises sup Ef ^f n(t 0 ), f(t 0 )g 2 C(n,1 log n) 2=(1+2) : (9) f2 (M ) Remark: As shown in Cai (1999b), the estimator of Hall, et al. (1999) does not achieve the optimal local adaptivity in terms of the rate of convergence; the block size L = C(log n) 1+ is too large. Therefore NeighBlock outperforms the estimator of Hall, et al. (1999) in terms of local adaptivity. 3.3 Denoising property In addition to the global and local estimation properties, the NeighBlock estimator enjoys an interesting smoothness property which should oer high visual quality of the reconstruction. The estimator, with high probability, removes pure noise completely. Theorem 4 If the target function is the zero function f 0, then, with probability tending to 1 as n!1, the NeighBlock estimator is also the zero function, i.e., there exist universal constants P n such that P ( ^f n 0) P n! 1; as n!1 (10) 9

10 4 Properties of the NeighCoe estimator The NeighCoe estimator is intuitively appealing and easy to implement. It can be shown that the estimator is locally adaptive and is within a logarithmic factor of being minimax over a wide range of Besov classes. This is the same asymptotic performance as VisuShrink. We summarize the results without proof in the following theorems. We shall see subsequently that, in most cases, the estimator enjoys superior numerical performance to many classical wavelet estimators, such as VisuShrink, SureShrink and TI-denoising estimators in most cases. It even outperforms NeighBlock, even though that estimator apparently has better asymptotic properties. The rst result shows that the global performance of the NeighCoe estimator is simultaneously within a logarithmic factor of minimax over a wide range of Besov classes, and the second result shows that NeighCoe has the same good pointwise behavior as NeighBlock. Theorem 5 Assume that the wavelet is r-regular. Then the NeighCoe estimator satises sup Ek ^f f2bp;q (M n, fk 2 C(n,1 log n) 2=(1+2) (11) ) for all M 2 (0; 1);2 (0;r);q 2 [1; 1] and p 2 [1; 1]. Theorem 6 Assume that the wavelet is r-regular and has r vanishing moments with r. Let t 0 2 (0; 1) be xed. Then the NeighCoe estimator satises sup E( ^f n(t 0 ), f(t 0 )) 2 C(n,1 log n) 2=(1+2) : (12) f2 (M ) 5 Numerical comparison A simulation study was conducted to compare the numerical performance of the NeighBlock and NeighCoe estimators with Donoho and Johnstone's VisuShrink and SureShrink as well as Coifman and Donoho's Translation-Invariant (TI) denoising method. SureShrink selects the threshold at each resolution level by minimizing Stein's (1981) unbiased estimate of risk. In the simulation, we use the hybrid method proposed in Donoho and Johnstone (1995b). The TI-denoising method was introduced by Coifman and Donoho (1995), and is equivalent to averaging over estimators based on all the shifts of the original data. This method has various advantages over the universal thresholding methods. For further details see the original papers. We implement the NeighBlock and NeighCoe estimators in the software package S+Wavelets. The programs are available from the web site Cai and Silverman (1999). We compare the numerical performance of the methods using eight test functions representing dierent level of spatial variability. The test functions are plotted in Figure 5. Sample sizes ranging from n = 512 to n = 8192 and root-signal-to-noise ratios (RSNR) from 3 to 7 were considered. The RSNR is the ratio of the standard deviation of the function values to the standard deviation of the noise. Several dierent wavelets were used. 10

11 For reasons of space, we only report in detail the results for one particular case, using Daubechies' compactly supported wavelet Symmlet 8 and RSNR equal to 3. Table 1 reports the average squared errors over 60 replications with sample sizes ranging from n = 512 to n = A graphical presentation is given in Figure 3. Dierent combinations of wavelets and signal-to-noise ratios yield basically the same results; for details see the web site Cai and Silverman (1999). NeighCoeff vs VisuShrink NeighCoeff vs SureShrink Doppler HeaviSine Bumps Blocks Spikes Blip Corner Wave Doppler HeaviSine Bumps Blocks Spikes Blip Corner Wave NeighCoeff vs TI NeighCoeff vs NeighBlock Doppler HeaviSine Bumps Blocks Spikes Blip Corner Wave Doppler HeaviSine Bumps Blocks Spikes Blip Corner Wave Figure 3: RSNR=3. The vertical bars represent the log ratios of the MSEs of various estimators to the corresponding MSE of the NeighCoe estimator. The higher the bar the better the relative performance of the NeighCoe estimator, and a value of zero means that the estimators have equal performance. For each signal the bars are ordered from left to right by the sample sizes (n = 512; 1024; 2048; 4096; 8192). The NeighBlock and NeighCoe methods both uniformly outperform VisuShrink in all examples. For ve of the eight test functions, Doppler, Bumps, Blocks, Spikes and Blip, our methods have better precision with sample size n than VisuShrink with sample size 2n for all sample sizes where the comparison is possible. The NeighCoe method is slightly better than NeighBlock in almost all cases, and outperforms the other methods as well. The NeighCoe method is also better than TI-denoising in most cases, especially when the underlying function is of signicant spatial variability. In terms of the mean square error criterion, the only conceivable competitor among the standard methods is SureShrink. Apart from being somewhat superior to SureShrink in mean square error, our methods yield noticeably better results visually. Our estimates do not contain the spurious ne-scale eects that are often contained in the SureShrink estimator. The curious behavior of some of the methods with the Waves signal calls for some explanation. Throughout, the primary resolution level j 0 = [log 2 log n]+1was used for all 11

12 methods. Thus, j 0 = 3 for n 2048, and j 0 = 4 for n = 4096 and This change in the value of j 0 aects whether or not the high frequency eect in the Waves signal is felt in the lowest level of wavelet coecients. For j 0 = 3, the standard methods all smooth out the high frequency eect to some extent, because of applying a soft threshold with xed threshold. An attractive feature of the NeighCoe and NeighBlock methods is that they are not sensitive to the choice of primary resolution level in this way, because the threshold adapts to the presence of signal in all the coecients. Figure 4 shows a typical segment of the result of the four methods applied to the inductance plethysmography data analyzed, for example, by Abramovich et al. (1998). It can be seen that the SureShrink estimator allows through high frequency eects that are almost certainly spurious while both the standard and the translation-invariant versions of VisuShrink smooth out features of the curve NeighBlock VisuShrink TI denoising NeighBlock NeighCoeff SureShrink Figure 4: A segment of the data and curve estimates for the inductance plethysmography data. Left gure: NeighBlock (solid), VisuShrink (dotted), TI denoising (dashed). Right gure: Neigh- Block (solid), NeighCoe (dotted), SureShrink (dashed). Both the VisuShrink and TI denoising estimates smooth out the broad features, while the SureShrink estimate contains high frequency eects near times 300 and 335, both of which are presumably spurious. It would be interesting to include practical comparisons with the block thresholding estimator of Hall et al. (1999). Their method requires the selection of smoothing parameters block length and threshold level neither of which is completely specied and no criterion is given for choosing the parameters objectively in nite sample cases, so any detailed comparison would have to be left for future work. However, simulation results by Hall et al. (1997) show that even the translation-averaged version of the estimator has little advantage over VisuShrink when the signal to noise ratio is high. Our simulation shows that NeighBlock uniformly outperforms VisuShrink in all examples, and indeed the relative performance of VisuShrink is even worse for values of RSNR higher than the one presented in detail. Therefore we expect our estimator to perform favorably over the estimator of Hall et al. in terms of mean squared error, at least in the case of high signal-to-noise-ratio. 12

13 6 Proofs 6.1 Sequence space approximation We shall prove Theorems 1, 2 and 5 by using the sequence space approach introduced by Donoho and Johnstone (1998). A key step is to use the asymptotic equivalence results presented by Brown and Low (1996a) and to approximate the problem of estimating f from the noisy observations in (1) by the problem of estimating the wavelet coecient sequence of f contaminated with i.i.d. Gaussian noise. Donoho and Johnstone (1998) show a strong equivalence result on the nonparametric regression and the white noise model over the Besov classes B p;q(m). When the wavelet is r-regular with r > and p; q 1, then a simultaneously near-optimal estimator in the sequence estimation problem can be applied to the empirical wavelet coecients in the function estimation problem in (1), and will be a simultaneously near-optimal estimator in the function estimation problem. For further details about the equivalence and approximation arguments, the readers are referred to Donoho and Johnstone (1995b), (1998) and (1999) and Brown and Low (1996a). For approximation results, see also Chambolle et al. (1998). Under the correspondence between the estimation problem in function space and the estimation problem in sequence space, it suces to consider the following sequence estimation problem. 6.2 Estimation in sequence space by NeighBlock and NeighCoe Suppose we observe sequence data y jk = jk + n,1=2 z jk ; j 0; k =1; 2; ; 2 j (13) where z jk are i.i.d. N(0; 1). The mean array is the object that we wish to estimate. We assume that is in some Besov Body s p;q(m) =f : kk b s p;q Mg, where the norm is as dened in (5) above. Make the usual calibration s = +1=2, 1=p. Donoho and Johnstone (1998) show that the minimax rate of convergence for estimating over the Besov body s p;q(m) isn,2=(1+2) as n!1. The accuracy of estimation is measured by the expected squared error R(^; ) =E P j;k(^ j;k, j;k ) 2. We now approach this sequence estimation problem using a procedure corresponding to NeighBlock. Let J = [log 2 n]. Divide each resolution level j 0 j <J into nonoverlapping blocks of length L 0 = [(log n)=2]. Again denote by b j i the i-th block at level j and similarly dene B j i to be the larger block obtained by extending b j i by [(log n)=4] elements in each direction. Dene Sji 2 to be the sum of the yjk 2 over B j i,by analogy with (3). Now estimate by ^ with ^ jk = 8 >< >: y jk for j j 0 (1, n,1 L 2 =Sji) 2 + y jk for (j; k) 2 b j i ; j 0 j <J 0 for j J (14) 13

14 Similarly, for each (j; i), dene ~B j i to be the large block containing (j; i, 1), (j; i), and (j; i + 1), and dene ~S ji 2 = y 2 j;i,1 + yj;i 2 + y 2 j;i+1 to be the sum of squares over ~B j i. W dene a corresponding procedure to NeighCoe: ^ jk = 8 >< >: y jk for j j 0 (1, 2 2 n,1 log n= ~S ji) 2 + y jk for j 0 j <J 0 for j J For the estimator (14), we have the following minimax results, demonstrating that the estimator attains the exact minimax rate over all the Besov Bodies s p;q(m) with p 2, and the exact minimax rate up to a logarithmic term for p<2. Theorem 7 Dene ^ as in (14). Then, as n!1, sup Ek^, k 2 s p;q (M 2 ) For the estimator (15), we have ( Cn,2=(1+2) for p 2 Cn,2=(1+2) (log n) (2,p)=fp(1+2)g for p<2 and p 1: Theorem 8 Dene ^ as in (15). Then, for p 1, as n!1, sup Ek^, k 2 s p;q (M 2 C(log n=n) 2=(1+2) : ) The results of Theorem 1 and 2 follow from Theorem 7 and the equivalence and the approximation arguments discussed in Section 6.1. And similarly, Theorem 5 follows from Theorem Proof of the main results We will prove Theorems 7 and 8 as well as Theorems 3 and 6. The proof of Theorem 4 is straightforward. The key results used in the proofs are the following oracle inequalities. Lemma 1 Assume that y j;k, ^ j;k and ^ j;k are given as in (13), (14), and (15) respectively. Then, dening > 1 by, log =3, for each i and j 0 j <J E(^ j;k, j;k ) 2 ( 2 n,1 log n ^ j;k)+2n 2,2 2 : (16) (j;k)2b j i (j;k)2b j i E(^ j;i, j;i ) 2 (2 2 n,1 log n) ^ (j;k)2 ~ B j i (15) 2 j;k +2 2 n,2 (log n) 1=2 : (17) The proof of this lemma is an extension of the proof of Theorem 1 of Cai (1999a). First consider (16). For j; k in B j i dene ^ y j;k =(1, n,1 L 2 =S 2 ji) + y jk : Since ^ y j;k = ^ j;k for (j; k) in b j i, extending the sum from b j i to B j i, and replacing ^ by ^ y, can only increase the left hand side of (16). The argument of Theorem 1 and Lemma 2 of Cai (1999a) shows that the inequality holds with these changes, completing the proof of (16). The proof of (17) follows from Theorem 1 in Cai (1999) and the following upper bound for the tail probability of the 2 distribution. 14

15 Lemma 2 Let >1. Then P ( 2 m >m),1=2 (, 1),1 m,1=2 e, m 2 (,log,1) : (18) We also recall two elementary inequalities between two dierent `p norms, and a bound for a certain sum. Lemma 3 Let x 2 IR m, and 0 <p 1 p 2 1. Then the following inequalities hold: kxk p2 kxk p1 m 1 p, 1 1 p 2 kxk p2 (19) P Lemma 4 Let 0 <a<1 and S = fx 2 IR k : ki=1 x a i B; x i 0; i =1; ;kg. Then for > 0, sup x2s k i=1 (x i ^ ) B 1,a : Proof of Theorems 7 and 8: First consider Theorem 7. Let y and ^ be given as in (13) and (14) respectively. Then, Ek^, k 2 2 = j<j 0 k E(^ jk, jk ) 2 + J,1 j=j 0 say. We bound the term S 2 by using Lemma 1. Let A j i = k E(^ jk, jk ) 2 + (j;k)2b j i 2 jk; 1 j=j k 2 jk S 1 + S 2 + S 3 ; (20) the sum of squared coecients within the block B j i. We then split up the sum dening S 2 into sums over the individual blocks b j i, and apply the oracle inequality (16). Since L = log n and the number of blocks is denitely less than n, this yields S 2 = J,1 j=j 0 k E(^ jk, jk ) 2 C J,1 j=j 0 i (A j i ^ 2 n,1 L)+2n,1 2 : (21) Note also that, since 2 s p;q(m), we have 2 js k j k p M for each j. We now complete the proof for the two cases separately. The case p 2 For 2 s p;q(m), Lemma 3 implies that k j k 2 2 (2 j ) 2( 1 2, 1 p ) k j k 2 p M 2 2 2j( 1 2, 1 p,s) = M 2 2,2j : (22) It follows that S 1 + S 3 2 j 0 n,1 2 + so that S 1 + S 3 can be neglected. 1 j=j M 2 2,2j = o(n,2=(1+2) ); (23) 15

16 We divide the sum in (21) into two parts. Choose J 1 such that 2 J 1 n 1=(1+2). Then, J 1,1 j=j 0 i (A j i ^ 2 n,1 L) and, making use of the bound (22), J,1 j=j 1 i (A j i ^ 2 n,1 L) J 1,1 j=j 0 i J,1 j=j 1 i 2 n,1 L C2 J 1 n,1 Cn,2=(1+2) ; (24) J,1 A j i 2 k j k 2 2 Cn,2=(1+2) : (25) j=j 1 Combining (24) and (25) demonstrates that S 2 Cn,2=(1+2), completing the proof for this case. The case p < 2 with p 1: For 2 s p;q(m), Lemma 3 now yields k j k 2 2 k j k 2 p M 2 2,2js. The assumption p 1 implies that s 1, so that 2 S 3 C 1 j=j 2,2js Cn,2s Cn,1 : Thus S 1 + S 3 = o(n,2=(1+2) )asbefore. Now let J 2 be an integer satisfying 2 J 2 n 1=(1+2) (log n),(2,p)=p(1+2). Then, by an argument analogous to that leading to (24), J 2,1 j=j 0 i (A j i ^ 2 n,1 L) J 2,1 j=j 0 i 2 n,1 L Cn,2=(1+2) (log n),(2,p)=p(1+2) : (26) Turning to the other part of S 2, it follows by convexity that, for each j, i (A j i ) p=2 i (j;k)2b j i ( 2 jk) p=2 2 k Applying Lemma 4 with a = p=2, we have, after some algebra, J,1 j=j 2 i ( 2 jk) p=2 2M p 2,jsp : (A j i ^ 2 n,1 L) Cn,2=(1+2) (log n),(2,p)=p(1+2) : (27) We complete the proof of Theorem 7 by combining the bounds (26) and (27), as in the case p 2: The proof of Theorem 8 is similar, using the oracle inequality (17) instead of (16). Proof of Theorems 3 and 6: The proofs of Theorems 3 and 6 are similar to each other. We prove Theorem 6 rst. Suppose we observe the data Y = fy i g as in (1). Let ~ =W n,1=2 Y be the discrete wavelet transform of n,1=2 Y. Write ~ =( ~ j0 1; ; ~ j0 2 j 0 ; ~ j0 1; ; ~ j0 2 j 0 ; ; ~ J,1;1 ; ; ~ J,1;2 J,1) T 16

17 where ~ j0 k are the gross structure terms at the lowest level, and the coecients ~ jk (j = 1; ;J, 1;k = 1; ; 2 j ) are ne structure wavelet terms. Denote j;k hf; j;k i and j;k hf; j;k i. One may write ~ jk = jk + a j;k + n,1=2 z jk (28) where a j;k are the approximation errors and z jk 's are i.i.d. N(0; 1). Lemma 5 below is a consequence of the vanishing moments conditions on and. Lemma 5 Under the conditions of Theorem 6, there exists a constant C > 0 such that for all f 2 (M), j j;k jc2,j(1=2+) and ja j;k jcn, 2,j=2 : (29) It follows from the simple inequality, E( P n i=1 i ) 2 ( P n i=1 (Ei 2 ) 1=2 ) 2, 2 4 2j j E( ^f n(t 0 ), f(t 0 )) 2 = E 2 4 2j 0 k=1 k=1 (E(^ j0 k, j0 k) 2 2 j 0 k(t 0 )) 1=2 + (Q 1 + Q 2 + Q 3 ) 2 : (^ j0 k, j0 k) j0 k(t 0 )+ J,1 2 j j=j 0 k=1 j=j 0 k=1 (^ jk, jk ) jk (t 0 ) (E(^ jk, jk ) 2 2 jk(t 0 )) 1= j j=j k=1 j jk jk (t 0 )j It is obvious that Q 1 = O(n,1 ). Note that there are at most N jk at level j that are nonvanishing at t 0, where N is the length of the support of. Denote K(t 0 ; j) = fk : j;k(t 0 ) 6= 0g. Then jk(t 0 ; j)j N: Now (29) yields Q 3 = 1 2 j j=j k=1 j jk jj jk (t 0 )j 1 j=j For Q 2, we use the oracle inequality (17) as well as (29). Q 2 J,1 j=j 0 k2k(t 0 ;j) J,1 Nk k 1 2 j=2 C2,j(1=2+) Cn, : (30) 2 j=2 k k 1 (E(^ jk, jk ) 2 ) 1=2 C 2 j=2 [(2,j(1+2) +2,j n,2 ) ^ (n,1 log n)+n,2 (log n) 1=2 ] 1=2 j=j 0 = C(log n=n) =(1+2) : (31) Theorem 6 is proved by putting together the three terms, Q 1, Q 2 and Q 3. To prove Theorem 3, we use the same bounds for Q 1 and Q 3, and use the oracle inequality (16) to bound the term Q 2. Acknowledgment Part of this work was carried out while Bernard Silverman was a Fellow at the Center for Advanced Study in the Behavioral Sciences, Stanford, supported by National Science Foundation Grant number SBR He is also grateful for the nancial support of the British Engineering and Physical Sciences Research Council

18 References [1] Abramovich, F., Sapatinas, T. and Silverman, B.W. (1998). Wavelet thresholding via a Bayesian approach. J. Roy. Statist. Soc. Ser. B, 60, 725{749. [2] Brown, L.D. and Low, M.G. (1996a). Asymptotic equivalence of nonparametric regression and white noise. Ann. Statist., 24, 2384{2398. [3] Brown, L.D. and Low, M.G. (1996b). A constrained risk inequality with applications to nonparametric functional estimation. Ann. Statist., 24, 2524{2335. [4] Cai, T. (1996). Minimax wavelet estimation via block thresholding. Technical Report #96-41, Department of Statistics, Purdue University. [5] Cai, T. (1999a). Adaptive wavelet estimation: a block thresholding and oracle inequality approach. Ann. Statist., in press. [6] Cai, T. (1999b). Wavelet regression via block thresholding: adaptivity and the choice of block size and threshold level. Technical Report #99-14, Department of Statistics, Purdue University. [7] Cai, T. and Silverman, B. W. (1999). Incorporating information on neighboring coef- cients into wavelet estimation. Web page available at [8] Chambolle, A., DeVore, R., Lee, N. and Lucier, B. (1998). Nonlinear wavelet image processing: Variational problems, compression, and noise removal through wavelet shrinkage. IEEE Transactions on Image Processing, 70, 319{335. [9] Coifman, R.R. and Donoho, D.L. (1995). Translation invariant denoising. In A. Antoniadis and G. Oppenheim (eds), Wavelets and Statistics, Lecture Notes in Statistics 103. New York: Springer-Verlag, pp. 125{150. [10] Daubechies, I. (1992). Ten Lectures on Wavelets. Philadelphia: SIAM. [11] DeVore, R. and Popov, V. (1988). Interpolation of Besov spaces. Trans. Amer. Math. Soc., 305, 397{414. [12] Donoho, D.L. and Johnstone, I.M. (1995a). Ideal spatial adaptation via wavelet shrinkage. Biometrika, 81, 425{455. [13] Donoho, D.L. and Johnstone, I.M. (1995b). Adapting to unknown smoothness via wavelet shrinkage, J. Amer. Statist. Assoc., 90, 1200{1224. [14] Donoho, D.L. and Johnstone, I.M. (1998). Minimax Estimation via Wavelet Shrinkage, Ann. Statist., 26, 879{921. [15] Donoho, D.L. and Johnstone, I.M. (1999). Asymptotic Minimaxity of Wavelet Estimators with Sampled Data. Statist. Sinica, 9, 1{32. 18

19 [16] Downie, T.R. and Silverman, B.W. (1998). The discrete multiple wavelet transform and thresholding methods. IEEE Transactions in Signal Processing, 46, 2558{2561. [17] Hall, P., Kerkyacharian, G. and Picard, D. (1999). On the minimax optimality of block thresholded wavelet estimators. Statist. Sinica, 9, 33{50. [18] Hall, P., Penev, S., Kerkyacharian, G. and Picard, D. (1997). Numerical performance of block thresholded wavelet estimators. Statist. Comput., 7, 115{124. [19] Lepski, O.V. (1990). On a problem of adaptive estimation in white gaussian noise. Theory of Probability and Appl., 35, 454{466. [20] Marron, J.S., Adak, S., Johnstone, I.M., Neumann, M.H. and Patil, P. (1998). Exact risk analysis of wavelet regression. J. Comput. Graph. Statist., 7, 278{309. [21] Meyer, Y. (1992). Wavelets and Operators. Cambridge: Cambridge University Press. [22] Stein, C. (1981). Estimation of the mean of a multivariate normal distribution. Ann. Statist., 9, 1135{1151. [23] Strang, G. (1992). Wavelet and dilation equations: a brief introduction. SIAM Review, 31, 614{

20 Doppler HeaviSine Bumps Blocks Spikes Blip Corner Wave Figure 5: Test functions. Doppler, HeaviSine, Bumps and Blocks are from Donoho and Johnstone (1995a). Blip and Wave are from Marron et al. (1995). The test functions are normalized so that every function has standard deviation 10. Formulae for Spikes and Corner are given in Cai (1999a). 20

21 Table 1: Mean Squared Error From 60 Replications (RSNR=3) n NeighCoe NeighBlock SureShrink TI-denoising VisuShrink Doppler HeaviSine Bumps Blocks Spikes Blip Corner Wave

Introduction Wavelet shrinage methods have been very successful in nonparametric regression. But so far most of the wavelet regression methods have be

Introduction Wavelet shrinage methods have been very successful in nonparametric regression. But so far most of the wavelet regression methods have be Wavelet Estimation For Samples With Random Uniform Design T. Tony Cai Department of Statistics, Purdue University Lawrence D. Brown Department of Statistics, University of Pennsylvania Abstract We show

More information

Adaptive Wavelet Estimation: A Block Thresholding and Oracle Inequality Approach

Adaptive Wavelet Estimation: A Block Thresholding and Oracle Inequality Approach University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1999 Adaptive Wavelet Estimation: A Block Thresholding and Oracle Inequality Approach T. Tony Cai University of Pennsylvania

More information

Wavelet Shrinkage for Nonequispaced Samples

Wavelet Shrinkage for Nonequispaced Samples University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Wavelet Shrinkage for Nonequispaced Samples T. Tony Cai University of Pennsylvania Lawrence D. Brown University

More information

Discussion of Regularization of Wavelets Approximations by A. Antoniadis and J. Fan

Discussion of Regularization of Wavelets Approximations by A. Antoniadis and J. Fan Discussion of Regularization of Wavelets Approximations by A. Antoniadis and J. Fan T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania Professors Antoniadis and Fan are

More information

ON BLOCK THRESHOLDING IN WAVELET REGRESSION: ADAPTIVITY, BLOCK SIZE, AND THRESHOLD LEVEL

ON BLOCK THRESHOLDING IN WAVELET REGRESSION: ADAPTIVITY, BLOCK SIZE, AND THRESHOLD LEVEL Statistica Sinica 12(2002), 1241-1273 ON BLOCK THRESHOLDING IN WAVELET REGRESSION: ADAPTIVITY, BLOCK SIZE, AND THRESHOLD LEVEL T. Tony Cai University of Pennsylvania Abstract: In this article we investigate

More information

A Data-Driven Block Thresholding Approach To Wavelet Estimation

A Data-Driven Block Thresholding Approach To Wavelet Estimation A Data-Driven Block Thresholding Approach To Wavelet Estimation T. Tony Cai 1 and Harrison H. Zhou University of Pennsylvania and Yale University Abstract A data-driven block thresholding procedure for

More information

On the Estimation of the Function and Its Derivatives in Nonparametric Regression: A Bayesian Testimation Approach

On the Estimation of the Function and Its Derivatives in Nonparametric Regression: A Bayesian Testimation Approach Sankhyā : The Indian Journal of Statistics 2011, Volume 73-A, Part 2, pp. 231-244 2011, Indian Statistical Institute On the Estimation of the Function and Its Derivatives in Nonparametric Regression: A

More information

Comments on \Wavelets in Statistics: A Review" by. A. Antoniadis. Jianqing Fan. University of North Carolina, Chapel Hill

Comments on \Wavelets in Statistics: A Review by. A. Antoniadis. Jianqing Fan. University of North Carolina, Chapel Hill Comments on \Wavelets in Statistics: A Review" by A. Antoniadis Jianqing Fan University of North Carolina, Chapel Hill and University of California, Los Angeles I would like to congratulate Professor Antoniadis

More information

Nonparametric Regression

Nonparametric Regression Adaptive Variance Function Estimation in Heteroscedastic Nonparametric Regression T. Tony Cai and Lie Wang Abstract We consider a wavelet thresholding approach to adaptive variance function estimation

More information

The Root-Unroot Algorithm for Density Estimation as Implemented. via Wavelet Block Thresholding

The Root-Unroot Algorithm for Density Estimation as Implemented. via Wavelet Block Thresholding The Root-Unroot Algorithm for Density Estimation as Implemented via Wavelet Block Thresholding Lawrence Brown, Tony Cai, Ren Zhang, Linda Zhao and Harrison Zhou Abstract We propose and implement a density

More information

An Introduction to Wavelets and some Applications

An Introduction to Wavelets and some Applications An Introduction to Wavelets and some Applications Milan, May 2003 Anestis Antoniadis Laboratoire IMAG-LMC University Joseph Fourier Grenoble, France An Introduction to Wavelets and some Applications p.1/54

More information

Design of Image Adaptive Wavelets for Denoising Applications

Design of Image Adaptive Wavelets for Denoising Applications Design of Image Adaptive Wavelets for Denoising Applications Sanjeev Pragada and Jayanthi Sivaswamy Center for Visual Information Technology International Institute of Information Technology - Hyderabad,

More information

Asymptotic Equivalence and Adaptive Estimation for Robust Nonparametric Regression

Asymptotic Equivalence and Adaptive Estimation for Robust Nonparametric Regression Asymptotic Equivalence and Adaptive Estimation for Robust Nonparametric Regression T. Tony Cai 1 and Harrison H. Zhou 2 University of Pennsylvania and Yale University Abstract Asymptotic equivalence theory

More information

DE-NOISING BY. David L. Donoho. Department of Statistics. Stanford University. Abstract

DE-NOISING BY. David L. Donoho. Department of Statistics. Stanford University. Abstract DE-NOSNG BY SOFT-THRESHOLDNG David L. Donoho Department of Statistics Stanford University Abstract Donoho and Johnstone (1992a) proposed a method for reconstructing an unknown function f on [0; 1] from

More information

Wavelets and Image Compression. Bradley J. Lucier

Wavelets and Image Compression. Bradley J. Lucier Wavelets and Image Compression Bradley J. Lucier Abstract. In this paper we present certain results about the compression of images using wavelets. We concentrate on the simplest case of the Haar decomposition

More information

Choice of Thresholds. for Wavelet Shrinkage Estimate of the Spectrum. Hong-ye Gao. MathSoft, Inc Westlake Ave. N., Suite 500.

Choice of Thresholds. for Wavelet Shrinkage Estimate of the Spectrum. Hong-ye Gao. MathSoft, Inc Westlake Ave. N., Suite 500. Choice of Thresholds for Wavelet Shrinkage Estimate of the Spectrum Hong-ye Gao Statistical Sciences Division MathSoft, Inc. 1700 Westlake Ave. N., Suite 500 Seattle, WA 98109 Abstract We study the problem

More information

Smooth functions and local extreme values

Smooth functions and local extreme values Smooth functions and local extreme values A. Kovac 1 Department of Mathematics University of Bristol Abstract Given a sample of n observations y 1,..., y n at time points t 1,..., t n we consider the problem

More information

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 011/17 BLOCK-THRESHOLD-ADAPTED

More information

Interpolating Scaling Vectors: Application to Signal and Image Denoising

Interpolating Scaling Vectors: Application to Signal and Image Denoising Interpolating Scaling Vectors: Application to Signal and Image Denoising Karsten Koch Philipps Universität Marburg FB 12 Mathematik und Informatik Hans-Meerwein Str, Lahnberge 35032 Marburg Germany Abstract

More information

A DATA-DRIVEN BLOCK THRESHOLDING APPROACH TO WAVELET ESTIMATION

A DATA-DRIVEN BLOCK THRESHOLDING APPROACH TO WAVELET ESTIMATION The Annals of Statistics 2009, Vol. 37, No. 2, 569 595 DOI: 10.1214/07-AOS538 Institute of Mathematical Statistics, 2009 A DATA-DRIVEN BLOCK THRESHOLDING APPROACH TO WAVELET ESTIMATION BY T. TONY CAI 1

More information

Adaptive Bayesian Wavelet Shrinkage. Hugh A. Chipman, Eric D. Kolaczyk, Robert E. McCulloch

Adaptive Bayesian Wavelet Shrinkage. Hugh A. Chipman, Eric D. Kolaczyk, Robert E. McCulloch Adaptive Bayesian Wavelet Shrinkage Hugh A. Chipman, Eric D. Kolaczyk, Robert E. McCulloch To appear in the Journal of the American Statistical Association. Hugh A. Chipman and Robert E. McCulloch are

More information

How much should we rely on Besov spaces as a framework for the mathematical study of images?

How much should we rely on Besov spaces as a framework for the mathematical study of images? How much should we rely on Besov spaces as a framework for the mathematical study of images? C. Sinan Güntürk Princeton University, Program in Applied and Computational Mathematics Abstract Relations between

More information

EXTENDING THE SCOPE OF WAVELET REGRESSION METHODS BY COEFFICIENT-DEPENDENT THRESHOLDING

EXTENDING THE SCOPE OF WAVELET REGRESSION METHODS BY COEFFICIENT-DEPENDENT THRESHOLDING EXTENDING THE SCOPE OF WAVELET REGRESSION METHODS BY COEFFICIENT-DEPENDENT THRESHOLDING ARNE KOVAC, BERNARD W. SILVERMAN ABSTRACT. Various aspects of the wavelet approach to nonparametric regression are

More information

Multivariate Bayes Wavelet Shrinkage and Applications

Multivariate Bayes Wavelet Shrinkage and Applications Journal of Applied Statistics Vol. 32, No. 5, 529 542, July 2005 Multivariate Bayes Wavelet Shrinkage and Applications GABRIEL HUERTA Department of Mathematics and Statistics, University of New Mexico

More information

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean

More information

FAST WAVELET TECHNIQUES FOR NEAR-OPTIMAL IMAGE PROCESSING

FAST WAVELET TECHNIQUES FOR NEAR-OPTIMAL IMAGE PROCESSING FAST WAVELET TECHNIQUES FOR NEAR-OPTIMAL IMAGE PROCESSING RONALD A. DeVORE BRADLEY J. LUCIER Department of Mathematics Department of Mathematics University of South Carolina Purdue University Columbia,

More information

Block thresholding for density estimation: local and global adaptivity

Block thresholding for density estimation: local and global adaptivity Journal of Multivariate Analysis 95 005) 76 106 www.elsevier.com/locate/jmva Bloc thresholding for density estimation: local and global adaptivity Eric Chicen a,,t.tonycai b,1 a Department of Statistics,

More information

Lecture Notes 5: Multiresolution Analysis

Lecture Notes 5: Multiresolution Analysis Optimization-based data analysis Fall 2017 Lecture Notes 5: Multiresolution Analysis 1 Frames A frame is a generalization of an orthonormal basis. The inner products between the vectors in a frame and

More information

Ole Christensen 3. October 20, Abstract. We point out some connections between the existing theories for

Ole Christensen 3. October 20, Abstract. We point out some connections between the existing theories for Frames and pseudo-inverses. Ole Christensen 3 October 20, 1994 Abstract We point out some connections between the existing theories for frames and pseudo-inverses. In particular, using the pseudo-inverse

More information

Which wavelet bases are the best for image denoising?

Which wavelet bases are the best for image denoising? Which wavelet bases are the best for image denoising? Florian Luisier a, Thierry Blu a, Brigitte Forster b and Michael Unser a a Biomedical Imaging Group (BIG), Ecole Polytechnique Fédérale de Lausanne

More information

Denoising Gabor Transforms

Denoising Gabor Transforms 1 Denoising Gabor Transforms James S. Walker Abstract We describe denoising one-dimensional signals by thresholding Blackman windowed Gabor transforms. This method is compared with Gauss-windowed Gabor

More information

Optimal Estimation of a Nonsmooth Functional

Optimal Estimation of a Nonsmooth Functional Optimal Estimation of a Nonsmooth Functional T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania http://stat.wharton.upenn.edu/ tcai Joint work with Mark Low 1 Question Suppose

More information

How smooth is the smoothest function in a given refinable space? Albert Cohen, Ingrid Daubechies, Amos Ron

How smooth is the smoothest function in a given refinable space? Albert Cohen, Ingrid Daubechies, Amos Ron How smooth is the smoothest function in a given refinable space? Albert Cohen, Ingrid Daubechies, Amos Ron A closed subspace V of L 2 := L 2 (IR d ) is called PSI (principal shift-invariant) if it is the

More information

WAVELETS and other multiscale analysis tools underlie

WAVELETS and other multiscale analysis tools underlie 1322 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 10, NO. 9, SEPTEMBER 2001 Wavelet-Based Image Estimation: An Empirical Bayes Approach Using Jeffreys Noninformative Prior Mário A. T. Figueiredo, Senior

More information

Moment Computation in Shift Invariant Spaces. Abstract. An algorithm is given for the computation of moments of f 2 S, where S is either

Moment Computation in Shift Invariant Spaces. Abstract. An algorithm is given for the computation of moments of f 2 S, where S is either Moment Computation in Shift Invariant Spaces David A. Eubanks Patrick J.Van Fleet y Jianzhong Wang ẓ Abstract An algorithm is given for the computation of moments of f 2 S, where S is either a principal

More information

Wavelet analysis and its statistical applications

Wavelet analysis and its statistical applications The Statistician (2000) 49, Part 1, pp. 1±29 Wavelet analysis and its statistical applications Felix Abramovich, Tel Aviv University, Israel Trevor C. Bailey University of Exeter, UK and Theofanis Sapatinas

More information

Bayesian Nonparametric Point Estimation Under a Conjugate Prior

Bayesian Nonparametric Point Estimation Under a Conjugate Prior University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 5-15-2002 Bayesian Nonparametric Point Estimation Under a Conjugate Prior Xuefeng Li University of Pennsylvania Linda

More information

Plugin Confidence Intervals in Discrete Distributions

Plugin Confidence Intervals in Discrete Distributions Plugin Confidence Intervals in Discrete Distributions T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania Philadelphia, PA 19104 Abstract The standard Wald interval is widely

More information

OPTIMAL POINTWISE ADAPTIVE METHODS IN NONPARAMETRIC ESTIMATION 1

OPTIMAL POINTWISE ADAPTIVE METHODS IN NONPARAMETRIC ESTIMATION 1 The Annals of Statistics 1997, Vol. 25, No. 6, 2512 2546 OPTIMAL POINTWISE ADAPTIVE METHODS IN NONPARAMETRIC ESTIMATION 1 By O. V. Lepski and V. G. Spokoiny Humboldt University and Weierstrass Institute

More information

2 Nason where and X f(x) = c k 0k (x)+ k2z c k = d jk = Z Z X j<j k2z R f(x) 0k(x) dx R f(x) jk (x) dx d jk jk (x) (1) where J controls the maximum re

2 Nason where and X f(x) = c k 0k (x)+ k2z c k = d jk = Z Z X j<j k2z R f(x) 0k(x) dx R f(x) jk (x) dx d jk jk (x) (1) where J controls the maximum re Choice of the threshold parameter in wavelet function estimation G.P. Nason 1 Department of Mathematics, University of Bristol, Bristol BS8 1TW, UK Abstract The procedures of Donoho, Johnstone, Kerkyacharian

More information

Nonparametric estimation using wavelet methods. Dominique Picard. Laboratoire Probabilités et Modèles Aléatoires Université Paris VII

Nonparametric estimation using wavelet methods. Dominique Picard. Laboratoire Probabilités et Modèles Aléatoires Université Paris VII Nonparametric estimation using wavelet methods Dominique Picard Laboratoire Probabilités et Modèles Aléatoires Université Paris VII http ://www.proba.jussieu.fr/mathdoc/preprints/index.html 1 Nonparametric

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published

More information

MULTI-SCALE IMAGE DENOISING BASED ON GOODNESS OF FIT (GOF) TESTS

MULTI-SCALE IMAGE DENOISING BASED ON GOODNESS OF FIT (GOF) TESTS MULTI-SCALE IMAGE DENOISING BASED ON GOODNESS OF FIT (GOF) TESTS Naveed ur Rehman 1, Khuram Naveed 1, Shoaib Ehsan 2, Klaus McDonald-Maier 2 1 Department of Electrical Engineering, COMSATS Institute of

More information

A Stable Finite Dierence Ansatz for Higher Order Dierentiation of Non-Exact. Data. Bob Anderssen and Frank de Hoog,

A Stable Finite Dierence Ansatz for Higher Order Dierentiation of Non-Exact. Data. Bob Anderssen and Frank de Hoog, A Stable Finite Dierence Ansatz for Higher Order Dierentiation of Non-Exact Data Bob Anderssen and Frank de Hoog, CSIRO Division of Mathematics and Statistics, GPO Box 1965, Canberra, ACT 2601, Australia

More information

저작권법에따른이용자의권리는위의내용에의하여영향을받지않습니다.

저작권법에따른이용자의권리는위의내용에의하여영향을받지않습니다. 저작자표시 - 비영리 - 변경금지 2.0 대한민국 이용자는아래의조건을따르는경우에한하여자유롭게 이저작물을복제, 배포, 전송, 전시, 공연및방송할수있습니다. 다음과같은조건을따라야합니다 : 저작자표시. 귀하는원저작자를표시하여야합니다. 비영리. 귀하는이저작물을영리목적으로이용할수없습니다. 변경금지. 귀하는이저작물을개작, 변형또는가공할수없습니다. 귀하는, 이저작물의재이용이나배포의경우,

More information

Functions: A Fourier Approach. Universitat Rostock. Germany. Dedicated to Prof. L. Berg on the occasion of his 65th birthday.

Functions: A Fourier Approach. Universitat Rostock. Germany. Dedicated to Prof. L. Berg on the occasion of his 65th birthday. Approximation Properties of Multi{Scaling Functions: A Fourier Approach Gerlind Plona Fachbereich Mathemati Universitat Rostoc 1851 Rostoc Germany Dedicated to Prof. L. Berg on the occasion of his 65th

More information

SYSTEM RECONSTRUCTION FROM SELECTED HOS REGIONS. Haralambos Pozidis and Athina P. Petropulu. Drexel University, Philadelphia, PA 19104

SYSTEM RECONSTRUCTION FROM SELECTED HOS REGIONS. Haralambos Pozidis and Athina P. Petropulu. Drexel University, Philadelphia, PA 19104 SYSTEM RECOSTRUCTIO FROM SELECTED HOS REGIOS Haralambos Pozidis and Athina P. Petropulu Electrical and Computer Engineering Department Drexel University, Philadelphia, PA 94 Tel. (25) 895-2358 Fax. (25)

More information

ROYAL INSTITUTE OF TECHNOLOGY KUNGL TEKNISKA HÖGSKOLAN. Department of Signals, Sensors & Systems

ROYAL INSTITUTE OF TECHNOLOGY KUNGL TEKNISKA HÖGSKOLAN. Department of Signals, Sensors & Systems The Evil of Supereciency P. Stoica B. Ottersten To appear as a Fast Communication in Signal Processing IR-S3-SB-9633 ROYAL INSTITUTE OF TECHNOLOGY Department of Signals, Sensors & Systems Signal Processing

More information

ADAPTIVE CONFIDENCE BALLS 1. BY T. TONY CAI AND MARK G. LOW University of Pennsylvania

ADAPTIVE CONFIDENCE BALLS 1. BY T. TONY CAI AND MARK G. LOW University of Pennsylvania The Annals of Statistics 2006, Vol. 34, No. 1, 202 228 DOI: 10.1214/009053606000000146 Institute of Mathematical Statistics, 2006 ADAPTIVE CONFIDENCE BALLS 1 BY T. TONY CAI AND MARK G. LOW University of

More information

A New Method for Varying Adaptive Bandwidth Selection

A New Method for Varying Adaptive Bandwidth Selection IEEE TRASACTIOS O SIGAL PROCESSIG, VOL. 47, O. 9, SEPTEMBER 1999 2567 TABLE I SQUARE ROOT MEA SQUARED ERRORS (SRMSE) OF ESTIMATIO USIG THE LPA AD VARIOUS WAVELET METHODS A ew Method for Varying Adaptive

More information

Wavelet-Based Statistical Signal Processing. Using Hidden Markov Models. Rice University.

Wavelet-Based Statistical Signal Processing. Using Hidden Markov Models. Rice University. Wavelet-Based Statistical Signal Processing Using Hidden Markov Models Matthew S. Crouse, Robert D. Nowak, 2 and Richard G. Baraniuk Department of Electrical and Computer Engineering Rice University 600

More information

Research Statement. Harrison H. Zhou Cornell University

Research Statement. Harrison H. Zhou Cornell University Research Statement Harrison H. Zhou Cornell University My research interests and contributions are in the areas of model selection, asymptotic decision theory, nonparametric function estimation and machine

More information

On least squares estimators under Bernoulli-Laplacian mixture priors

On least squares estimators under Bernoulli-Laplacian mixture priors On least squares estimators under Bernoulli-Laplacian mixture priors Aleksandra Pižurica and Wilfried Philips Ghent University Image Processing and Interpretation Group Sint-Pietersnieuwstraat 41, B9000

More information

Signal Denoising with Wavelets

Signal Denoising with Wavelets Signal Denoising with Wavelets Selin Aviyente Department of Electrical and Computer Engineering Michigan State University March 30, 2010 Introduction Assume an additive noise model: x[n] = f [n] + w[n]

More information

Optimal global rates of convergence for interpolation problems with random design

Optimal global rates of convergence for interpolation problems with random design Optimal global rates of convergence for interpolation problems with random design Michael Kohler 1 and Adam Krzyżak 2, 1 Fachbereich Mathematik, Technische Universität Darmstadt, Schlossgartenstr. 7, 64289

More information

OPTIMAL SURE PARAMETERS FOR SIGMOIDAL WAVELET SHRINKAGE

OPTIMAL SURE PARAMETERS FOR SIGMOIDAL WAVELET SHRINKAGE 17th European Signal Processing Conference (EUSIPCO 009) Glasgow, Scotland, August 4-8, 009 OPTIMAL SURE PARAMETERS FOR SIGMOIDAL WAVELET SHRINKAGE Abdourrahmane M. Atto 1, Dominique Pastor, Gregoire Mercier

More information

J. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis

J. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis A Combined Adaptive-Mixtures/Plug-In Estimator of Multivariate Probability Densities 1 J. Cwik and J. Koronacki Institute of Computer Science, Polish Academy of Sciences Ordona 21, 01-237 Warsaw, Poland

More information

Some new method of approximation and estimation on sphere

Some new method of approximation and estimation on sphere Some new method of approximation and estimation on sphere N. Jarzebkowska B. Cmiel K. Dziedziul September 17-23, 2017, in Bedlewo (Poland) [BD] Bownik, M.; D.K. Smooth orthogonal projections on sphere.

More information

Optimal testing in functional analysis of variance. models

Optimal testing in functional analysis of variance. models Optimal testing in functional analysis of variance models Felix Abramovich, Department of Statistics and Operations Research, Tel Aviv University, Tel Aviv 69978, Israel. Anestis Antoniadis, Laboratoire

More information

Wavelet Thresholding for Non Necessarily Gaussian Noise: Functionality

Wavelet Thresholding for Non Necessarily Gaussian Noise: Functionality Wavelet Thresholding for Non Necessarily Gaussian Noise: Functionality By R. Averkamp 1 and C. Houdré 2 Freiburg University and Université Paris XII and Georgia Institute of Technology For signals belonging

More information

2 FRED J. HICKERNELL the sample mean of the y (i) : (2) ^ 1 N The mean square error of this estimate may be written as a sum of two parts, a bias term

2 FRED J. HICKERNELL the sample mean of the y (i) : (2) ^ 1 N The mean square error of this estimate may be written as a sum of two parts, a bias term GOODNESS OF FIT STATISTICS, DISCREPANCIES AND ROBUST DESIGNS FRED J. HICKERNELL Abstract. The Cramer{Von Mises goodness-of-t statistic, also known as the L 2 -star discrepancy, is the optimality criterion

More information

Wavelet Estimators in Nonparametric Regression: A Comparative Simulation Study

Wavelet Estimators in Nonparametric Regression: A Comparative Simulation Study Wavelet Estimators in Nonparametric Regression: A Comparative Simulation Study Anestis Antoniadis, Jeremie Bigot Laboratoire IMAG-LMC, University Joseph Fourier, BP 53, 384 Grenoble Cedex 9, France. and

More information

The choice of weights in kernel regression estimation

The choice of weights in kernel regression estimation Biometrika (1990), 77, 2, pp. 377-81 Printed in Great Britain The choice of weights in kernel regression estimation BY THEO GASSER Zentralinstitut fur Seelische Gesundheit, J5, P.O.B. 5970, 6800 Mannheim

More information

Support Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee

More information

Ideal Spatial Adaptation by Wavelet Shrinkage. David L. Donoho. June Abstract

Ideal Spatial Adaptation by Wavelet Shrinkage. David L. Donoho. June Abstract Ideal Spatial Adaptation by Wavelet Shrinkage David L. Donoho Iain M. Johnstone Department of Statistics, Stanford University, Stanford, CA, 94305-4065, U.S.A. June 1992 Revised April 1993 Abstract With

More information

Widths. Center Fluctuations. Centers. Centers. Widths

Widths. Center Fluctuations. Centers. Centers. Widths Radial Basis Functions: a Bayesian treatment David Barber Bernhard Schottky Neural Computing Research Group Department of Applied Mathematics and Computer Science Aston University, Birmingham B4 7ET, U.K.

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Computation Of Asymptotic Distribution. For Semiparametric GMM Estimators. Hidehiko Ichimura. Graduate School of Public Policy

Computation Of Asymptotic Distribution. For Semiparametric GMM Estimators. Hidehiko Ichimura. Graduate School of Public Policy Computation Of Asymptotic Distribution For Semiparametric GMM Estimators Hidehiko Ichimura Graduate School of Public Policy and Graduate School of Economics University of Tokyo A Conference in honor of

More information

ECE 901 Lecture 16: Wavelet Approximation Theory

ECE 901 Lecture 16: Wavelet Approximation Theory ECE 91 Lecture 16: Wavelet Approximation Theory R. Nowak 5/17/29 1 Introduction In Lecture 4 and 15, we investigated the problem of denoising a smooth signal in additive white noise. In Lecture 4, we considered

More information

CLASSIFYING THE SMOOTHNESS OF IMAGES: THEORY AND APPLICATIONS TO WAVELET IMAGE PROCESSING

CLASSIFYING THE SMOOTHNESS OF IMAGES: THEORY AND APPLICATIONS TO WAVELET IMAGE PROCESSING CLASSIFYING THE SMOOTHNESS OF IMAGES: THEORY AND APPLICATIONS TO WAVELET IMAGE PROCESSING RONALD A. DeVORE 1, University of South Carolina, and BRADLEY J. LUCIER 2, Purdue University Abstract Devore, Jawerth,

More information

MINIMAX ESTIMATION WITH THRESHOLDING AND ITS APPLICATION TO WAVELET ANALYSIS. Harrison H. Zhou* and J. T. Gene Hwang** Cornell University

MINIMAX ESTIMATION WITH THRESHOLDING AND ITS APPLICATION TO WAVELET ANALYSIS. Harrison H. Zhou* and J. T. Gene Hwang** Cornell University MINIMAX ESTIMATION WITH THRESHOLDING AND ITS APPLICATION TO WAVELET ANALYSIS Harrison H. Zhou* and J. T. Gene Hwang** Cornell University March 3, Abstract. Many statistical practices involve choosing between

More information

Learning with Ensembles: How. over-tting can be useful. Anders Krogh Copenhagen, Denmark. Abstract

Learning with Ensembles: How. over-tting can be useful. Anders Krogh Copenhagen, Denmark. Abstract Published in: Advances in Neural Information Processing Systems 8, D S Touretzky, M C Mozer, and M E Hasselmo (eds.), MIT Press, Cambridge, MA, pages 190-196, 1996. Learning with Ensembles: How over-tting

More information

Two Denoising Methods by Wavelet Transform

Two Denoising Methods by Wavelet Transform IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 47, NO. 12, DECEMBER 1999 3401 Two Denoising Methods by Wavelet Transform Quan Pan, Lei Zhang, Guanzhong Dai, and Hongcai Zhang Abstract Two wavelet-based noise

More information

Symmetric Wavelet Tight Frames with Two Generators

Symmetric Wavelet Tight Frames with Two Generators Symmetric Wavelet Tight Frames with Two Generators Ivan W. Selesnick Electrical and Computer Engineering Polytechnic University 6 Metrotech Center, Brooklyn, NY 11201, USA tel: 718 260-3416, fax: 718 260-3906

More information

Biorthogonal Spline Type Wavelets

Biorthogonal Spline Type Wavelets PERGAMON Computers and Mathematics with Applications 0 (00 1 0 www.elsevier.com/locate/camwa Biorthogonal Spline Type Wavelets Tian-Xiao He Department of Mathematics and Computer Science Illinois Wesleyan

More information

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Bodduluri Asha, B. Leela kumari Abstract: It is well known that in a real world signals do not exist without noise, which may be negligible

More information

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f Numer. Math. 67: 289{301 (1994) Numerische Mathematik c Springer-Verlag 1994 Electronic Edition Least supported bases and local linear independence J.M. Carnicer, J.M. Pe~na? Departamento de Matematica

More information

Nonparametric Estimation of Functional-Coefficient Autoregressive Models

Nonparametric Estimation of Functional-Coefficient Autoregressive Models Nonparametric Estimation of Functional-Coefficient Autoregressive Models PEDRO A. MORETTIN and CHANG CHIANN Department of Statistics, University of São Paulo Introduction Nonlinear Models: - Exponential

More information

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter:

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter: Toutenburg, Shalabh: Estimation of Regression Coefficients Subject to Exact Linear Restrictions when some Observations are Missing and Balanced Loss Function is Used Sonderforschungsbereich 386, Paper

More information

A NOTE ON MATRIX REFINEMENT EQUATIONS. Abstract. Renement equations involving matrix masks are receiving a lot of attention these days.

A NOTE ON MATRIX REFINEMENT EQUATIONS. Abstract. Renement equations involving matrix masks are receiving a lot of attention these days. A NOTE ON MATRI REFINEMENT EQUATIONS THOMAS A. HOGAN y Abstract. Renement equations involving matrix masks are receiving a lot of attention these days. They can play a central role in the study of renable

More information

Nets Hawk Katz Theorem. There existsaconstant C>so that for any number >, whenever E [ ] [ ] is a set which does not contain the vertices of any axis

Nets Hawk Katz Theorem. There existsaconstant C>so that for any number >, whenever E [ ] [ ] is a set which does not contain the vertices of any axis New York Journal of Mathematics New York J. Math. 5 999) {3. On the Self Crossing Six Sided Figure Problem Nets Hawk Katz Abstract. It was shown by Carbery, Christ, and Wright that any measurable set E

More information

WAVELET EXPANSIONS OF DISTRIBUTIONS

WAVELET EXPANSIONS OF DISTRIBUTIONS WAVELET EXPANSIONS OF DISTRIBUTIONS JASSON VINDAS Abstract. These are lecture notes of a talk at the School of Mathematics of the National University of Costa Rica. The aim is to present a wavelet expansion

More information

Quadrature Prefilters for the Discrete Wavelet Transform. Bruce R. Johnson. James L. Kinsey. Abstract

Quadrature Prefilters for the Discrete Wavelet Transform. Bruce R. Johnson. James L. Kinsey. Abstract Quadrature Prefilters for the Discrete Wavelet Transform Bruce R. Johnson James L. Kinsey Abstract Discrepancies between the Discrete Wavelet Transform and the coefficients of the Wavelet Series are known

More information

MLISP: Machine Learning in Signal Processing Spring Lecture 10 May 11

MLISP: Machine Learning in Signal Processing Spring Lecture 10 May 11 MLISP: Machine Learning in Signal Processing Spring 2018 Lecture 10 May 11 Prof. Venia Morgenshtern Scribe: Mohamed Elshawi Illustrations: The elements of statistical learning, Hastie, Tibshirani, Friedman

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

Constrained Leja points and the numerical solution of the constrained energy problem

Constrained Leja points and the numerical solution of the constrained energy problem Journal of Computational and Applied Mathematics 131 (2001) 427 444 www.elsevier.nl/locate/cam Constrained Leja points and the numerical solution of the constrained energy problem Dan I. Coroian, Peter

More information

Construction of Orthogonal Wavelets Using. Fractal Interpolation Functions. Abstract

Construction of Orthogonal Wavelets Using. Fractal Interpolation Functions. Abstract Construction of Orthogonal Wavelets Using Fractal Interpolation Functions George Donovan, Jerey S. Geronimo Douglas P. Hardin and Peter R. Massopust 3 Abstract Fractal interpolation functions are used

More information

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses Statistica Sinica 5(1995), 459-473 OPTIMAL DESIGNS FOR POLYNOMIAL REGRESSION WHEN THE DEGREE IS NOT KNOWN Holger Dette and William J Studden Technische Universitat Dresden and Purdue University Abstract:

More information

[km]

[km] SEGMENTATION INDEPENDENT ESTIMATES OF TURBULENCE PARAMETERS G. C. Papanicolaou a, K. Solna b and D. Washburn c adepartment of Mathematics, Stanford University, Stanford, CA 9435 bdepartment of Mathematics,

More information

Shrinkage and Denoising by Minimum Message Length

Shrinkage and Denoising by Minimum Message Length 1 Shrinkage and Denoising by Minimum Message Length Daniel F. Schmidt and Enes Makalic Abstract This paper examines orthonormal regression and wavelet denoising within the Minimum Message Length (MML framework.

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

1 Introduction 1 INTRODUCTION 1

1 Introduction 1 INTRODUCTION 1 1 INTRODUCTION 1 Audio Denoising by Time-Frequency Block Thresholding Guoshen Yu, Stéphane Mallat and Emmanuel Bacry CMAP, Ecole Polytechnique, 91128 Palaiseau, France March 27 Abstract For audio denoising,

More information

Penalty Methods for Bivariate Smoothing and Chicago Land Values

Penalty Methods for Bivariate Smoothing and Chicago Land Values Penalty Methods for Bivariate Smoothing and Chicago Land Values Roger Koenker University of Illinois, Urbana-Champaign Ivan Mizera University of Alberta, Edmonton Northwestern University: October 2001

More information

ARTIFICIAL INTELLIGENCE LABORATORY. and CENTER FOR BIOLOGICAL INFORMATION PROCESSING. A.I. Memo No August Federico Girosi.

ARTIFICIAL INTELLIGENCE LABORATORY. and CENTER FOR BIOLOGICAL INFORMATION PROCESSING. A.I. Memo No August Federico Girosi. MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL INFORMATION PROCESSING WHITAKER COLLEGE A.I. Memo No. 1287 August 1991 C.B.I.P. Paper No. 66 Models of

More information

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing Kernel PCA Pattern Reconstruction via Approximate Pre-Images Bernhard Scholkopf, Sebastian Mika, Alex Smola, Gunnar Ratsch, & Klaus-Robert Muller GMD FIRST, Rudower Chaussee 5, 12489 Berlin, Germany fbs,

More information

THE LUBRICATION APPROXIMATION FOR THIN VISCOUS FILMS: REGULARITY AND LONG TIME BEHAVIOR OF WEAK SOLUTIONS A.L. BERTOZZI AND M. PUGH.

THE LUBRICATION APPROXIMATION FOR THIN VISCOUS FILMS: REGULARITY AND LONG TIME BEHAVIOR OF WEAK SOLUTIONS A.L. BERTOZZI AND M. PUGH. THE LUBRICATION APPROXIMATION FOR THIN VISCOUS FILMS: REGULARITY AND LONG TIME BEHAVIOR OF WEAK SOLUTIONS A.L. BERTOI AND M. PUGH April 1994 Abstract. We consider the fourth order degenerate diusion equation

More information

One of the basic techniques in the vast wavelet literature is the wavelet transform. Advantages provided by such a transform have been demonstrated in

One of the basic techniques in the vast wavelet literature is the wavelet transform. Advantages provided by such a transform have been demonstrated in K-Stationarity and Wavelets Bing Cheng and Howell Tong Institute of Mathematics and Statistics University of Kent at Canterbury Kent CT2 7NF, UK April 29, 1996 1 Introduction Wavelets oer exciting possibilities

More information

RECONSTRUCTION OF WAVELET COEFFICIENTS USING TOTAL VARIATION MINIMIZATION

RECONSTRUCTION OF WAVELET COEFFICIENTS USING TOTAL VARIATION MINIMIZATION RECONSTRUCTION OF WAVELET COEFFICIENTS USING TOTAL VARIATION MINIMIATION SYLVAIN DURAND AND JACQUES FROMENT y Abstract. We propose a model to reconstruct wavelet coecients using a total variation minimization

More information

Sparse linear models

Sparse linear models Sparse linear models Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 2/22/2016 Introduction Linear transforms Frequency representation Short-time

More information