An Empirical Bayesian interpretation and generalization of NL-means
|
|
- Gyles Ferguson
- 5 years ago
- Views:
Transcription
1 Computer Science Tecnical Report TR , October 2010 Courant Institute of Matematical Sciences, New York University ttp://cs.nyu.edu/web/researc/tecreports/reports.tml An Empirical Bayesian interpretation and generalization of NL-means Martin Rapan Courant Inst. of Matematical Sciences New York University Eero P. Simoncelli Howard Huges Medical Institute, Center for Neural Science, and Courant Inst. of Matematical Sciences New York University Abstract A number of recent algoritms in signal and image processing are based on te empirical distribution of localized patces. Here, we develop a nonparametric empirical Bayesian estimator for recovering an image corrupted by additive Gaussian noise, based on fitting te density over image patces wit a local exponential model. Te resulting solution is in te form of an adaptively weigted average of te observed patc wit te mean of a set of similar patces, and tus bot justifies and generalizes te recently proposed nonlocalmeans (NL-means) metod for image denoising. Unlike NL-means, our estimator includes a dependency on te size of te patc similarity neigborood, and we sow tat tis neigborood size can be cosen in suc a way tat te estimator converges to te optimal Bayes least squares estimator as te amount of data grows. We demonstrate te increase in performance of our metod compared to NL-means on a set of simulated examples. 1 Introduction Local patces witin a potograpic image are often igly structured. Moreover, it is quite common to find many patces witin te same image tat ave similar or even identical content. Tis fact as been exploited in developing new forms of signal representation [10], and in applications suc as denoising [5], clustering [3], inpainting [4], and texture syntesis [15, 7, 8]. Most of tese metods are based on assumptions (sometimes implicit) about te probability distribution over te intensity vectors associated wit te patces. A recent example is a patced-based inference metodology, known as nonlocal means (NL-means), in wic eac pixel in a noisy image is replaced by a weigted average over oter pixels in te image tat occur in similar image patces [1, 2]. Te weigting is determined by te similarity of te patces, as measured wit a Gaussian kernel. Tis metod is intuitively appealing, and gives good (altoug less tan state-of-te-art) denoising performance on natural images. More recent variants ave been developed tat acieve state-ofte-art [5]. But as we sow ere, te NL-means as te drawback tat te estimate depends crucially on te Gaussian kernel widt, and even in te limit of infinite data, tere is no way to coose tis widt so as to guarantee convergence to te optimal answer. In tis paper, we develop an alternative patc-based denoising algoritm as an empirical Bayes () estimator, and sow tat it generalizes and improves on te NL-means solution. We use an exponential density model for local patces of a noisy image, and derive a maximum likeliood metod for estimating te parameters from data. We ten combine tis wit a nonparametric form of te optimal least-squares estimator for additive Gaussian noise [13]. Tis simple, but often overlooked form expresses te usual Bayes least squares estimator directly in terms of te density of noisy observations, witout reference to te (clean) prior density. Te resulting denoising metod resembles NL-means, in tat te denoised value is a function of te
2 weigted average over oter pixels in te same image tat come from similar patces. But te function is nonlinear, and is scaled according to te size of te neigborood over wic intensities are averaged. As a result, we prove tat te estimator converges to te Bayes least squares optimum as te amount of data increases, a feature not present in te standard NL-means approac. 2 NL-means estimation Suppose Y is a noise-corrupted measurement of X, were eiter or bot variables can be finite-dimensional vectors. Given tis measurement, te estimate of X tat minimizes te expected squared error is te Bayes Least Squared (BLS) estimator, wic is te expected value of X conditioned on Y, E{X Y }. In particular, for an observation y 0, te estimate is ˆx BLS (y 0 )=E{X Y = y 0 }. In te NL-means approac, a vector observation Y = y 0 is denoised by averaging it togeter wit patces tat are similar, i.e., tat are in some neigborood, N, of size about te vector y 0. As te number of data points available increases to infinity, te NL-means estimator will converge to ˆx NL (y 0 ) E{Y Y N }. (1) Notice tat tis is a conditional expectation of Y, not X (as required for te BLS estimate). Te original presentation of NL-means [1] justifies tis by considering te case of estimating te center pixel, X c of te patc based on te surrounding pixels. In tis case, and assuming zero-mean noise, we can write E{Y c Y c = y c } = E{E{Y c X c,y c = y c } Y c = y c } = E{X c Y c = y c }, (2) were Y c indicates te patc witout te center pixel. Tus, by setting N to be a neigborood wic does not depend on te central coefficient of te vector, Eq. (1) will tend to te optimal estimator of te central coefficient as a function of te surrounding coefficients, Eq. (2). In practice, NL-means is implemented to denoise te central coefficient in te patc, but includes te central coefficient in determining similarity of patces. Under tese conditions, te solution does not, in general, converge to te correct BLS estimator, regardless of te size of te neigborood. To see tis, consider te simplified situation were te neigborood is defined as N = {y : y y 0 }. (3) Clearly, ˆx NL (y 0 ) will depend on te value of. For a fixed value of, we end up wit a conditional expectation of Y instead of X, as mentioned above. And, in te limit as 0, te estimate will converge to te identity, i.e. ˆx NL (y 0 ) y 0. Intuitively, tis is because te mean of data witin a neigborood must also lie witin tat neigborood. Tus, te best we can ope for is to find some binwidt,, tat works reasonably well asymptotically, but tis coice will be igly dependent on te (unknown) signal prior. Tis is illustrated for a simple scalar example in Fig. 1. Te prior ere is made up of delta functions wic are well separated, relative to te standard deviation of te noise. Wit te rigt coice of binwidt (about twice te widt of te noise distribution), for any amount of data, NL-means provides a good approximation to te optimal denoiser. But note tat te NL-means beavior depends critically on te coice of binwidt: Fig. 1 also sows a result for a smaller binwidt, wic is clearly suboptimal. In general, tere is no simple way to coose an optimal binwidt, and to do so in a way tat makes te NL-means estimator converge to te BLS estimator. 3 Patc-based empirical Bayes least squares estimator We wis to develop a patc-based estimator tat makes explicit te dependence on neigborood size,, and for wic we can demonstrate proper convergence in te limit as 0. For te particular case of additive Gaussian noise, Miyasawa [13] developed an nonparametric empirical form of te Bayes least squares estimate tat is written directly in terms of te measurement (noisy) density P Y (Y ), wit no explicit reference to te prior: E{X Y = y 0 } = y 0 +Λ y ln(p Y (y 0 )), (4) were Λ is te covariance matrix of te noise (assumed known). Note tat tis expression is not an approximation: te equality is exact, assuming te noise is additive and Gaussian (wit zero mean and covariance Λ), and assuming te measurement density is known.
3 P Y (y) y E{X Y=y} 0 y E{Y Y y } 0 y Fig. 1: Observed density and optimal estimator for prior made up of tree isolated delta functions. Top panel: observed density (arrows represent prior). Middle panel: BLS estimator (assuming true prior is known). Bottom: NL-means estimates calculated for two different binsizes (tick solid and dased lines). Note also tat te NL-means estimator approaces te identity function as te binsize goes to zero. 3.1 Learning te estimator function from data Te formulation of Eq. (4) offers a means of computing te BLS estimator from noisy samples, if one can construct an approximation of te noisy distribution, P Y (Y ). Simple istograms will not suffice for tis approximation, because Eq.(4) requires us to compute te logaritmic derivative of te estimated density. Instead, we use an exponential model for te local density centered on an observation Y k = y 0 witin some neigborood, N of size, centered on y 0 : P Y Y N (y) = ea(y y0) Z(a,) 1 N (5) were 1 N denotes te indicator function of N, and Z(a,) normalizes te density over te interval: Z(a,)= e a (y y0) dy. (6) N Tis type of local exponential approximation is similar to tat used in [11] for density estimation. Te sape of te neigborood, N, is unspecified, but te linear size sould scale wit. For example, it could be a sperical ball of radius around y 0 or te ypercube centered on y 0 of lengt 2 on eac side. In any case, te logaritmic gradient of tis density evaluated at y 0 is simply a. We can ten fit te parameter a by maximizing te likeliood of te conditional density over te neigborood: â = arg max ln(p Y Y N (Y n )) (7) a {n:y n N } Substituting Eq. (5), taking te gradient of te sum, and setting it equal to zero gives a ln(z(â,)) = Ȳ y 0 (8) were Ȳ is te average of te data tat lie in N. Note tat substituting Eq. (6)inforZ(â, ) and adding y 0 to bot sides gives E{Y Y N } = Ȳ. Tus, te ML solution for â matces te local mean of te fitted density to te local empirical mean. Equation (8) provides an implicit definition of â wic involves te function Z(a,). Noting tat Z scales according to Z(a,)= d Z(a, 1), we can solve explicitly for â: â = 1 F 1 (Ȳ y 0 ) (9)
4 were F (a) = a ln(z(a, 1)), (10) does not depend on. Finally, replacing y ln(p Y (y)) in Eq. (4) byâ gives te empirical Bayes estimator: ˆx (y 0 )=y 0 +Λ 1 F 1 (Ȳ y 0 ). (11) Note tat it may be difficult to calculate F 1 for a general neigborood. We ll return to tis issue in section Convergence wit proper coice of binwidt Te empirical Bayes estimator of Eq. (11) depends an te coice of binwidt,, wic not only appears explicitly in te expression, but also implicitly affects te coice of wic data vectors are included in calculating te local mean, Ȳ. Tis binwidt can vary for eac data point, but in tis paper, we will consider only te simpler case of a constant binwidt for all data. In order for our estimator to converge to te BLS estimator, we need â, asdefined in Eq. (9), to converge to te logaritmic derivative of te true density, and tis means te binwidt must srink to zero as te total amount of data grows. Te rate of srinkage sould be set in suc a way tat te amount of data witin eac neigborood goes to infinity. In tis subsection, we calculate te rate at wic te binwidts sould srink, in order to ensure convergence. First, we note tat te factor of 1 in Eq. (9) implies tat, for â to converge at all, it is necessary tat (Ȳ ) F 1 y0 0. (12) Since it can be sown tat 0 is te only zero of F, we need only sow tat Ȳ y 0 0. (13) Looking again at Eq. (9), and considering a Taylor approximation of F about zero, we see tat for for â to converge, it is furter necessary tat Ȳ y0 ave some finite limit 2 Ȳ y 0 2 v, (14) for some value v. In tis case, a bit of calculation sows tat te Taylor approximation of Eq. (9) may be written â V 1 C1 1 v, were C 1 /V 1 is te Jacobian matrix of F evaluated at zero, wit V 1 denoting te volume of te neigborood N 1 (N, wit size =1), and C 1 = (ỹ y 0 )(ỹ y 0 ) T dỹ. N 1 Finally, combining te Taylor approximation wit te definition of â tells us tat our estimator converges to te optimal BLS estimator if and only if we can find a coice of binwidt so tat 1 2 V 1C1 1 (Ȳ y 0) ln(p Y (y 0 )). (15) We now discuss ow te binwidt sould srink as te amount of data increases in order to ensure tis condition is met. First, we decompose te asymptotic mean squared error into variance and squared bias terms: E { (Ȳ y0 2 C 1 {Ȳ y0 = Var 2 ) 2 } ln(p Y (y 0 )) V 1 } + ( E {Ȳ y0 2 } C ) 2 1 ln(p Y (y 0 )), (16) V 1
5 We expect (and will sow) tat te variance term sould decrease as grows (since more data will fall into eac bin), wereas te squared bias sould increase as te neigborood grows. Tis is illustrated in Fig. 2. Te trick is to find a way to srink te binwidts as te amount of data increases, so tat te bias goes to zero, but slowly enoug to ensure tat te variance still goes to zero as well. For large amounts of data, we expect to be small, and so we may use small approximations for te bias and variance. Since te local average, Ȳ is calculated using only data wic fall in N, we may write te expected value as E{Ȳ } = E{Y Y N N } = ỹp Y (ỹ)dỹ (17) N P Y (ỹ)dỹ To get an idea of ow tis beaves asymptotically, we can take te local Taylor approximation of te density and substitute into Eq. (17). Since te neigboroods are centered on y 0, te integrals of terms wit odd degree will vanis, leaving E{Ȳ } y 0 C P Y (y 0 )+O( d+4 ) P Y (y 0 )V + O( d+2 (18) ) were V denotes te volume of te neigborood and C = (ỹ y 0 )(ỹ y 0 ) T dỹ N For symmetric neigboroods, C will be a diagonal matrix. If te N as te same sape for all just rescaled by te factor, a simple cange of variables sows tat V = d V 1 and C = d+2 C 1 so tat E{Ȳ y 0 2 } C 1 ln(p Y (y 0 )) + O( 2 ) (19) V 1 so tat te bias term in Eq. (16) will srink to zero as long as te binwidt srinks to zero. Now consider te variance term in Eq. (16). We need to coose binwidts tat srink to zero slowly enoug to allow {Ȳ } y0 Var 0. 2 First note tat E{ Y y 0 2 Y N } = O( 2 ) (20) Combining tis wit Eq. (19) gives Var(Ȳ y ) = n 4 Var(Y y 0 Y N ) = O( 1 n 2 )+O( 1 n ) (21) were n is te number of data points wic fall in te bin. Tus, we see tat as long as n 2 (22) te variance of our approximation will also tend to zero. If tere are N total data points, we can approximate te number falling in N as n P Y (y 0 )NV 1 d (23) and, inserting tis into Eq. (22) gives For example, we migt use N d+2, 0 (24) N 1 m+d+2,m>0 (25)
6 variance increasing N bias Fig. 2: Bias-variance tradeoff, as a function of binwidt,. Solid blue line indicates squared bias. Dased lines indicate variance, for different values of N (number of data points). Solid black line (wit black points) sows te variance resulting wen is cosen as a function of N, as indicated in Eq. (25) wit m =2. Under tese conditions, te variance falls to zero as srinks. Te variance resulting from tis coice (for m =2) is illustrated in Fig. 2, and can be seen to approac zero as decreases. Wit tis coice of binwidt, it now follows tat te estimator in Eq. (11) converges asymptotically to te optimal BLS solution. Note tat a consequence of our asymptotic analysis is tat, wen te estimator in Eq. (11) converges, we can use te Taylor series approximation of te estimator, so tat for te same coice of binwidts ˆx (y 0 ) y 0 +ΛV 1 C1 1 (Ȳ y 0 2 ) (26) will also converge asymptotically to te BLS estimator. In te case of sperical neigboroods we get an estimate of te logaritmic derivative very similar to tat derived for te mean-sift metod of clustering[3]. Note, toug, tat te derivation used for tat approac uses te derivative of te Epanecnikov kernel, wic would only seem to work for sperical neigboroods. Our metod, on te oter and, works for general neigboroods. Note also tat mean-sift is meant to be used as an iterative gradient ascent, moving data towards peaks in te density. For our purposes, owever, we ave sown tat wen te logaritmic gradient is scaled by te noise variance, we acieve BLS denoising in one step. 4 Empirical Bayes versus NL means We ave now introduced bot te Empirical Bayes metod for approximating te BLS estimator, Eq. (11), and te NL-means estimator Eq. (1). For pedagogical reasons, we will compare and contrast te two using low dimensional examples wic allow for bot easier visualization and easier computation. As discussed in te last section, we expect our estimator to converge to te ideal solution as te amount of data goes to infinity. NL-means, on te oter and does not ave tis asymptotic property in general, since if te binwidts were to srink te NL-means metod would do no denoising. Terefore, we expect te optimal optimal binwidt to asymptote to some finite, nonzero optimal value tat depends eavily on bot te prior and te amount of noise. For tis reason, tis metod may be able to give us improved performance for low to moderate amounts of data, but will in general be asymptotically biased. To illustrate te convergence beavior as a function of te amount of data, we ave simulated a scalar estimation problem. Signal values are drawn from a generalized Gaussian density, p(x) exp( x β ), wit exponent β =0.5. For eac N, we draw N samples from tis distribution, and add Gaussian wite noise to eac. Te noise was cosen so tat te noisy Signal to Noise Ratio (SNR) (10 log of squared error divided by signal variance) was 7.8 db. We ten apply our estimator, wit binwidt cosen according to Eq. (25). As a reference, we also sow te performance of te ideal BLS estimator (wic knows te true prior). We also sow te result obtained using NL-means (Eq. (1)), wit binwidt cosen to optimize performance (i.e., by comparing te denoising result to te clean signal). For eac N, tis simulation is performed 1000 times, so as to obtain estimates of te average performance, as well as te variability. Te SNR performance is plotted, as a function of N, in Fig. 3. Note tat te average performance of NL-means is reasonable for small N
7 SNR improvement(db) NL-means 1.3 BLS # samples (a) SNR improvement(db) NL-means 1.6 BLS # samples (b) Fig. 3: Performance of Empirical Bayes metod compared wit NL-means, and te optimal BLS estimator, for (a) a generalized Gaussian prior; (b) te sum of two generalized Gaussians. Lines (solid, dased, and dotted) indicate mean performance of eac estimator over 1000 simulations. Gray regions (for and NLmeans) indicate plus/minus one standard deviation. See text. True Observation NL means True Observation NL means True Observation NL means Fig. 4: Beavior of metod compared wit NL-means, for two-dimensional data, wit tree different neigborood coices. Prior is a delta at te origin (indicated by cross). Only a subsample of te data points are sown, for visual clarity. Boxes indicate neigboroods. Note tat in te tird example, te neigborood extends beyond te display. See text. (altoug te variability is quite ig), but saturates as N grows, and tus does not converge to te optimal performance of te BLS estimator. By comparison, our estimator starts out wit a lower SNR at small N, but asymptotically converges (in bot mean and variance) to te optimal solution. Figure 3 sows anoter example, for a signal drawn from a mixture of two generalized Gaussian variables, wit mean values ±5. Here, we see a more substantial asymptotic bias for te NL-means estimator. Figure 4 sows a comparison of our metod to NL-means in two dimensions. Te signal samples in tis example are drawn from a delta function at te origin. We see tat te NL-means solution is igly dependent on te size and sape of te neigborood. It is possible to acieve good results in tis particular case of a delta function prior by using a very large neigborood (3rd panel). But for te smaller neigboroods, te NL-means estimate is restricted to lie witin te neigborood, and is tus eavily biased, wile te metod is not. 5 Discussion We ve derived a patc-based metod for signal denoising by assuming a local exponential density model for te noisy data. We ve proven tat, despite te fact tat we do not include any assumption about te
8 overall form of te signal prior, our estimator converges to te optimal (Bayes) least squares estimator as te amount of data increases, assuming te binsize is reduced appropriately. Te estimator is a nonlinear function of te average of similar patces, and tus may be viewed as a generalization of te NL-means and mean-sift algoritms. Te main drawback of te metod (as wit NL-means) is te ig computational cost associated wit gatering te data patces tat are witin an -neigborood of te patc being denoised. We believe tat some of te metods tat ave been developed for accelerating NL-means [18, 14, 6] (or te oter patc-based applications mentioned in te introduction) may be utilized in our estimator as well, and we are currently working on suc an implementation so as to examine empirical performance of our metod on image denoising. We envision a number of ways in wic our metod may be extended. First, te binsizes need not be uniform for all patces, but can be adapted according to te local sample density. Specifically, for eac patc tat is to be denoised, a rule suc as tat of Eq. (25) could be used to adjust te binsize to as to acieve te best bias/variance tradeoff (i.e., so as to best approximate te BLS estimator). Tis extension necessarily increases te computational cost (since te binsize must now be optimized for eac patc) but may lead to substantial increases in performance. Second, te NL-means metod computes a weigted average of patces, were te weigting is a function of bot te patc similarity, and te patc proximity witin te signal/image (for example, bilateral filtering may be viewed as an NL-means metod, wit one-pixel patces). We are exploring means by wic our metod can be generalized to include bot of tese forms of weigting. Tird, estimators ave been developed for observation processes oter tan additive Gaussian noise [17, 12, 9, 16], and eac of tese could be developed into a patc-based estimator (altoug most of tese would not be written in terms of local means). And finally, we believe our local exponential density estimate will prove useful as a substrate for generalizing oter patc-based empirical metods tat ave become popular in computer vision, grapics, and image processing. References [1] A. Buades, B. Coll, and J. M. Morel. A review of image denoising algoritms, wit a new one. Multiscale Modeling and Simulation, 4(2): , July , 2 [2] A. Buades, B. Coll, and J. M. Morel. Nonlocal image and movie denoising. Intl Journal of Computer Vision, 76(2): , [3] D. Comaniciu and P. Meer. Mean sift: A robust approac toward feature space analysis. IEEE Pat. Anal. Mac. Intell., 24: , , 6 [4] A. Criminisi, P. Perez, and K. Toyama. Region filling and object removal by exemplar-based inpainting. IEEE Trand Image Proc, 13: , [5] K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian. Image denoising by sparse 3D transform-domain collaborative filtering. IEEE Trans. Image Proc., 16(8), August [6] A. Dauwe, B. Goossens, H. Luong, and W. Pilips. A fast non-local image denoising algoritm. In Proc SPIE Electronic Imaging, volume 6812, [7] J. S. De Bonet. Multiresolution sampling procedure for analysis and syntesis of texture images. In Computer Grapics. ACM SIGGRAPH, [8] A. A. Efros and T. K. Leung. Texture syntesis by non-parameteric sampling. In Proc. Int l Conference on Computer Vision, Corfu, [9] Y. C. Eldar. Generalized SURE for exponential families: Applications to regularization. IEEE Trans. on Signal Processing, 57(2): , Feb [10] N. Jojic, B.Frey, and A. Kannan. Epitomic analysis of appearance and sape. In ICCV, [11] C. R. Loader. Local likeliood density estimation. Annals of Statistics, 24(4): , [12] J. S. Maritz and T. Lwin. Empirical Bayes Metods. Capman & Hall, 2nd edition, [13] K. Miyasawa. An empirical Bayes estimator of te mean of a normal population. Bull. Inst. Internat. Statist., 38: , , 2 [14] J. Orcarda, M. Ebraimi, and A. Wong. Efcient nonlocal-means denoising using te SVD. In Proc IEEE Intl Conf on Image Processing, [15] K. Popat and R. W. Picard. Cluster-based probability model and its application to image and texture processing. IEEE Trans Im Proc, 6(2): , [16] M. Rapan and E. P. Simoncelli. Least squares estimation witout priors or supervision. Neural Computation, In Press. 8 [17] H. Robbins. An empirical Bayes approac to statistics. Proc. Tird Berkley Symposium on Matematcal Statistics, 1: , [18] J. Wang, Y. Guo, Y. Ying, Y. Liu, and Q. Peng. Fast non-local algoritm for image denoising. In Proc IEEE Intl Conf on Image Processing,
Regularized Regression
Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize
More informationLecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines
Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to
More informationA = h w (1) Error Analysis Physics 141
Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.
More informationlecture 26: Richardson extrapolation
43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)
More informationChapter 1. Density Estimation
Capter 1 Density Estimation Let X 1, X,..., X n be observations from a density f X x. Te aim is to use only tis data to obtain an estimate ˆf X x of f X x. Properties of f f X x x, Parametric metods f
More informationREVIEW LAB ANSWER KEY
REVIEW LAB ANSWER KEY. Witout using SN, find te derivative of eac of te following (you do not need to simplify your answers): a. f x 3x 3 5x x 6 f x 3 3x 5 x 0 b. g x 4 x x x notice te trick ere! x x g
More informationCombining functions: algebraic methods
Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)
More informationConsider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.
Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions
More informationINTRODUCTION AND MATHEMATICAL CONCEPTS
INTODUCTION ND MTHEMTICL CONCEPTS PEVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips of sine,
More informationExam 1 Review Solutions
Exam Review Solutions Please also review te old quizzes, and be sure tat you understand te omework problems. General notes: () Always give an algebraic reason for your answer (graps are not sufficient),
More informationINTRODUCTION AND MATHEMATICAL CONCEPTS
Capter 1 INTRODUCTION ND MTHEMTICL CONCEPTS PREVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips
More informationThe Verlet Algorithm for Molecular Dynamics Simulations
Cemistry 380.37 Fall 2015 Dr. Jean M. Standard November 9, 2015 Te Verlet Algoritm for Molecular Dynamics Simulations Equations of motion For a many-body system consisting of N particles, Newton's classical
More informationThe Priestley-Chao Estimator
Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are
More informationSECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY
(Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative
More informationOrder of Accuracy. ũ h u Ch p, (1)
Order of Accuracy 1 Terminology We consider a numerical approximation of an exact value u. Te approximation depends on a small parameter, wic can be for instance te grid size or time step in a numerical
More information2.8 The Derivative as a Function
.8 Te Derivative as a Function Typically, we can find te derivative of a function f at many points of its domain: Definition. Suppose tat f is a function wic is differentiable at every point of an open
More informationIEOR 165 Lecture 10 Distribution Estimation
IEOR 165 Lecture 10 Distribution Estimation 1 Motivating Problem Consider a situation were we ave iid data x i from some unknown distribution. One problem of interest is estimating te distribution tat
More informationSin, Cos and All That
Sin, Cos and All Tat James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 9, 2017 Outline Sin, Cos and all tat! A New Power Rule Derivatives
More informationNew Distribution Theory for the Estimation of Structural Break Point in Mean
New Distribution Teory for te Estimation of Structural Break Point in Mean Liang Jiang Singapore Management University Xiaou Wang Te Cinese University of Hong Kong Jun Yu Singapore Management University
More informationLIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION
LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LAURA EVANS.. Introduction Not all differential equations can be explicitly solved for y. Tis can be problematic if we need to know te value of y
More informationProbabilistic Graphical Models Homework 1: Due January 29, 2014 at 4 pm
Probabilistic Grapical Models 10-708 Homework 1: Due January 29, 2014 at 4 pm Directions. Tis omework assignment covers te material presented in Lectures 1-3. You must complete all four problems to obtain
More informationNumerical Differentiation
Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function
More informationHOMEWORK HELP 2 FOR MATH 151
HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,
More informationThe total error in numerical differentiation
AMS 147 Computational Metods and Applications Lecture 08 Copyrigt by Hongyun Wang, UCSC Recap: Loss of accuracy due to numerical cancellation A B 3, 3 ~10 16 In calculating te difference between A and
More information2.11 That s So Derivative
2.11 Tat s So Derivative Introduction to Differential Calculus Just as one defines instantaneous velocity in terms of average velocity, we now define te instantaneous rate of cange of a function at a point
More informationTe comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab
To appear in: Advances in Neural Information Processing Systems 9, eds. M. C. Mozer, M. I. Jordan and T. Petsce. MIT Press, 997 Bayesian Model Comparison by Monte Carlo Caining David Barber D.Barber@aston.ac.uk
More informationHomework 1 Due: Wednesday, September 28, 2016
0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q
More information1 1. Rationalize the denominator and fully simplify the radical expression 3 3. Solution: = 1 = 3 3 = 2
MTH - Spring 04 Exam Review (Solutions) Exam : February 5t 6:00-7:0 Tis exam review contains questions similar to tose you sould expect to see on Exam. Te questions included in tis review, owever, are
More informationDifferential Calculus (The basics) Prepared by Mr. C. Hull
Differential Calculus Te basics) A : Limits In tis work on limits, we will deal only wit functions i.e. tose relationsips in wic an input variable ) defines a unique output variable y). Wen we work wit
More informationMVT and Rolle s Theorem
AP Calculus CHAPTER 4 WORKSHEET APPLICATIONS OF DIFFERENTIATION MVT and Rolle s Teorem Name Seat # Date UNLESS INDICATED, DO NOT USE YOUR CALCULATOR FOR ANY OF THESE QUESTIONS In problems 1 and, state
More informationEFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS
Statistica Sinica 24 2014, 395-414 doi:ttp://dx.doi.org/10.5705/ss.2012.064 EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Jun Sao 1,2 and Seng Wang 3 1 East Cina Normal University,
More informationHow to Find the Derivative of a Function: Calculus 1
Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te
More informationLecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.
Lecture XVII Abstract We introduce te concept of directional derivative of a scalar function and discuss its relation wit te gradient operator. Directional derivative and gradient Te directional derivative
More informationRecall from our discussion of continuity in lecture a function is continuous at a point x = a if and only if
Computational Aspects of its. Keeping te simple simple. Recall by elementary functions we mean :Polynomials (including linear and quadratic equations) Eponentials Logaritms Trig Functions Rational Functions
More informationDifferential equations. Differential equations
Differential equations A differential equation (DE) describes ow a quantity canges (as a function of time, position, ) d - A ball dropped from a building: t gt () dt d S qx - Uniformly loaded beam: wx
More informationKernel Density Based Linear Regression Estimate
Kernel Density Based Linear Regression Estimate Weixin Yao and Zibiao Zao Abstract For linear regression models wit non-normally distributed errors, te least squares estimate (LSE will lose some efficiency
More informationThe Complexity of Computing the MCD-Estimator
Te Complexity of Computing te MCD-Estimator Torsten Bernolt Lerstul Informatik 2 Universität Dortmund, Germany torstenbernolt@uni-dortmundde Paul Fiscer IMM, Danisc Tecnical University Kongens Lyngby,
More informationPhysically Based Modeling: Principles and Practice Implicit Methods for Differential Equations
Pysically Based Modeling: Principles and Practice Implicit Metods for Differential Equations David Baraff Robotics Institute Carnegie Mellon University Please note: Tis document is 997 by David Baraff
More informationTime (hours) Morphine sulfate (mg)
Mat Xa Fall 2002 Review Notes Limits and Definition of Derivative Important Information: 1 According to te most recent information from te Registrar, te Xa final exam will be eld from 9:15 am to 12:15
More informationLecture 21. Numerical differentiation. f ( x+h) f ( x) h h
Lecture Numerical differentiation Introduction We can analytically calculate te derivative of any elementary function, so tere migt seem to be no motivation for calculating derivatives numerically. However
More informationDigital Filter Structures
Digital Filter Structures Te convolution sum description of an LTI discrete-time system can, in principle, be used to implement te system For an IIR finite-dimensional system tis approac is not practical
More information1 Proving the Fundamental Theorem of Statistical Learning
THEORETICAL MACHINE LEARNING COS 5 LECTURE #7 APRIL 5, 6 LECTURER: ELAD HAZAN NAME: FERMI MA ANDDANIEL SUO oving te Fundaental Teore of Statistical Learning In tis section, we prove te following: Teore.
More informationMaterial for Difference Quotient
Material for Difference Quotient Prepared by Stepanie Quintal, graduate student and Marvin Stick, professor Dept. of Matematical Sciences, UMass Lowell Summer 05 Preface Te following difference quotient
More informationHOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS
HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS Po-Ceng Cang National Standard Time & Frequency Lab., TL, Taiwan 1, Lane 551, Min-Tsu Road, Sec. 5, Yang-Mei, Taoyuan, Taiwan 36 Tel: 886 3
More informationDerivation Of The Schwarzschild Radius Without General Relativity
Derivation Of Te Scwarzscild Radius Witout General Relativity In tis paper I present an alternative metod of deriving te Scwarzscild radius of a black ole. Te metod uses tree of te Planck units formulas:
More informationFinancial Econometrics Prof. Massimo Guidolin
CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis
More informationFast optimal bandwidth selection for kernel density estimation
Fast optimal bandwidt selection for kernel density estimation Vikas Candrakant Raykar and Ramani Duraiswami Dept of computer science and UMIACS, University of Maryland, CollegePark {vikas,ramani}@csumdedu
More informationChapter 5 FINITE DIFFERENCE METHOD (FDM)
MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential
More informationMath 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006
Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2
More informationBasic Nonparametric Estimation Spring 2002
Basic Nonparametric Estimation Spring 2002 Te following topics are covered today: Basic Nonparametric Regression. Tere are four books tat you can find reference: Silverman986, Wand and Jones995, Hardle990,
More informationAverage Rate of Change
Te Derivative Tis can be tougt of as an attempt to draw a parallel (pysically and metaporically) between a line and a curve, applying te concept of slope to someting tat isn't actually straigt. Te slope
More informationWYSE Academic Challenge 2004 Sectional Mathematics Solution Set
WYSE Academic Callenge 00 Sectional Matematics Solution Set. Answer: B. Since te equation can be written in te form x + y, we ave a major 5 semi-axis of lengt 5 and minor semi-axis of lengt. Tis means
More informationNotes on wavefunctions II: momentum wavefunctions
Notes on wavefunctions II: momentum wavefunctions and uncertainty Te state of a particle at any time is described by a wavefunction ψ(x). Tese wavefunction must cange wit time, since we know tat particles
More informationIntroduction to Derivatives
Introduction to Derivatives 5-Minute Review: Instantaneous Rates and Tangent Slope Recall te analogy tat we developed earlier First we saw tat te secant slope of te line troug te two points (a, f (a))
More informationCopyright c 2008 Kevin Long
Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula
More informationDifferentiation in higher dimensions
Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends
More informationTHE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5
THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION NEW MODULAR SCHEME introduced from te examinations in 009 MODULE 5 SOLUTIONS FOR SPECIMEN PAPER B THE QUESTIONS ARE CONTAINED IN A SEPARATE FILE
More informationMAT244 - Ordinary Di erential Equations - Summer 2016 Assignment 2 Due: July 20, 2016
MAT244 - Ordinary Di erential Equations - Summer 206 Assignment 2 Due: July 20, 206 Full Name: Student #: Last First Indicate wic Tutorial Section you attend by filling in te appropriate circle: Tut 0
More informationCubic Functions: Local Analysis
Cubic function cubing coefficient Capter 13 Cubic Functions: Local Analysis Input-Output Pairs, 378 Normalized Input-Output Rule, 380 Local I-O Rule Near, 382 Local Grap Near, 384 Types of Local Graps
More informationThe derivative function
Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative
More informationPolynomial Interpolation
Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc
More informationContinuity and Differentiability Worksheet
Continuity and Differentiability Workseet (Be sure tat you can also do te grapical eercises from te tet- Tese were not included below! Typical problems are like problems -3, p. 6; -3, p. 7; 33-34, p. 7;
More informationBootstrap confidence intervals in nonparametric regression without an additive model
Bootstrap confidence intervals in nonparametric regression witout an additive model Dimitris N. Politis Abstract Te problem of confidence interval construction in nonparametric regression via te bootstrap
More informationPolynomial Interpolation
Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x
More informationPolynomials 3: Powers of x 0 + h
near small binomial Capter 17 Polynomials 3: Powers of + Wile it is easy to compute wit powers of a counting-numerator, it is a lot more difficult to compute wit powers of a decimal-numerator. EXAMPLE
More information4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.
Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra
More informationNotes on Neural Networks
Artificial neurons otes on eural etwors Paulo Eduardo Rauber 205 Consider te data set D {(x i y i ) i { n} x i R m y i R d } Te tas of supervised learning consists on finding a function f : R m R d tat
More informationA MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES
A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties
More information1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x)
Calculus. Gradients and te Derivative Q f(x+) δy P T δx R f(x) 0 x x+ Let P (x, f(x)) and Q(x+, f(x+)) denote two points on te curve of te function y = f(x) and let R denote te point of intersection of
More informationMathematics 105 Calculus I. Exam 1. February 13, Solution Guide
Matematics 05 Calculus I Exam February, 009 Your Name: Solution Guide Tere are 6 total problems in tis exam. On eac problem, you must sow all your work, or oterwise torougly explain your conclusions. Tere
More informationQuaternion Dynamics, Part 1 Functions, Derivatives, and Integrals. Gary D. Simpson. rev 01 Aug 08, 2016.
Quaternion Dynamics, Part 1 Functions, Derivatives, and Integrals Gary D. Simpson gsim1887@aol.com rev 1 Aug 8, 216 Summary Definitions are presented for "quaternion functions" of a quaternion. Polynomial
More informationLogarithmic functions
Roberto s Notes on Differential Calculus Capter 5: Derivatives of transcendental functions Section Derivatives of Logaritmic functions Wat ou need to know alread: Definition of derivative and all basic
More informationarxiv: v1 [math.pr] 28 Dec 2018
Approximating Sepp s constants for te Slepian process Jack Noonan a, Anatoly Zigljavsky a, a Scool of Matematics, Cardiff University, Cardiff, CF4 4AG, UK arxiv:8.0v [mat.pr] 8 Dec 08 Abstract Slepian
More informationChapter 2 Limits and Continuity
4 Section. Capter Limits and Continuity Section. Rates of Cange and Limits (pp. 6) Quick Review.. f () ( ) () 4 0. f () 4( ) 4. f () sin sin 0 4. f (). 4 4 4 6. c c c 7. 8. c d d c d d c d c 9. 8 ( )(
More information1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point
MA00 Capter 6 Calculus and Basic Linear Algebra I Limits, Continuity and Differentiability Te concept of its (p.7 p.9, p.4 p.49, p.55 p.56). Limits Consider te function determined by te formula f Note
More information. If lim. x 2 x 1. f(x+h) f(x)
Review of Differential Calculus Wen te value of one variable y is uniquely determined by te value of anoter variable x, ten te relationsip between x and y is described by a function f tat assigns a value
More informationDEPARTMENT MATHEMATIK SCHWERPUNKT MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE
U N I V E R S I T Ä T H A M B U R G A note on residual-based empirical likeliood kernel density estimation Birte Musal and Natalie Neumeyer Preprint No. 2010-05 May 2010 DEPARTMENT MATHEMATIK SCHWERPUNKT
More informationMath 31A Discussion Notes Week 4 October 20 and October 22, 2015
Mat 3A Discussion Notes Week 4 October 20 and October 22, 205 To prepare for te first midterm, we ll spend tis week working eamples resembling te various problems you ve seen so far tis term. In tese notes
More informationNUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,
NUMERICAL DIFFERENTIATION James T Smit San Francisco State University In calculus classes, you compute derivatives algebraically: for example, f( x) = x + x f ( x) = x x Tis tecnique requires your knowing
More informationDerivatives. By: OpenStaxCollege
By: OpenStaxCollege Te average teen in te United States opens a refrigerator door an estimated 25 times per day. Supposedly, tis average is up from 10 years ago wen te average teenager opened a refrigerator
More informationPoisson Equation in Sobolev Spaces
Poisson Equation in Sobolev Spaces OcMountain Dayligt Time. 6, 011 Today we discuss te Poisson equation in Sobolev spaces. It s existence, uniqueness, and regularity. Weak Solution. u = f in, u = g on
More informationTaylor Series and the Mean Value Theorem of Derivatives
1 - Taylor Series and te Mean Value Teorem o Derivatives Te numerical solution o engineering and scientiic problems described by matematical models oten requires solving dierential equations. Dierential
More informationEmpirical Bayes Least Squares Estimation without an Explicit Prior
Computer Science Technical Report TR2007-900, May 2007 Courant Institute of Mathematical Sciences, New York University http://cs.nyu.edu/web/research/techreports/reports.html Empirical Bayes Least Squares
More informationThe Laplace equation, cylindrically or spherically symmetric case
Numerisce Metoden II, 7 4, und Übungen, 7 5 Course Notes, Summer Term 7 Some material and exercises Te Laplace equation, cylindrically or sperically symmetric case Electric and gravitational potential,
More informationWeek #15 - Word Problems & Differential Equations Section 8.2
Week #1 - Word Problems & Differential Equations Section 8. From Calculus, Single Variable by Huges-Hallett, Gleason, McCallum et. al. Copyrigt 00 by Jon Wiley & Sons, Inc. Tis material is used by permission
More information7 Semiparametric Methods and Partially Linear Regression
7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).
More informationChapter Seven The Quantum Mechanical Simple Harmonic Oscillator
Capter Seven Te Quantum Mecanical Simple Harmonic Oscillator Introduction Te potential energy function for a classical, simple armonic oscillator is given by ZÐBÑ œ 5B were 5 is te spring constant. Suc
More informationLIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT
LIMITS AND DERIVATIVES Te limit of a function is defined as te value of y tat te curve approaces, as x approaces a particular value. Te limit of f (x) as x approaces a is written as f (x) approaces, as
More informationOptimal parameters for a hierarchical grid data structure for contact detection in arbitrarily polydisperse particle systems
Comp. Part. Mec. 04) :357 37 DOI 0.007/s4057-04-000-9 Optimal parameters for a ierarcical grid data structure for contact detection in arbitrarily polydisperse particle systems Dinant Krijgsman Vitaliy
More information7.1 Using Antiderivatives to find Area
7.1 Using Antiderivatives to find Area Introduction finding te area under te grap of a nonnegative, continuous function f In tis section a formula is obtained for finding te area of te region bounded between
More informationImpact of Lightning Strikes on National Airspace System (NAS) Outages
Impact of Ligtning Strikes on National Airspace System (NAS) Outages A Statistical Approac Aurélien Vidal University of California at Berkeley NEXTOR Berkeley, CA, USA aurelien.vidal@berkeley.edu Jasenka
More informationA Numerical Scheme for Particle-Laden Thin Film Flow in Two Dimensions
A Numerical Sceme for Particle-Laden Tin Film Flow in Two Dimensions Mattew R. Mata a,, Andrea L. Bertozzi a a Department of Matematics, University of California Los Angeles, 520 Portola Plaza, Los Angeles,
More informationI. INTRODUCTION. A. Motivation
Approximate Message Passing Algoritm wit Universal Denoising and Gaussian Mixture Learning Yanting Ma, Student Member, IEEE, Junan Zu, Student Member, IEEE, and Dror Baron, Senior Member, IEEE Abstract
More informationTHE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein
Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul
More informationSolution for the Homework 4
Solution for te Homework 4 Problem 6.5: In tis section we computed te single-particle translational partition function, tr, by summing over all definite-energy wavefunctions. An alternative approac, owever,
More informationDiffraction. S.M.Lea. Fall 1998
Diffraction.M.Lea Fall 1998 Diffraction occurs wen EM waves approac an aperture (or an obstacle) wit dimension d > λ. We sall refer to te region containing te source of te waves as region I and te region
More informationModel development for the beveling of quartz crystal blanks
9t International Congress on Modelling and Simulation, Pert, Australia, 6 December 0 ttp://mssanz.org.au/modsim0 Model development for te beveling of quartz crystal blanks C. Dong a a Department of Mecanical
More informationFinite Difference Methods Assignments
Finite Difference Metods Assignments Anders Söberg and Aay Saxena, Micael Tuné, and Maria Westermarck Revised: Jarmo Rantakokko June 6, 1999 Teknisk databeandling Assignment 1: A one-dimensional eat equation
More informationFast Exact Univariate Kernel Density Estimation
Fast Exact Univariate Kernel Density Estimation David P. Hofmeyr Department of Statistics and Actuarial Science, Stellenbosc University arxiv:1806.00690v2 [stat.co] 12 Jul 2018 July 13, 2018 Abstract Tis
More informationAN IMPROVED WEIGHTED TOTAL HARMONIC DISTORTION INDEX FOR INDUCTION MOTOR DRIVES
AN IMPROVED WEIGHTED TOTA HARMONIC DISTORTION INDEX FOR INDUCTION MOTOR DRIVES Tomas A. IPO University of Wisconsin, 45 Engineering Drive, Madison WI, USA P: -(608)-6-087, Fax: -(608)-6-5559, lipo@engr.wisc.edu
More information