IEEE Transactions on Signal Processing, accepted October 2005, in press
|
|
- Stephanie Lambert
- 5 years ago
- Views:
Transcription
1 IEEE Personal use of this material is permitted However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE IEEE Transactions on Signal Processing, accepted October 5, in press
2 The aussian Transform of distributions: definition, computation and application Teodor Iulian Alecu*, Sviatoslav Voloshynovskiy and Thierry Pun Abstract This paper introduces the general purpose aussian Transform of distributions, which aims at representing a generic symmetric distribution as an infinite mixture of aussian distributions We start by the mathematical formulation of the problem and continue with the investigation of the conditions of existence of such a transform Our analysis leads to the derivation of analytical and numerical tools for the computation of the aussian Transform, mainly based on the Laplace and Fourier transforms, as well as of the afferent properties set (eg the transform of sums of independent variables) The aussian Transform of distributions is then analytically derived for the aussian and Laplacian distributions, and obtained numerically for the eneralized aussian and the eneralized Cauchy distribution families In order to illustrate the usage of the proposed transform we further show how an infinite mixture of aussians model can be used to estimate/denoise non-aussian data with linear estimators based on the Wiener filter The decomposition of the data into aussian components is straightforwardly computed with the aussian Transform, previously derived The estimation is then based on a two-step procedure, the first step consisting in variance estimation, and the second step in data estimation through Wiener filtering To this purpose we propose new generic variance estimators based on the Infinite Mixture of aussians prior It is shown that the proposed estimators compare favorably in terms of distortion with the shrinkage denoising technique, and that the distortion lower bound under this framework is lower than the classical MMSE bound Index Terms aussian mixture, aussian Transform of distributions, eneralized aussian, denoising, shrinkage I INTRODUCTION aussian distributions are extensively used in the (broad sense) signal processing community, mainly for computational benefits For instance, in an estimation problem aussian priors yield quadratic functionals and linear solutions In ratedistortion and coding theories, closed form results are mostly available for aussian source and channel descriptions [] However, real data is most often non aussian distributed, This work is supported in part by the Swiss NCCR IM (Interactive Multimodal Information Management, the EU Network of Excellence Similar ( and the SNF grant N PP-68653/ T I Alecu, S Voloshynovskiy, T Pun are with the Computer Vision and Multimedia Laboratory, University of eneva, 4 Rue énéral-dufour, 4 eneva, Switzerland (corresponding author: TeodorAlecu@cuiunigech, phone: +4()37984; fax: +4()379778) and is best described by other types of distribution (eg in image processing, the most commonly used model for the wavelet coefficients distribution is the eneralized aussian distribution [6]) The goal of the work presented in this paper is to describe non-aussian distributions as an infinite mixture of aussian distributions, through direct computation of the mixing functions from the non-aussian distribution In a related work [4], it was proven that any distribution can be approximated through a mixture of aussian up to an arbitrary level of precision However, no hint was given by the author on how to obtain the desired mixture in the general case In the radar community the problem of modeling non- aussian radar clutters led to the theory of Spherically Invariant Random Processes (SIRP) [3], which aimed at generating non-aussian multivariate distributions using univariate aussian distributions However, the proposed solutions bypassed the mixing function ([] and references therein), performing direct computation of the multivariate non-aussian distribution from the marginal probability distribution function and the desired covariance matrix In [5], an analytical formula is given for an infinite mixture of aussians equivalent to the Laplacian distribution, and used in a source coding application Unfortunately, no generalization was attempted by the authors The work presented here has the roots in their proof and extends the concept through the introduced aussian Transform of distributions, which permits straightforward derivation of exact mixing functions for a wide range of symmetric distributions, including the eneralized Cauchy and the eneralized aussian distributions In order to exemplify the possible usage of the transform we consider then the problem of estimation/denoising, motivated by works such as [9] and [], where, in order to preserve the computational advantages provided by aussian modelling, non-aussian distributions are approximated with finite aussian mixtures obtained through iterative numerical optimization techniques Relying on the aussian Transform, we propose new generic algorithms for denoising non- aussian data, based on the description of non-aussian distributions as an Infinite Mixture of aussians (IM) The rest of the paper is divided in three main sections In Section II we define the aussian Transform, analyze its existence, investigate its properties and derive the mathematical tools for analytical and/or numerical computation In Section III we exemplify both the transform for some typical distributions such as eneralized aussian and eneralized Cauchy, and some of the properties deduced in Section II The last section tackles the problem of
3 estimation/denoising and proposes new aussian Transform based denoising schemes, exemplifying them for the eneralized aussian Distribution (D) family The obtained results are compared in terms of distortion with the state-of-the-art denoising method of Moulin and Liu [7] II THE AUSSIAN TRANSFORM OF DISTRIBUTIONS A Definition and existence We consider a generic symmetric continuous distribution p(x) As we are aiming at representing it through an infinite mixture of aussians, we can safely disregard the mean, and assume for simplicity reasons that p(x) is zero-mean We are looking for an integral representation in the form: ( ) N ( x ) d = p( x), () where N ( x ) is the zero-mean aussian distribution: x N ( x ) = e, π and ( ) is the mixing function that should reproduce the original p(x) We can now introduce the aussian Transform Definition : aussian Transform The direct aussian Transform is defined as the operator which transforms p( x ) into ( ), and the Inverse aussian Transform - is defined as the operator which maps ( ) back to p( x ) : : p( x) ( ); : ( ) p( x) Obviously, - is simply given by (): ( ( )) = ( ) ( x ) d = p( x) N () The aussian Transform of a distribution exists if, given p( x ), a mixing distribution ( ) can be found such as to comply with () This can be summarized in three conditions Condition For a given p(x), a function ( ) defined according to () exists Condition This function is non-negative Condition 3 Its integral ( ) d is equal to The last condition is a consequence of Condition Indeed, if ( ) exists, then integrating both sides of () with respect to x and inverting the integration order one obtains: ( ) N ( x ) d dx = p( x) dx, x dx d = N Finally, since ( x ) N is a distribution: ( ) d = In order to investigate Condition, the existence of ( ), we perform the following variable substitutions: s = x and t = Since p(x) is symmetric, it can be rewritten as: p( x) = p( x ) = p( s) The left hand side of () transforms to: ( ) ( ) st N x d = e dt tt π t According to the definition of the Laplace Transform L [], equation () finally takes the form: L = p( s) (3) tt π t Thus, ( ) is linked to the original probability distribution p(x) through the Laplace Transform and can be computed using the Inverse Laplace Transform L - The direct aussian Transform is therefore given by: = π ( p( x) ) = ( p( s) )() t L, (4) t and the Inverse aussian Transform can be computed as: ( ( )) = L ( s) (5) tt π t s= x Consequently, the existence of the aussian Transform is conditioned by the existence of the Inverse Laplace Transform of p( s ) Using the general properties of the Laplace Transform, it is sufficient [] to prove that the limit at infinity of s p( s) is bounded, or equivalently : lim x x p x < (6) The above condition is satisfied by all the distributions from the exponential family, as well as by all the distributions with finite variance The equation (4) allows for straightforward identification of aussian Transforms for distributions whose Laplace Transforms are known, by simply using handbook tables Unfortunately, the condition (6) does not guarantee compliance with the Condition : non-negativity As it is rather difficult to verify a priori this constraint, the test should be performed a posteriori, either analytically or numerically However, by imposing the non-negativity constraint in (), one can see that it is necessary to have a strictly decreasing distribution function p(x) on positive x, as p(x) is described by a sum of strictly decreasing zero-mean aussians functions From the same reason, but not related to non-negativity, the distribution function p(x) should also be of class C B Properties of the aussian Transform Property : Uniqueness If two functions and
4 ( ) exist such that ( ( )) = p( x) ( ( )) = p( x) ( ) ( ) and then they are identical = Conversely, if two distribution p x have the same aussian probabilities p x and Transform ( ) then they are identical Proof The second proposition of the uniqueness property is proved straightforwardly using () In what concerns the first proposition, it is a direct consequence of the uniqueness of the Laplace Transform [], which states that if two functions have the same Laplace transform, then they are identical It is then sufficient to observe that is computed as a Laplace Transform (5) Property : Mean Value (variance conservation) The mean value of the aussian Transform is equal to the variance of the original distribution x p( x) dx = ( ) d (7) x p x dx = x x d dx = Proof N ( ) N = x x dx d = d qed Property 3: Value at infinity If lim x x p( x) < then the aussian Transform tends to when tends to infinity: lim ( ) = (8) Proof We use the initial value theorem for the Laplace Transform [] and the direct formula (3): limt = lims s p s = lim x x p x, tt π t then lim < lim = 3 ( )( ) ( ) We study now the influence of basic operations on the aussian Transform, such as scalar multiplication and addition of independent variables Property 4: Scaling If and are random variables such that = α and ( p ( x) ) =, then the aussian Transform corresponding to the scaled variable is a scaled version of : = (9) α α Proof We begin by observing that the distribution probability of the random variable can be expressed as: x p ( x) = p α α Then we compute the Inverse aussian Transform of the expression proposed in (9) using (): ( ) = N x d ; α α x α * α ( ( )) e d α α α = π α x ( ( )) = p p = x α α Finally using the uniqueness property we conclude that the aussian Transform ( p ( x) ) is unique and given by (9) Property 5: Convolution If and are two independent p x = and random variables with ( p ( x) ) = ( ) ( ), then the aussian Transform of their sum is the convolution of their respective aussian Transforms (the result can be generalized for the sum of multiple variables): p + ( x) = () Proof Consider the random variable = + Since exists, is a random aussian variable with variance distributed according to the distribution probability Similarly, is a random aussian variable with variance distributed according to Then is also a random aussian variable with variance = + But and are independent variables drawn from and It follows that is a random variable described by the probability distribution =, qed Corollary If and are independent random variables and is aussian distributed p ( x) = N ( x ), then the aussian transform of their sum is a shifted version of :, < + = () ( ), C Numerical computation The computation of the aussian Transform for distributions not available in handbooks is still possible through the complex inversion method for Laplace Transforms known as the Bromwich integral [3]: ( ) ε + i ut L f s = f ( u) e du π i, () ε i where ε is a real positive constant satisfying ε > sup( Re( poles ( f ))) and u is an auxiliary variable The equation () can be rearranged as a Fourier Transform, allowing the use of the numerous numerical and/or symbolical packages available ( ω is the variable in the Fourier space): εt L ( f ( s) ) = e F ( f ( ε + iω )) (3) Very often p(x) has no poles, being a continuous and bounded
5 function, and in this case it might be very practical to evaluate (3) in the limit case ε Using (4) and (3): π ( ) = F ( p( iω) )() t (4) t = When the original distribution is only numerically known, approximation of ( ) is still possible either through analytical approximations of p(x), followed by (4) or (4), or through solving the inverse problem yielded by (), or through real inversion of the Laplace transform using (4) The accuracy of the obtained transforms can then be assessed using an appropriate metric or distance such as Kullback- Leibler However, the abovementioned approximation methods and the accuracy study constitute stand-alone problems out of the scope of the present paper C b p x b = p s b = π + π + b C ( ) ; ( ) C ( ) b x b s b = (7) p x b e π The results (5), (6) and (7) are plotted in Fig To exemplify the Corollary to the Convolution Property, consider now Cauchy data contaminated with independent additive white aussian (AW) noise Then the aussian Transform of the measured data (Fig ) is a shifted version of the original data transform () III EAMPLES OF AUSSIAN TRANSFORMS A Analytic aussian transforms The most obvious and natural example is the aussian Transform of a aussian distribution N ( x ) One would expect to have (with δ the Dirac function): ( N ( x )) = δ ( ) (5) Proof It is trivial that (5) verifies the equation () Then, by using the uniqueness of the aussian Transform one can conclude that the aussian Transform of a aussian distribution N ( x ) is a Dirac function centered in The convolution property () can now be used to prove a well known result in statistics: the sum of two independent aussian variables, with respective probability laws x N x, is another aussian variable with N ( ) and ( ) probability distribution ( x + ) N (the extension to non-zero mean distributions is trivial) Proof: ( p ( x) ) = δ ( ); ( ) p x = δ p x = p x * p x = δ + ( + ) ( ) Then, inverting the aussian Transform: p ( ( )) + = δ + = N x + Similarly to (5), it is possible to compute the aussian Transforms of other usual symmetric distributions using the Laplace transform tables [] We exemplify with the Laplacian and Cauchy distributions Laplacian distribution: L λ λx L λ λ s p ( x λ) = e ; p ( s λ) = e (6) λ L λ ( p ( s )) λ = e As mentioned, the result (6) was already proven in [5] Cauchy distribution: Figure aussian Transforms of the Laplacian, aussian and Cauchy distributions Figure Shift of the aussian Transform following data contamination with AW noise The previous analytical results will be generalized in subsection B through numerical computations B Numerical computation of aussian Transforms This part illustrates the numerical computation of the aussian Transform through (4) for the eneralized aussian and eneralized Cauchy distributions families ) eneralized aussian Distribution (D) The D family is described by an exponential probability density function with parameters and : η ( ) x p ( x, ) = e η, (8) Γ ( ) where η( ) 3 ( ) ( ) = Γ Γ For = the D particularizes to the Laplacian distribution, while for = one obtains the aussian distribution The aussian Transform of the eneralized aussian distribution can not be obtained in analytical form using (4) However, it does exist (6) and can be calculated numerically through (4): ( ) π ( ) = F ( ω), p i t, t = η ( ) η ( ) ( ω) ( ) ( ω) p i = e i, Γ
6 The aussian Transforms for ranging from 5 to with fixed = are plotted in Fig 3 The transforms evolve from a Dirac-like distribution centered on for small to exponential for =, then Rayleigh-like for =, bellshaped for = 5 and again Dirac-like centered at for = As expected, the aussian Transform of the Laplacian distribution ( = ) is exponential (6) Figure 3 aussian Transforms of D Unfortunately, the real part of the complex probability function diverges for periodical values of, which impedes the computation of the transform through this method for > However, real data (in transform domains such as discrete cosine and wavelet) is mostly confined to << [6] ) eneralized Cauchy Distribution (CD) The CD probability density function is given by Γ ( ν + ) ν C b px x ν, b =, ν + 5 πγ ( ν) ( b + x ) and it particularizes to the Cauchy distribution for ν = 5 Its aussian Transform can be computed through: ν b Γ ( ν + 5) ν 5 ν, b( ) = ( F (( b + iω) ) ) Γ( ν) t = Corresponding plots are given in Fig 4 Figure 4 aussian Transforms of CD As a remark, the variance of the Cauchy distribution being infinite, its aussian Transform has infinite mean value (7) The CD family possesses finite variance only for ν > IV DENOISIN We are interested in the generic denoising problem of estimating the original scalar data x from the measured data y, degraded by additive noise z: y = x+ z (9) We assume that the noise random variable is zero-mean aussian distributed with variance : z p ( z) = e () π We further assume that the probability density function p (x), which describes the source random variable, is a zeromean symmetric distribution, and that its aussian Transform ( ) = ( p ( x) ) exists For exemplification of the proposed techniques we consider p (x) to be the eneralized aussian distribution (8) We base our estimation on the Maximum a Posteriori (MAP) principle [4]: xˆ = arg max x p Y ( x y), which yields for iid zero-mean aussian data and noise (with variances and ) the classical Wiener filter: xˆ = y () + We also note that the estimate () is also the Minimum Mean Square Error (MMSE) estimate for iid zero-mean aussian data and noise: xˆmmse = arg min ( ˆ x xx) p Y ( x y) dx x () xˆ = xp x y dx MMSE Y x According to the IM model, each data sample is interpreted to be drawn from a aussian distribution with variance distributed accordingly to the aussian Consequently, using the prior Transform knowledge of ( ), we can estimate the variance from y and use Wiener filtering () Thus, our estimator is given by: ( y) xˆ = y (3) y + The initial problem is now reduced to a variance estimation problem For this purpose, we begin by introducing the cumulative estimation technique A Cumulative Estimation We are looking for an estimator that provides variance estimates consistent with the original data distribution, in the sense that the description of the data as a aussian mixture should be identical to the non-aussian original description We set the above assertion as a consistency principle
7 Principle : Consistency The probability distribution function of the variance estimate should be identical to the aussian Transform of the prior distribution p (x) Moreover, since the distributions p (x), p (z) are symmetric and zero-mean, we assume an estimator of the form: ˆ ( y) = ξ M ( M ( y) ), where ξ M denotes the consistent variance estimator depending on the symmetric real non-negative function: + M: ; M( y) = M( y), + where and denote real, respectively real nonnegative numbers As an example, a possible M function is the absolute value of y We further assume that this estimator is monotonically increasing with the function M: y < y ξ M y M M M M < ξ M y (4) ( ) ( ) If one chooses the M function as the absolute value y, this simply means that if the absolute value of y is larger, then the estimated variance ˆ ( y) is also larger If the variables are vectors (eg local estimation paradigm) one can similarly use as the M function the variance of the measurement vector, and (4) would mean that if the variance of the measurement is larger, then the estimated variance is also larger The assumption in (4) implies that the cumulative distribution function of the variance estimates is equal to the cumulative distribution function of M: p( ˆ ˆ ( y) ) = p( M M ( y) ) (5) Considering continuous probability density functions and incorporating the consistency condition (the distribution of the estimates should be identical to ), (5) can be rewritten as: M PM ( M( y) ) = P ( ξ ( M( y) )), M( y) x (6) P M y = p u du; P = u du, M ( ) M where p M is the probability density function describing M ( y), and P M and P are cumulative probability functions, as defined in (6) Then the estimator we are looking for is simply given by: M ( M( y) ) = P PM ( M ( y) ), (7) ξ where P is the mathematical inverse of the function P : ( if P ) then = u P u = We denote (7) as the cumulative estimator, simply because it uses cumulative probability functions for the estimation We further infer that the cumulative estimator is robust with respect to monotonically increasing transformations of M Theorem If M and M are two symmetric real nonnegative functions M, M + :, and if M( y) < M( y) M( y) < M ( y), then the associated cumulative estimators are identical: M ( ( y) ) = ξ ( ( y) ) M ξ M M Proof Since the inequalities are preserved (theorem hypothesis), then the cumulative distribution functions P M and P M are equal: p( M M( y) ) = p( M M( y) ) then M ( y) M ( y) P M y = p u du = p u du = P M y ( ) M M M M ( ) The proof is ended by corroborating the above equality with the cumulative estimation formula (7) B IM cumulative estimator The cumulative estimator (7) can be directly used in (3) by setting the M function as M ( y) = y, and the IM cumulative estimator (IM CE) is simply: ˆ _ CE ( y) xˆ IM _ CE = y with ˆ _ CE ( y) + (8) y ˆ _ CE ( y) = P ( PM( y) ), PM( y) = py ( u) du As and are assumed to be independent, p Y (y) can be computed as the convolution of p (x) and p (z): py = p p In the general case the operations in (8) have to be performed numerically However, if the distribution p (x) is Laplacian (D with = ), its aussian Transform is (6), (8): ( ), = e, (9) and the cumulative estimator from (8) reduces to: _ CE, = ( y) = ln ( P M ( y) ) (3) As tends to, the aussian Transform of the D tends asymptotically to a Dirac function centered at (5), and the estimator (8) to the classical Wiener filter () In order to asses the performances of the proposed estimators we measure the accuracy of the estimate according to the L norm: e( x, xˆ) = ( x xˆ) The global distortion induced by an estimator is therefore: D = e( x, xˆ ) p ( x) p ( z) dxdz (3) We compare our estimator in terms of distortion with the estimators known as shrinkage functions from [7] The idea of shrinking originates from [8], but [7] provides both improved results and statistical interpretation of the functions In fact, the shrinkage functions from [7] are the result of direct MAP estimation, which in the case of D data and aussian noise yields ˆx as the solution of the equation in x: y = x η( ) x (3) Equation (3) has analytical solutions for = 5 (cubic), = or (linear) and = 5 (quadratic) The resulting shrinkage functions have thresholding characteristics for
8 and yield the Wiener filter for = The equivalent IM CE functions do not present thresholding characteristics for any, behaving in continuous manner raphic illustration of the shrinkage functions and of the equivalent IM CE functions obtained via (8) are presented in Fig 5 of separable IM components In fact, for a given y, the MMSE estimator is unique (injective estimation), which is not the case for the ideal estimator that would lead to the proposed IM lower bound, and that would yield, for the same value of y, different estimates for each isolated aussian component Figure 5 Shrinkage functions and IM CE estimation curves One can also compute the lower bound on the distortion induced by an estimation procedure based on signal decomposition into an IM Assuming that the aussian components of the IM are separable, and that the variance estimate is constant over each separate component, the equation () is the MMSE estimate for each component The bound is attained when the estimator "knows" the aussian component of the mixture that has generated the estimated sample The ensuing distortion lower bound is: Dlower bound = ( ) d, (33) + where the ratio + is the theoretical distortion of the MMSE filter () We also compare our results with the MMSE distortion bound, given by the MMSE estimator (), which can be rewritten as follows: xˆ MMSE = xp( x y) dx = x p( x y, ) p( y) d dx x x Using Bayes rule and switching integrals, one obtains: p( y ) p( ) xˆ MMSE = xp ( x y, ) dx d x p( y) Considering aussian noise the first parenthesis yields the MMSE estimate (), and by integrating the aussian p y one finds: Transform and developing xˆ p y and N( + ) ( y + ) ( ) d y d + MMSE = N y (34) The MMSE distortion bound is then computed by plugging the result (34) into (3) The IM lower bound (33) and the MMSE bound (34) are superposed in Fig 6, which displays the distortion of the estimation (3) as a function of the signal to noise ratio SNR = log (the variance of the D family is given by, see the moments method from [6] and equation (8)) One may notice that the IM lower bound is lower than the MMSE bound, a result due to the assumption Figure 6 Distortion for D data as a function of SNR (db) A considerable difference in performance in the favor of IM CE with respect to the shrinkage functions is observed for = 5 The difference tends to decrease for larger, and inverts for = 5, albeit with a very small relative module However, the IM CE estimator being unique for a given y, the IM CE distortion curve necessarily remains above the MMSE bound, as for any other injective estimator All curves tend to the distortion of the Wiener filter when approaches, plotted dashed-dotted in Fig on the graph corresponding to = 5 C IM ML and MAP variance estimation The simplest way of obtaining the variances in a statistical framework is through Maximum Likelihood (ML) estimation: ˆ _ ML = arg max p( y ), (35) which for zero-mean independent aussian source and noise data yields the solution: ˆ _ ML = max ( y,) (36) If the variance distribution were known, one could improve (36) through MAP estimation Using the aussian Transform the MAP variance estimate is: ˆ _ MAP = arg max ( ) p( y ) (37)
9 enerally, the expression (37) requires numerical maximization, but an analytical form exists for = and aussian noise (using (9) and differentiation of (37)): y ˆ _ MAP, = = max, (38) 8y + + One would expect the IM MAP estimator to behave similarly to the original shrinkage functions of Moulin and Liu, as it applies the MAP principle two times (37), (3) to obtain an indirect estimate, whereas the shrinkage functions are the result of direct MAP estimation (3) This is confirmed in Fig 7, which compares the equivalent shrinkage functions of the IM ML and IM MAP estimators with the IM CE curve and with the Moulin and Liu curve for = Figure 7 IM ML and MAP equivalent shrinkage functions Unlike the IM CE estimator, the two proposed IM ML and MAP estimators do present thresholding characteristics, and, as expected, the equivalent IM MAP curve behaves very similarly to the original shrinkage curve We tested empirically the MAP (38) estimator by generating D eneralized aussian data of length N= 6 with = 5 and =, adding independent white aussian noise, and performing denoising according to () The source and noise probability densities were assumed to be known, and the corresponding distortion was computed as the mean square error (Fig 8) We also tested the cumulative estimator (8) and the shrinkage function (3) Figure 8 Empirical distortion curves As expected, empirical results reproduce the theoretical results obtained for the scalar estimators (Fig 6) The IM MAP results are close, but above the direct MAP estimation of the shrinkage functions for = (this result was predictable from the curves in Fig7), however they notably improve on the Moulin and Liu results for = 5 We expect the IM MAP curves to remain above the Moulin and Liu curves for, both falling slowly down to the aussian-aussian curve in Fig6 as the IM MAP estimator tends to the Wiener filter for = While the IM MAP estimator does not outperform the cumulative estimator, nor the shrinkage estimator for, and while the IM CE estimator does not outperform the classical MMSE estimate, the linear nature of the estimator (3), coupled with the simple analytical forms (38) and (3) for =, allowed the successful use of the IM MAP and IM CE estimators in the highly underdetermined EE (electroencephalogram) inverse problem [5] with Laplacian prior, where a direct application of the MAP principle would lead to an underdetermined quadratic programming problem, and where a direct application of the MMSE principle would lead to highly expensive computations V CONCLUSION AND FUTURE WORK We introduced in this paper the aussian Transform, which allows for the representation of symmetric distributions as an infinite aussian mixture The scope of applicability of the aussian Transform is potentially very broad, from denoising and regularization to filtering, coding, compression, watermarking etc However, an extension of the concept to non-symmetric distribution would be required for some specific applications Further investigation of the existence conditions, especially non-negativity, is also necessary Finally, one would need adapted numerical packages (most likely based on existing Laplace and Fourier transform computational packages) for the computation of aussian Transforms of both analytically and numerically defined distributions We also presented in this paper various denoising algorithms based on the Infinite Mixture of aussians model Their application to the denoising of eneralized aussian data showed significant improvements in terms of distortion, both theoretically and empirically, with respect to the shrinkage functions of Moulin&Liu Also, the lower bound under the proposed IM estimation framework is lower than the classical MMSE bound However the potential of IM based denoising is not yet fully explored, as the lower bound presented in this paper is still below current results The estimation methods proposed may reach this lower bound in the case where the data is generated locally from the same aussian component, which would allow for exact variance estimation ACKNOWLEDMENT The authors thank Dr Oleksiy Koval for careful reading
10 and fruitful discussions The authors also thank anonymous reviewers for valuable comments and suggestions REFERENCES A D Poularikas, S Seely, Laplace Transforms The Transforms and Applications Handbook: Second Edition Ed Alexander D Poularikas Boca Raton: CRC Press LLC, Arfken, Mathematical Methods for Physicists, 3rd ed Orlando, FL: Academic Press, pp , Arfken, "Inverse Laplace Transformation" 5 in Mathematical Methods for Physicists, 3rd ed Orlando, FL: Academic Press, pp , R Wilson, MMM: Multiresolution aussian Mixture Models for Computer Vision, in IEEE International Conference on Pattern Recognition, Barcelona, Spain, September 5 A Hjørungnes, J M Lervik, and T A Ramstad, Entropy Coding of Composite Sources Modeled by Infinite Mixture aussian Distributions, in Proc 996 IEEE Digital Signal Processing Workshop, Loen, Norway, September S Mallat, A Theory for Multiresolution Signal Decomposition: The Wavelet Representation, in IEEE Transactions on Pattern Analysis and Machine Intelligence Volume, Issue 7 (July 989) p P Moulin and J Liu, "Analysis of Multiresolution Image Denoising Schemes Using eneralized - aussian and Complexity Priors", IEEE Trans Info Theory, Special Issue on Multiscale Analysis, Vol 45, No 3, pp 99-99, Apr D L Donoho and I M Johnstone, "Ideal Spatial Adaptation Via Wavelet Shrinkage", Biometrika, Vol 8, pp , A Bijaoui, "Wavelets, aussian mixtures and Wiener filtering'', Signal Processing, Vol8, No4, April, pp 79-7 Benaroya, L; Bimbot, F; "Wiener-based source separation with HMM/MM using a single sensor'', Proc 4th Int Symp on Independent Component Anal and Blind Signal Separation, April 3, Nara, Japan T M Cover, J A Thomas, "Elements of Information Theory", John Wiley & Sons Inc, 99 M Rangaswamy, D Weiner, A Ozturk, "Non-aussian Random Vector Identification Using Spherically Invariant Random Processes", in IEEE Transactions on Aerospace and Electronic Systems, vol 9, No, January E Conte and M Longo, "Characterization of radar clutter as a spherically invariant random process", in IEEE Proceedings, Pt F (Communication, Radar and Signal Processing), 34, (987), MBertero, PBoccacci, "Introduction to Inverse Problems in Imaging", IOP Publishing, Bristol, T I Alecu, P Missionnier, S Voloshynovskiy, P iannakopoulos and T Pun, "Soft/hard focalization in the EE inverse problem", in IEEE Workshop on Statistical Signal Processing (SSP'5), Bordeaux, France, July, 7-5
The Gaussian transform. ALECU, Teodor, VOLOSHYNOVSKYY, Svyatoslav, PUN, Thierry. Abstract
Proceedings Chapter The aussian transform ALECU, Teodor, VOLOSHYNOVSKYY, Svyatoslav, PUN, Thierry Abstract This paper introduces the general purpose aussian Transform, which aims at representing a generic
More informationBayesian Paradigm. Maximum A Posteriori Estimation
Bayesian Paradigm Maximum A Posteriori Estimation Simple acquisition model noise + degradation Constraint minimization or Equivalent formulation Constraint minimization Lagrangian (unconstraint minimization)
More informationA GLRT FOR RADAR DETECTION IN THE PRESENCE OF COMPOUND-GAUSSIAN CLUTTER AND ADDITIVE WHITE GAUSSIAN NOISE. James H. Michels. Bin Liu, Biao Chen
A GLRT FOR RADAR DETECTION IN THE PRESENCE OF COMPOUND-GAUSSIAN CLUTTER AND ADDITIVE WHITE GAUSSIAN NOISE Bin Liu, Biao Chen Syracuse University Dept of EECS, Syracuse, NY 3244 email : biliu{bichen}@ecs.syr.edu
More informationLow-Complexity Image Denoising via Analytical Form of Generalized Gaussian Random Vectors in AWGN
Low-Complexity Image Denoising via Analytical Form of Generalized Gaussian Random Vectors in AWGN PICHID KITTISUWAN Rajamangala University of Technology (Ratanakosin), Department of Telecommunication Engineering,
More informationWavelet Based Image Restoration Using Cross-Band Operators
1 Wavelet Based Image Restoration Using Cross-Band Operators Erez Cohen Electrical Engineering Department Technion - Israel Institute of Technology Supervised by Prof. Israel Cohen 2 Layout Introduction
More informationEVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER
EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER Zhen Zhen 1, Jun Young Lee 2, and Abdus Saboor 3 1 Mingde College, Guizhou University, China zhenz2000@21cn.com 2 Department
More informationMMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING
MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING Yichuan Hu (), Javier Garcia-Frias () () Dept. of Elec. and Comp. Engineering University of Delaware Newark, DE
More informationIndependent Component Analysis. Contents
Contents Preface xvii 1 Introduction 1 1.1 Linear representation of multivariate data 1 1.1.1 The general statistical setting 1 1.1.2 Dimension reduction methods 2 1.1.3 Independence as a guiding principle
More informationMachine Learning Lecture Notes
Machine Learning Lecture Notes Predrag Radivojac January 25, 205 Basic Principles of Parameter Estimation In probabilistic modeling, we are typically presented with a set of observations and the objective
More informationLecture Note 1: Probability Theory and Statistics
Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would
More informationPerformance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels
Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels Jilei Hou, Paul H. Siegel and Laurence B. Milstein Department of Electrical and Computer Engineering
More informationWavelet Footprints: Theory, Algorithms, and Applications
1306 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 51, NO. 5, MAY 2003 Wavelet Footprints: Theory, Algorithms, and Applications Pier Luigi Dragotti, Member, IEEE, and Martin Vetterli, Fellow, IEEE Abstract
More informationBLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS
BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS M. Kawamoto 1,2, Y. Inouye 1, A. Mansour 2, and R.-W. Liu 3 1. Department of Electronic and Control Systems Engineering,
More informationUnderdetermined Instantaneous Audio Source Separation via Local Gaussian Modeling
Underdetermined Instantaneous Audio Source Separation via Local Gaussian Modeling Emmanuel Vincent, Simon Arberet, and Rémi Gribonval METISS Group, IRISA-INRIA Campus de Beaulieu, 35042 Rennes Cedex, France
More informationLecture Note 2: Estimation and Information Theory
Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 2: Estimation and Information Theory Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 2.1 Estimation A static estimation problem
More informationOn the Cost of Worst-Case Coding Length Constraints
On the Cost of Worst-Case Coding Length Constraints Dror Baron and Andrew C. Singer Abstract We investigate the redundancy that arises from adding a worst-case length-constraint to uniquely decodable fixed
More informationS. R. Tate. Stable Computation of the Complex Roots of Unity, IEEE Transactions on Signal Processing, Vol. 43, No. 7, 1995, pp
Stable Computation of the Complex Roots of Unity By: Stephen R. Tate S. R. Tate. Stable Computation of the Complex Roots of Unity, IEEE Transactions on Signal Processing, Vol. 43, No. 7, 1995, pp. 1709
More informationModeling Multiscale Differential Pixel Statistics
Modeling Multiscale Differential Pixel Statistics David Odom a and Peyman Milanfar a a Electrical Engineering Department, University of California, Santa Cruz CA. 95064 USA ABSTRACT The statistics of natural
More informationLatent Tree Approximation in Linear Model
Latent Tree Approximation in Linear Model Navid Tafaghodi Khajavi Dept. of Electrical Engineering, University of Hawaii, Honolulu, HI 96822 Email: navidt@hawaii.edu ariv:1710.01838v1 [cs.it] 5 Oct 2017
More informationLogarithmic quantisation of wavelet coefficients for improved texture classification performance
Logarithmic quantisation of wavelet coefficients for improved texture classification performance Author Busch, Andrew, W. Boles, Wageeh, Sridharan, Sridha Published 2004 Conference Title 2004 IEEE International
More informationThe Fading Number of a Multiple-Access Rician Fading Channel
The Fading Number of a Multiple-Access Rician Fading Channel Intermediate Report of NSC Project Capacity Analysis of Various Multiple-Antenna Multiple-Users Communication Channels with Joint Estimation
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationEstimation Error Bounds for Frame Denoising
Estimation Error Bounds for Frame Denoising Alyson K. Fletcher and Kannan Ramchandran {alyson,kannanr}@eecs.berkeley.edu Berkeley Audio-Visual Signal Processing and Communication Systems group Department
More informationA Strict Stability Limit for Adaptive Gradient Type Algorithms
c 009 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional A Strict Stability Limit for Adaptive Gradient Type Algorithms
More informationCS 195-5: Machine Learning Problem Set 1
CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of
More informationNUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS
NUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS Justin Dauwels Dept. of Information Technology and Electrical Engineering ETH, CH-8092 Zürich, Switzerland dauwels@isi.ee.ethz.ch
More informationInformation Theory in Intelligent Decision Making
Information Theory in Intelligent Decision Making Adaptive Systems and Algorithms Research Groups School of Computer Science University of Hertfordshire, United Kingdom June 7, 2015 Information Theory
More informationLegendre-Fenchel transforms in a nutshell
1 2 3 Legendre-Fenchel transforms in a nutshell Hugo Touchette School of Mathematical Sciences, Queen Mary, University of London, London E1 4NS, UK Started: July 11, 2005; last compiled: August 14, 2007
More informationSatellite image deconvolution using complex wavelet packets
Satellite image deconvolution using complex wavelet packets André Jalobeanu, Laure Blanc-Féraud, Josiane Zerubia ARIANA research group INRIA Sophia Antipolis, France CNRS / INRIA / UNSA www.inria.fr/ariana
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationDiscriminant Analysis with High Dimensional. von Mises-Fisher distribution and
Athens Journal of Sciences December 2014 Discriminant Analysis with High Dimensional von Mises - Fisher Distributions By Mario Romanazzi This paper extends previous work in discriminant analysis with von
More informationBlind Spectral-GMM Estimation for Underdetermined Instantaneous Audio Source Separation
Blind Spectral-GMM Estimation for Underdetermined Instantaneous Audio Source Separation Simon Arberet 1, Alexey Ozerov 2, Rémi Gribonval 1, and Frédéric Bimbot 1 1 METISS Group, IRISA-INRIA Campus de Beaulieu,
More informationRobust Inference. A central concern in robust statistics is how a functional of a CDF behaves as the distribution is perturbed.
Robust Inference Although the statistical functions we have considered have intuitive interpretations, the question remains as to what are the most useful distributional measures by which to describe a
More information01 Probability Theory and Statistics Review
NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement
More informationUnsupervised Learning with Permuted Data
Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University
More informationA Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases
2558 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 48, NO 9, SEPTEMBER 2002 A Generalized Uncertainty Principle Sparse Representation in Pairs of Bases Michael Elad Alfred M Bruckstein Abstract An elementary
More informationImpact of channel-state information on coded transmission over fading channels with diversity reception
Impact of channel-state information on coded transmission over fading channels with diversity reception Giorgio Taricco Ezio Biglieri Giuseppe Caire September 4, 1998 Abstract We study the synergy between
More informationDETECTION theory deals primarily with techniques for
ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for
More informationWiener Measure and Brownian Motion
Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u
More informationSome Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom
Some Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom Stefan M. Moser April 7, 007 Abstract The non-central chi-square distribution plays an important role
More informationImage Denoising using Uniform Curvelet Transform and Complex Gaussian Scale Mixture
EE 5359 Multimedia Processing Project Report Image Denoising using Uniform Curvelet Transform and Complex Gaussian Scale Mixture By An Vo ISTRUCTOR: Dr. K. R. Rao Summer 008 Image Denoising using Uniform
More informationExpect Values and Probability Density Functions
Intelligent Systems: Reasoning and Recognition James L. Crowley ESIAG / osig Second Semester 00/0 Lesson 5 8 april 0 Expect Values and Probability Density Functions otation... Bayesian Classification (Reminder...3
More informationA Statistical Analysis of Fukunaga Koontz Transform
1 A Statistical Analysis of Fukunaga Koontz Transform Xiaoming Huo Dr. Xiaoming Huo is an assistant professor at the School of Industrial and System Engineering of the Georgia Institute of Technology,
More informationA Partial Differential Equation Approach to Image Zoom
A Partial Differential Equation Approach to Image Zoom Abdelmounim Belahmidi and Frédéric Guichard January 2004 Abstract We propose a new model for zooming digital image. This model, driven by a partial
More informationEXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME. Xavier Mestre 1, Pascal Vallet 2
EXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME Xavier Mestre, Pascal Vallet 2 Centre Tecnològic de Telecomunicacions de Catalunya, Castelldefels, Barcelona (Spain) 2 Institut
More informationIEEE copyright notice
IEEE copyright notice Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for
More informationWhich wavelet bases are the best for image denoising?
Which wavelet bases are the best for image denoising? Florian Luisier a, Thierry Blu a, Brigitte Forster b and Michael Unser a a Biomedical Imaging Group (BIG), Ecole Polytechnique Fédérale de Lausanne
More informationPROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata
' / PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE Noboru Murata Waseda University Department of Electrical Electronics and Computer Engineering 3--
More informationInformation geometry for bivariate distribution control
Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationParametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory
Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007
More informationA View on Extension of Utility-Based on Links with Information Measures
Communications of the Korean Statistical Society 2009, Vol. 16, No. 5, 813 820 A View on Extension of Utility-Based on Links with Information Measures A.R. Hoseinzadeh a, G.R. Mohtashami Borzadaran 1,b,
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationIntegrating Correlated Bayesian Networks Using Maximum Entropy
Applied Mathematical Sciences, Vol. 5, 2011, no. 48, 2361-2371 Integrating Correlated Bayesian Networks Using Maximum Entropy Kenneth D. Jarman Pacific Northwest National Laboratory PO Box 999, MSIN K7-90
More informationMULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING. Kaitlyn Beaudet and Douglas Cochran
MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING Kaitlyn Beaudet and Douglas Cochran School of Electrical, Computer and Energy Engineering Arizona State University, Tempe AZ 85287-576 USA ABSTRACT The problem
More informationMathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector
On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean
More informationMeasure-Transformed Quasi Maximum Likelihood Estimation
Measure-Transformed Quasi Maximum Likelihood Estimation 1 Koby Todros and Alfred O. Hero Abstract In this paper, we consider the problem of estimating a deterministic vector parameter when the likelihood
More informationEmpirical Bayes Least Squares Estimation without an Explicit Prior
Computer Science Technical Report TR2007-900, May 2007 Courant Institute of Mathematical Sciences, New York University http://cs.nyu.edu/web/research/techreports/reports.html Empirical Bayes Least Squares
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationLecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary
ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood
More informationDeakin Research Online
Deakin Research Online This is the published version: Phat, V. N. and Trinh, H. 1, Exponential stabilization of neural networks with various activation functions and mixed time-varying delays, IEEE transactions
More informationLecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions
DD2431 Autumn, 2014 1 2 3 Classification with Probability Distributions Estimation Theory Classification in the last lecture we assumed we new: P(y) Prior P(x y) Lielihood x2 x features y {ω 1,..., ω K
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationBlock 2: Introduction to Information Theory
Block 2: Introduction to Information Theory Francisco J. Escribano April 26, 2015 Francisco J. Escribano Block 2: Introduction to Information Theory April 26, 2015 1 / 51 Table of contents 1 Motivation
More informationThe Method of Types and Its Application to Information Hiding
The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,
More informationHypothesis Testing with Communication Constraints
Hypothesis Testing with Communication Constraints Dinesh Krithivasan EECS 750 April 17, 2006 Dinesh Krithivasan (EECS 750) Hyp. testing with comm. constraints April 17, 2006 1 / 21 Presentation Outline
More informationVerifying Regularity Conditions for Logit-Normal GLMM
Verifying Regularity Conditions for Logit-Normal GLMM Yun Ju Sung Charles J. Geyer January 10, 2006 In this note we verify the conditions of the theorems in Sung and Geyer (submitted) for the Logit-Normal
More informationPoisson Image Denoising Using Best Linear Prediction: A Post-processing Framework
Poisson Image Denoising Using Best Linear Prediction: A Post-processing Framework Milad Niknejad, Mário A.T. Figueiredo Instituto de Telecomunicações, and Instituto Superior Técnico, Universidade de Lisboa,
More informationLegendre-Fenchel transforms in a nutshell
1 2 3 Legendre-Fenchel transforms in a nutshell Hugo Touchette School of Mathematical Sciences, Queen Mary, University of London, London E1 4NS, UK Started: July 11, 2005; last compiled: October 16, 2014
More informationIEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm
IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.
More informationIEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 3, MARCH
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 57, NO 3, MARCH 2011 1587 Universal and Composite Hypothesis Testing via Mismatched Divergence Jayakrishnan Unnikrishnan, Member, IEEE, Dayu Huang, Student
More informationis a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.
Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable
More informationA SIRV-CFAR Adaptive Detector Exploiting Persymmetric Clutter Covariance Structure
A SIRV-CFAR Adaptive Detector Exploiting Persymmetric Clutter Covariance Structure Guilhem Pailloux 2 Jean-Philippe Ovarlez Frédéric Pascal 3 and Philippe Forster 2 ONERA - DEMR/TSI Chemin de la Hunière
More informationcertain class of distributions, any SFQ can be expressed as a set of thresholds on the sufficient statistic. For distributions
Score-Function Quantization for Distributed Estimation Parvathinathan Venkitasubramaniam and Lang Tong School of Electrical and Computer Engineering Cornell University Ithaca, NY 4853 Email: {pv45, lt35}@cornell.edu
More informationRiccati difference equations to non linear extended Kalman filter constraints
International Journal of Scientific & Engineering Research Volume 3, Issue 12, December-2012 1 Riccati difference equations to non linear extended Kalman filter constraints Abstract Elizabeth.S 1 & Jothilakshmi.R
More informationWorst case analysis for a general class of on-line lot-sizing heuristics
Worst case analysis for a general class of on-line lot-sizing heuristics Wilco van den Heuvel a, Albert P.M. Wagelmans a a Econometric Institute and Erasmus Research Institute of Management, Erasmus University
More informationDesign of Image Adaptive Wavelets for Denoising Applications
Design of Image Adaptive Wavelets for Denoising Applications Sanjeev Pragada and Jayanthi Sivaswamy Center for Visual Information Technology International Institute of Information Technology - Hyderabad,
More informationBlind separation of instantaneous mixtures of dependent sources
Blind separation of instantaneous mixtures of dependent sources Marc Castella and Pierre Comon GET/INT, UMR-CNRS 7, 9 rue Charles Fourier, 9 Évry Cedex, France marc.castella@int-evry.fr, CNRS, I3S, UMR
More informationPhenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012
Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM
More informationIMPROVEMENTS IN MODAL PARAMETER EXTRACTION THROUGH POST-PROCESSING FREQUENCY RESPONSE FUNCTION ESTIMATES
IMPROVEMENTS IN MODAL PARAMETER EXTRACTION THROUGH POST-PROCESSING FREQUENCY RESPONSE FUNCTION ESTIMATES Bere M. Gur Prof. Christopher Niezreci Prof. Peter Avitabile Structural Dynamics and Acoustic Systems
More informationEECS 750. Hypothesis Testing with Communication Constraints
EECS 750 Hypothesis Testing with Communication Constraints Name: Dinesh Krithivasan Abstract In this report, we study a modification of the classical statistical problem of bivariate hypothesis testing.
More informationThe Minimum Message Length Principle for Inductive Inference
The Principle for Inductive Inference Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health University of Melbourne University of Helsinki, August 25,
More informationUWB Geolocation Techniques for IEEE a Personal Area Networks
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com UWB Geolocation Techniques for IEEE 802.15.4a Personal Area Networks Sinan Gezici Zafer Sahinoglu TR-2004-110 August 2004 Abstract A UWB positioning
More informationA New Method of Generating Multivariate Weibull Distributed Data
A New Method of Generating Multivariate Weibull Distributed Data Justin G Metcalf, K. James Sangston, Muralidhar Rangaswamy, Shannon D Blunt, Braham Himed Sensors Directorate, Air Force Research Laboratory
More informationNOISE can sometimes improve nonlinear signal processing
488 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 59, NO. 2, FEBRUARY 2011 Noise Benefits in Quantizer-Array Correlation Detection and Watermark Decoding Ashok Patel, Member, IEEE, and Bart Kosko, Fellow,
More informationExpectation propagation for signal detection in flat-fading channels
Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA
More informationEM Algorithm. Expectation-maximization (EM) algorithm.
EM Algorithm Outline: Expectation-maximization (EM) algorithm. Examples. Reading: A.P. Dempster, N.M. Laird, and D.B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, J. R. Stat. Soc.,
More informationThe Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models
The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health
More informationConvex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs)
ORF 523 Lecture 8 Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Any typos should be emailed to a a a@princeton.edu. 1 Outline Convexity-preserving operations Convex envelopes, cardinality
More informationGeneralized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses
Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:
More informationMMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only
MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University
More informationNonlinear Diffusion. Journal Club Presentation. Xiaowei Zhou
1 / 41 Journal Club Presentation Xiaowei Zhou Department of Electronic and Computer Engineering The Hong Kong University of Science and Technology 2009-12-11 2 / 41 Outline 1 Motivation Diffusion process
More informationVariational Principal Components
Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationEquivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 12, DECEMBER 2008 2009 Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation Yuanqing Li, Member, IEEE, Andrzej Cichocki,
More informationTranslation-Invariant Shrinkage/Thresholding of Group Sparse. Signals
Translation-Invariant Shrinkage/Thresholding of Group Sparse Signals Po-Yu Chen and Ivan W. Selesnick Polytechnic Institute of New York University, 6 Metrotech Center, Brooklyn, NY 11201, USA Email: poyupaulchen@gmail.com,
More informationMMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm
Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Bodduluri Asha, B. Leela kumari Abstract: It is well known that in a real world signals do not exist without noise, which may be negligible
More informationProbability and statistics; Rehearsal for pattern recognition
Probability and statistics; Rehearsal for pattern recognition Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských
More informationOn the Hilbert Transform of Wavelets
On the Hilbert Transform of Wavelets Kunal Narayan Chaudhury and Michael Unser Abstract A wavelet is a localized function having a prescribed number of vanishing moments. In this correspondence, we provide
More informationAn Investigation of 3D Dual-Tree Wavelet Transform for Video Coding
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding Beibei Wang, Yao Wang, Ivan Selesnick and Anthony Vetro TR2004-132 December
More information