IEEE Transactions on Signal Processing, accepted October 2005, in press

Size: px
Start display at page:

Download "IEEE Transactions on Signal Processing, accepted October 2005, in press"

Transcription

1 IEEE Personal use of this material is permitted However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE IEEE Transactions on Signal Processing, accepted October 5, in press

2 The aussian Transform of distributions: definition, computation and application Teodor Iulian Alecu*, Sviatoslav Voloshynovskiy and Thierry Pun Abstract This paper introduces the general purpose aussian Transform of distributions, which aims at representing a generic symmetric distribution as an infinite mixture of aussian distributions We start by the mathematical formulation of the problem and continue with the investigation of the conditions of existence of such a transform Our analysis leads to the derivation of analytical and numerical tools for the computation of the aussian Transform, mainly based on the Laplace and Fourier transforms, as well as of the afferent properties set (eg the transform of sums of independent variables) The aussian Transform of distributions is then analytically derived for the aussian and Laplacian distributions, and obtained numerically for the eneralized aussian and the eneralized Cauchy distribution families In order to illustrate the usage of the proposed transform we further show how an infinite mixture of aussians model can be used to estimate/denoise non-aussian data with linear estimators based on the Wiener filter The decomposition of the data into aussian components is straightforwardly computed with the aussian Transform, previously derived The estimation is then based on a two-step procedure, the first step consisting in variance estimation, and the second step in data estimation through Wiener filtering To this purpose we propose new generic variance estimators based on the Infinite Mixture of aussians prior It is shown that the proposed estimators compare favorably in terms of distortion with the shrinkage denoising technique, and that the distortion lower bound under this framework is lower than the classical MMSE bound Index Terms aussian mixture, aussian Transform of distributions, eneralized aussian, denoising, shrinkage I INTRODUCTION aussian distributions are extensively used in the (broad sense) signal processing community, mainly for computational benefits For instance, in an estimation problem aussian priors yield quadratic functionals and linear solutions In ratedistortion and coding theories, closed form results are mostly available for aussian source and channel descriptions [] However, real data is most often non aussian distributed, This work is supported in part by the Swiss NCCR IM (Interactive Multimodal Information Management, the EU Network of Excellence Similar ( and the SNF grant N PP-68653/ T I Alecu, S Voloshynovskiy, T Pun are with the Computer Vision and Multimedia Laboratory, University of eneva, 4 Rue énéral-dufour, 4 eneva, Switzerland (corresponding author: TeodorAlecu@cuiunigech, phone: +4()37984; fax: +4()379778) and is best described by other types of distribution (eg in image processing, the most commonly used model for the wavelet coefficients distribution is the eneralized aussian distribution [6]) The goal of the work presented in this paper is to describe non-aussian distributions as an infinite mixture of aussian distributions, through direct computation of the mixing functions from the non-aussian distribution In a related work [4], it was proven that any distribution can be approximated through a mixture of aussian up to an arbitrary level of precision However, no hint was given by the author on how to obtain the desired mixture in the general case In the radar community the problem of modeling non- aussian radar clutters led to the theory of Spherically Invariant Random Processes (SIRP) [3], which aimed at generating non-aussian multivariate distributions using univariate aussian distributions However, the proposed solutions bypassed the mixing function ([] and references therein), performing direct computation of the multivariate non-aussian distribution from the marginal probability distribution function and the desired covariance matrix In [5], an analytical formula is given for an infinite mixture of aussians equivalent to the Laplacian distribution, and used in a source coding application Unfortunately, no generalization was attempted by the authors The work presented here has the roots in their proof and extends the concept through the introduced aussian Transform of distributions, which permits straightforward derivation of exact mixing functions for a wide range of symmetric distributions, including the eneralized Cauchy and the eneralized aussian distributions In order to exemplify the possible usage of the transform we consider then the problem of estimation/denoising, motivated by works such as [9] and [], where, in order to preserve the computational advantages provided by aussian modelling, non-aussian distributions are approximated with finite aussian mixtures obtained through iterative numerical optimization techniques Relying on the aussian Transform, we propose new generic algorithms for denoising non- aussian data, based on the description of non-aussian distributions as an Infinite Mixture of aussians (IM) The rest of the paper is divided in three main sections In Section II we define the aussian Transform, analyze its existence, investigate its properties and derive the mathematical tools for analytical and/or numerical computation In Section III we exemplify both the transform for some typical distributions such as eneralized aussian and eneralized Cauchy, and some of the properties deduced in Section II The last section tackles the problem of

3 estimation/denoising and proposes new aussian Transform based denoising schemes, exemplifying them for the eneralized aussian Distribution (D) family The obtained results are compared in terms of distortion with the state-of-the-art denoising method of Moulin and Liu [7] II THE AUSSIAN TRANSFORM OF DISTRIBUTIONS A Definition and existence We consider a generic symmetric continuous distribution p(x) As we are aiming at representing it through an infinite mixture of aussians, we can safely disregard the mean, and assume for simplicity reasons that p(x) is zero-mean We are looking for an integral representation in the form: ( ) N ( x ) d = p( x), () where N ( x ) is the zero-mean aussian distribution: x N ( x ) = e, π and ( ) is the mixing function that should reproduce the original p(x) We can now introduce the aussian Transform Definition : aussian Transform The direct aussian Transform is defined as the operator which transforms p( x ) into ( ), and the Inverse aussian Transform - is defined as the operator which maps ( ) back to p( x ) : : p( x) ( ); : ( ) p( x) Obviously, - is simply given by (): ( ( )) = ( ) ( x ) d = p( x) N () The aussian Transform of a distribution exists if, given p( x ), a mixing distribution ( ) can be found such as to comply with () This can be summarized in three conditions Condition For a given p(x), a function ( ) defined according to () exists Condition This function is non-negative Condition 3 Its integral ( ) d is equal to The last condition is a consequence of Condition Indeed, if ( ) exists, then integrating both sides of () with respect to x and inverting the integration order one obtains: ( ) N ( x ) d dx = p( x) dx, x dx d = N Finally, since ( x ) N is a distribution: ( ) d = In order to investigate Condition, the existence of ( ), we perform the following variable substitutions: s = x and t = Since p(x) is symmetric, it can be rewritten as: p( x) = p( x ) = p( s) The left hand side of () transforms to: ( ) ( ) st N x d = e dt tt π t According to the definition of the Laplace Transform L [], equation () finally takes the form: L = p( s) (3) tt π t Thus, ( ) is linked to the original probability distribution p(x) through the Laplace Transform and can be computed using the Inverse Laplace Transform L - The direct aussian Transform is therefore given by: = π ( p( x) ) = ( p( s) )() t L, (4) t and the Inverse aussian Transform can be computed as: ( ( )) = L ( s) (5) tt π t s= x Consequently, the existence of the aussian Transform is conditioned by the existence of the Inverse Laplace Transform of p( s ) Using the general properties of the Laplace Transform, it is sufficient [] to prove that the limit at infinity of s p( s) is bounded, or equivalently : lim x x p x < (6) The above condition is satisfied by all the distributions from the exponential family, as well as by all the distributions with finite variance The equation (4) allows for straightforward identification of aussian Transforms for distributions whose Laplace Transforms are known, by simply using handbook tables Unfortunately, the condition (6) does not guarantee compliance with the Condition : non-negativity As it is rather difficult to verify a priori this constraint, the test should be performed a posteriori, either analytically or numerically However, by imposing the non-negativity constraint in (), one can see that it is necessary to have a strictly decreasing distribution function p(x) on positive x, as p(x) is described by a sum of strictly decreasing zero-mean aussians functions From the same reason, but not related to non-negativity, the distribution function p(x) should also be of class C B Properties of the aussian Transform Property : Uniqueness If two functions and

4 ( ) exist such that ( ( )) = p( x) ( ( )) = p( x) ( ) ( ) and then they are identical = Conversely, if two distribution p x have the same aussian probabilities p x and Transform ( ) then they are identical Proof The second proposition of the uniqueness property is proved straightforwardly using () In what concerns the first proposition, it is a direct consequence of the uniqueness of the Laplace Transform [], which states that if two functions have the same Laplace transform, then they are identical It is then sufficient to observe that is computed as a Laplace Transform (5) Property : Mean Value (variance conservation) The mean value of the aussian Transform is equal to the variance of the original distribution x p( x) dx = ( ) d (7) x p x dx = x x d dx = Proof N ( ) N = x x dx d = d qed Property 3: Value at infinity If lim x x p( x) < then the aussian Transform tends to when tends to infinity: lim ( ) = (8) Proof We use the initial value theorem for the Laplace Transform [] and the direct formula (3): limt = lims s p s = lim x x p x, tt π t then lim < lim = 3 ( )( ) ( ) We study now the influence of basic operations on the aussian Transform, such as scalar multiplication and addition of independent variables Property 4: Scaling If and are random variables such that = α and ( p ( x) ) =, then the aussian Transform corresponding to the scaled variable is a scaled version of : = (9) α α Proof We begin by observing that the distribution probability of the random variable can be expressed as: x p ( x) = p α α Then we compute the Inverse aussian Transform of the expression proposed in (9) using (): ( ) = N x d ; α α x α * α ( ( )) e d α α α = π α x ( ( )) = p p = x α α Finally using the uniqueness property we conclude that the aussian Transform ( p ( x) ) is unique and given by (9) Property 5: Convolution If and are two independent p x = and random variables with ( p ( x) ) = ( ) ( ), then the aussian Transform of their sum is the convolution of their respective aussian Transforms (the result can be generalized for the sum of multiple variables): p + ( x) = () Proof Consider the random variable = + Since exists, is a random aussian variable with variance distributed according to the distribution probability Similarly, is a random aussian variable with variance distributed according to Then is also a random aussian variable with variance = + But and are independent variables drawn from and It follows that is a random variable described by the probability distribution =, qed Corollary If and are independent random variables and is aussian distributed p ( x) = N ( x ), then the aussian transform of their sum is a shifted version of :, < + = () ( ), C Numerical computation The computation of the aussian Transform for distributions not available in handbooks is still possible through the complex inversion method for Laplace Transforms known as the Bromwich integral [3]: ( ) ε + i ut L f s = f ( u) e du π i, () ε i where ε is a real positive constant satisfying ε > sup( Re( poles ( f ))) and u is an auxiliary variable The equation () can be rearranged as a Fourier Transform, allowing the use of the numerous numerical and/or symbolical packages available ( ω is the variable in the Fourier space): εt L ( f ( s) ) = e F ( f ( ε + iω )) (3) Very often p(x) has no poles, being a continuous and bounded

5 function, and in this case it might be very practical to evaluate (3) in the limit case ε Using (4) and (3): π ( ) = F ( p( iω) )() t (4) t = When the original distribution is only numerically known, approximation of ( ) is still possible either through analytical approximations of p(x), followed by (4) or (4), or through solving the inverse problem yielded by (), or through real inversion of the Laplace transform using (4) The accuracy of the obtained transforms can then be assessed using an appropriate metric or distance such as Kullback- Leibler However, the abovementioned approximation methods and the accuracy study constitute stand-alone problems out of the scope of the present paper C b p x b = p s b = π + π + b C ( ) ; ( ) C ( ) b x b s b = (7) p x b e π The results (5), (6) and (7) are plotted in Fig To exemplify the Corollary to the Convolution Property, consider now Cauchy data contaminated with independent additive white aussian (AW) noise Then the aussian Transform of the measured data (Fig ) is a shifted version of the original data transform () III EAMPLES OF AUSSIAN TRANSFORMS A Analytic aussian transforms The most obvious and natural example is the aussian Transform of a aussian distribution N ( x ) One would expect to have (with δ the Dirac function): ( N ( x )) = δ ( ) (5) Proof It is trivial that (5) verifies the equation () Then, by using the uniqueness of the aussian Transform one can conclude that the aussian Transform of a aussian distribution N ( x ) is a Dirac function centered in The convolution property () can now be used to prove a well known result in statistics: the sum of two independent aussian variables, with respective probability laws x N x, is another aussian variable with N ( ) and ( ) probability distribution ( x + ) N (the extension to non-zero mean distributions is trivial) Proof: ( p ( x) ) = δ ( ); ( ) p x = δ p x = p x * p x = δ + ( + ) ( ) Then, inverting the aussian Transform: p ( ( )) + = δ + = N x + Similarly to (5), it is possible to compute the aussian Transforms of other usual symmetric distributions using the Laplace transform tables [] We exemplify with the Laplacian and Cauchy distributions Laplacian distribution: L λ λx L λ λ s p ( x λ) = e ; p ( s λ) = e (6) λ L λ ( p ( s )) λ = e As mentioned, the result (6) was already proven in [5] Cauchy distribution: Figure aussian Transforms of the Laplacian, aussian and Cauchy distributions Figure Shift of the aussian Transform following data contamination with AW noise The previous analytical results will be generalized in subsection B through numerical computations B Numerical computation of aussian Transforms This part illustrates the numerical computation of the aussian Transform through (4) for the eneralized aussian and eneralized Cauchy distributions families ) eneralized aussian Distribution (D) The D family is described by an exponential probability density function with parameters and : η ( ) x p ( x, ) = e η, (8) Γ ( ) where η( ) 3 ( ) ( ) = Γ Γ For = the D particularizes to the Laplacian distribution, while for = one obtains the aussian distribution The aussian Transform of the eneralized aussian distribution can not be obtained in analytical form using (4) However, it does exist (6) and can be calculated numerically through (4): ( ) π ( ) = F ( ω), p i t, t = η ( ) η ( ) ( ω) ( ) ( ω) p i = e i, Γ

6 The aussian Transforms for ranging from 5 to with fixed = are plotted in Fig 3 The transforms evolve from a Dirac-like distribution centered on for small to exponential for =, then Rayleigh-like for =, bellshaped for = 5 and again Dirac-like centered at for = As expected, the aussian Transform of the Laplacian distribution ( = ) is exponential (6) Figure 3 aussian Transforms of D Unfortunately, the real part of the complex probability function diverges for periodical values of, which impedes the computation of the transform through this method for > However, real data (in transform domains such as discrete cosine and wavelet) is mostly confined to << [6] ) eneralized Cauchy Distribution (CD) The CD probability density function is given by Γ ( ν + ) ν C b px x ν, b =, ν + 5 πγ ( ν) ( b + x ) and it particularizes to the Cauchy distribution for ν = 5 Its aussian Transform can be computed through: ν b Γ ( ν + 5) ν 5 ν, b( ) = ( F (( b + iω) ) ) Γ( ν) t = Corresponding plots are given in Fig 4 Figure 4 aussian Transforms of CD As a remark, the variance of the Cauchy distribution being infinite, its aussian Transform has infinite mean value (7) The CD family possesses finite variance only for ν > IV DENOISIN We are interested in the generic denoising problem of estimating the original scalar data x from the measured data y, degraded by additive noise z: y = x+ z (9) We assume that the noise random variable is zero-mean aussian distributed with variance : z p ( z) = e () π We further assume that the probability density function p (x), which describes the source random variable, is a zeromean symmetric distribution, and that its aussian Transform ( ) = ( p ( x) ) exists For exemplification of the proposed techniques we consider p (x) to be the eneralized aussian distribution (8) We base our estimation on the Maximum a Posteriori (MAP) principle [4]: xˆ = arg max x p Y ( x y), which yields for iid zero-mean aussian data and noise (with variances and ) the classical Wiener filter: xˆ = y () + We also note that the estimate () is also the Minimum Mean Square Error (MMSE) estimate for iid zero-mean aussian data and noise: xˆmmse = arg min ( ˆ x xx) p Y ( x y) dx x () xˆ = xp x y dx MMSE Y x According to the IM model, each data sample is interpreted to be drawn from a aussian distribution with variance distributed accordingly to the aussian Consequently, using the prior Transform knowledge of ( ), we can estimate the variance from y and use Wiener filtering () Thus, our estimator is given by: ( y) xˆ = y (3) y + The initial problem is now reduced to a variance estimation problem For this purpose, we begin by introducing the cumulative estimation technique A Cumulative Estimation We are looking for an estimator that provides variance estimates consistent with the original data distribution, in the sense that the description of the data as a aussian mixture should be identical to the non-aussian original description We set the above assertion as a consistency principle

7 Principle : Consistency The probability distribution function of the variance estimate should be identical to the aussian Transform of the prior distribution p (x) Moreover, since the distributions p (x), p (z) are symmetric and zero-mean, we assume an estimator of the form: ˆ ( y) = ξ M ( M ( y) ), where ξ M denotes the consistent variance estimator depending on the symmetric real non-negative function: + M: ; M( y) = M( y), + where and denote real, respectively real nonnegative numbers As an example, a possible M function is the absolute value of y We further assume that this estimator is monotonically increasing with the function M: y < y ξ M y M M M M < ξ M y (4) ( ) ( ) If one chooses the M function as the absolute value y, this simply means that if the absolute value of y is larger, then the estimated variance ˆ ( y) is also larger If the variables are vectors (eg local estimation paradigm) one can similarly use as the M function the variance of the measurement vector, and (4) would mean that if the variance of the measurement is larger, then the estimated variance is also larger The assumption in (4) implies that the cumulative distribution function of the variance estimates is equal to the cumulative distribution function of M: p( ˆ ˆ ( y) ) = p( M M ( y) ) (5) Considering continuous probability density functions and incorporating the consistency condition (the distribution of the estimates should be identical to ), (5) can be rewritten as: M PM ( M( y) ) = P ( ξ ( M( y) )), M( y) x (6) P M y = p u du; P = u du, M ( ) M where p M is the probability density function describing M ( y), and P M and P are cumulative probability functions, as defined in (6) Then the estimator we are looking for is simply given by: M ( M( y) ) = P PM ( M ( y) ), (7) ξ where P is the mathematical inverse of the function P : ( if P ) then = u P u = We denote (7) as the cumulative estimator, simply because it uses cumulative probability functions for the estimation We further infer that the cumulative estimator is robust with respect to monotonically increasing transformations of M Theorem If M and M are two symmetric real nonnegative functions M, M + :, and if M( y) < M( y) M( y) < M ( y), then the associated cumulative estimators are identical: M ( ( y) ) = ξ ( ( y) ) M ξ M M Proof Since the inequalities are preserved (theorem hypothesis), then the cumulative distribution functions P M and P M are equal: p( M M( y) ) = p( M M( y) ) then M ( y) M ( y) P M y = p u du = p u du = P M y ( ) M M M M ( ) The proof is ended by corroborating the above equality with the cumulative estimation formula (7) B IM cumulative estimator The cumulative estimator (7) can be directly used in (3) by setting the M function as M ( y) = y, and the IM cumulative estimator (IM CE) is simply: ˆ _ CE ( y) xˆ IM _ CE = y with ˆ _ CE ( y) + (8) y ˆ _ CE ( y) = P ( PM( y) ), PM( y) = py ( u) du As and are assumed to be independent, p Y (y) can be computed as the convolution of p (x) and p (z): py = p p In the general case the operations in (8) have to be performed numerically However, if the distribution p (x) is Laplacian (D with = ), its aussian Transform is (6), (8): ( ), = e, (9) and the cumulative estimator from (8) reduces to: _ CE, = ( y) = ln ( P M ( y) ) (3) As tends to, the aussian Transform of the D tends asymptotically to a Dirac function centered at (5), and the estimator (8) to the classical Wiener filter () In order to asses the performances of the proposed estimators we measure the accuracy of the estimate according to the L norm: e( x, xˆ) = ( x xˆ) The global distortion induced by an estimator is therefore: D = e( x, xˆ ) p ( x) p ( z) dxdz (3) We compare our estimator in terms of distortion with the estimators known as shrinkage functions from [7] The idea of shrinking originates from [8], but [7] provides both improved results and statistical interpretation of the functions In fact, the shrinkage functions from [7] are the result of direct MAP estimation, which in the case of D data and aussian noise yields ˆx as the solution of the equation in x: y = x η( ) x (3) Equation (3) has analytical solutions for = 5 (cubic), = or (linear) and = 5 (quadratic) The resulting shrinkage functions have thresholding characteristics for

8 and yield the Wiener filter for = The equivalent IM CE functions do not present thresholding characteristics for any, behaving in continuous manner raphic illustration of the shrinkage functions and of the equivalent IM CE functions obtained via (8) are presented in Fig 5 of separable IM components In fact, for a given y, the MMSE estimator is unique (injective estimation), which is not the case for the ideal estimator that would lead to the proposed IM lower bound, and that would yield, for the same value of y, different estimates for each isolated aussian component Figure 5 Shrinkage functions and IM CE estimation curves One can also compute the lower bound on the distortion induced by an estimation procedure based on signal decomposition into an IM Assuming that the aussian components of the IM are separable, and that the variance estimate is constant over each separate component, the equation () is the MMSE estimate for each component The bound is attained when the estimator "knows" the aussian component of the mixture that has generated the estimated sample The ensuing distortion lower bound is: Dlower bound = ( ) d, (33) + where the ratio + is the theoretical distortion of the MMSE filter () We also compare our results with the MMSE distortion bound, given by the MMSE estimator (), which can be rewritten as follows: xˆ MMSE = xp( x y) dx = x p( x y, ) p( y) d dx x x Using Bayes rule and switching integrals, one obtains: p( y ) p( ) xˆ MMSE = xp ( x y, ) dx d x p( y) Considering aussian noise the first parenthesis yields the MMSE estimate (), and by integrating the aussian p y one finds: Transform and developing xˆ p y and N( + ) ( y + ) ( ) d y d + MMSE = N y (34) The MMSE distortion bound is then computed by plugging the result (34) into (3) The IM lower bound (33) and the MMSE bound (34) are superposed in Fig 6, which displays the distortion of the estimation (3) as a function of the signal to noise ratio SNR = log (the variance of the D family is given by, see the moments method from [6] and equation (8)) One may notice that the IM lower bound is lower than the MMSE bound, a result due to the assumption Figure 6 Distortion for D data as a function of SNR (db) A considerable difference in performance in the favor of IM CE with respect to the shrinkage functions is observed for = 5 The difference tends to decrease for larger, and inverts for = 5, albeit with a very small relative module However, the IM CE estimator being unique for a given y, the IM CE distortion curve necessarily remains above the MMSE bound, as for any other injective estimator All curves tend to the distortion of the Wiener filter when approaches, plotted dashed-dotted in Fig on the graph corresponding to = 5 C IM ML and MAP variance estimation The simplest way of obtaining the variances in a statistical framework is through Maximum Likelihood (ML) estimation: ˆ _ ML = arg max p( y ), (35) which for zero-mean independent aussian source and noise data yields the solution: ˆ _ ML = max ( y,) (36) If the variance distribution were known, one could improve (36) through MAP estimation Using the aussian Transform the MAP variance estimate is: ˆ _ MAP = arg max ( ) p( y ) (37)

9 enerally, the expression (37) requires numerical maximization, but an analytical form exists for = and aussian noise (using (9) and differentiation of (37)): y ˆ _ MAP, = = max, (38) 8y + + One would expect the IM MAP estimator to behave similarly to the original shrinkage functions of Moulin and Liu, as it applies the MAP principle two times (37), (3) to obtain an indirect estimate, whereas the shrinkage functions are the result of direct MAP estimation (3) This is confirmed in Fig 7, which compares the equivalent shrinkage functions of the IM ML and IM MAP estimators with the IM CE curve and with the Moulin and Liu curve for = Figure 7 IM ML and MAP equivalent shrinkage functions Unlike the IM CE estimator, the two proposed IM ML and MAP estimators do present thresholding characteristics, and, as expected, the equivalent IM MAP curve behaves very similarly to the original shrinkage curve We tested empirically the MAP (38) estimator by generating D eneralized aussian data of length N= 6 with = 5 and =, adding independent white aussian noise, and performing denoising according to () The source and noise probability densities were assumed to be known, and the corresponding distortion was computed as the mean square error (Fig 8) We also tested the cumulative estimator (8) and the shrinkage function (3) Figure 8 Empirical distortion curves As expected, empirical results reproduce the theoretical results obtained for the scalar estimators (Fig 6) The IM MAP results are close, but above the direct MAP estimation of the shrinkage functions for = (this result was predictable from the curves in Fig7), however they notably improve on the Moulin and Liu results for = 5 We expect the IM MAP curves to remain above the Moulin and Liu curves for, both falling slowly down to the aussian-aussian curve in Fig6 as the IM MAP estimator tends to the Wiener filter for = While the IM MAP estimator does not outperform the cumulative estimator, nor the shrinkage estimator for, and while the IM CE estimator does not outperform the classical MMSE estimate, the linear nature of the estimator (3), coupled with the simple analytical forms (38) and (3) for =, allowed the successful use of the IM MAP and IM CE estimators in the highly underdetermined EE (electroencephalogram) inverse problem [5] with Laplacian prior, where a direct application of the MAP principle would lead to an underdetermined quadratic programming problem, and where a direct application of the MMSE principle would lead to highly expensive computations V CONCLUSION AND FUTURE WORK We introduced in this paper the aussian Transform, which allows for the representation of symmetric distributions as an infinite aussian mixture The scope of applicability of the aussian Transform is potentially very broad, from denoising and regularization to filtering, coding, compression, watermarking etc However, an extension of the concept to non-symmetric distribution would be required for some specific applications Further investigation of the existence conditions, especially non-negativity, is also necessary Finally, one would need adapted numerical packages (most likely based on existing Laplace and Fourier transform computational packages) for the computation of aussian Transforms of both analytically and numerically defined distributions We also presented in this paper various denoising algorithms based on the Infinite Mixture of aussians model Their application to the denoising of eneralized aussian data showed significant improvements in terms of distortion, both theoretically and empirically, with respect to the shrinkage functions of Moulin&Liu Also, the lower bound under the proposed IM estimation framework is lower than the classical MMSE bound However the potential of IM based denoising is not yet fully explored, as the lower bound presented in this paper is still below current results The estimation methods proposed may reach this lower bound in the case where the data is generated locally from the same aussian component, which would allow for exact variance estimation ACKNOWLEDMENT The authors thank Dr Oleksiy Koval for careful reading

10 and fruitful discussions The authors also thank anonymous reviewers for valuable comments and suggestions REFERENCES A D Poularikas, S Seely, Laplace Transforms The Transforms and Applications Handbook: Second Edition Ed Alexander D Poularikas Boca Raton: CRC Press LLC, Arfken, Mathematical Methods for Physicists, 3rd ed Orlando, FL: Academic Press, pp , Arfken, "Inverse Laplace Transformation" 5 in Mathematical Methods for Physicists, 3rd ed Orlando, FL: Academic Press, pp , R Wilson, MMM: Multiresolution aussian Mixture Models for Computer Vision, in IEEE International Conference on Pattern Recognition, Barcelona, Spain, September 5 A Hjørungnes, J M Lervik, and T A Ramstad, Entropy Coding of Composite Sources Modeled by Infinite Mixture aussian Distributions, in Proc 996 IEEE Digital Signal Processing Workshop, Loen, Norway, September S Mallat, A Theory for Multiresolution Signal Decomposition: The Wavelet Representation, in IEEE Transactions on Pattern Analysis and Machine Intelligence Volume, Issue 7 (July 989) p P Moulin and J Liu, "Analysis of Multiresolution Image Denoising Schemes Using eneralized - aussian and Complexity Priors", IEEE Trans Info Theory, Special Issue on Multiscale Analysis, Vol 45, No 3, pp 99-99, Apr D L Donoho and I M Johnstone, "Ideal Spatial Adaptation Via Wavelet Shrinkage", Biometrika, Vol 8, pp , A Bijaoui, "Wavelets, aussian mixtures and Wiener filtering'', Signal Processing, Vol8, No4, April, pp 79-7 Benaroya, L; Bimbot, F; "Wiener-based source separation with HMM/MM using a single sensor'', Proc 4th Int Symp on Independent Component Anal and Blind Signal Separation, April 3, Nara, Japan T M Cover, J A Thomas, "Elements of Information Theory", John Wiley & Sons Inc, 99 M Rangaswamy, D Weiner, A Ozturk, "Non-aussian Random Vector Identification Using Spherically Invariant Random Processes", in IEEE Transactions on Aerospace and Electronic Systems, vol 9, No, January E Conte and M Longo, "Characterization of radar clutter as a spherically invariant random process", in IEEE Proceedings, Pt F (Communication, Radar and Signal Processing), 34, (987), MBertero, PBoccacci, "Introduction to Inverse Problems in Imaging", IOP Publishing, Bristol, T I Alecu, P Missionnier, S Voloshynovskiy, P iannakopoulos and T Pun, "Soft/hard focalization in the EE inverse problem", in IEEE Workshop on Statistical Signal Processing (SSP'5), Bordeaux, France, July, 7-5

The Gaussian transform. ALECU, Teodor, VOLOSHYNOVSKYY, Svyatoslav, PUN, Thierry. Abstract

The Gaussian transform. ALECU, Teodor, VOLOSHYNOVSKYY, Svyatoslav, PUN, Thierry. Abstract Proceedings Chapter The aussian transform ALECU, Teodor, VOLOSHYNOVSKYY, Svyatoslav, PUN, Thierry Abstract This paper introduces the general purpose aussian Transform, which aims at representing a generic

More information

Bayesian Paradigm. Maximum A Posteriori Estimation

Bayesian Paradigm. Maximum A Posteriori Estimation Bayesian Paradigm Maximum A Posteriori Estimation Simple acquisition model noise + degradation Constraint minimization or Equivalent formulation Constraint minimization Lagrangian (unconstraint minimization)

More information

A GLRT FOR RADAR DETECTION IN THE PRESENCE OF COMPOUND-GAUSSIAN CLUTTER AND ADDITIVE WHITE GAUSSIAN NOISE. James H. Michels. Bin Liu, Biao Chen

A GLRT FOR RADAR DETECTION IN THE PRESENCE OF COMPOUND-GAUSSIAN CLUTTER AND ADDITIVE WHITE GAUSSIAN NOISE. James H. Michels. Bin Liu, Biao Chen A GLRT FOR RADAR DETECTION IN THE PRESENCE OF COMPOUND-GAUSSIAN CLUTTER AND ADDITIVE WHITE GAUSSIAN NOISE Bin Liu, Biao Chen Syracuse University Dept of EECS, Syracuse, NY 3244 email : biliu{bichen}@ecs.syr.edu

More information

Low-Complexity Image Denoising via Analytical Form of Generalized Gaussian Random Vectors in AWGN

Low-Complexity Image Denoising via Analytical Form of Generalized Gaussian Random Vectors in AWGN Low-Complexity Image Denoising via Analytical Form of Generalized Gaussian Random Vectors in AWGN PICHID KITTISUWAN Rajamangala University of Technology (Ratanakosin), Department of Telecommunication Engineering,

More information

Wavelet Based Image Restoration Using Cross-Band Operators

Wavelet Based Image Restoration Using Cross-Band Operators 1 Wavelet Based Image Restoration Using Cross-Band Operators Erez Cohen Electrical Engineering Department Technion - Israel Institute of Technology Supervised by Prof. Israel Cohen 2 Layout Introduction

More information

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER Zhen Zhen 1, Jun Young Lee 2, and Abdus Saboor 3 1 Mingde College, Guizhou University, China zhenz2000@21cn.com 2 Department

More information

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING Yichuan Hu (), Javier Garcia-Frias () () Dept. of Elec. and Comp. Engineering University of Delaware Newark, DE

More information

Independent Component Analysis. Contents

Independent Component Analysis. Contents Contents Preface xvii 1 Introduction 1 1.1 Linear representation of multivariate data 1 1.1.1 The general statistical setting 1 1.1.2 Dimension reduction methods 2 1.1.3 Independence as a guiding principle

More information

Machine Learning Lecture Notes

Machine Learning Lecture Notes Machine Learning Lecture Notes Predrag Radivojac January 25, 205 Basic Principles of Parameter Estimation In probabilistic modeling, we are typically presented with a set of observations and the objective

More information

Lecture Note 1: Probability Theory and Statistics

Lecture Note 1: Probability Theory and Statistics Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would

More information

Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels

Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels Jilei Hou, Paul H. Siegel and Laurence B. Milstein Department of Electrical and Computer Engineering

More information

Wavelet Footprints: Theory, Algorithms, and Applications

Wavelet Footprints: Theory, Algorithms, and Applications 1306 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 51, NO. 5, MAY 2003 Wavelet Footprints: Theory, Algorithms, and Applications Pier Luigi Dragotti, Member, IEEE, and Martin Vetterli, Fellow, IEEE Abstract

More information

BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS

BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS M. Kawamoto 1,2, Y. Inouye 1, A. Mansour 2, and R.-W. Liu 3 1. Department of Electronic and Control Systems Engineering,

More information

Underdetermined Instantaneous Audio Source Separation via Local Gaussian Modeling

Underdetermined Instantaneous Audio Source Separation via Local Gaussian Modeling Underdetermined Instantaneous Audio Source Separation via Local Gaussian Modeling Emmanuel Vincent, Simon Arberet, and Rémi Gribonval METISS Group, IRISA-INRIA Campus de Beaulieu, 35042 Rennes Cedex, France

More information

Lecture Note 2: Estimation and Information Theory

Lecture Note 2: Estimation and Information Theory Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 2: Estimation and Information Theory Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 2.1 Estimation A static estimation problem

More information

On the Cost of Worst-Case Coding Length Constraints

On the Cost of Worst-Case Coding Length Constraints On the Cost of Worst-Case Coding Length Constraints Dror Baron and Andrew C. Singer Abstract We investigate the redundancy that arises from adding a worst-case length-constraint to uniquely decodable fixed

More information

S. R. Tate. Stable Computation of the Complex Roots of Unity, IEEE Transactions on Signal Processing, Vol. 43, No. 7, 1995, pp

S. R. Tate. Stable Computation of the Complex Roots of Unity, IEEE Transactions on Signal Processing, Vol. 43, No. 7, 1995, pp Stable Computation of the Complex Roots of Unity By: Stephen R. Tate S. R. Tate. Stable Computation of the Complex Roots of Unity, IEEE Transactions on Signal Processing, Vol. 43, No. 7, 1995, pp. 1709

More information

Modeling Multiscale Differential Pixel Statistics

Modeling Multiscale Differential Pixel Statistics Modeling Multiscale Differential Pixel Statistics David Odom a and Peyman Milanfar a a Electrical Engineering Department, University of California, Santa Cruz CA. 95064 USA ABSTRACT The statistics of natural

More information

Latent Tree Approximation in Linear Model

Latent Tree Approximation in Linear Model Latent Tree Approximation in Linear Model Navid Tafaghodi Khajavi Dept. of Electrical Engineering, University of Hawaii, Honolulu, HI 96822 Email: navidt@hawaii.edu ariv:1710.01838v1 [cs.it] 5 Oct 2017

More information

Logarithmic quantisation of wavelet coefficients for improved texture classification performance

Logarithmic quantisation of wavelet coefficients for improved texture classification performance Logarithmic quantisation of wavelet coefficients for improved texture classification performance Author Busch, Andrew, W. Boles, Wageeh, Sridharan, Sridha Published 2004 Conference Title 2004 IEEE International

More information

The Fading Number of a Multiple-Access Rician Fading Channel

The Fading Number of a Multiple-Access Rician Fading Channel The Fading Number of a Multiple-Access Rician Fading Channel Intermediate Report of NSC Project Capacity Analysis of Various Multiple-Antenna Multiple-Users Communication Channels with Joint Estimation

More information

Lecture 5 Channel Coding over Continuous Channels

Lecture 5 Channel Coding over Continuous Channels Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From

More information

Estimation Error Bounds for Frame Denoising

Estimation Error Bounds for Frame Denoising Estimation Error Bounds for Frame Denoising Alyson K. Fletcher and Kannan Ramchandran {alyson,kannanr}@eecs.berkeley.edu Berkeley Audio-Visual Signal Processing and Communication Systems group Department

More information

A Strict Stability Limit for Adaptive Gradient Type Algorithms

A Strict Stability Limit for Adaptive Gradient Type Algorithms c 009 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional A Strict Stability Limit for Adaptive Gradient Type Algorithms

More information

CS 195-5: Machine Learning Problem Set 1

CS 195-5: Machine Learning Problem Set 1 CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of

More information

NUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS

NUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS NUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS Justin Dauwels Dept. of Information Technology and Electrical Engineering ETH, CH-8092 Zürich, Switzerland dauwels@isi.ee.ethz.ch

More information

Information Theory in Intelligent Decision Making

Information Theory in Intelligent Decision Making Information Theory in Intelligent Decision Making Adaptive Systems and Algorithms Research Groups School of Computer Science University of Hertfordshire, United Kingdom June 7, 2015 Information Theory

More information

Legendre-Fenchel transforms in a nutshell

Legendre-Fenchel transforms in a nutshell 1 2 3 Legendre-Fenchel transforms in a nutshell Hugo Touchette School of Mathematical Sciences, Queen Mary, University of London, London E1 4NS, UK Started: July 11, 2005; last compiled: August 14, 2007

More information

Satellite image deconvolution using complex wavelet packets

Satellite image deconvolution using complex wavelet packets Satellite image deconvolution using complex wavelet packets André Jalobeanu, Laure Blanc-Féraud, Josiane Zerubia ARIANA research group INRIA Sophia Antipolis, France CNRS / INRIA / UNSA www.inria.fr/ariana

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and Athens Journal of Sciences December 2014 Discriminant Analysis with High Dimensional von Mises - Fisher Distributions By Mario Romanazzi This paper extends previous work in discriminant analysis with von

More information

Blind Spectral-GMM Estimation for Underdetermined Instantaneous Audio Source Separation

Blind Spectral-GMM Estimation for Underdetermined Instantaneous Audio Source Separation Blind Spectral-GMM Estimation for Underdetermined Instantaneous Audio Source Separation Simon Arberet 1, Alexey Ozerov 2, Rémi Gribonval 1, and Frédéric Bimbot 1 1 METISS Group, IRISA-INRIA Campus de Beaulieu,

More information

Robust Inference. A central concern in robust statistics is how a functional of a CDF behaves as the distribution is perturbed.

Robust Inference. A central concern in robust statistics is how a functional of a CDF behaves as the distribution is perturbed. Robust Inference Although the statistical functions we have considered have intuitive interpretations, the question remains as to what are the most useful distributional measures by which to describe a

More information

01 Probability Theory and Statistics Review

01 Probability Theory and Statistics Review NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement

More information

Unsupervised Learning with Permuted Data

Unsupervised Learning with Permuted Data Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University

More information

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases 2558 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 48, NO 9, SEPTEMBER 2002 A Generalized Uncertainty Principle Sparse Representation in Pairs of Bases Michael Elad Alfred M Bruckstein Abstract An elementary

More information

Impact of channel-state information on coded transmission over fading channels with diversity reception

Impact of channel-state information on coded transmission over fading channels with diversity reception Impact of channel-state information on coded transmission over fading channels with diversity reception Giorgio Taricco Ezio Biglieri Giuseppe Caire September 4, 1998 Abstract We study the synergy between

More information

DETECTION theory deals primarily with techniques for

DETECTION theory deals primarily with techniques for ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for

More information

Wiener Measure and Brownian Motion

Wiener Measure and Brownian Motion Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u

More information

Some Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom

Some Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom Some Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom Stefan M. Moser April 7, 007 Abstract The non-central chi-square distribution plays an important role

More information

Image Denoising using Uniform Curvelet Transform and Complex Gaussian Scale Mixture

Image Denoising using Uniform Curvelet Transform and Complex Gaussian Scale Mixture EE 5359 Multimedia Processing Project Report Image Denoising using Uniform Curvelet Transform and Complex Gaussian Scale Mixture By An Vo ISTRUCTOR: Dr. K. R. Rao Summer 008 Image Denoising using Uniform

More information

Expect Values and Probability Density Functions

Expect Values and Probability Density Functions Intelligent Systems: Reasoning and Recognition James L. Crowley ESIAG / osig Second Semester 00/0 Lesson 5 8 april 0 Expect Values and Probability Density Functions otation... Bayesian Classification (Reminder...3

More information

A Statistical Analysis of Fukunaga Koontz Transform

A Statistical Analysis of Fukunaga Koontz Transform 1 A Statistical Analysis of Fukunaga Koontz Transform Xiaoming Huo Dr. Xiaoming Huo is an assistant professor at the School of Industrial and System Engineering of the Georgia Institute of Technology,

More information

A Partial Differential Equation Approach to Image Zoom

A Partial Differential Equation Approach to Image Zoom A Partial Differential Equation Approach to Image Zoom Abdelmounim Belahmidi and Frédéric Guichard January 2004 Abstract We propose a new model for zooming digital image. This model, driven by a partial

More information

EXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME. Xavier Mestre 1, Pascal Vallet 2

EXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME. Xavier Mestre 1, Pascal Vallet 2 EXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME Xavier Mestre, Pascal Vallet 2 Centre Tecnològic de Telecomunicacions de Catalunya, Castelldefels, Barcelona (Spain) 2 Institut

More information

IEEE copyright notice

IEEE copyright notice IEEE copyright notice Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for

More information

Which wavelet bases are the best for image denoising?

Which wavelet bases are the best for image denoising? Which wavelet bases are the best for image denoising? Florian Luisier a, Thierry Blu a, Brigitte Forster b and Michael Unser a a Biomedical Imaging Group (BIG), Ecole Polytechnique Fédérale de Lausanne

More information

PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata

PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata ' / PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE Noboru Murata Waseda University Department of Electrical Electronics and Computer Engineering 3--

More information

Information geometry for bivariate distribution control

Information geometry for bivariate distribution control Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic

More information

Statistics for scientists and engineers

Statistics for scientists and engineers Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3

More information

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007

More information

A View on Extension of Utility-Based on Links with Information Measures

A View on Extension of Utility-Based on Links with Information Measures Communications of the Korean Statistical Society 2009, Vol. 16, No. 5, 813 820 A View on Extension of Utility-Based on Links with Information Measures A.R. Hoseinzadeh a, G.R. Mohtashami Borzadaran 1,b,

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

Integrating Correlated Bayesian Networks Using Maximum Entropy

Integrating Correlated Bayesian Networks Using Maximum Entropy Applied Mathematical Sciences, Vol. 5, 2011, no. 48, 2361-2371 Integrating Correlated Bayesian Networks Using Maximum Entropy Kenneth D. Jarman Pacific Northwest National Laboratory PO Box 999, MSIN K7-90

More information

MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING. Kaitlyn Beaudet and Douglas Cochran

MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING. Kaitlyn Beaudet and Douglas Cochran MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING Kaitlyn Beaudet and Douglas Cochran School of Electrical, Computer and Energy Engineering Arizona State University, Tempe AZ 85287-576 USA ABSTRACT The problem

More information

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean

More information

Measure-Transformed Quasi Maximum Likelihood Estimation

Measure-Transformed Quasi Maximum Likelihood Estimation Measure-Transformed Quasi Maximum Likelihood Estimation 1 Koby Todros and Alfred O. Hero Abstract In this paper, we consider the problem of estimating a deterministic vector parameter when the likelihood

More information

Empirical Bayes Least Squares Estimation without an Explicit Prior

Empirical Bayes Least Squares Estimation without an Explicit Prior Computer Science Technical Report TR2007-900, May 2007 Courant Institute of Mathematical Sciences, New York University http://cs.nyu.edu/web/research/techreports/reports.html Empirical Bayes Least Squares

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood

More information

Deakin Research Online

Deakin Research Online Deakin Research Online This is the published version: Phat, V. N. and Trinh, H. 1, Exponential stabilization of neural networks with various activation functions and mixed time-varying delays, IEEE transactions

More information

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions DD2431 Autumn, 2014 1 2 3 Classification with Probability Distributions Estimation Theory Classification in the last lecture we assumed we new: P(y) Prior P(x y) Lielihood x2 x features y {ω 1,..., ω K

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Block 2: Introduction to Information Theory

Block 2: Introduction to Information Theory Block 2: Introduction to Information Theory Francisco J. Escribano April 26, 2015 Francisco J. Escribano Block 2: Introduction to Information Theory April 26, 2015 1 / 51 Table of contents 1 Motivation

More information

The Method of Types and Its Application to Information Hiding

The Method of Types and Its Application to Information Hiding The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,

More information

Hypothesis Testing with Communication Constraints

Hypothesis Testing with Communication Constraints Hypothesis Testing with Communication Constraints Dinesh Krithivasan EECS 750 April 17, 2006 Dinesh Krithivasan (EECS 750) Hyp. testing with comm. constraints April 17, 2006 1 / 21 Presentation Outline

More information

Verifying Regularity Conditions for Logit-Normal GLMM

Verifying Regularity Conditions for Logit-Normal GLMM Verifying Regularity Conditions for Logit-Normal GLMM Yun Ju Sung Charles J. Geyer January 10, 2006 In this note we verify the conditions of the theorems in Sung and Geyer (submitted) for the Logit-Normal

More information

Poisson Image Denoising Using Best Linear Prediction: A Post-processing Framework

Poisson Image Denoising Using Best Linear Prediction: A Post-processing Framework Poisson Image Denoising Using Best Linear Prediction: A Post-processing Framework Milad Niknejad, Mário A.T. Figueiredo Instituto de Telecomunicações, and Instituto Superior Técnico, Universidade de Lisboa,

More information

Legendre-Fenchel transforms in a nutshell

Legendre-Fenchel transforms in a nutshell 1 2 3 Legendre-Fenchel transforms in a nutshell Hugo Touchette School of Mathematical Sciences, Queen Mary, University of London, London E1 4NS, UK Started: July 11, 2005; last compiled: October 16, 2014

More information

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 3, MARCH

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 3, MARCH IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 57, NO 3, MARCH 2011 1587 Universal and Composite Hypothesis Testing via Mismatched Divergence Jayakrishnan Unnikrishnan, Member, IEEE, Dayu Huang, Student

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

A SIRV-CFAR Adaptive Detector Exploiting Persymmetric Clutter Covariance Structure

A SIRV-CFAR Adaptive Detector Exploiting Persymmetric Clutter Covariance Structure A SIRV-CFAR Adaptive Detector Exploiting Persymmetric Clutter Covariance Structure Guilhem Pailloux 2 Jean-Philippe Ovarlez Frédéric Pascal 3 and Philippe Forster 2 ONERA - DEMR/TSI Chemin de la Hunière

More information

certain class of distributions, any SFQ can be expressed as a set of thresholds on the sufficient statistic. For distributions

certain class of distributions, any SFQ can be expressed as a set of thresholds on the sufficient statistic. For distributions Score-Function Quantization for Distributed Estimation Parvathinathan Venkitasubramaniam and Lang Tong School of Electrical and Computer Engineering Cornell University Ithaca, NY 4853 Email: {pv45, lt35}@cornell.edu

More information

Riccati difference equations to non linear extended Kalman filter constraints

Riccati difference equations to non linear extended Kalman filter constraints International Journal of Scientific & Engineering Research Volume 3, Issue 12, December-2012 1 Riccati difference equations to non linear extended Kalman filter constraints Abstract Elizabeth.S 1 & Jothilakshmi.R

More information

Worst case analysis for a general class of on-line lot-sizing heuristics

Worst case analysis for a general class of on-line lot-sizing heuristics Worst case analysis for a general class of on-line lot-sizing heuristics Wilco van den Heuvel a, Albert P.M. Wagelmans a a Econometric Institute and Erasmus Research Institute of Management, Erasmus University

More information

Design of Image Adaptive Wavelets for Denoising Applications

Design of Image Adaptive Wavelets for Denoising Applications Design of Image Adaptive Wavelets for Denoising Applications Sanjeev Pragada and Jayanthi Sivaswamy Center for Visual Information Technology International Institute of Information Technology - Hyderabad,

More information

Blind separation of instantaneous mixtures of dependent sources

Blind separation of instantaneous mixtures of dependent sources Blind separation of instantaneous mixtures of dependent sources Marc Castella and Pierre Comon GET/INT, UMR-CNRS 7, 9 rue Charles Fourier, 9 Évry Cedex, France marc.castella@int-evry.fr, CNRS, I3S, UMR

More information

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012 Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM

More information

IMPROVEMENTS IN MODAL PARAMETER EXTRACTION THROUGH POST-PROCESSING FREQUENCY RESPONSE FUNCTION ESTIMATES

IMPROVEMENTS IN MODAL PARAMETER EXTRACTION THROUGH POST-PROCESSING FREQUENCY RESPONSE FUNCTION ESTIMATES IMPROVEMENTS IN MODAL PARAMETER EXTRACTION THROUGH POST-PROCESSING FREQUENCY RESPONSE FUNCTION ESTIMATES Bere M. Gur Prof. Christopher Niezreci Prof. Peter Avitabile Structural Dynamics and Acoustic Systems

More information

EECS 750. Hypothesis Testing with Communication Constraints

EECS 750. Hypothesis Testing with Communication Constraints EECS 750 Hypothesis Testing with Communication Constraints Name: Dinesh Krithivasan Abstract In this report, we study a modification of the classical statistical problem of bivariate hypothesis testing.

More information

The Minimum Message Length Principle for Inductive Inference

The Minimum Message Length Principle for Inductive Inference The Principle for Inductive Inference Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health University of Melbourne University of Helsinki, August 25,

More information

UWB Geolocation Techniques for IEEE a Personal Area Networks

UWB Geolocation Techniques for IEEE a Personal Area Networks MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com UWB Geolocation Techniques for IEEE 802.15.4a Personal Area Networks Sinan Gezici Zafer Sahinoglu TR-2004-110 August 2004 Abstract A UWB positioning

More information

A New Method of Generating Multivariate Weibull Distributed Data

A New Method of Generating Multivariate Weibull Distributed Data A New Method of Generating Multivariate Weibull Distributed Data Justin G Metcalf, K. James Sangston, Muralidhar Rangaswamy, Shannon D Blunt, Braham Himed Sensors Directorate, Air Force Research Laboratory

More information

NOISE can sometimes improve nonlinear signal processing

NOISE can sometimes improve nonlinear signal processing 488 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 59, NO. 2, FEBRUARY 2011 Noise Benefits in Quantizer-Array Correlation Detection and Watermark Decoding Ashok Patel, Member, IEEE, and Bart Kosko, Fellow,

More information

Expectation propagation for signal detection in flat-fading channels

Expectation propagation for signal detection in flat-fading channels Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA

More information

EM Algorithm. Expectation-maximization (EM) algorithm.

EM Algorithm. Expectation-maximization (EM) algorithm. EM Algorithm Outline: Expectation-maximization (EM) algorithm. Examples. Reading: A.P. Dempster, N.M. Laird, and D.B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, J. R. Stat. Soc.,

More information

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health

More information

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs)

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs) ORF 523 Lecture 8 Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Any typos should be emailed to a a a@princeton.edu. 1 Outline Convexity-preserving operations Convex envelopes, cardinality

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University

More information

Nonlinear Diffusion. Journal Club Presentation. Xiaowei Zhou

Nonlinear Diffusion. Journal Club Presentation. Xiaowei Zhou 1 / 41 Journal Club Presentation Xiaowei Zhou Department of Electronic and Computer Engineering The Hong Kong University of Science and Technology 2009-12-11 2 / 41 Outline 1 Motivation Diffusion process

More information

Variational Principal Components

Variational Principal Components Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 12, DECEMBER 2008 2009 Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation Yuanqing Li, Member, IEEE, Andrzej Cichocki,

More information

Translation-Invariant Shrinkage/Thresholding of Group Sparse. Signals

Translation-Invariant Shrinkage/Thresholding of Group Sparse. Signals Translation-Invariant Shrinkage/Thresholding of Group Sparse Signals Po-Yu Chen and Ivan W. Selesnick Polytechnic Institute of New York University, 6 Metrotech Center, Brooklyn, NY 11201, USA Email: poyupaulchen@gmail.com,

More information

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Bodduluri Asha, B. Leela kumari Abstract: It is well known that in a real world signals do not exist without noise, which may be negligible

More information

Probability and statistics; Rehearsal for pattern recognition

Probability and statistics; Rehearsal for pattern recognition Probability and statistics; Rehearsal for pattern recognition Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských

More information

On the Hilbert Transform of Wavelets

On the Hilbert Transform of Wavelets On the Hilbert Transform of Wavelets Kunal Narayan Chaudhury and Michael Unser Abstract A wavelet is a localized function having a prescribed number of vanishing moments. In this correspondence, we provide

More information

An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding

An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding Beibei Wang, Yao Wang, Ivan Selesnick and Anthony Vetro TR2004-132 December

More information