Artifact-free Wavelet Denoising: Non-convex Sparse Regularization, Convex Optimization

Size: px
Start display at page:

Download "Artifact-free Wavelet Denoising: Non-convex Sparse Regularization, Convex Optimization"

Transcription

1 IEEE SIGNAL PROCESSING LETTERS. 22(9): , SEPTEMBER 215. (PREPRINT) 1 Artifact-free Wavelet Denoising: Non-convex Sparse Regularization, Convex Optimization Yin Ding and Ivan W. Selesnick Abstract Algorithms for signal denoising that combine avelet-domain sparsity and total variation (TV) regularization are relatively free of artifacts, such as pseudo-gibbs oscillations, normally introduced by pure avelet thresholding. This paper formulates avelet-tv (WATV) denoising as a unified problem. To strongly induce avelet sparsity, the proposed approach uses non-convex penalty functions. At the same time, in order to dra on the advantages of convex optimization (unique minimum, reliable algorithms, simplified regularization parameter selection), the non-convex penalties are chosen so as to ensure the convexity of the total objective function. A computationally efficient, fast converging algorithm is derived. I. INTRODUCTION Simple thresholding of avelet coefficients is a reasonably effective procedure for noise reduction hen the signal of interest possesses a sparse avelet representation. Hoever, avelet thresholding often introduces artifacts such as spurious noise spikes and pseudo-gibbs oscillations around discontinuities [1]. In contrast, total variation (TV) denoising [4] does not introduce such artifacts; but it often produces undesirable staircase artifacts. Roughly stated, noise spikes resulting from avelet thresholding are due to noisy avelet coefficients exceeding the threshold, hereas pseudo-gibbs artifacts are due to nonzero coefficients being erroneously set to zero. An approach to alleviate pseudo-gibbs artifacts is to estimate thresholded coefficients by variational means [5], [14]. The use of TV minimization as a ay to recover thresholded coefficients is particularly effective [1], [11], [18], [19]. In this approach, thresholding is performed first. Those coefficients that survive thresholding (i.e., significant coefficients) are maintained. Those coefficients that do not survive (i.e., insignificant coefficients) are then optimized to minimize a TV objective function. TV regularization as subsequently used ith avelet packets [26], curvelets [7], complex avelets [25], shearlets [21], tetrolets [2], [7], and generalized in [2]. Poerful proximal algorithms have been developed for signal restoration ith general hybrid regularization (including avelet-tv) alloing constraints and non-gaussian noise [2]. In this letter, e propose a unified avelet-tv (WATV) approach that estimates all avelet coefficients (significant and insignificant) simultaneously via the minimization of a Copyright (c) 215 IEEE. Personal use of this material is permitted. Hoever, permission to use this material for any other purposes must be obtained from the IEEE by sending a request to pubs-permissions@ieee.org. The authors are ith the Department of Electrical and Computer Engineering, School of Engineering, Ne York University, 6 Metrotech Center, Brooklyn, NY yd72@nyu.edu and selesi@nyu.edu. MATLAB softare available at single objective function. To induce avelet-domain sparsity, e employ non-convex penalties due to their strong sparsityinducing properties [29] [1]. In general practice, hen nonconvex penalties are used, the convexity of the objective function is usually sacrificed. Hoever, in our approach, e restrict the non-convex penalty so as to ensure the strict convexity of the objective function; then, the minimizer is unique and can be reliably obtained through convex optimization. The proposed WATV denoising method is relatively resistant to pseudo-gibbs oscillations and spurious noise spikes. We derive a computationally efficient, fast converging optimization algorithm for the proposed objective function. The formulation of convex optimization problems for linear inverse problems, here a non-convex regularizer is designed in such a ay that the total objective function is convex, as proposed by Blake and Zimmerman [4] and Nikolova [27], [28], [1]. The use of semidefinite programming (SDP) to attain such a regularizer has been considered [5]. Recently, the concept has been applied to group-sparse denoising [12], TV denoising [6], and non-convex fused-lasso []. II. PROBLEM FORMULATION We consider the estimation of a signal x R N observed in additive hite Gaussian noise (AWGN), y n = x n + v n, n =, 1,..., N 1. (1) We denote the avelet transform by W and the avelet coefficients of signal x as = W x. We index the coefficients as j,k here j and k are the scale and time indices, respectively. In this ork, e use the translational-invariant (i.e., undecimated) avelet transform [1], [24], hich satisfies the Parseval frame condition, W T W = I. (2) But the proposed WATV denoising algorithm can be used ith any transform W satisfying (2). The total variation (TV) of signal x R N is defined as TV(x) := Dx 1 here D is the first-order difference matrix, D =......, () and x 1 denotes the l 1 norm of x, i.e., x 1 = n x n. We also denote x 2 2 := n x n 2. For a set of doubly-indexed avelet coefficients, e denote 2 2 := j,k j,k 2.

2 IEEE SIGNAL PROCESSING LETTERS. 22(9): , SEPTEMBER 215. (PREPRINT) 2 The proposed WATV denoising method finds avelet coefficients by solving the optimization problem, ŵ F () = 1 2 W y 2 2 (a) x φ(x) 2 (b) + j,k λ j φ( j,k ; a j ) + β DW T 1. (4) 2 1 The estimate of the signal x is then given by the inverse avelet transform of ŵ, i.e., ˆx = W T ŵ. The penalty term DW T 1 is the total variation of the signal estimate ˆx. The regularization parameters are λ j > and β >. We allo the avelet regularization and penalty parameters (λ j and a j ) to vary ith scale j. A. Parameterized Penalty Function In the proposed approach, the function φ( ; a): R R is a non-convex sparsity-inducing penalty function ith parameter a. We assume φ satisfies the folloing conditions: 1) φ is continuous on R 2) φ is tice continuously differentiable, increasing, and concave on R + ) φ(x; ) = x 4) φ(; a) = 5) φ( x; a) = φ(x; a) 6) φ ( + ; a) = 1 7) φ (x; a) a for all x The folloing proposition indicates a suitable range of values for the parameter a. This proposition combines Propositions 1 and 2 of [12]. Proposition 1. Suppose the parameterized penalty function φ satisfies the above-listed properties and λ >. If then the function f : R R, a < 1 λ, (5) f(x) = 1 2 (y x)2 + λ φ(x; a), (6) is strictly convex and the function θ : R R, θ(y; λ, a) x R 1 2 (y x)2 + λ φ(x; a), (7) is a continuous nonlinear threshold function ith threshold value λ. That is, θ(y; λ, a) = for all y < λ. Note that (7) is the definition of the proximity operator, or proximal mapping, of φ [2], [15]. Several common penalties satisfy the above-listed properties, for example, the logarithmic and arctangent penalties (hen suitably normalized), both of hich have been advocated for sparse regularization [8], [29]. We use the arctangent penalty as it induces sparsity more strongly [12], [5], ( ( ) ) 2 a φ(x; a) = atan 1+2a x π 6, a > (8) x, a =. Figure 1 illustrates this penalty and its corresponding threshold function θ, hich is given by the solution to a cubic polynomial; see Eqn. (21) in [5] θ(y) soft(y) 1 2 Fig. 1. (a) Non-convex penalty. (b) Threshold function (λ = 1, a =.95). The parameter a (ith a < 1/λ) controls the nonconvexity of the penalty φ; specifically, φ ( + ; a) = a. This parameter also controls the shape of the threshold function θ; namely, the right-sided derivative at the threshold λ, is θ (λ + ) = 1/(1 aλ). If a =, then θ (λ + ) = 1 and θ is the soft-threshold function. If a 1/λ (and a < 1/λ), then θ is a continuous approximation of the hard-threshold function. But, if a > 1/λ, then f in (6) is not convex, the associated threshold function is not continuous, and the threshold value is not λ. As discussed in [5], the value a = 1/λ is the critical value that maximizes the sparsity-inducing behavior of the penalty hile ensuring convexity of f. B. Convexity of the Objective Function In the WATV denoising problem (4), e use the non-convex penalty φ to strongly promote sparsity, but e seek to ensure the strict convexity of the objective function F. Proposition 2. Suppose the parameterized penalty function φ satisfies the properties listed in Sec. II-A. The objective function (4) is strictly convex if for each scale j. a j < 1 λ j (9) Proof. We rite the objective function in (4) as F () = 1 2 ([W y] j,k j,k ) 2 + λ j φ( j,k ; a j ) j,k + β DW T 1. (1) By Proposition 1, condition (9) ensures each term in the curly braces is strictly convex. Also, the l 1 norm term is convex. It follos that F, being the sum of strictly convex and convex functions, is strictly convex. When β = in (4) and a j < 1/λ j for all j, then ŵ j,k = θ([w y] j,k ; λ j, a j ). That is, the solution is obtained by aveletdomain thresholding. C. Parameters The proposed formulation (4) requires parameters λ j, a j, and β. We suggest using a j = 1/λ j or slightly less (e.g.,

3 IEEE SIGNAL PROCESSING LETTERS. 22(9): , SEPTEMBER 215. (PREPRINT).95/λ j ) to maximally induce sparsity hile keeping F convex. To set λ j and β, e consider separately the special cases of (i) β = and (ii) λ j = for all j. In case (i), problem (4) reduces to pure avelet denoising, for hich e presume suitable parameters, denoted λ W j, are knon. In case (ii), problem (4) reduces to pure TV denoising, for hich e presume a suitable parameter, denoted β TV, is knon. We suggest setting λ j = η λ W j, β = (1 η) β TV, (11) here < η < 1 controls the relative eight of avelet and TV regularization. We further suggest the smaller range of.9 < η < 1., so that TV regularization makes a relatively minor adjustment of the avelet solution, to reduce spurious noise spikes and pseudo-gibbs phenomena. In this ork, e set the avelet thresholds λ W j = 2.5σ j, here σj 2 denotes the noise variance in avelet scale j. For an undecimated avelet transform satisfying (2), e have σ j = σ/2 j/2 here σ 2 is the noise variance in the signal domain. For pure TV denoising, e use β TV = Nσ/4 in accordance ith [17]. Therefore, e suggest using λ j = 2.5 η σ/2 j/2, β = (1 η) Nσ/4, (12) ith a nominal value of η =.95. Additionally, e omit regularization of lo-pass avelet coefficients. III. OPTIMIZATION ALGORITHM To solve the WATV denoising problem (4) e use the split augmented Lagrangian shrinkage algorithm (SALSA) [1] hich in turn is based on the alternating direction method of multipliers (ADMM) [6], [22]. Proximal methods (e.g., Douglas-Rachford) could also be used [15]. We assume that a j satisfies (9) for each j, so that F is strictly convex. With variable splitting, problem (4) is expressed as a constrained problem, here arg min g 1() + g 2 (u) u, subject to u =, g 1 () = 1 2 W y j,k g 2 (u) = β DW T u 1. λ j φ( j,k ; a j ) (1) (14a) (14b) Both g 1 and g 2 are strictly convex, hence e may utilize the convergence theory of [22]. Even though φ is non-convex, the function g 1 is strictly convex because a j satisfies (9). The augmented Lagrangian is given by L(, u, µ) = g 1 () + g 2 (u) + µ 2 u d 2 2. (15) The solution of (1) can be found by iteratively minimizing ith respect to and u alternately, as proven in [22]. Thereby, e obtain an iterative algorithm to solve (4), here each iteration consists of three steps: g1 () + µ 2 u d 2 2, (16a) u g2 (u) + µ u 2 u d 2 2 d = d (u ),, (16b) (16c) for some µ >. We initialize u = W y and d =. Note that both (16a) and (16b) are strictly convex. (Sub-problem (16a) is strictly convex because a j satisfies (9).) Sub-problem (16a) can be solved exactly and explicitly. Combining the quadratic terms, (16a) may be expressed as j,k 1 2 (p j,k j,k ) 2 + λ j µ + 1 φ( j,k; a j ), here p = (W y + µ(u d))/(µ + 1). Since this problem is separable, it can be further ritten as 1 j,k v R 2 (p j,k v) 2 + λ j µ + 1 φ(v; a j), (17) for each j and k. This is the same form as (6). Consequently, the solution to sub-problem (16a) is given by ( ) λ j j,k = θ p j,k ; µ + 1, a j (18) here θ is the threshold function (7). Sub-problem (16b) can also be solved exactly. We use properties of proximity operators to express (16b) in terms of a TV denoising problem [15]. In turn, the TV denoising problem can be solved exactly by a fast finite-time algorithm [16]. Define function q as q(v) β DW T u 1 + µ u 2 u v 2 2 (19) ψ(w T u) + 1 u 2 u v 2 2 (2) here h(u) u u v 2 2 (21) = prox h (v) (22) = v + W ( prox ψ (W T v) W T v ) (2) ψ(x) = β µ Dx 1, h(u) = ψ(w T u), (24) and here e have used the semi-orthogonal linear transform property of proximity operators to obtain (2) from (22), hich depends on (2). See Table 1.1 and Example in [15], and see [9], [] for details. Note that prox ψ (W T v) x ψ(x) x W T v 2 2 x β µ Dx x W T v 2 2 (25) = tvd ( W T v, β/µ ) (26) is total variation denoising. We use the fast finite-time exact softare of [16]. Hence, (16b) can be implemented as v = d + (27) u = v + W ( tvd(w T v, β/µ) W T v ). (28) Hence, solving problem (4) by (16a)-(16c), consists of iteratively thresholding (18) [to solve (16a)] and performing total variation denoising (28) [to solve (16b)]. Table I summarizes the hole algorithm. Each iteration involves one avelet transform, one inverse avelet transform, and one total variation denoising problem. (The term W y in the algorithm

4 IEEE SIGNAL PROCESSING LETTERS. 22(9): , SEPTEMBER 215. (PREPRINT) 4 TABLE I ITERATIVE ALGORITHM FOR WATV DENOISING (4). 6 4 (a) Simulated data (σ = 4.) Input: y, λ j, β, µ Initialization: u = W y, d = a j = 1/λ j Repeat: p = (W y + µ(u d))/(µ + 1) j,k = θ(p j,k ; λ j /(µ + 1), a j ) for all j, k v = d + u = v + W ( tvd(w T v, β/µ) W T v ) d = d (u ) Until convergence Return: x = W T (b) Wavelet hard thresholding (RMSE = 1.68) TABLE II AVERAGE RMSE FOR DENOISING EXAMPLE 6 4 (c) To step WATV (RMSE = 1.58) σ Thresholding To-step WATV Proposed WATV may be precomputed.) Since both minimization problems, (16a) and (16b), are solved exactly, the algorithm is guaranteed to converge to the unique global minimizer according to the convergence theory of [22]. IV. EXAMPLE We illustrate the proposed WATV denoising algorithm using noisy data of length 124 (Fig. 2a), generated by Wavelab ( ith AWGN of σ = 4.. We use a 5-scale undecimated avelet transform ith to vanishing moments. Hard-thresholding (ith λ W j = 2.5σ j ) results in noise spikes due to avelet coefficients exceeding the threshold (Fig. 2b). To-step WATV denoising [19] (avelet hard-thresholding folloed by TV optimization of coefficients thresholded to zero) does not correct noise spikes due to noisy coefficients that exceed the threshold, so e use higher thresholds of σ j 2 log N to avoid noise spikes from occurring in the first step. The result (Fig. 2c) is relatively free of noise spikes and psuedo-gibbs oscillations. The proposed WATV algorithm [ith parameters (12), η =.95, and a j = 1/λ j ] is resistant to spurious noise spikes, preserves sharp discontinuities, and results in a loer rootmean-square-error (RMSE) (Fig. 2d). The proposed algorithm as run for 2 iterations in a time of.15 seconds using Matlab 7.14 on a MacBook Air (1. GHz CPU). Table II shos the RMSE for several values of the noise standard deviation (σ) for each method. Each RMSE value is obtained by averaging 2 independent realizations. The higher RMSE of to-step WATV appears to be due to distortion induced by the greater threshold value (d) Proposed WATV (RMSE = 1.41) Fig. 2. (a) Simulated data. (b) Wavelet denoising (hard-thresholding). (c) To-step WATV denoising [19]. (d) Proposed WATV denoising. V. CONCLUSION Inspired by [18], [19], this paper describes a avelet- TV (WATV) denoising method that is relatively free of artifacts (pseudo-gibbs oscillations and noise spikes) normally introduced by simple avelet thresholding. The method is formulated as an optimization problem incorporating both avelet sparsity and TV regularization. To strongly induce avelet sparsity, the approach uses parameterized non-convex penalty functions. Proposition 2 gives the critical parameter value to maximally promote avelet sparsity hile ensuring convexity of the objective function as a hole. A fast iterative implementation is derived. We expect the approach to be useful ith transforms other than the avelet transform, provided they satisfy the Parseval frame property (2). ACKNOWLEDGMENT The authors gratefully acknoledge constructive comments by anonymous revieers. In particular, a revieer noted that (16b) can be expressed explicitly in terms of total variation denoising, an improvement upon the inexact solution in the original manuscript.

5 IEEE SIGNAL PROCESSING LETTERS. 22(9): , SEPTEMBER 215. (PREPRINT) 5 REFERENCES [1] M. V. Afonso, J. M. Bioucas-Dias, and M. A. T. Figueiredo. Fast image recovery using variable splitting and constrained optimization. IEEE Trans. Image Process., 19(9): , September 21. [2] H. H. Bauschke and P. L. Combettes. Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer, 211. [] I. Bayram, P.-Y. Chen, and I. Selesnick. Fused lasso ith a non-convex sparsity inducing penalty. In Proc. IEEE Int. Conf. Acoust., Speech, Signal Processing (ICASSP), May 214. [4] A. Blake and A. Zisserman. Visual Reconstruction. MIT Press, [5] Y. Bobichon and A. Bijaoui. Regularized multiresolution methods for astronomical image enhancement. In Proc. SPIE, volume 169 (Wavelet Applications in Signal and Image Processing V), pages 6 45, October [6] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning, (1):1 122, 211. [7] E. J. Candés and F. Guo. Ne multiscale transforms, minimum total variation synthesis: applications to edge-preserving image reconstruction. Signal Processing, 82(11): , 22. [8] E. J. Candès, M. B. Wakin, and S. Boyd. Enhancing sparsity by reeighted l1 minimization. J. Fourier Anal. Appl., 14(5):877 95, December 28. [9] L. Chaâri, N. Pustelnik, C. Chaux, and J.-C. Pesquet. Solving inverse problems ith overcomplete transforms and convex optimization techniques. In Proceedings of SPIE, volume 7446 (Wavelets XIII), August 29. [1] T. F. Chan and H. M. Zhou. Total variation improved avelet thresholding in image compression. In Proc. IEEE Int. Conf. Image Processing (ICIP), volume 2, pages 91 94, September 2. [11] T. F. Chan and H.-M. Zhou. Total variation avelet thresholding. J. of Scientific Computing, 2(2):15 41, August 27. [12] P.-Y. Chen and I. W. Selesnick. Group-sparse signal denoising: Nonconvex regularization, convex optimization. IEEE Trans. Signal Process., 62(1): , July 214. [1] R. R. Coifman and D. L. Donoho. Translation-invariant de-noising. In A. Antoniadis and G. Oppenheim, editors, Wavelet and statistics, pages Springer-Verlag, [14] R. R. Coifman and A. Soa. Combining the calculus of variations and avelets for image enhancement. J. of Appl. and Comp. Harm. Analysis, 9(1):1 18, 2. [15] P. L. Combettes and J.-C. Pesquet. Proximal splitting methods in signal processing. In H. H. Bauschke et al., editors, Fixed-Point Algorithms for Inverse Problems in Science and Engineering, pages Springer- Verlag, 211. [16] L. Condat. A direct algorithm for 1-D total variation denoising. IEEE Signal Processing Letters, 2(11): , November 21. [17] L. Dümbgen and A. Kovac. Extensions of smoothing via taut strings. Electron. J. Statist., :41 75, 29. [18] S. Durand and J. Froment. Artifact free signal denoising ith avelets. In Proc. IEEE Int. Conf. Acoust., Speech, Signal Processing (ICASSP), 21. [19] S. Durand and J. Froment. Reconstruction of avelet coefficients using total variation minimization. SIAM J. Sci. Comput., 24(5): , 2. [2] S. Durand and M. Nikolova. Denoising of frame coefficients using l 1 data-fidelity term and edge-preserving regularization. Multiscale Modeling & Simulation, 6(2): , 27. [21] G. R. Easley, D. Labate, and F. Colonna. Shearlet-based total variation diffusion for denoising. IEEE Trans. Image Process., 18(2):26 268, February 29. [22] J. Eckstein and D. Bertsekas. On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators. Math. Program., 5:29 18, [2] J. Krommeh and J. Ma. Tetrolet shrinkage ith anisotropic total variation minimization for image approximation. Signal Processing, 9(8): , 21. [24] M. Lang, H. Guo, J. E. Odegard, C. S. Burrus, and R. O. Wells, Jr. Noise reduction using an undecimated discrete avelet transform. IEEE Signal Processing Letters, (1):1 12, January [25] J. Ma. Toards artifact-free characterization of surface topography using complex avelets and total variation minimization. Applied Mathematics and Computation, 17(2):114 1, 25. [26] F. Malgouyres. Minimizing the total variation under a general convex constraint for image restoration. IEEE Trans. Image Process., 11(12): , December 22. [27] M. Nikolova. Estimation of binary images by minimizing convex criteria. In Proc. IEEE Int. Conf. Image Processing (ICIP), pages vol. 2, [28] M. Nikolova. Markovian reconstruction using a GNC approach. IEEE Trans. Image Process., 8(9): , [29] M. Nikolova. Energy minimization methods. In O. Scherzer, editor, Handbook of Mathematical Methods in Imaging, chapter 5, pages Springer, 211. [] M. Nikolova, M. K. Ng, and C. Tam. On l 1 data fitting and concave regularization for image recovery. SIAM J. Sci. Comput., 5(1):A97 A4, 21. [1] M. Nikolova, M. K. Ng, and C.-P. Tam. Fast nonconvex nonsmooth minimization methods for image restoration and reconstruction. IEEE Trans. Image Process., 19(12):7 88, December 21. [2] N. Pustelnik, C. Chaux, and J. Pesquet. Parallel proximal algorithm for image restoration using hybrid regularization. IEEE Trans. Image Process., 2(9): , September 211. [] N. Pustelnik, J.-C. Pesquet, and C. Chaux. Relaxing tight frame condition in parallel proximal methods for signal restoration. IEEE Trans. Signal Process., 6(2):968 97, February 212. [4] L. Rudin, S. Osher, and E. Fatemi. Nonlinear total variation based noise removal algorithms. Physica D, 6: , [5] I. W. Selesnick and I. Bayram. Sparse signal estimation by maximally sparse convex optimization. IEEE Trans. Signal Process., 62(5): , March 214. [6] I. W. Selesnick, A. Parekh, and I. Bayram. Convex 1-D total variation denoising ith non-convex regularization. IEEE Signal Processing Letters, 22(2): , February 215. [7] L. Wang, L. Xiao, J. Zhang, and Z. Wei. Ne image restoration method associated ith tetrolets shrinkage and eighted anisotropic total variation. Signal Processing, 9(4):661 67, 21.

Denoising of NIRS Measured Biomedical Signals

Denoising of NIRS Measured Biomedical Signals Denoising of NIRS Measured Biomedical Signals Y V Rami reddy 1, Dr.D.VishnuVardhan 2 1 M. Tech, Dept of ECE, JNTUA College of Engineering, Pulivendula, A.P, India 2 Assistant Professor, Dept of ECE, JNTUA

More information

Minimizing Isotropic Total Variation without Subiterations

Minimizing Isotropic Total Variation without Subiterations MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Minimizing Isotropic Total Variation without Subiterations Kamilov, U. S. TR206-09 August 206 Abstract Total variation (TV) is one of the most

More information

Sparse Signal Estimation by Maximally Sparse Convex Optimization

Sparse Signal Estimation by Maximally Sparse Convex Optimization IEEE TRANSACTIONS ON SIGNAL PROCESSING. (PREPRINT) Sparse Signal Estimation by Maximally Sparse Convex Optimization Ivan W. Selesnick and Ilker Bayram Abstract This paper addresses the problem of sparsity

More information

Sparse Regularization via Convex Analysis

Sparse Regularization via Convex Analysis Sparse Regularization via Convex Analysis Ivan Selesnick Electrical and Computer Engineering Tandon School of Engineering New York University Brooklyn, New York, USA 29 / 66 Convex or non-convex: Which

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in

More information

SPARSE SIGNAL RESTORATION. 1. Introduction

SPARSE SIGNAL RESTORATION. 1. Introduction SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful

More information

A NONCONVEX ADMM ALGORITHM FOR GROUP SPARSITY WITH SPARSE GROUPS. Rick Chartrand and Brendt Wohlberg

A NONCONVEX ADMM ALGORITHM FOR GROUP SPARSITY WITH SPARSE GROUPS. Rick Chartrand and Brendt Wohlberg A NONCONVEX ADMM ALGORITHM FOR GROUP SPARSITY WITH SPARSE GROUPS Rick Chartrand and Brendt Wohlberg Los Alamos National Laboratory Los Alamos, NM 87545, USA ABSTRACT We present an efficient algorithm for

More information

Signal Restoration with Overcomplete Wavelet Transforms: Comparison of Analysis and Synthesis Priors

Signal Restoration with Overcomplete Wavelet Transforms: Comparison of Analysis and Synthesis Priors Signal Restoration with Overcomplete Wavelet Transforms: Comparison of Analysis and Synthesis Priors Ivan W. Selesnick a and Mário A. T. Figueiredo b a Polytechnic Institute of New York University, Brooklyn,

More information

On the Convergence of the Iterative Shrinkage/Thresholding Algorithm With a Weakly Convex Penalty

On the Convergence of the Iterative Shrinkage/Thresholding Algorithm With a Weakly Convex Penalty On the Convergence of the Iterative Shrinkage/Thresholding Algorithm With a Weakly Convex Penalty İlker Bayram arxiv:5.78v [math.oc] Nov 5 Abstract We consider the iterative shrinkage/thresholding algorithm

More information

SPARSITY-ASSISTED SIGNAL SMOOTHING (REVISITED) Ivan Selesnick

SPARSITY-ASSISTED SIGNAL SMOOTHING (REVISITED) Ivan Selesnick SPARSITY-ASSISTED SIGNAL SMOOTHING (REVISITED) Ivan Selesnick Electrical and Computer Engineering Tandon School of Engineering, New York University Brooklyn, New York, USA ABSTRACT This paper proposes

More information

Sparsity-Assisted Signal Smoothing

Sparsity-Assisted Signal Smoothing Sparsity-Assisted Signal Smoothing Ivan W. Selesnick Electrical and Computer Engineering NYU Polytechnic School of Engineering Brooklyn, NY, 1121 selesi@poly.edu Summary. This chapter describes a method

More information

About Split Proximal Algorithms for the Q-Lasso

About Split Proximal Algorithms for the Q-Lasso Thai Journal of Mathematics Volume 5 (207) Number : 7 http://thaijmath.in.cmu.ac.th ISSN 686-0209 About Split Proximal Algorithms for the Q-Lasso Abdellatif Moudafi Aix Marseille Université, CNRS-L.S.I.S

More information

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Bodduluri Asha, B. Leela kumari Abstract: It is well known that in a real world signals do not exist without noise, which may be negligible

More information

Symmetric Wavelet Tight Frames with Two Generators

Symmetric Wavelet Tight Frames with Two Generators Symmetric Wavelet Tight Frames with Two Generators Ivan W. Selesnick Electrical and Computer Engineering Polytechnic University 6 Metrotech Center, Brooklyn, NY 11201, USA tel: 718 260-3416, fax: 718 260-3906

More information

A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT

A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Progress In Electromagnetics Research Letters, Vol. 16, 53 60, 2010 A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Y. P. Liu and Q. Wan School of Electronic Engineering University of Electronic

More information

Sparse signal representation and the tunable Q-factor wavelet transform

Sparse signal representation and the tunable Q-factor wavelet transform Sparse signal representation and the tunable Q-factor wavelet transform Ivan Selesnick Polytechnic Institute of New York University Brooklyn, New York Introduction Problem: Decomposition of a signal into

More information

Sparse signal representation and the tunable Q-factor wavelet transform

Sparse signal representation and the tunable Q-factor wavelet transform Sparse signal representation and the tunable Q-factor wavelet transform Ivan Selesnick Polytechnic Institute of New York University Brooklyn, New York Introduction Problem: Decomposition of a signal into

More information

TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS

TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS Martin Kleinsteuber and Simon Hawe Department of Electrical Engineering and Information Technology, Technische Universität München, München, Arcistraße

More information

Error Analysis for H 1 Based Wavelet Interpolations

Error Analysis for H 1 Based Wavelet Interpolations Error Analysis for H 1 Based Wavelet Interpolations Tony F. Chan Hao-Min Zhou Tie Zhou Abstract We rigorously study the error bound for the H 1 wavelet interpolation problem, which aims to recover missing

More information

Learning MMSE Optimal Thresholds for FISTA

Learning MMSE Optimal Thresholds for FISTA MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Learning MMSE Optimal Thresholds for FISTA Kamilov, U.; Mansour, H. TR2016-111 August 2016 Abstract Fast iterative shrinkage/thresholding algorithm

More information

446 SCIENCE IN CHINA (Series F) Vol. 46 introduced in refs. [6, ]. Based on this inequality, we add normalization condition, symmetric conditions and

446 SCIENCE IN CHINA (Series F) Vol. 46 introduced in refs. [6, ]. Based on this inequality, we add normalization condition, symmetric conditions and Vol. 46 No. 6 SCIENCE IN CHINA (Series F) December 003 Construction for a class of smooth wavelet tight frames PENG Lizhong (Λ Π) & WANG Haihui (Ξ ) LMAM, School of Mathematical Sciences, Peking University,

More information

Design of Projection Matrix for Compressive Sensing by Nonsmooth Optimization

Design of Projection Matrix for Compressive Sensing by Nonsmooth Optimization Design of Proection Matrix for Compressive Sensing by Nonsmooth Optimization W.-S. Lu T. Hinamoto Dept. of Electrical & Computer Engineering Graduate School of Engineering University of Victoria Hiroshima

More information

Introduction to Sparsity in Signal Processing 1

Introduction to Sparsity in Signal Processing 1 Introduction to Sparsity in Signal Processing Ivan Selesnick June, NYU-Poly Introduction These notes describe how sparsity can be used in several signal processing problems. A common theme throughout these

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein December 2, 2015 1 / 25 Background The Dual Problem Consider

More information

Sparse linear models and denoising

Sparse linear models and denoising Lecture notes 4 February 22, 2016 Sparse linear models and denoising 1 Introduction 1.1 Definition and motivation Finding representations of signals that allow to process them more effectively is a central

More information

The Suppression of Transient Artifacts in Time Series via Convex Analysis

The Suppression of Transient Artifacts in Time Series via Convex Analysis The Suppression of Transient Artifacts in Time Series via Convex Analysis Yining Feng 1, Harry Graber 2, and Ivan Selesnick 1 (1) Tandon School of Engineering, New York University, Brooklyn, New York (2)

More information

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Minru Bai(x T) College of Mathematics and Econometrics Hunan University Joint work with Xiongjun Zhang, Qianqian Shao June 30,

More information

Adaptive Primal Dual Optimization for Image Processing and Learning

Adaptive Primal Dual Optimization for Image Processing and Learning Adaptive Primal Dual Optimization for Image Processing and Learning Tom Goldstein Rice University tag7@rice.edu Ernie Esser University of British Columbia eesser@eos.ubc.ca Richard Baraniuk Rice University

More information

A Primal-dual Three-operator Splitting Scheme

A Primal-dual Three-operator Splitting Scheme Noname manuscript No. (will be inserted by the editor) A Primal-dual Three-operator Splitting Scheme Ming Yan Received: date / Accepted: date Abstract In this paper, we propose a new primal-dual algorithm

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Recent developments on sparse representation

Recent developments on sparse representation Recent developments on sparse representation Zeng Tieyong Department of Mathematics, Hong Kong Baptist University Email: zeng@hkbu.edu.hk Hong Kong Baptist University Dec. 8, 2008 First Previous Next Last

More information

Sparse linear models

Sparse linear models Sparse linear models Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 2/22/2016 Introduction Linear transforms Frequency representation Short-time

More information

Continuous State MRF s

Continuous State MRF s EE64 Digital Image Processing II: Purdue University VISE - December 4, Continuous State MRF s Topics to be covered: Quadratic functions Non-Convex functions Continuous MAP estimation Convex functions EE64

More information

EMPLOYING PHASE INFORMATION FOR AUDIO DENOISING. İlker Bayram. Istanbul Technical University, Istanbul, Turkey

EMPLOYING PHASE INFORMATION FOR AUDIO DENOISING. İlker Bayram. Istanbul Technical University, Istanbul, Turkey EMPLOYING PHASE INFORMATION FOR AUDIO DENOISING İlker Bayram Istanbul Technical University, Istanbul, Turkey ABSTRACT Spectral audio denoising methods usually make use of the magnitudes of a time-frequency

More information

Solving DC Programs that Promote Group 1-Sparsity

Solving DC Programs that Promote Group 1-Sparsity Solving DC Programs that Promote Group 1-Sparsity Ernie Esser Contains joint work with Xiaoqun Zhang, Yifei Lou and Jack Xin SIAM Conference on Imaging Science Hong Kong Baptist University May 14 2014

More information

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases 2558 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 48, NO 9, SEPTEMBER 2002 A Generalized Uncertainty Principle Sparse Representation in Pairs of Bases Michael Elad Alfred M Bruckstein Abstract An elementary

More information

Translation-Invariant Shrinkage/Thresholding of Group Sparse. Signals

Translation-Invariant Shrinkage/Thresholding of Group Sparse. Signals Translation-Invariant Shrinkage/Thresholding of Group Sparse Signals Po-Yu Chen and Ivan W. Selesnick Polytechnic Institute of New York University, 6 Metrotech Center, Brooklyn, NY 11201, USA Email: poyupaulchen@gmail.com,

More information

Variational image restoration by means of wavelets: simultaneous decomposition, deblurring and denoising

Variational image restoration by means of wavelets: simultaneous decomposition, deblurring and denoising Variational image restoration by means of wavelets: simultaneous decomposition, deblurring and denoising I. Daubechies and G. Teschke December 2, 2004 Abstract Inspired by papers of Vese Osher [20] and

More information

POISSON noise, also known as photon noise, is a basic

POISSON noise, also known as photon noise, is a basic IEEE SIGNAL PROCESSING LETTERS, VOL. N, NO. N, JUNE 2016 1 A fast and effective method for a Poisson denoising model with total variation Wei Wang and Chuanjiang He arxiv:1609.05035v1 [math.oc] 16 Sep

More information

A Higher-Density Discrete Wavelet Transform

A Higher-Density Discrete Wavelet Transform A Higher-Density Discrete Wavelet Transform Ivan W. Selesnick Abstract In this paper, we describe a new set of dyadic wavelet frames with three generators, ψ i (t), i =,, 3. The construction is simple,

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 XI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 Alternating direction methods of multipliers for separable convex programming Bingsheng He Department of Mathematics

More information

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems)

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Donghwan Kim and Jeffrey A. Fessler EECS Department, University of Michigan

More information

A Brief Overview of Practical Optimization Algorithms in the Context of Relaxation

A Brief Overview of Practical Optimization Algorithms in the Context of Relaxation A Brief Overview of Practical Optimization Algorithms in the Context of Relaxation Zhouchen Lin Peking University April 22, 2018 Too Many Opt. Problems! Too Many Opt. Algorithms! Zero-th order algorithms:

More information

Generalized Orthogonal Matching Pursuit- A Review and Some

Generalized Orthogonal Matching Pursuit- A Review and Some Generalized Orthogonal Matching Pursuit- A Review and Some New Results Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur, INDIA Table of Contents

More information

Mathematical analysis of a model which combines total variation and wavelet for image restoration 1

Mathematical analysis of a model which combines total variation and wavelet for image restoration 1 Информационные процессы, Том 2, 1, 2002, стр. 1 10 c 2002 Malgouyres. WORKSHOP ON IMAGE PROCESSING AND RELAED MAHEMAICAL OPICS Mathematical analysis of a model which combines total variation and wavelet

More information

In collaboration with J.-C. Pesquet A. Repetti EC (UPE) IFPEN 16 Dec / 29

In collaboration with J.-C. Pesquet A. Repetti EC (UPE) IFPEN 16 Dec / 29 A Random block-coordinate primal-dual proximal algorithm with application to 3D mesh denoising Emilie CHOUZENOUX Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est, France Horizon Maths 2014

More information

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection EUSIPCO 2015 1/19 A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection Jean-Christophe Pesquet Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est

More information

Sparse Regularization via Convex Analysis

Sparse Regularization via Convex Analysis IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 65, NO. 7, PP. 448-4494, SEPTEMBER 7 (PREPRINT) Sparse Regularization via Convex Analysis Ivan Selesnick Abstract Sparse approximate solutions to linear equations

More information

Sparse and Regularized Optimization

Sparse and Regularized Optimization Sparse and Regularized Optimization In many applications, we seek not an exact minimizer of the underlying objective, but rather an approximate minimizer that satisfies certain desirable properties: sparsity

More information

Bayesian-based Wavelet Shrinkage for SAR Image Despeckling Using Cycle Spinning *

Bayesian-based Wavelet Shrinkage for SAR Image Despeckling Using Cycle Spinning * Jun. 006 Journal of Electronic Science and Technology of China Vol.4 No. ayesian-based Wavelet Shrinkage for SAR Image Despeckling Using Cycle Spinning * ZHANG De-xiang,, GAO Qing-wei,, CHEN Jun-ning.

More information

Edge-preserving wavelet thresholding for image denoising

Edge-preserving wavelet thresholding for image denoising Journal of Computational and Applied Mathematics ( ) www.elsevier.com/locate/cam Edge-preserving wavelet thresholding for image denoising D. Lazzaro, L.B. Montefusco Departement of Mathematics, University

More information

A General Framework for a Class of Primal-Dual Algorithms for TV Minimization

A General Framework for a Class of Primal-Dual Algorithms for TV Minimization A General Framework for a Class of Primal-Dual Algorithms for TV Minimization Ernie Esser UCLA 1 Outline A Model Convex Minimization Problem Main Idea Behind the Primal Dual Hybrid Gradient (PDHG) Method

More information

Variational methods for restoration of phase or orientation data

Variational methods for restoration of phase or orientation data Variational methods for restoration of phase or orientation data Martin Storath joint works with Laurent Demaret, Michael Unser, Andreas Weinmann Image Analysis and Learning Group Universität Heidelberg

More information

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Math 273a: Optimization Overview of First-Order Optimization Algorithms Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization

More information

Combining multiresolution analysis and non-smooth optimization for texture segmentation

Combining multiresolution analysis and non-smooth optimization for texture segmentation Combining multiresolution analysis and non-smooth optimization for texture segmentation Nelly Pustelnik CNRS, Laboratoire de Physique de l ENS de Lyon Intro Basics Segmentation Two-step texture segmentation

More information

Gauge optimization and duality

Gauge optimization and duality 1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel

More information

arxiv: v1 [cs.lg] 20 Feb 2014

arxiv: v1 [cs.lg] 20 Feb 2014 GROUP-SPARSE MATRI RECOVERY iangrong Zeng and Mário A. T. Figueiredo Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal ariv:1402.5077v1 [cs.lg] 20 Feb 2014 ABSTRACT We apply the

More information

IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 9, SEPTEMBER

IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 9, SEPTEMBER IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 9, SEPTEMBER 2015 1239 Preconditioning for Underdetermined Linear Systems with Sparse Solutions Evaggelia Tsiligianni, StudentMember,IEEE, Lisimachos P. Kondi,

More information

Introduction to Sparsity in Signal Processing

Introduction to Sparsity in Signal Processing 1 Introduction to Sparsity in Signal Processing Ivan Selesnick Polytechnic Institute of New York University Brooklyn, New York selesi@poly.edu 212 2 Under-determined linear equations Consider a system

More information

On the Dual-Tree Complex Wavelet Packet and. Abstract. Index Terms I. INTRODUCTION

On the Dual-Tree Complex Wavelet Packet and. Abstract. Index Terms I. INTRODUCTION On the Dual-Tree Complex Wavelet Packet and M-Band Transforms İlker Bayram, Student Member, IEEE, and Ivan W. Selesnick, Member, IEEE Abstract The -band discrete wavelet transform (DWT) provides an octave-band

More information

Introduction to Wavelets and Wavelet Transforms

Introduction to Wavelets and Wavelet Transforms Introduction to Wavelets and Wavelet Transforms A Primer C. Sidney Burrus, Ramesh A. Gopinath, and Haitao Guo with additional material and programs by Jan E. Odegard and Ivan W. Selesnick Electrical and

More information

A discretized Newton flow for time varying linear inverse problems

A discretized Newton flow for time varying linear inverse problems A discretized Newton flow for time varying linear inverse problems Martin Kleinsteuber and Simon Hawe Department of Electrical Engineering and Information Technology, Technische Universität München Arcisstrasse

More information

Multiscale Geometric Analysis: Thoughts and Applications (a summary)

Multiscale Geometric Analysis: Thoughts and Applications (a summary) Multiscale Geometric Analysis: Thoughts and Applications (a summary) Anestis Antoniadis, University Joseph Fourier Assimage 2005,Chamrousse, February 2005 Classical Multiscale Analysis Wavelets: Enormous

More information

Enhanced Compressive Sensing and More

Enhanced Compressive Sensing and More Enhanced Compressive Sensing and More Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Nonlinear Approximation Techniques Using L1 Texas A & M University

More information

Introduction to Sparsity in Signal Processing

Introduction to Sparsity in Signal Processing 1 Introduction to Sparsity in Signal Processing Ivan Selesnick Polytechnic Institute of New York University Brooklyn, New York selesi@poly.edu 212 2 Under-determined linear equations Consider a system

More information

PART II: Basic Theory of Half-quadratic Minimization

PART II: Basic Theory of Half-quadratic Minimization PART II: Basic Theory of Half-quadratic Minimization Ran He, Wei-Shi Zheng and Liang Wang 013-1-01 Outline History of half-quadratic (HQ) minimization Half-quadratic minimization HQ loss functions HQ in

More information

y(x) = x w + ε(x), (1)

y(x) = x w + ε(x), (1) Linear regression We are ready to consider our first machine-learning problem: linear regression. Suppose that e are interested in the values of a function y(x): R d R, here x is a d-dimensional vector-valued

More information

Bayesian Paradigm. Maximum A Posteriori Estimation

Bayesian Paradigm. Maximum A Posteriori Estimation Bayesian Paradigm Maximum A Posteriori Estimation Simple acquisition model noise + degradation Constraint minimization or Equivalent formulation Constraint minimization Lagrangian (unconstraint minimization)

More information

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization Panos Parpas Department of Computing Imperial College London www.doc.ic.ac.uk/ pp500 p.parpas@imperial.ac.uk jointly with D.V.

More information

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications Class 19: Data Representation by Design What is data representation? Let X be a data-space X M (M) F (M) X A data representation

More information

An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding

An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com An Investigation of 3D Dual-Tree Wavelet Transform for Video Coding Beibei Wang, Yao Wang, Ivan Selesnick and Anthony Vetro TR2004-132 December

More information

GREEDY SIGNAL RECOVERY REVIEW

GREEDY SIGNAL RECOVERY REVIEW GREEDY SIGNAL RECOVERY REVIEW DEANNA NEEDELL, JOEL A. TROPP, ROMAN VERSHYNIN Abstract. The two major approaches to sparse recovery are L 1-minimization and greedy methods. Recently, Needell and Vershynin

More information

Inpainting for Compressed Images

Inpainting for Compressed Images Inpainting for Compressed Images Jian-Feng Cai a, Hui Ji,b, Fuchun Shang c, Zuowei Shen b a Department of Mathematics, University of California, Los Angeles, CA 90095 b Department of Mathematics, National

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

Universität des Saarlandes. Fachrichtung 6.1 Mathematik

Universität des Saarlandes. Fachrichtung 6.1 Mathematik Universität des Saarlandes U N I V E R S I T A S S A R A V I E N I S S Fachrichtung 6. Mathematik Preprint Nr. 94 On the Equivalence of Soft Wavelet Shrinkage, Total Variation Diffusion, Total Variation

More information

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING JIAN-FENG CAI, STANLEY OSHER, AND ZUOWEI SHEN Abstract. Real images usually have sparse approximations under some tight frame systems derived

More information

AN INTRODUCTION TO COMPRESSIVE SENSING

AN INTRODUCTION TO COMPRESSIVE SENSING AN INTRODUCTION TO COMPRESSIVE SENSING Rodrigo B. Platte School of Mathematical and Statistical Sciences APM/EEE598 Reverse Engineering of Complex Dynamical Networks OUTLINE 1 INTRODUCTION 2 INCOHERENCE

More information

Contraction Methods for Convex Optimization and monotone variational inequalities No.12

Contraction Methods for Convex Optimization and monotone variational inequalities No.12 XII - 1 Contraction Methods for Convex Optimization and monotone variational inequalities No.12 Linearized alternating direction methods of multipliers for separable convex programming Bingsheng He Department

More information

EUSIPCO

EUSIPCO EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,

More information

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Gongguo Tang and Arye Nehorai Department of Electrical and Systems Engineering Washington University in St Louis

More information

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem Kangkang Deng, Zheng Peng Abstract: The main task of genetic regulatory networks is to construct a

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

1 Sparsity and l 1 relaxation

1 Sparsity and l 1 relaxation 6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the

More information

An ADMM algorithm for optimal sensor and actuator selection

An ADMM algorithm for optimal sensor and actuator selection An ADMM algorithm for optimal sensor and actuator selection Neil K. Dhingra, Mihailo R. Jovanović, and Zhi-Quan Luo 53rd Conference on Decision and Control, Los Angeles, California, 2014 1 / 25 2 / 25

More information

Estimation Error Bounds for Frame Denoising

Estimation Error Bounds for Frame Denoising Estimation Error Bounds for Frame Denoising Alyson K. Fletcher and Kannan Ramchandran {alyson,kannanr}@eecs.berkeley.edu Berkeley Audio-Visual Signal Processing and Communication Systems group Department

More information

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar

More information

Denosing Using Wavelets and Projections onto the l 1 -Ball

Denosing Using Wavelets and Projections onto the l 1 -Ball 1 Denosing Using Wavelets and Projections onto the l 1 -Ball October 6, 2014 A. Enis Cetin, M. Tofighi Dept. of Electrical and Electronic Engineering, Bilkent University, Ankara, Turkey cetin@bilkent.edu.tr,

More information

Coordinate Update Algorithm Short Course Operator Splitting

Coordinate Update Algorithm Short Course Operator Splitting Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators

More information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department

More information

Generalized ADMM with Optimal Indefinite Proximal Term for Linearly Constrained Convex Optimization

Generalized ADMM with Optimal Indefinite Proximal Term for Linearly Constrained Convex Optimization Generalized ADMM with Optimal Indefinite Proximal Term for Linearly Constrained Convex Optimization Fan Jiang 1 Zhongming Wu Xingju Cai 3 Abstract. We consider the generalized alternating direction method

More information

arxiv: v1 [cs.it] 21 Feb 2013

arxiv: v1 [cs.it] 21 Feb 2013 q-ary Compressive Sensing arxiv:30.568v [cs.it] Feb 03 Youssef Mroueh,, Lorenzo Rosasco, CBCL, CSAIL, Massachusetts Institute of Technology LCSL, Istituto Italiano di Tecnologia and IIT@MIT lab, Istituto

More information

VARIOUS types of wavelet transform are available for

VARIOUS types of wavelet transform are available for IEEE TRANSACTIONS ON SIGNAL PROCESSING A Higher-Density Discrete Wavelet Transform Ivan W. Selesnick, Member, IEEE Abstract This paper describes a new set of dyadic wavelet frames with two generators.

More information

AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION

AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION New Horizon of Mathematical Sciences RIKEN ithes - Tohoku AIMR - Tokyo IIS Joint Symposium April 28 2016 AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION T. Takeuchi Aihara Lab. Laboratories for Mathematics,

More information

Sparse Optimization Lecture: Basic Sparse Optimization Models

Sparse Optimization Lecture: Basic Sparse Optimization Models Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm

More information

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March

More information

A Lower Bound Theorem. Lin Hu.

A Lower Bound Theorem. Lin Hu. American J. of Mathematics and Sciences Vol. 3, No -1,(January 014) Copyright Mind Reader Publications ISSN No: 50-310 A Lower Bound Theorem Department of Applied Mathematics, Beijing University of Technology,

More information

Optimal Linearized Alternating Direction Method of Multipliers for Convex Programming 1

Optimal Linearized Alternating Direction Method of Multipliers for Convex Programming 1 Optimal Linearized Alternating Direction Method of Multipliers for Convex Programming Bingsheng He 2 Feng Ma 3 Xiaoming Yuan 4 October 4, 207 Abstract. The alternating direction method of multipliers ADMM

More information

A Variational Approach to Reconstructing Images Corrupted by Poisson Noise

A Variational Approach to Reconstructing Images Corrupted by Poisson Noise J Math Imaging Vis c 27 Springer Science + Business Media, LLC. Manufactured in The Netherlands. DOI: 1.7/s1851-7-652-y A Variational Approach to Reconstructing Images Corrupted by Poisson Noise TRIET

More information

Diffusion based Projection Method for Distributed Source Localization in Wireless Sensor Networks

Diffusion based Projection Method for Distributed Source Localization in Wireless Sensor Networks The Third International Workshop on Wireless Sensor, Actuator and Robot Networks Diffusion based Projection Method for Distributed Source Localization in Wireless Sensor Networks Wei Meng 1, Wendong Xiao,

More information