Optimisation in imaging

Size: px
Start display at page:

Download "Optimisation in imaging"

Transcription

1 Optimisation in imaging Hugues Talbot ICIP 2014 Tutorial Optimisation on Hierarchies H. Talbot : Optimisation 1/59

2 Outline of the lecture 1 Concepts in optimization Cost function Constraints Duality 2 Inverse problems in imaging Useful formulation in imaging 3 Formulations in imaging 4 DCTV model 5 Generalization of TV models Algorithms 6 Applications in image processing 7 Non-convex regularization 8 Conclusion H. Talbot : Optimisation 2/59

3 Concepts in optimization Introduction Mathematical optimization is a domain of applied mathematics relevant to many areas including statistics, mechanics, signal and image processing. Generalizes many well known techniques such as least squares, linear programming, convex programming, integer programming, combinatorial optimization and others. In this talk we will overview both the continuous and discrete formulations. We follow the notations of [1]. H. Talbot : Optimisation 3/59

4 Concepts in optimization General form Cost function and constraints An optimization problem generally has the following form minimize f 0 (x) subject to f i (x) b i, i = 1,..., m (1) x = (x 1,..., x n ) is a vector of R n called the optimization variable of the problem; f 0 : R n R is the cost function functional; the f i : R n R are the constraints and the b i are the bounds (or limits). A vector x is is optimal, or is a solution to the problem, if it has the smallest objective value among all vectors that satisfy the constraints. H. Talbot : Optimisation 4/59

5 Concepts in optimization Types of optimization problems The type of the variables, the cost function and the constraints determine the type of problems we are dealing with. Optimization problems, in their most general form, are usually unsolvable in practice. NP-complete problems (traveling salesperson, subset-sum, etc) can classically be put in this form and so can many NP-hard problems. Some mathematical regularity is necessary to be able to find a solution: for example, linearity or convexity in all the functions. Requiring integer solutions usually, but not always, makes things much harder: Diophantine vs linear equations for instance. H. Talbot : Optimisation 5/59

6 Concepts in optimization Example: least-squares Least squares with no constraints minimize f 0 (x) = Ax b 2 2 = k a x i b i (2) i=1 The system is quadratic, so convex and differentiable. The solution to (2) is unique and reduces to the linear equation and so the analytical solution is (A A)x = A b (3) x = (A A) 1 A b (4) H. Talbot : Optimisation 6/59

7 Concepts in optimization Regularization: Tikhonov Even with something as simple as least-squares, if A A is ill-conditioned, the solution will be very sensitive to noise, e.g. in the example of deconvolution or tomography. One solution is to use regularization. Ill-posed least-squares problems The simplest regularization strategy is due to Tikhonov [3]. minimize f 0 (x) = Ax b Γx 2 2, (5) where Γ is a well-chosen operator, e.g. λi or x or a wavelet operator. The solution is given analytically by x = (A A + Γ Γ) 1 A b (6) H. Talbot : Optimisation 7/59

8 Concepts in optimization Linear programming Linear programming with constraints minimize c x subject to a i x b i; i = 1,..., n (7) No analytical solution. Well established family of algorithms: the Simplexe (Dantzig 1948) ; interior-point (Karmarkar 1984) Not always easy to recognize. Important for compressive sensing. H. Talbot : Optimisation 8/59

9 Concepts in optimization Duality in the LP case Primal / Dual linear programs Primal minimize c x subject to a i x b i; i = 1,..., n (8) Dual maximize b y subject to a i x c i ; i = 1,..., m (9) A primal/dual pair of LP problems can be obtained by transposing the constraint matrix and swapping cost function and constraint bounds. The primal and dual optima, if they exist, are the same, and can be easily deducted from each other. H. Talbot : Optimisation 9/59

10 Concepts in optimization Duality in convex optimization The same concept of duality applies in convex optimization Duality allows one to swap constraints for terms in the objective function Two concepts of duality : Lagrange and Fenchel. Both are equivalent. H. Talbot : Optimisation 10/59

11 Inverse problems in imaging Motivation: inverse problems in imaging Images we observe are nearly always blurred, noisy, projected versions of some reality. We wish to dispel the fog of acquisition by removing all the artefacts as much as possible to observe the real data. This is an inverse problem. H. Talbot : Optimisation 11/59

12 Inverse problems in imaging Maximum Likelihood We want to estimate some statistical parameter θ on the basis of some observation x. If f is the sampling distribution, f (x θ) is the probability of x when the population parameter is θ. The function θ f (x θ) is the likelihood. The Maximum Likelihood estimate is ˆθ ML (x) = argmax θ f (x θ) E.g, if we have a linear operator H (in matrix form) and Gaussian deviates, then argmax x f (x) = Hx y 2 2 = x H Hx + 2y Hx y y is a quadratic form with a unique maximum, provided by f (x) = 2H Hx + 2H y = 0 θ = (H H) 1 H y H. Talbot : Optimisation 12/59

13 Inverse problems in imaging Strengths and drawbacks of MLE When possible, MLE is fast and effective. Many imaging operators have a MLE interpretation: Gaussian smoothing ; Wiener filtering ; Filtered back projection for tomography ; Principal component analysis... However these require a very descriptive model (with few degrees of freedom) and a lot of data, typically unsuitable for images because we do not have a suitable model for natural images. When we do not have all these hypotheses, sometimes the Bayesian Maximum A Posteriori approach can be used instead. H. Talbot : Optimisation 13/59

14 Inverse problems in imaging Maximum A Posteriori If we assume that we know a prior distribution g over θ, i.e. some a-priori information. Following Bayesian statistics, we can treat θ as a random variable and compute the posterior distribution of θ: (i.e. the Bayes theorem). f (x θ)g(θ) θ f (θ x) = ϑ Θ f (x ϑ)g(ϑ)dϑ Then the Maximum a Posteriori is the estimate ˆθ MAP (x) = argmax θ MAP is a regularization of ML. f (θ x) = argmax θ f (x θ)g(θ) H. Talbot : Optimisation 14/59

15 Inverse problems in imaging Markov Random Fields So far this is statistics theory. What is the link between MAP and imaging? We need an imaging model. A Markov Random Field is a model made of a set of sites (a.k.a. pixels) S = {s 1,..., s n }, a set of random variables y = {y 1,..., y n } associated with each pixel, and a set of neighbours N 1,...,n at each pixel location. N p describes the neighborhood at pixel p. Obeys the Markov condition, i.e. Pr(y p y S\p ) = Pr(y p N p ) I.e.: the probability of a pixel p depends only on its immediate neighbours. H. Talbot : Optimisation 15/59

16 Inverse problems in imaging Formulating the MAP of an MRF Now let us express a MAP formulation for an MRF Given a set of observables x = {x 1,..., x n }, We derive a MAP ŷ = argmax Pr(y 1...n x) (10) y 1...n = argmax n Pr(x n y n ) Pr(y 1...n ) (11) y 1...n n=1 = argmax n log[pr(x n y n )] + log[pr(y 1...n )] (12) y 1...n n=1 n = argmin U p (y p ) + P u,p (y u, y p ) y 1...n p=1 u N p (13) (Geman & Geman, PAMI 1984). H. Talbot : Optimisation 16/59

17 Inverse problems in imaging Solving the MAP-MRF formulation This last sum is an energy contains unary terms U p (y p ) and pairwise terms P u,p (y u, y p ). We now have an optimization problem. Depending on the expression of the probability functions, can solve it by i: statistical means, e.g. EM, ii: physical analogies, e.g. simulated annealing or iii: via linear/convex optimization techniques. With some restrictions, graph cuts are able to optimize these energies. H. Talbot : Optimisation 17/59

18 Inverse problems in imaging MRF and Graph Cuts For instance, consider the binary segmentation problem. With unary weights the above can be written: argmin Ê(G) = v i V w i (V i ) + λ w ij δ Vi V j (14) e ij E V i is 1 if v i V s and 0 if v i V t, i.e. it is 1 if pixel i belongs to the partition containing s and 0 otherwise. δ Vi V j is 1 if the corresponding e ij is on the cut, and 0 otherwise. The first sum contains the pairwise terms, and sums the cost of the cut in the image plane. The second sum contains the unary terms, and adds the cost of a pixel to belong to either the partition containing s or the partition containing t. H. Talbot : Optimisation 18/59

19 Inverse problems in imaging Illustration S S T (a) T (b) Figure : Segmentation with unary weights. In this case weighted edges link the source and the sink to all the pixels in the image (a). The min-cut is a surface separating s from t (b). Some strong edge weights can ensure the surface crosses the pixel plane, enforcing topology constraints. H. Talbot : Optimisation 19/59

20 Inverse problems in imaging Segmentation example (a) (b) Figure : Binary segmentation with unary weights and no markers (Boykov-Jolly segmentation model, ICCV 2001). H. Talbot : Optimisation 20/59

21 Inverse problems in imaging Image restoration and graph cuts GC are able to optimize some MRF energies exactly (globally) in the binary case More generally, submodular (e.g. discrete-convex) energies can be at least locally optimized using graph cuts Using various constructions, e.g. Ishikawa PAMI 2003, it is possible to map restoration (denoisng) problems to GC. Many GC optimization approaches have been invented to solve the corresponding energies: α-expansions, α β moves, convex moves, etc (Veksler 1999). They were essentially known before in other communities (Murota 2003). More recent approaches are able to optimize the same kind of energies using different techniques: Belief propagation, Primal-dual Tree-Reweighted, etc (Kolmogorov PAMI 2006). H. Talbot : Optimisation 21/59

22 Inverse problems in imaging Graph-based energies These formulation are very useful but suffer from the purely discrete graph framework Formulations and solutions are not isotropic (grid bias) Graph based formulation can be resource-intensive (memory and speed) They are hard to parallelize Hard to incorporate extra constraints and projection/linear operators. H. Talbot : Optimisation 22/59

23 Formulations in imaging Continuous image restoration model We suppose there exists some unknown image x R N. However we do observe some data y R Q via some linear operator H, which is corrupted by some noise: y = Hx + u, H R Q N x y H. Talbot : Optimisation 23/59

24 Formulations in imaging Recovery We seek to recover a good approximation ˆx of x from H and y. H can be: Model for camera, including defocus and motion blur MRI, PET, X-Ray tomography... u often modeled by Additive White Gaussian Noise, but can be Poisson, Poisson Gauss, Rician, etc. Simplest case: least squares: ˆx = argmin x Hx y 2 2 analytical, simple, effective, but not robust to outliers. H. Talbot : Optimisation 24/59

25 Formulations in imaging Recovery We seek to recover a good approximation ˆx of x from H and y. H can be: Model for camera, including defocus and motion blur MRI, PET, X-Ray tomography... u often modeled by Additive White Gaussian Noise, but can be Poisson, Poisson Gauss, Rician, etc. Tikhonov regularization: ˆx = argmin x x λ Hx y 2 2 reflect the prior assumption that we want to avoid large x. Also analytical and more robust but not sparse. H. Talbot : Optimisation 24/59

26 Formulations in imaging Recovery We seek to recover a good approximation ˆx of x from H and y. H can be: Model for camera, including defocus and motion blur MRI, PET, X-Ray tomography... u often modeled by Additive White Gaussian Noise, but can be Poisson, Poisson Gauss, Rician, etc. Enforced sparsity: ˆx = argmin x x 0 + λ Hx y 2 If we know x to be sparse (many zero elements) in some space (e.g. Wavelets). Highly non-convex. H. Talbot : Optimisation 24/59

27 Formulations in imaging Recovery We seek to recover a good approximation ˆx of x from H and y. H can be: Model for camera, including defocus and motion blur MRI, PET, X-Ray tomography... u often modeled by Additive White Gaussian Noise, but can be Poisson, Poisson Gauss, Rician, etc. Compressive sensing: ˆx = argmin x x 1 + λ Hx y 2 If we know x to be sparse (many zero elements) in some space (e.g. Wavelets). Smallest convex approximation of the l 0 pseudo-norm. H. Talbot : Optimisation 24/59

28 Formulations in imaging Formal context Penalized optimization problem Find where ( ) min F (x) = Φ(Hx y) + λr(x), x R N Φ Fidelity to data term, related to noise R Regularization term, related to some a priori assumptions λ Regularization weight Here, x is sparse in a dictionary V of analysis vectors in R N F 0 (x) = Φ(Hx y)+λ l 0 (V x) H. Talbot : Optimisation 25/59

29 Formulations in imaging Formal context Penalized optimization problem Find where ( ) min F (x) = Φ(Hx y) + λr(x), x R N Φ Fidelity to data term, related to noise R Regularization term, related to some a priori assumptions λ Regularization weight Here, x is sparse in a dictionary V of analysis vectors in R N F δ (x) = Φ(Hx y)+λ C c=1 ψ δ (V c x) H. Talbot where : Optimisation ψ δ is a differentiable, non-convex approximation of the 25/59 l 0 norm.

30 Formulations in imaging Benefits and drawbacks of the continuous approach pros flexible theory (not just denoising; deblurring, tomography, MRI reconstruction, etc) large library of algorithms, many more than in the discrete case isotropic convergence proofs and characterization of solutions. cons non-explicit discretization non-flexible structure deriving projections operators sometimes inefficient or impossible conditions for convergence. H. Talbot : Optimisation 26/59

31 Formulations in imaging Discrete and continuous approaches Both the previous discrete and continuous formulation have a MAP interpretation. Total Variation (TV) minimization: good regularization tool Weighted TV : penalization of the gradient leading to improved results Our contribution General combinatorial formulation of the dual TV problem : easily suitable to various graphs Generic constraint in the dual problem : more flexible penalization of the gradient sharper results H. Talbot : Optimisation 27/59

32 Formulations in imaging Outline 1 Generalization of TV models 2 Parallel Proximal Algorithm as an efficient solver 3 Results H. Talbot : Optimisation 28/59

33 DCTV model Total variation regularization Given an original image f Deduce a restored image u Weighted anisotropic TV model [Gilboa and Osher 2007] ( 1/2dx min w x,y (u y u x ) dy) (u x f x ) 2 dx u 2λ }{{}}{{} regularization R(u) data fidelity Φ(u) where λ ]0, + [ regularization parameter H. Talbot : Optimisation 29/59

34 DCTV model Equivalent dual formulation Weighted anisotropic TV model [Gilboa and Osher 2007] ( 1/2dx min w x,y (u y u x ) dy) 2 + Φ(u) u is equivalent [Chan, Golub, Mulet 1999] to the min-max problem min max wx,y 1/2 (u y u x )p x,y dxdy + Φ(u) u p 1 with p a projection vector field. Main idea p was introduced in practice to compute a faster solution constraining p can promote better results H. Talbot : Optimisation 30/59

35 DCTV model Segmentation Same model as denoising, with a labeled fidelity term Same regularisation. This includes very widespread models such as watershed, region growing, minimal curves and surfaces, geodesic active contours, and more. H. Talbot : Optimisation 31/59

36 DCTV model Deblurring, tomography Deblurring / tomography simply composes a linear term within the fidelity. Same model for regularization as before Possible to do very advanced applications: local tomography, angular integration tomography, dual image deblurring, etc. Also applicable with wavelets, etc. Any linear operator can serve. H. Talbot : Optimisation 32/59

37 Generalization of TV models Discrete formulation on graphs - notations Graph of N vertices, M edges Incidence matrix A R M N A = p 1 p 2 p 3 p 4 e e e e e A gradient operator A divergence operator allows general formulation of problems on arbitrary graphs For more details: L. Grady and J.R. Polimeni, Discrete Calculus: Applied Analysis on Graphs for Computational Science, Springer, H. Talbot : Optimisation 33/59

38 Generalization of TV models Discrete formulations of TV and its dual Let u R N be the restored image. [Bougleux et al. 2007] min u n ( ) 1/2 w i,j (u j u i ) 2 + Φ(u) j N i i=1 where N i = {j {1,..., n} e i,j E}. We introduce the following combinatorial formulation for the primal dual problem min u max p ((Au) w) + Φ(u) p 1, p R M H. Talbot : Optimisation 34/59

39 Generalization of TV models Dual constrained TV based formulation Constraining the projection vector Introducing the projection vector F R M = p w Constraining F to belong to a convex set C min sup u R N F C F (Au) }{{} regularization + 1 2λ u f 2 2 }{{} data fidelity C = m 1 i=1 C i where C 1,..., C m 1 closed convex sets of R M. Given g R N, θ i R M, α 1, C i = {F R M θ i F α g i }. H. Talbot : Optimisation 35/59

40 Generalization of TV models Dual constrained TV based formulation min sup u R N F C F (Au) }{{} regularization + 1 2λ u f 2 2 }{{} data fidelity C = m 1 i=1 C i, C i = {F R M θ i F α g i }, α 1. Example adapted to image denoising g i R N weight on vertex i, inversely function of the gradient of f at node i. Flat area : weak gradient strong g i strong F i,j weak local variations of u. Contours : strong gradient weak g i weak F i,j large local variations of u allowed. g j4 Fj4,i Fj1,i g j1 g i g j3 Fj3,i Fj2,i g j2 C i = {F R M Fj,i 2 g i } j N i H. Talbot : Optimisation 36/59

41 Generalization of TV models Illustration of constraining flow Illustration of constraining flow. H. Talbot : Optimisation 37/59

42 Generalization of TV models Sharper results Noisy image DCTV Weighted TV H. Talbot : Optimisation 38/59

43 Generalization of TV models Extension of our DCTV based formulation min sup u R N F C f R Q, observed image u R N, restored image F (Au) }{{} regularization + 1 2λ u f 2 2 }{{} data fidelity F R M, dual solution : projection vector H. Talbot : Optimisation 39/59

44 Generalization of TV models Extension of our DCTV based formulation min sup u R N F C f R Q, observed image u R N, restored image F (Au) }{{} regularization + 1 2λ Hu f 2 2 }{{} data fidelity F R M, dual solution : projection vector H R Q N, degradation matrix H. Talbot : Optimisation 39/59

45 Generalization of TV models Extension of our DCTV based formulation min sup u R N F C F (Au) }{{} regularization f R Q, observed image u R N, restored image + 1 2λ Hu f η 2 Ku 2 }{{} data fidelity F R M, dual solution : projection vector H R Q N, degradation matrix K R N N : projection onto Ker H, η 0 H. Talbot : Optimisation 39/59

46 Generalization of TV models Extension of our DCTV based formulation min sup u R N F C F (Au) }{{} regularization f R Q, observed image u R N, restored image (Hu f ) Λ 1 (Hu f ) + η 2 Ku 2 }{{} data fidelity F R M, dual solution : projection vector H R Q N, degradation matrix K R N N, projection onto Ker H, η 0 Λ R Q Q, matrix of weights, positive definite H. Talbot : Optimisation 39/59

47 Generalization of TV models Primal formulation min u R N σ C (Au) + 1 }{{} 2 (Hu f ) Λ 1 (Hu f ) + η 2 Ku 2 }{{} regularization data fidelity C = m 1 i=1 C i where C 1,..., C m 1 closed convex sets of R M. σ C support function of the convex set C σ C : R M ], + ]: a sup F a. F C H. Talbot : Optimisation 40/59

48 Generalization of TV models Dual problem The problem admits a unique solution û. Fenchel-Rockafellar dual problem: min F R M m 1 i=1 ι Ci (F ) +f m (F ) }{{} f i (F ) where ι C is the indicator function of the convex C (equal to 0 inside C and + outside), f m : F 1 2 F AΓA F F AΓH Λ 1 f, and Γ = (H Λ 1 H + ηk) 1. If F is a solution to the dual problem, ( û = Γ H Λ 1 f A F ). H. Talbot : Optimisation 41/59

49 Generalization of TV models Families of algorithms in continuous optimization Contour-based algorithms Snakes Level sets Region-based algorithms Primal only algorithms Primal-dual algorithms H. Talbot : Optimisation 42/59

50 Generalization of TV models Parallel ProXimal Algorithm (PPXA) optimizing DCTV [2] γ > 0, ν ]0, 2[. Repeat until convergence For (in parallel) { r = 1,..., s + 1 P Cr (y r ) if r s π r = (γaγa + I ) 1 (γaγh Λ 1 f + y s+1) otherwise z = 2 s+1 (π π s+1 ) F For (in parallel) r = 1,..., s + 1 yr = y r + ν(z p r ) F = F + ν 2 (z F ) H. Talbot : Optimisation 43/59

51 Generalization of TV models Parallel ProXimal Algorithm (PPXA) optimizing DCTV [2] γ > 0, ν ]0, 2[. Repeat until convergence For (in parallel) { r = 1,..., s + 1 P Cr (y r ) if r s π r = (γaγa + I ) 1 (γaγh Λ 1 f + y s+1) otherwise z = 2 s+1 (π π s+1 ) F For (in parallel) r = 1,..., s + 1 yr = y r + ν(z p r ) F = F + ν 2 (z F ) Simple projections onto hyperspheres H. Talbot : Optimisation 43/59

52 Generalization of TV models Parallel ProXimal Algorithm (PPXA) optimizing DCTV [2] γ > 0, ν ]0, 2[. Repeat until convergence For (in parallel) { r = 1,..., s + 1 P Cr (y r ) if r s π r = (γaγa + I ) 1 (γaγh Λ 1 f + y s+1) otherwise z = 2 s+1 (π π s+1 ) F For (in parallel) r = 1,..., s + 1 yr = y r + ν(z p r ) F = F + ν 2 (z F ) Linear system resolution H. Talbot : Optimisation 43/59

53 Generalization of TV models Quantitative perfomances Speed : competitive with the most efficient algorithm for optimizing weighted TV Denoising a image with an Alternated Direction of Multiplier Method: 0.4 seconds with the Parallel Proximal Algorithm: 0.7 seconds Quantitative denoising experiments on standard images show improvements of SNR (from 0.2 to 0.5 db) for images corrupted with Gaussian noise of variance σ 2 from 5 to 25. H. Talbot : Optimisation 44/59

54 Applications in image processing Results in image denoising Original image Noisy SNR=10.1dB Weighted TV SNR=13.4dB DCTV SNR=13.8dB H. Talbot : Optimisation 45/59

55 Applications in image processing Results in image denoising Weighted TV SNR=13.4dB DCTV SNR=13.8dB H. Talbot : Optimisation 46/59

56 Applications in image processing Comparison with more standard TV Figure : Left hand side: Standard deviation of each test image compared with the standard deviation of the denoising results, averaged results with (σ 2 = 5, 10, 15, 20, 25, 50). Right hand side: mean SNR over the experiments, H. Talbot : Optimisation 47/59

57 Applications in image processing Image denoising and deconvolution Original Noisy, blurred DCTV image image SNR=12.3dB result SNR=17.2dB H. Talbot : Optimisation 48/59

58 Applications in image processing Image fusion Original Noisy blurry DCTV image SNR=7.2dB SNR=11.6dB SNR=16.3dB H. Talbot : Optimisation 49/59

59 Applications in image processing Mesh denoising Original Noisy DCTV regularization mesh mesh on spatial coordinates H. Talbot : Optimisation 50/59

60 Applications in image processing (d) Noisy sampled (e) Taubin filtered (f) DCTV result H. Talbot : SNR Optimisation = 22.1 db result [?] SNR = 19.4 db (λ = 0.5) SNR 51/59 = 23.3 db Irregular graph (a) Original (b) Bottlenosed (c) Sampled image dolphin structure image

61 Applications in image processing Non-local regularization Original image Noisy PSNR=28.1dB (a) Nonlocal graph (figure P. Coupé, [?] Figure : Example of Non-Local Graph. Nonlocal DCTV PSNR=35 db H. Talbot : Optimisation 52/59

62 Non-convex regularization Non-convex optimization The current frontier. Many interesting applications thought to be very hard to solve: blind deblurring Many current methods extend to the Non-Convex case Generally only a local minimum is reached, but this might be OK. The miimum might be of high quality : stochastic optimization. For instance: see results achieved by deep-learning methods. H. Talbot : Optimisation 53/59

63 Non-convex regularization l 2 -l 0 regularization functions We consider the following class of potential functions: 1 ( δ (0, + )) ψ δ is differentiable. 2 ( δ (0, + )) lim t ψ δ (t) = 1. 3 ( δ (0, + )) ψ δ (t) = O(t 2 ) for small t. Examples: ψ δ (t) = t2 2δ 2 +t 2 ψ δ (t) = 1 exp( t2 2δ 2 ) H. Talbot : Optimisation 54/59

64 Non-convex regularization Majorize-Minimize principle [Hunter04] Objective: Find ˆx Arg min x F δ (x) For all x, let Q(., x ) a tangent majorant of F δ at x i.e., Q(x, x ) F δ (x), Q(x, x ) = F δ (x ) x, MM algorithm: j {0,..., J}, x j+1 Arg min x Q(x, x j ) H. Talbot : Optimisation 55/59

65 Non-convex regularization Image reconstruction Original image x Noisy sinogram y SNR=25 db { H Radon projection matrix y = H x + u with u Gaussian noise ( ˆx Arg min 1 x 2 Hx y 2 + λ c ψ δ(vc x) ) Non convex penalty / convex penalty H. Talbot : Optimisation 56/59

66 Non-convex regularization Results: Non convex penalty Reconstructed image SNR = 20.4 db MM-MG algorithm: Convergence in 134 s H. Talbot : Optimisation 57/59

67 Non-convex regularization Results: Convex penalty Reconstructed image SNR = 18.4 db MM-MG algorithm: Convergence in 60 s H. Talbot : Optimisation 58/59

68 Conclusion Conclusion Optimization is a very powerful, general methodology We ve drawn a panorama of interesting methodologies in image processing Extension of TV models via dual formulations Many applications in inverse problems including segmentation Proposed algorithm efficiently solves convex and non-convex problems Application to arbitrary graphs Generally optimization problems are unsolvable without some regularity assumptions. There exist a trade-off between the generality of a framework and the efficiency of associated algorithms. On to new things: hierarchies of partitions. H. Talbot : Optimisation 59/59

69 Conclusion S.P. Boyd and L. Vandenberghe. Convex optimization. Cambridge Univ Press, P. L. Combettes and J.-C. Pesquet. Proximal splitting methods in signal processing. In H. H. Bauschke, R. Burachik, P. L. Combettes, V. Elser, D. R. Luke, and H. Wolkowicz, editors, Fixed-Point Algorithms for Inverse Problems in Science and Engineering. Springer-Verlag, New York, Andrey Nikolayevich Tikhonov. On the stability of inverse problems. In Dokl. Akad. Nauk SSSR, volume 39, pages , H. Talbot : Optimisation 59/59

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard

More information

A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing

A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing Emilie Chouzenoux emilie.chouzenoux@univ-mlv.fr Université Paris-Est Lab. d Informatique Gaspard

More information

Markov Random Fields

Markov Random Fields Markov Random Fields Umamahesh Srinivas ipal Group Meeting February 25, 2011 Outline 1 Basic graph-theoretic concepts 2 Markov chain 3 Markov random field (MRF) 4 Gauss-Markov random field (GMRF), and

More information

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection EUSIPCO 2015 1/19 A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection Jean-Christophe Pesquet Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est

More information

In collaboration with J.-C. Pesquet A. Repetti EC (UPE) IFPEN 16 Dec / 29

In collaboration with J.-C. Pesquet A. Repetti EC (UPE) IFPEN 16 Dec / 29 A Random block-coordinate primal-dual proximal algorithm with application to 3D mesh denoising Emilie CHOUZENOUX Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est, France Horizon Maths 2014

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Markov Random Fields for Computer Vision (Part 1)

Markov Random Fields for Computer Vision (Part 1) Markov Random Fields for Computer Vision (Part 1) Machine Learning Summer School (MLSS 2011) Stephen Gould stephen.gould@anu.edu.au Australian National University 13 17 June, 2011 Stephen Gould 1/23 Pixel

More information

6. Approximation and fitting

6. Approximation and fitting 6. Approximation and fitting Convex Optimization Boyd & Vandenberghe norm approximation least-norm problems regularized approximation robust approximation 6 Norm approximation minimize Ax b (A R m n with

More information

Inverse problem and optimization

Inverse problem and optimization Inverse problem and optimization Laurent Condat, Nelly Pustelnik CNRS, Gipsa-lab CNRS, Laboratoire de Physique de l ENS de Lyon Decembre, 15th 2016 Inverse problem and optimization 2/36 Plan 1. Examples

More information

Gauge optimization and duality

Gauge optimization and duality 1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel

More information

Discrete Inference and Learning Lecture 3

Discrete Inference and Learning Lecture 3 Discrete Inference and Learning Lecture 3 MVA 2017 2018 h

More information

Bayesian Paradigm. Maximum A Posteriori Estimation

Bayesian Paradigm. Maximum A Posteriori Estimation Bayesian Paradigm Maximum A Posteriori Estimation Simple acquisition model noise + degradation Constraint minimization or Equivalent formulation Constraint minimization Lagrangian (unconstraint minimization)

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in

More information

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Minru Bai(x T) College of Mathematics and Econometrics Hunan University Joint work with Xiongjun Zhang, Qianqian Shao June 30,

More information

Efficient Variational Inference in Large-Scale Bayesian Compressed Sensing

Efficient Variational Inference in Large-Scale Bayesian Compressed Sensing Efficient Variational Inference in Large-Scale Bayesian Compressed Sensing George Papandreou and Alan Yuille Department of Statistics University of California, Los Angeles ICCV Workshop on Information

More information

Convex Optimization: Applications

Convex Optimization: Applications Convex Optimization: Applications Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh Convex Optimization 1-75/36-75 Based on material from Boyd, Vandenberghe Norm Approximation minimize Ax b (A R m

More information

Energy Minimization via Graph Cuts

Energy Minimization via Graph Cuts Energy Minimization via Graph Cuts Xiaowei Zhou, June 11, 2010, Journal Club Presentation 1 outline Introduction MAP formulation for vision problems Min-cut and Max-flow Problem Energy Minimization via

More information

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March

More information

Part 7: Structured Prediction and Energy Minimization (2/2)

Part 7: Structured Prediction and Energy Minimization (2/2) Part 7: Structured Prediction and Energy Minimization (2/2) Colorado Springs, 25th June 2011 G: Worst-case Complexity Hard problem Generality Optimality Worst-case complexity Integrality Determinism G:

More information

Enhanced Compressive Sensing and More

Enhanced Compressive Sensing and More Enhanced Compressive Sensing and More Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Nonlinear Approximation Techniques Using L1 Texas A & M University

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

Solving DC Programs that Promote Group 1-Sparsity

Solving DC Programs that Promote Group 1-Sparsity Solving DC Programs that Promote Group 1-Sparsity Ernie Esser Contains joint work with Xiaoqun Zhang, Yifei Lou and Jack Xin SIAM Conference on Imaging Science Hong Kong Baptist University May 14 2014

More information

SPARSE SIGNAL RESTORATION. 1. Introduction

SPARSE SIGNAL RESTORATION. 1. Introduction SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

1 Sparsity and l 1 relaxation

1 Sparsity and l 1 relaxation 6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the

More information

ENERGY METHODS IN IMAGE PROCESSING WITH EDGE ENHANCEMENT

ENERGY METHODS IN IMAGE PROCESSING WITH EDGE ENHANCEMENT ENERGY METHODS IN IMAGE PROCESSING WITH EDGE ENHANCEMENT PRASHANT ATHAVALE Abstract. Digital images are can be realized as L 2 (R 2 objects. Noise is introduced in a digital image due to various reasons.

More information

Accelerated MRI Image Reconstruction

Accelerated MRI Image Reconstruction IMAGING DATA EVALUATION AND ANALYTICS LAB (IDEAL) CS5540: Computational Techniques for Analyzing Clinical Data Lecture 15: Accelerated MRI Image Reconstruction Ashish Raj, PhD Image Data Evaluation and

More information

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems)

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Donghwan Kim and Jeffrey A. Fessler EECS Department, University of Michigan

More information

Submodularity in Machine Learning

Submodularity in Machine Learning Saifuddin Syed MLRG Summer 2016 1 / 39 What are submodular functions Outline 1 What are submodular functions Motivation Submodularity and Concavity Examples 2 Properties of submodular functions Submodularity

More information

arxiv: v1 [cs.it] 21 Feb 2013

arxiv: v1 [cs.it] 21 Feb 2013 q-ary Compressive Sensing arxiv:30.568v [cs.it] Feb 03 Youssef Mroueh,, Lorenzo Rosasco, CBCL, CSAIL, Massachusetts Institute of Technology LCSL, Istituto Italiano di Tecnologia and IIT@MIT lab, Istituto

More information

Multiresolution Graph Cut Methods in Image Processing and Gibbs Estimation. B. A. Zalesky

Multiresolution Graph Cut Methods in Image Processing and Gibbs Estimation. B. A. Zalesky Multiresolution Graph Cut Methods in Image Processing and Gibbs Estimation B. A. Zalesky 1 2 1. Plan of Talk 1. Introduction 2. Multiresolution Network Flow Minimum Cut Algorithm 3. Integer Minimization

More information

Markov Random Fields and Bayesian Image Analysis. Wei Liu Advisor: Tom Fletcher

Markov Random Fields and Bayesian Image Analysis. Wei Liu Advisor: Tom Fletcher Markov Random Fields and Bayesian Image Analysis Wei Liu Advisor: Tom Fletcher 1 Markov Random Field: Application Overview Awate and Whitaker 2006 2 Markov Random Field: Application Overview 3 Markov Random

More information

Convex Optimization and l 1 -minimization

Convex Optimization and l 1 -minimization Convex Optimization and l 1 -minimization Sangwoon Yun Computational Sciences Korea Institute for Advanced Study December 11, 2009 2009 NIMS Thematic Winter School Outline I. Convex Optimization II. l

More information

LPA-ICI Applications in Image Processing

LPA-ICI Applications in Image Processing LPA-ICI Applications in Image Processing Denoising Deblurring Derivative estimation Edge detection Inverse halftoning Denoising Consider z (x) =y (x)+η (x), wherey is noise-free image and η is noise. assume

More information

Sparsifying Transform Learning for Compressed Sensing MRI

Sparsifying Transform Learning for Compressed Sensing MRI Sparsifying Transform Learning for Compressed Sensing MRI Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laborarory University of Illinois

More information

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING JIAN-FENG CAI, STANLEY OSHER, AND ZUOWEI SHEN Abstract. Real images usually have sparse approximations under some tight frame systems derived

More information

Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials

Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials by Phillip Krahenbuhl and Vladlen Koltun Presented by Adam Stambler Multi-class image segmentation Assign a class label to each

More information

Part 6: Structured Prediction and Energy Minimization (1/2)

Part 6: Structured Prediction and Energy Minimization (1/2) Part 6: Structured Prediction and Energy Minimization (1/2) Providence, 21st June 2012 Prediction Problem Prediction Problem y = f (x) = argmax y Y g(x, y) g(x, y) = p(y x), factor graphs/mrf/crf, g(x,

More information

Dual methods for the minimization of the total variation

Dual methods for the minimization of the total variation 1 / 30 Dual methods for the minimization of the total variation Rémy Abergel supervisor Lionel Moisan MAP5 - CNRS UMR 8145 Different Learning Seminar, LTCI Thursday 21st April 2016 2 / 30 Plan 1 Introduction

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample

More information

A General Framework for a Class of Primal-Dual Algorithms for TV Minimization

A General Framework for a Class of Primal-Dual Algorithms for TV Minimization A General Framework for a Class of Primal-Dual Algorithms for TV Minimization Ernie Esser UCLA 1 Outline A Model Convex Minimization Problem Main Idea Behind the Primal Dual Hybrid Gradient (PDHG) Method

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with H. Bauschke and J. Bolte Optimization

More information

CS-E4830 Kernel Methods in Machine Learning

CS-E4830 Kernel Methods in Machine Learning CS-E4830 Kernel Methods in Machine Learning Lecture 3: Convex optimization and duality Juho Rousu 27. September, 2017 Juho Rousu 27. September, 2017 1 / 45 Convex optimization Convex optimisation This

More information

10. Multi-objective least squares

10. Multi-objective least squares L Vandenberghe ECE133A (Winter 2018) 10 Multi-objective least squares multi-objective least squares regularized data fitting control estimation and inversion 10-1 Multi-objective least squares we have

More information

Functionals for signal and image reconstruction: properties of their minimizers and applications

Functionals for signal and image reconstruction: properties of their minimizers and applications Functionals for signal and image reconstruction: properties of their minimizers and applications Research report Mila NIKOLOVA Centre de Mathématiques et de Leurs Applications (CMLA) - CNRS-UMR 8536 -

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Image restoration: edge preserving

Image restoration: edge preserving Image restoration: edge preserving Convex penalties Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP 1 / 32 Topics Image restoration,

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information

(4D) Variational Models Preserving Sharp Edges. Martin Burger. Institute for Computational and Applied Mathematics

(4D) Variational Models Preserving Sharp Edges. Martin Burger. Institute for Computational and Applied Mathematics (4D) Variational Models Preserving Sharp Edges Institute for Computational and Applied Mathematics Intensity (cnt) Mathematical Imaging Workgroup @WWU 2 0.65 0.60 DNA Akrosom Flagellum Glass 0.55 0.50

More information

CSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1

CSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1 CSCI5654 (Linear Programming, Fall 2013) Lectures 10-12 Lectures 10,11 Slide# 1 Today s Lecture 1. Introduction to norms: L 1,L 2,L. 2. Casting absolute value and max operators. 3. Norm minimization problems.

More information

Lecture Notes 9: Constrained Optimization

Lecture Notes 9: Constrained Optimization Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form

More information

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization / Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear

More information

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem

More information

A Graph Cut Algorithm for Generalized Image Deconvolution

A Graph Cut Algorithm for Generalized Image Deconvolution A Graph Cut Algorithm for Generalized Image Deconvolution Ashish Raj UC San Francisco San Francisco, CA 94143 Ramin Zabih Cornell University Ithaca, NY 14853 Abstract The goal of deconvolution is to recover

More information

Image Noise: Detection, Measurement and Removal Techniques. Zhifei Zhang

Image Noise: Detection, Measurement and Removal Techniques. Zhifei Zhang Image Noise: Detection, Measurement and Removal Techniques Zhifei Zhang Outline Noise measurement Filter-based Block-based Wavelet-based Noise removal Spatial domain Transform domain Non-local methods

More information

Does Better Inference mean Better Learning?

Does Better Inference mean Better Learning? Does Better Inference mean Better Learning? Andrew E. Gelfand, Rina Dechter & Alexander Ihler Department of Computer Science University of California, Irvine {agelfand,dechter,ihler}@ics.uci.edu Abstract

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)

More information

Adaptive Primal Dual Optimization for Image Processing and Learning

Adaptive Primal Dual Optimization for Image Processing and Learning Adaptive Primal Dual Optimization for Image Processing and Learning Tom Goldstein Rice University tag7@rice.edu Ernie Esser University of British Columbia eesser@eos.ubc.ca Richard Baraniuk Rice University

More information

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique Master 2 MathBigData S. Gaïffas 1 3 novembre 2014 1 CMAP - Ecole Polytechnique 1 Supervised learning recap Introduction Loss functions, linearity 2 Penalization Introduction Ridge Sparsity Lasso 3 Some

More information

Continuous State MRF s

Continuous State MRF s EE64 Digital Image Processing II: Purdue University VISE - December 4, Continuous State MRF s Topics to be covered: Quadratic functions Non-Convex functions Continuous MAP estimation Convex functions EE64

More information

Graphical Models for Collaborative Filtering

Graphical Models for Collaborative Filtering Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,

More information

Variable Metric Forward-Backward Algorithm

Variable Metric Forward-Backward Algorithm Variable Metric Forward-Backward Algorithm 1/37 Variable Metric Forward-Backward Algorithm for minimizing the sum of a differentiable function and a convex function E. Chouzenoux in collaboration with

More information

Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization

Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization 1/33 Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization Xiaoqun Zhang xqzhang@sjtu.edu.cn Department of Mathematics/Institute of Natural Sciences, Shanghai Jiao Tong University

More information

Probabilistic Graphical Models Lecture Notes Fall 2009

Probabilistic Graphical Models Lecture Notes Fall 2009 Probabilistic Graphical Models Lecture Notes Fall 2009 October 28, 2009 Byoung-Tak Zhang School of omputer Science and Engineering & ognitive Science, Brain Science, and Bioinformatics Seoul National University

More information

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization Hongchao Zhang hozhang@math.lsu.edu Department of Mathematics Center for Computation and Technology Louisiana State

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Noise, Denoising, and Image Reconstruction with Noise (lecture 10)

EE 367 / CS 448I Computational Imaging and Display Notes: Noise, Denoising, and Image Reconstruction with Noise (lecture 10) EE 367 / CS 448I Computational Imaging and Display Notes: Noise, Denoising, and Image Reconstruction with Noise (lecture 0) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

Recent Advances in Bayesian Inference for Inverse Problems

Recent Advances in Bayesian Inference for Inverse Problems Recent Advances in Bayesian Inference for Inverse Problems Felix Lucka University College London, UK f.lucka@ucl.ac.uk Applied Inverse Problems Helsinki, May 25, 2015 Bayesian Inference for Inverse Problems

More information

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting

More information

Pushmeet Kohli Microsoft Research

Pushmeet Kohli Microsoft Research Pushmeet Kohli Microsoft Research E(x) x in {0,1} n Image (D) [Boykov and Jolly 01] [Blake et al. 04] E(x) = c i x i Pixel Colour x in {0,1} n Unary Cost (c i ) Dark (Bg) Bright (Fg) x* = arg min E(x)

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Max Margin-Classifier

Max Margin-Classifier Max Margin-Classifier Oliver Schulte - CMPT 726 Bishop PRML Ch. 7 Outline Maximum Margin Criterion Math Maximizing the Margin Non-Separable Data Kernels and Non-linear Mappings Where does the maximization

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 27 Introduction Fredholm first kind integral equation of convolution type in one space dimension: g(x) = 1 k(x x )f(x

More information

Tightness of LP Relaxations for Almost Balanced Models

Tightness of LP Relaxations for Almost Balanced Models Tightness of LP Relaxations for Almost Balanced Models Adrian Weller University of Cambridge AISTATS May 10, 2016 Joint work with Mark Rowland and David Sontag For more information, see http://mlg.eng.cam.ac.uk/adrian/

More information

Variational Image Restoration

Variational Image Restoration Variational Image Restoration Yuling Jiao yljiaostatistics@znufe.edu.cn School of and Statistics and Mathematics ZNUFE Dec 30, 2014 Outline 1 1 Classical Variational Restoration Models and Algorithms 1.1

More information

TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS

TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS Martin Kleinsteuber and Simon Hawe Department of Electrical Engineering and Information Technology, Technische Universität München, München, Arcistraße

More information

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm

MMSE Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Denoising of 2-D Signals Using Consistent Cycle Spinning Algorithm Bodduluri Asha, B. Leela kumari Abstract: It is well known that in a real world signals do not exist without noise, which may be negligible

More information

Minimizing Isotropic Total Variation without Subiterations

Minimizing Isotropic Total Variation without Subiterations MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Minimizing Isotropic Total Variation without Subiterations Kamilov, U. S. TR206-09 August 206 Abstract Total variation (TV) is one of the most

More information

Sparse molecular image representation

Sparse molecular image representation Sparse molecular image representation Sofia Karygianni a, Pascal Frossard a a Ecole Polytechnique Fédérale de Lausanne (EPFL), Signal Processing Laboratory (LTS4), CH-115, Lausanne, Switzerland Abstract

More information

Statistical Data Mining and Machine Learning Hilary Term 2016

Statistical Data Mining and Machine Learning Hilary Term 2016 Statistical Data Mining and Machine Learning Hilary Term 2016 Dino Sejdinovic Department of Statistics Oxford Slides and other materials available at: http://www.stats.ox.ac.uk/~sejdinov/sdmml Naïve Bayes

More information

Compressed Sensing: Extending CLEAN and NNLS

Compressed Sensing: Extending CLEAN and NNLS Compressed Sensing: Extending CLEAN and NNLS Ludwig Schwardt SKA South Africa (KAT Project) Calibration & Imaging Workshop Socorro, NM, USA 31 March 2009 Outline 1 Compressed Sensing (CS) Introduction

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

Introduction to Machine Learning. Lecture 2

Introduction to Machine Learning. Lecture 2 Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for

More information

CSC 412 (Lecture 4): Undirected Graphical Models

CSC 412 (Lecture 4): Undirected Graphical Models CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:

More information

Discriminative Fields for Modeling Spatial Dependencies in Natural Images

Discriminative Fields for Modeling Spatial Dependencies in Natural Images Discriminative Fields for Modeling Spatial Dependencies in Natural Images Sanjiv Kumar and Martial Hebert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 {skumar,hebert}@ri.cmu.edu

More information

Investigating the Influence of Box-Constraints on the Solution of a Total Variation Model via an Efficient Primal-Dual Method

Investigating the Influence of Box-Constraints on the Solution of a Total Variation Model via an Efficient Primal-Dual Method Article Investigating the Influence of Box-Constraints on the Solution of a Total Variation Model via an Efficient Primal-Dual Method Andreas Langer Department of Mathematics, University of Stuttgart,

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a

More information

Logistic Regression. William Cohen

Logistic Regression. William Cohen Logistic Regression William Cohen 1 Outline Quick review classi5ication, naïve Bayes, perceptrons new result for naïve Bayes Learning as optimization Logistic regression via gradient ascent Over5itting

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Least squares regularized or constrained by L0: relationship between their global minimizers. Mila Nikolova

Least squares regularized or constrained by L0: relationship between their global minimizers. Mila Nikolova Least squares regularized or constrained by L0: relationship between their global minimizers Mila Nikolova CMLA, CNRS, ENS Cachan, Université Paris-Saclay, France nikolova@cmla.ens-cachan.fr SIAM Minisymposium

More information

Dictionary Learning for photo-z estimation

Dictionary Learning for photo-z estimation Dictionary Learning for photo-z estimation Joana Frontera-Pons, Florent Sureau, Jérôme Bobin 5th September 2017 - Workshop Dictionary Learning on Manifolds MOTIVATION! Goal : Measure the radial positions

More information

Noise, Image Reconstruction with Noise!

Noise, Image Reconstruction with Noise! Noise, Image Reconstruction with Noise! EE367/CS448I: Computational Imaging and Display! stanford.edu/class/ee367! Lecture 10! Gordon Wetzstein! Stanford University! What s a Pixel?! photon to electron

More information

Statistics 352: Spatial statistics. Jonathan Taylor. Department of Statistics. Models for discrete data. Stanford University.

Statistics 352: Spatial statistics. Jonathan Taylor. Department of Statistics. Models for discrete data. Stanford University. 352: 352: Models for discrete data April 28, 2009 1 / 33 Models for discrete data 352: Outline Dependent discrete data. Image data (binary). Ising models. Simulation: Gibbs sampling. Denoising. 2 / 33

More information

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen Advanced Machine Learning Lecture 3 Linear Regression II 02.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de This Lecture: Advanced Machine Learning Regression

More information

Link lecture - Lagrange Multipliers

Link lecture - Lagrange Multipliers Link lecture - Lagrange Multipliers Lagrange multipliers provide a method for finding a stationary point of a function, say f(x, y) when the variables are subject to constraints, say of the form g(x, y)

More information