Primal-dual fixed point algorithms for separable minimization problems and their applications in imaging

Similar documents
A primal dual fixed point algorithm for convex separable minimization with applications to image restoration

A General Framework for a Class of Primal-Dual Algorithms for TV Minimization

A Primal-dual Three-operator Splitting Scheme

Primal Dual Methods in Imaging Science

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR TV MINIMIZATION

Primal-dual algorithms for the sum of two and three functions 1

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR CONVEX OPTIMIZATION IN IMAGING SCIENCE

A primal-dual fixed point algorithm for multi-block convex minimization *

Coordinate Update Algorithm Short Course Operator Splitting

Solving DC Programs that Promote Group 1-Sparsity

A General Framework for a Class of First Order Primal Dual Algorithms for Convex Optimization in Imaging Science

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables

Adaptive Primal Dual Optimization for Image Processing and Learning

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Sparse Optimization Lecture: Dual Methods, Part I

On the equivalence of the primal-dual hybrid gradient method and Douglas Rachford splitting

Beyond Heuristics: Applying Alternating Direction Method of Multipliers in Nonconvex Territory

A Unified Primal-Dual Algorithm Framework Based on Bregman Iteration

Tight Rates and Equivalence Results of Operator Splitting Schemes

Accelerated primal-dual methods for linearly constrained convex problems

Proximal splitting methods on convex problems with a quadratic term: Relax!

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING

arxiv: v4 [math.oc] 29 Jan 2018

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

About Split Proximal Algorithms for the Q-Lasso

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization

Operator Splitting for Parallel and Distributed Optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Optimization for Learning and Big Data

Splitting methods for decomposing separable convex programs

A Tutorial on Primal-Dual Algorithm

Lecture: Algorithms for Compressed Sensing

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches

Dual and primal-dual methods

Minimizing the Difference of L 1 and L 2 Norms with Applications

An Algorithmic Framework of Generalized Primal-Dual Hybrid Gradient Methods for Saddle Point Problems

Asynchronous Algorithms for Conic Programs, including Optimal, Infeasible, and Unbounded Ones

On the equivalence of the primal-dual hybrid gradient method and Douglas-Rachford splitting

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise

Variational Image Restoration

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

A Unified Approach to Proximal Algorithms using Bregman Distance

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Conductivity imaging from one interior measurement

You should be able to...

Low Complexity Regularization 1 / 33

INERTIAL PRIMAL-DUAL ALGORITHMS FOR STRUCTURED CONVEX OPTIMIZATION

Enhanced Compressive Sensing and More

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION

A New Primal Dual Algorithm for Minimizing the Sum of Three Functions with a Linear Operator

A GENERAL INERTIAL PROXIMAL POINT METHOD FOR MIXED VARIATIONAL INEQUALITY PROBLEM

Convergence analysis for a primal-dual monotone + skew splitting algorithm with applications to total variation minimization

The Direct Extension of ADMM for Multi-block Convex Minimization Problems is Not Necessarily Convergent

The Alternating Direction Method of Multipliers

Does Alternating Direction Method of Multipliers Converge for Nonconvex Problems?

Contraction Methods for Convex Optimization and monotone variational inequalities No.12

INERTIAL ACCELERATED ALGORITHMS FOR SOLVING SPLIT FEASIBILITY PROBLEMS. Yazheng Dang. Jie Sun. Honglei Xu

First-order methods for structured nonsmooth optimization

Primal-dual coordinate descent

Constrained optimization

Distributed Optimization via Alternating Direction Method of Multipliers

A Linearly Convergent First-order Algorithm for Total Variation Minimization in Image Processing

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11

Iterative regularization of nonlinear ill-posed problems in Banach space

ADMM for monotone operators: convergence analysis and rates

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection

Gauge optimization and duality

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy

On the acceleration of the double smoothing technique for unconstrained convex optimization problems

M. Marques Alves Marina Geremia. November 30, 2017

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization

Douglas-Rachford Splitting: Complexity Estimates and Accelerated Variants

On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems)

Primal-dual coordinate descent A Coordinate Descent Primal-Dual Algorithm with Large Step Size and Possibly Non-Separable Functions

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

ALTERNATING DIRECTION METHOD OF MULTIPLIERS WITH VARIABLE STEP SIZES

Preconditioned ADMM with nonlinear operator constraint

SIAM Conference on Imaging Science, Bologna, Italy, Adaptive FISTA. Peter Ochs Saarland University

arxiv: v1 [math.na] 2 Nov 2015

arxiv: v1 [math.oc] 20 Jun 2014

Math 273a: Optimization Convex Conjugacy

Adaptive Primal-Dual Hybrid Gradient Methods for Saddle-Point Problems

Convergence of Fixed-Point Iterations

Linearized Alternating Direction Method of Multipliers via Positive-Indefinite Proximal Regularization for Convex Programming.

Optimal Linearized Alternating Direction Method of Multipliers for Convex Programming 1

arxiv: v2 [math.oc] 28 Jan 2019

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

Bias-free Sparse Regression with Guaranteed Consistency

Primal Dual Algorithms for Convex Optimization in Imaging Science

arxiv: v3 [math.oc] 29 Jun 2016

ACCELERATED LINEARIZED BREGMAN METHOD. June 21, Introduction. In this paper, we are interested in the following optimization problem.

On convergence rate of the Douglas-Rachford operator splitting method

Variable Metric Forward-Backward Algorithm

SPARSE SIGNAL RESTORATION. 1. Introduction

Image restoration by minimizing zero norm of wavelet frame coefficients

FAST ALTERNATING DIRECTION OPTIMIZATION METHODS

Transcription:

1/38 Primal-dual fixed point algorithms for separable minimization problems and their applications in imaging Xiaoqun Zhang Department of Mathematics and Institute of Natural Sciences Shanghai Jiao Tong University (SJTU) Joint work with Peijun Chen (SJTU), Jianguo Huang(SJTU) French-German Conference on Mathematical Image Analysis IHP, Paris, Jan 13-15, 2014

2/38 Outline Background Primal dual fixed point algorithm Extensions Conclusions

Background 3/38 Background

Background 4/38 General convex separable minimization model x = arg min x R n (f 1 B)(x) + f 2 (x) B : R n R m a linear transform. f 1, f 2 are proper l.s.c convex function defined in a Hilbert space. f 2 is differentiable on R n with a 1/β-Lipschitz continuous gradient for some β (0, + )

Background 5/38 Operator splitting methods min x X f(x) + h(x) where f, h are proper l.s.c. convex and h is differentiable on X with a 1/β-Lipschitz continuous gradient Define proximal operator prox f as prox f : X X x arg min f(y) + 1 2 x y 2 2, y X Proximal forward-backward splitting (PFBS) 1 for 0 < γ < 2β, x k+1 = prox γf (x k γ h(x k )), Many more other variants and related work (partial list: ISTA, FPCA, FISTA,GFB...). An important assumption: prox f (x) is an easy problem!! 1 Moreau 1962; Lions-Mercier; 1979, Combettes- Wajs, 2005; FPC

Background 6/38 Soft shrinkage For f(x) = µ x 1, prox f (x) = sign(x) max( x µ, 0) Componentwise thus efficient. Generally no easy form for prox Bx 1 (x) for non-invertible B. Largely used in compressive sensing and imaging sciences. Similar formula for matrix nuclear norm minimization (restore low rank matrix) or other matrix sparsity.

Background 7/38 Methods on splitting form (SP P ) Augmented Lagrangian: max inf y x,z {f 1 (z) + f 2 (x) + y, Bx z } L ν (x, z; y) = f 1 (z) + f 2 (x) + y, Bx z + ν Bx z 2 2 Alternating direction of multiplier method (ADMM) (Glowinski et al. 75, Gabay et al. 83) x k+1 = arg min x L ν (x, z k ; y k ) z k+1 = arg min z L ν (x k+1, z; y k ) y k+1 = y k + γν(bx k+1 z k+1 ) Split Inexact Uzawa (SIU/BOS) Method 2 x k+1 = arg min u L ν (x, z k ; y k ) + x x k 2 D 1 z k+1 = arg min z L ν (x k+1, z; y k ) + z z k 2 D 2 Cy k+1 = Cy k + (Bx k+1 z k+1 ) 2 Zhang-Burger-Osher,2011

Background 8/38 Methods for Primal-Dual from (PD) (P D) : min x sup f1 (y) + y, Bu + f 2 (x) y Primal-dual hybrid Gradient (PDHG) Method (Zhu-Chan, 2008) y k+1 = arg max f y 1 (y) + y, Bx k 1 y y k 2 2 2δ k x k+1 = arg min f 2 (x) + B T y k+1, x + 1 x x k 2 x 2 2α k Modified PDHG (PDHGMp, Esser-Zhang-Chan,2010, equivalent to SIU on (SPP); Pock-Cremers-Bischof-Chambolle, 2009, Chambolle-Pock, 2011(θ = 1)) Replace p k in first step of PDHG with 2p k p k 1 to get PDHGMp: x k+1 = arg min x f 2 (x) + B T ( 2y k y k 1), x y k+1 = arg min y f 1 (y) y, Bx k+1 + 1 2δ y yk 2 2 + 1 2α x xk 2 2

Background Connections 3 14 (P) minu FP (u) FP (u) =J(Au)+H(u) E. Esser, X. Zhang and T.F. Chan (D) maxp FD(p) FD(p) = J (p) H ( A T p) (PD) minu sup p LPD(u, p) LPD(u, p) = p, Au J (p)+h(u) (SPP) maxp infu,w LP (u, w, p) LP (u, w, p) =J(w)+H(u)+ p, Au w AMA on (SPP) PFBS on (D) (SPD) maxu infp,y LD(p, y, u) LD(p, y, u) =J (p)+h (y)+ u, A T p y PFBS on (P) AMA on (SPD) + 1 u 2α uk 2 2 + 1 p 2δ pk 2 2 Relaxed AMA Relaxed AMA on (SPP) on (SPD) + δ Au 2 w 2 2 + α 2 AT p + y 2 2 Douglas ADMM on Rachford on (SPP) (D) Primal-Dual Proximal Point on (PD) = PDHG Douglas ADMM Rachford on on (SPD) (P) + 1 u 2 uk, ( 1 α δat A)(u u k ) p k+1 u k + 1 p 2 pk, ( 1 δ αaat )(p p k ) 2p k+1 p k 2u k u k 1 Split Inexact Uzawa on (SPP) PDHGMp PDHGMu Split Inexact Uzawa on (SPD) Legend: (P): Primal (D): Dual (PD): Primal-Dual (SPP): Split Primal (SPD): Split Dual AMA: Alternating Minimization Algorithm (4.2.1) PFBS: Proximal Forward Backward Splitting (4.2.1) ADMM: Alternating Direction Method of Multipliers (4.2.2) PDHG: Primal Dual Hybrid Gradient (4.2) PDHGM: Modified PDHG (4.2.3) Bold: Well Understood Convergence Properties 9/38

Background 10/38 First order methods in imaging sciences Many efficient algorithms exist: augmented lagrangian, splitting methods, alternating methods, primal-dual, fixed point methods etc. Huge number of seemingly related methods. Many of them requires subproblem solving involving inner-iterations, ad-hoc parameter selections, which can not be clearly controlled in real implementation. Need for methods with simple, explicit iterations capable of solving large scale, often non-differentiable convex models with separable structure Convergence analysis are mainly for the objective function, or in ergodic sense. Most of them have sublinear convergence (O(1/N) or O(1/N 2 )).

Primal dual fixed point algorithm 11/38 Primal dual fixed point methods

Primal dual fixed point algorithm 12/38 Fixed point algorithm based on proximity operator (FP 2 O) For a given b R n, solve for prox f1 B(b) H(v) = (I prox f 1λ )(Bb + (I λbb T )v) for all v R m FP 2 O (Micchelli-Shen-Xu, 11 ) Step 1: Set v 0 R m, 0 < λ < 2/λ max (BB T ), κ (0, 1). Step 2: Calculate v, which is the fixed point of H, with iteration v k+1 = H κ (v k ) where H κ = κi + (1 κ)h for κ (0, 1) Step 3: prox f1 B(b) = b λb T v.

Primal dual fixed point algorithm 13/38 Solve for general problem Solve for x = arg min x R n (f 1 B)(x) + f 2 (x) PFBS FP 2 O (Argyriou-Micchelli-Pontil-Shen-Xu,11 ) Step 1: Set x 0 R n, 0 < γ < 2β. Step 2: for k = 0, 1, 2, Calculate x k+1 = prox γf1 B(x k γ f 2 (x k )) using FP 2 O. end for Note: Inner iterations are involved, no clear stopping criteria and error analysis!

Primal dual fixed point algorithm 14/38 Primal dual fixed point algorithm Primal-dual fixed points algorithm based on proximity operator, PDFP 2 O. Step 1: Set x 0 R n, v 0 R m, 0 < λ 1/λ max (BB T ), 0 < γ < 2β. Step 2: for k = 0, 1, 2, x k+1/2 = x k γ f 2 (x k ), v k+1 = (I prox γ λ f 1 )(Bx k+1/2 + (I λbb T )v k ), x k+1 = x k+1/2 λb T v k+1. end for No inner iterations if prox γ λ f (x) is an easy problem. 1 Extension of FP2 2 O and PFBS. Can be extended to κ average fixed point iteration

Primal dual fixed point algorithm 15/38 Fixed point operator notion Define T 1 : R m R n R m as T 1 (v, x) = (I prox γ λ f 1 )(B(x γ f 2(x)) + (I λbb T )v) and T 2 : R m R n R n as Denote T : R m R n R m R n as T 2 (v, x) = x γ f 2 (x) λb T T 1. T (v, x) = (T 1 (v, x), T 2 (v, x)).

Primal dual fixed point algorithm 16/38 Convergence of PDFP2O Theorem Let λ > 0, γ > 0. Suppose that x is a solution of x = arg min x R n (f 1 B)(x) + f 2 (x) if and only if there exists v R m such that u = (v, x ) is a fixed point of T. Theorem Suppose 0 < γ < 2β, 0 < λ 1/λ max (BB T ) and κ [0, 1). Let u k = (v k, x k ) be a sequence generated by PDFP 2 O. Then {u k } converges to a fixed point of T and {x k } converges to a solution of problem x = arg min x R n (f 1 B)(x) + f 2 (x)

Primal dual fixed point algorithm 17/38 Convergence rate analysis Condition For 0 < γ < 2β and 0 < λ 1/λ max (BB T ), there exist η 1, η 2 [0, 1) such that I λbb T 2 η 2 1 and where g(x) = x γ f 2 (x). Remarks g(x) g(y) 2 η 2 x y 2 for all x, y R n, If B has full row rank, f 2 is strongly convex, this condition can be satisfied. As a typical example, consider f 2 (x) = 1 2 Ax b 2 2 with AT A full rank.

Primal dual fixed point algorithm 18/38 Linear convergence rate Theorem Suppose the above condition holds true. Let u k = (v k, x k ) be a fixed point iteration sequence of operator T. Then the sequence {u k } must converge to the unique fixed point u = (v, x ) R m R n of T, with x the unique solution of the minimization problem. Furthermore, x k x 2 cθk 1 θ, where c = u 1 u 0 λ, θ = κ + (1 κ)η (0, 1) and η = max{η 1, η 2 }. Let B = I, λ = 1, f 2 (x) strongly convex, then PFBS converges linearly. If B is full row rank, then PFP 2 O converges linearly. Related work: linear Convergence of the ADMM methods (Luo 2012, Deng-Yin, CAM12-52, Goldstein -O Donoghue -Setzer,2012)

Primal dual fixed point algorithm 19/38 Connection with other primal dual algorithms Table : The comparison between CP (θ = 1) and PDFP 2 O. PDHGm/CP (θ = 1) PDFP 2 O v k+1 = (I + σ f1 ) 1 (v k + v k+1 Form σby k ) By k ) = (I + λ γ f 1 ) 1 ( λ γ v k + x k+1 = (I + τ f 2 ) 1 (x k x k+1 = x k γ f 2 (x k ) τb T v k+1 ) γb T v k+1 y k+1 = 2x k+1 x k y k+1 = x k+1 γ f 2 (x k+1 ) γb T v k+1 1 Condition 0 < στ < λ max(bb T ) 1 0 < γ < 2β, 0 < λ λ max(bb T ) Relation σ = λ/γ, τ = γ

Primal dual fixed point algorithm 20/38 Connection with Splitting types of method for ν > 0. (ASB 4 ) SIU 5 : (1a) is replaced by min µ Bx 1 + 1 x R n 2 Ax b 2 2 x k+1 = (A T A + νb T B) 1 (A T b + νb T (d k v k )), d k+1 = prox 1 ν f (Bx 1 k+1 + v k ), v k+1 = v k (d k+1 Bx k+1 ), (1a) x k+1 = x k δa T (Ax k b) δνb T (Bx k d k + v k ) PDFP2O: (1a) is replaced by x k+1 = x k δa T (Ax k b) δνb T (Bx k d k + v k ) δ 2 νa T AB T (d k Bx k ) 4 Alternating split Bregman (Goldstein-Osher,08 ) (ADMM) 5 Split Inexact Uzawa, Zhang-Burger-Osher, 10

Primal dual fixed point algorithm 21/38 Example: Image restoration with total variation Image superresolution: subsampling operator A is implemented by taking the average of every d d pixels and sampling the average, if a zoom-in ratio d is desired. CT reconstruction: A is parallel beam radon transform. Parallel MRI: Observation model: b j = DF S j x + n where b j is the vector of measured Fourier coefficients at receiver j, D is a diagonal downsampling operator, F is the Fourier transform, S j corresponds to diagonal coil sensitivity mapping for receiver j and n is gaussian noise. Let A j = DF S j, we can recover images x by solving x = arg min x R n µ Bx 1 + 1 2 N A j x b j 2 2 j=1

Primal dual fixed point algorithm 22/38 Super-resolution Results Figure : Super resolution results from 128 128 image to 512 512 image by ASB, CP, SIU and PDFP 2 O corresponding to tolerance error ε = 10 4 Original Zooming ASB, PSNR=29.37 CP, PSNR=29.32 SIU, PSNR=29.31 PDFP 2 O, PSNR=29.32

Primal dual fixed point algorithm 23/38 Image superresolution Table : Performance comparison among ASB, CP, SIU and PDFP 2 O for image superresolution. For ASB, N I = 2, ν = 0.01. For CP, N I = 1, σ = 0.005. For SIU, δ = 24, ν = 0.006. For PDFP 2 O, γ = 30, λ = 1/6. ε = 10 2 ε = 10 3 ε = 10 4 ASB (21, 1.58, 29.34) (74, 5.55, 29.38) (254, 19.14, 29.37) CP (46, 1.94, 28.97) (150, 6.35, 29.24) (481, 20.37, 29.32) SIU (42, 1.14, 28.91) (141, 3.78, 29.22) (446, 12.45, 29.31) PDFP 2 O (38, 0.97, 28.98) (128, 3.25, 29.25) (417, 10.59,29.32)

Primal dual fixed point algorithm 24/38 Computerized tomography (CT) reconstruction Figure : A tomographic reconstruction example for a 128 128 image, with 50 projections corresponding to tolerance error ε = 10 4 Original FBP ASB, PSNR=32.43 CP, PSNR=32.32 SIU, PSNR=32.23 PDFP 2 O, PSNR=32.23

Primal dual fixed point algorithm 25/38 Computerized Tomography (CT) reconstruction Table : Performance evaluation comparison among ASB, CP, SIU and PDFP 2 O in CT reconstruction. For ASB, CP, SIU, empirically best parameter sets are used. ε = 10 3 ε = 10 4 ε = 10 5 ASB 1 (129, 2.72, 31.97) (279, 5.87, 32.43) (1068, 24.49, 32.45) ASB 2 (229, 4.79, 31.75) (345, 7.24, 32.26) (492, 10.33, 32.41) CP (167, 3.37, 31.80) (267, 5.40, 32.32) (588, 12.73, 32.45) SIU (655, 4.61, 31.70) (971, 6.84, 32.23) (1307, 9.20, 32.38) PDFP 2 O (655, 4.61, 31.70) (971, 6.81, 32.23) (1307, 9.15, 32.38)

Primal dual fixed point algorithm Parallel MRI 6 Figure : In-vivo MR images acquired eight-channel head data Coil 1 Coil 2 Coil 3 Coil 4 (b) Coil 5 Coil 6 Coil 7 Coil 8 6 Ji J X, Son J B and Rane S D 2007 PULSAR: a MATLAB toolbox for parallel magnetic resonance imaging using array coils and multiple channel receivers 26/38

50 100 150 200 250 50 100 150 200 250 Primal dual fixed point algorithm Parallel MRI reconstruction SOS SENSE SPACE-RIP GRAPPA R O S R O I R R=2 O N AP 0 0.001939 0.001939 0.001624 SNR 36.08 31.08 31.08 29.06 time 0.19 5.94 155.88 44.85 ASB CP SIU PDFP 2 O R=2 AP 0.000811 0.000823 0.000823 0.000822 SNR 39.20 39.27 39.26 39.36 time 1.81 3.72 1.87 1.84 27/38

Extensions 28/38 Extensions

Extensions 29/38 Extensions With convex constraints: arg min x C (f 1 B)(x) + f 2 (x), Nonconvex optimization arg min (f 1 B)(x) + f 2 (x), where f 2 (x) is nonconvex (for example nonlinear least square). Application: nonlinear inverse problem Acceleration and preconditioning.

Extensions With extra convex constraints arg min (f 1 B)(x) + f 2 (x), x C Let Proj C denote the projection onto the (closed) convex set C and Prox f1 denote the proximal point solution. Convert the problem to arg min ( f 1 B)(x) + f 2 (x), x ( ) where B B =, ( I f 1 B)(x) = (f 1 B)(x) + χ C (x). Apply PDFP2O, we obtain the scheme (infeasible scheme) v k+1 = (I prox γ λ f )(B(x 1 k γ f 2 (x k )) + (I λbb T )v k λby k ), y k+1 = (I proj C )((x k γ f 2 (x k )) λb T v k + (1 λ)y k ), x k+1 = x k γ f 2 (x k ) λb T v k+1 λy k+1. Convergence is derived directly from PDFP2O. 30/38

Extensions 31/38 Feasible iterative scheme y k+1 =Proj C (x k γ f 2 (x k ) λb T v k ) v k+1 =(I Prox γ λ f 1 )(Byk+1 + v k ) x k+1 =Proj C (x k γ f 2 (x k ) λb T v k+1 ) Equivalent iteration: y k+1 =Proj C (x k γ L(x k, v k )) v k+1 = arg max L(y k+1, v) γ 2λ v v k 2 x k+1 =Proj C (x k γ L(x k, v k+1 )) where L(x, v) = f 2 (x) f 1 (v) + Bx, v, v k = λ γ (v k)

Extensions 32/38 Convergence for feasible iterative scheme For x C, v S, define T 1 (x, v) = Prox λ T (x, v) : γ f ( λ 1 γ B Proj C(x γ f 2 (x) γb T v) + v) T 2 (x, v) = Proj C (x γ f 2 (x) γb T T 1 (x, v)) The solution pair (x, v ) satisfy T (x, v ) = (v, x ). If 0 < γ < 2β, 0 < λ < 1, then the sequence converges. BB T Related reference: Krol A Li S Shen L and Xu Y 2012 Preconditioned Alternating Projection Algorithms for Maximum a Posteriori ECT Reconstruction Inverse Problem 28 115005.

Extensions 33/38 Example: CT reconstruction with real data PDFP2O PDFP2OC 50 steps 100 steps Figure : Real CT reconstruction with nonnegative constraint (40 projections )

Extensions 34/38 Example: Parallel MRI with nonnegative constraint SNR=38.06, 20.1s SNR=39.55, 10.3s Figure : Subsampling R = 4

Extensions 35/38 Nonconvex optimization for nonlinear inverse problem Let f(x) = F (x) y 2 2. min x F (x) y 2 2 + λ W x 1, subject to l x u. y k = Prox γc (x k τ ( x f(x k ) + λµw T b k) ), h k = b k + W y k, b k+1 = h k shrink(h k, 1/µ), x k+1 = Prox γc (x k τ ( x f(x k ) + λµw T b k+1 ) ),

Extensions Application for Quantitative photo-acoustic reconstruction 8 Reconstruction time: 180 s v.s 887s SplitBregman (L-BFGS for the subproblem) 7 Figure : Quantitative photo-acoustic reconstruction for the absorption and (reduced) scattering. 7 H. Zhao, S. Osher and H. Zhao: Quantitative Photoacoustic Tomography, 2012 8 Ongoing joint work with X. Zhang, W Zhou and H. Gao 36/38

Conclusions 37/38 Summary and perspectives First order methods are efficient and enjoy comparative numerical performance, especially when the parameters are properly chosen. For both ASB and CP, inexact solver such as CG method can be applied in practice. However the choice of inner iteration is rather ad-hoc. For both SIU and PDFP 2 O, the iteration schemes are straightforward since only one inner iteration is used and easy to implement in practice and can be easily, especially when one extra constraint is present. It can be also extended to low rank matrix or other more complex sparsity reconstruction. Parameter choices for all the methods. The rules of parameter selection in PDFP 2 O has some advantages. The convergence and the relation of convergence rate and condition number of the operators is clearly stated under this theoretical framework. Future work: convergence for nonconvex problems; acceleration and preconditioning; parellel and distributed, decentralized computing.

Conclusions 38/38 Reference P. Chen, J. Huang and X. Zhang, A Primal-Dual Fixed Point Algorithm for Convex Separable Minimization with Applications to Image Restoration. Inverse problems, 29 (2), 2013 Contact: xqzhang@sjtu.edu.cn Supported by China-Germany-Finland joint grant, NSFC. Thank you for your attention!