Primal-dual algorithms for next-generation radio-interferometric imaging

Size: px
Start display at page:

Download "Primal-dual algorithms for next-generation radio-interferometric imaging"

Transcription

1 1/17 Primal-dual algorithms for next-generation radio-interferometric imaging Alexandru Onose 1,RafaelE.Carrillo 2,AbdullahAbdulaziz 1, Arwa Dabbech 1,JasonD.McEwen 3 and Yves Wiaux 1 1 Institute of Sensors, Signals and Systems, Heriot-Watt University, United Kingdom 2 Signal Processing Laboratory, Ecole Polytechnique Fédérale de Lausanne, Switzerland 3 Mullard Space Science Laboratory, University College London, United Kingdom

2 Next-generation radio-interferometric challenges 2/17 Increase the resolution and sensitivity up to two orders of magnitude over current instruments gigapixel images huge dynamic range

3 Next-generation radio-interferometric challenges 2/17 Increase the resolution and sensitivity up to two orders of magnitude over current instruments gigapixel images huge dynamic range Unprecedented amount of data to be processed Good reconstruction quality with scalable algorithms employing parallel and distributed processing

4 Measurement model Measurement equation yu) = Dl, u)xl)e 2iπu l d 2 l 3/17 Discretised version of the inverse problem y = Φx + n with Φ = GFZ x R N + the intensity image of interest Φ C M N alinearmap;imagedomaintovisibilityspace y C M the measured visibilities G C M kn gridding matrix modelling DDEs FZ C kn N Fourier matrix with zero padding

5 Measurement model Measurement equation yu) = Dl, u)xl)e 2iπu l d 2 l 3/17 Discretised version of the inverse problem y = Φx + n with Φ = GFZ x R N + the intensity image of interest Φ C M N alinearmap;imagedomaintovisibilityspace y C M the measured visibilities G C M kn gridding matrix modelling DDEs FZ C kn N Fourier matrix with zero padding extremely large M and N

6 Measurement model Measurement equation yu) = Dl, u)xl)e 2iπu l d 2 l 3/17 Discretised version of the inverse problem y = Φx + n with Φ = GFZ x R N + the intensity image of interest Φ C M N alinearmap;imagedomaintovisibilityspace y C M the measured visibilities G C M kn gridding matrix modelling DDEs FZ C kn N Fourier matrix with zero padding extremely large M and N scalable algorithms; parallelisms and distributed processing

7 Convex optimisation and compressive sensing Problem formulation 0) 4/17 Discretised version of the ill-posed inverse problem y = Φx + n with Φ = GFZ Compressive sensing approach Constrain the solution to belong to the intersection of several convex sets enforcing data fidelity and positivity Add a convex regulariser on the solution to enforce low dimensionality, sparsity,inagiventransform Convex optimisation solvers Primal-dual forward-backward algorithm Distributed single-band RI imaging Wide-band RI imaging

8 Distributed single-band RI imaging Problem formulation 1) Split the large-scale inverse problem block wise 5/17 y = Φ x + n with Φ = G FZ y = y 1. Φ = Φ 1. = G 1 M 1. FZ y nd Φ nd G nd M nd Regularisation of the ill-posed problem Sparsity constraint for x in a collection of wavelet bases ] Ψ = [Ψ 1... Ψ nb

9 Distributed single-band RI imaging Problem formulation 2) 6/17 Convex optimisation task min x f x) + γ n b i=1 l i Ψ i x) + n d =1 h Φ x) Enforce positivity, sparsity and data fidelity f z) =ι C z), C = R N + l i z) = z 1 h z) =ι B z), B = {z C M : z y 2 ϵ }

10 Distributed single-band RI imaging The primal dual approach Primal problem 7/17 min f x)+γ x n b i=1 l i Ψ n d i x)+ h Φ x) =1 Dual formulation of the original convex optimisation task min f u i v n b i=1 Ψ i u i Primal dual algorithm n d =1 Φ v ) + 1 γ n b i=1 l i u i)+ n d =1 h v ) Alternate solving the primal problem and the dual problem Converges towards a Kuhn-Tucker point

11 Distributed single-band RI imaging Advantages of the primal dual approach 8/17 Full splitting of the operators and functions No inversion of the linear operators The updates are performed on the dual variables in parallel Non smooth convex functions Forward backward iterations Alternate between a gradient like step forward) and a proximal update backward) Randomised updates on the dual variables Reduce computational and memory requirements per iteration Require more iterations to converge

12 Distributed single-band RI imaging Primal dual algorithm given x 0), x 0), u 0), v 0), ũ 0), ṽ 0),γ,τ,σ i i i repeat for t =1,... generate sets P {1,...,n b } and D {1,...,n d } 9/17 b t) = M FZ x t 1), D run simultaneously Ddistribute b t) and do in parallel v t) = I P B ) v t 1) + G b t) ) ṽ t) = G vt) i end and gather ṽ t) i Pdo in parallel end end u t) i = I S γ σ i x t) = P C x t 1) τ x t) =2 x t) x t 1) until convergence ) u t 1) n b i=1 i + Ψ i x t)) ũ t) i σ i ũ t) + Z F i n d =1 = Ψ i u t) i ) ς M ṽt)

13 Distributed single-band RI imaging Computational complexity 10 / 17 b t) = M FZ x t 1), D Ddistribute b t) and do in parallel v t) = I P B ) v t 1) + G b t) ) ṽ t) = G vt) end and gather ṽ t) i Pdo in parallel end u t) i = I S γ σ i ) u t 1) x t) = P C... i ) + Ψ i x t)) ũ t) i = Ψ i u t) i central node O p Pi O ) O kn log kn + O kn log kn 2n b N ) ) n b + k)n + n d n v ) n d data fidelity nodes p D O 2 kns n d MN + M )

14 Distributed single-band RI imaging Some results 11 / SNR number of iterations PD,M = 10N ADMM,M = 10N SDMM,M = 10N PD,M =5N ADMM,M =5N SDMM,M =5N PD,M =2N ADMM,M =2N SDMM,M =2N SNR number of iterations PD PD R,pP =1,pD =0.9 PD R,pP =1,pD =0.8 PD R,pP =1,pD =0.7 PD R,pP =1,pD =0.6 PD R,pP =1,pD =0.5 PD R,pP =1,pD =0.4 PD R,pP =1,pD =0.3 PD R,pP =1,pD = R. E. Carrillo, J. D. McEwen, and Y. Wiaux, PURIFY: a new approach to radio-interferometric imaging, MNRAS, vol. 439, no. 4, pp , 2014.

15 Distributed single-band RI imaging Some results 12 / SNR Complexity normalised) number of iterations PD PD R,pP =1,pD =0.9 PD R,pP =1,pD =0.8 PD R,pP =1,pD =0.7 PD R,pP =1,pD =0.6 PD R,pP =1,pD =0.5 PD R,pP =1,pD =0.4 PD R,pP =1,pD =0.3 PD R,pP =1,pD =0.2 2 PD R PD R only FFT) 1.5 PD R only data term) pd

16 Wide-band RI imaging Problem formulation 0) 13 / 17 Multiple frequency bands; images x l at each band l =1,...,L Discretised version of the ill-posed inverse problem y l = Φ l x l + n l with Φ = G l FZ X =[x 1,...,x L ] R N L + the hyper-spectral image cube Y =[y 1,...,y L ] C M L the measured visibilities Φ l C M N alinearmap;imagedomaintovisibilityspace Compressive sensing approach Enforce data fidelity and positivity Convex regulariser to enforce low dimensionality Joint sparsity and low rank on X Primal-dual forward-backward algorithm

17 Wide-band RI imaging Problem formulation 1) 14 / 17 Minimisation problem min X f X) + µ g 1Ψ X) + g 2 X) + L l=1 h l Φl x l ) ) Enforce positivity, oint sparsity, low rank, data fidelity f Z) =ι D Z), D = R N L + g 1 Z) = Z l2,1, g 2 Z) = Z h I z) =ι Bl z), B l = {z C M : z y l 2 ϵ l }

18 Wide-band RI imaging Primal dual algorithm given X 0), X 0), V 0) 1, V0) 2, V0) 3,µ,τ,σ 1,σ 2,σ 3 repeat for t =1,... do in parallel end end V t) t 1) 1 =V 1 + Ψ X t 1) S l 2,1 V t 1) µ/σ Ψ X t 1)) V t) 2 =Vt 1) 2 + X t 1) S 1/σ 2 V t 1) 2 + X t 1)) l {1,...,L} do in parallel u t) l X t) =P C X t 1) τ =u t 1) + Φ l l x t 1) P l Bl u t 1) l ) + Φ l x t 1) l [ ])) σ 1 ΨV t) 1 + σ 2 V t) 2 + σ 3 Φ 1 ut) 1,...,Φ L ut) L 15 / 17 X t) =2X t) X t 1) until convergence

19 Distributed single-band RI imaging Some results 16 / SNR LRJS, 64-bands LRJS, 32-bands LRJS, 16-bands WDCT, 64-bands WDCT, 32-bands WDCT, 16-bands Single-band average, 64-bands Single-band average, 32-bands Single-band average, 16-bands M/N -3 A. Ferrari, J. Deguignet, C. Ferrari, D. Mary, A. Schutz, and O. Smirnov, Multi-frequency image reconstruction for radio interferometry. A regularized inverse problem approach, ArXiv e-prints, Apr

20 Summary 17 / 17 Full splitting of the operators and functions No inversion of the linear operators Broad range of convex prior functions Forward backward iterations for non smooth functions Randomisation Reduce computational and memory requirements Require more iterations to converge

Quantifying Uncertainty in High Dimensional Inverse Problems by Convex Optimisation

Quantifying Uncertainty in High Dimensional Inverse Problems by Convex Optimisation 1 Quantifying Uncertainty in High Dimensional Inverse Problems by Convex Optimisation Xiaohao Cai, Marcelo Pereyra, and Jason D. McEwen arxiv:1811.02514v1 [eess.sp] 4 Nov 2018 Abstract Inverse problems

More information

Image restoration. An example in astronomy

Image restoration. An example in astronomy Image restoration Convex approaches: penalties and constraints An example in astronomy Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS

More information

Compressed Sensing: Extending CLEAN and NNLS

Compressed Sensing: Extending CLEAN and NNLS Compressed Sensing: Extending CLEAN and NNLS Ludwig Schwardt SKA South Africa (KAT Project) Calibration & Imaging Workshop Socorro, NM, USA 31 March 2009 Outline 1 Compressed Sensing (CS) Introduction

More information

Compressed sensing for radio interferometry: spread spectrum imaging techniques

Compressed sensing for radio interferometry: spread spectrum imaging techniques Compressed sensing for radio interferometry: spread spectrum imaging techniques Y. Wiaux a,b, G. Puy a, Y. Boursier a and P. Vandergheynst a a Institute of Electrical Engineering, Ecole Polytechnique Fédérale

More information

Model Selection with Partly Smooth Functions

Model Selection with Partly Smooth Functions Model Selection with Partly Smooth Functions Samuel Vaiter, Gabriel Peyré and Jalal Fadili vaiter@ceremade.dauphine.fr August 27, 2014 ITWIST 14 Model Consistency of Partly Smooth Regularizers, arxiv:1405.1004,

More information

Image restoration: numerical optimisation

Image restoration: numerical optimisation Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context

More information

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization Hongchao Zhang hozhang@math.lsu.edu Department of Mathematics Center for Computation and Technology Louisiana State

More information

IMAGES are observed on a spherical manifold in many. Sparse image reconstruction on the sphere: implications of a new sampling theorem

IMAGES are observed on a spherical manifold in many. Sparse image reconstruction on the sphere: implications of a new sampling theorem IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL., NO., AUGUST 2012 1 Sparse image reconstruction on the sphere: implications of a new sampling theorem Jason D. McEwen, Member, IEEE, Gilles Puy, Jean-Philippe

More information

In collaboration with J.-C. Pesquet A. Repetti EC (UPE) IFPEN 16 Dec / 29

In collaboration with J.-C. Pesquet A. Repetti EC (UPE) IFPEN 16 Dec / 29 A Random block-coordinate primal-dual proximal algorithm with application to 3D mesh denoising Emilie CHOUZENOUX Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est, France Horizon Maths 2014

More information

Bayesian inference and mathematical imaging. Part III: probability & convex optimisation.

Bayesian inference and mathematical imaging. Part III: probability & convex optimisation. Bayesian inference and mathematical imaging. Part III: probability & convex optimisation. Dr. Marcelo Pereyra http://www.macs.hw.ac.uk/ mp71/ Maxwell Institute for Mathematical Sciences, Heriot-Watt University

More information

Heriot-Watt University. Spread spectrum for imaging techniques in radio interferometry Wiaux, Yves; Puy, Gilles; Boursier, Y.; Vandergheynst, Pierre

Heriot-Watt University. Spread spectrum for imaging techniques in radio interferometry Wiaux, Yves; Puy, Gilles; Boursier, Y.; Vandergheynst, Pierre Heriot-Watt University Heriot-Watt University Research Gateway Spread spectrum for imaging techniques in radio interferometry Wiaux, Yves; Puy, Gilles; Boursier, Y.; Vandergheynst, Pierre Published in:

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique

More information

New Applications of Sparse Methods in Physics. Ra Inta, Centre for Gravitational Physics, The Australian National University

New Applications of Sparse Methods in Physics. Ra Inta, Centre for Gravitational Physics, The Australian National University New Applications of Sparse Methods in Physics Ra Inta, Centre for Gravitational Physics, The Australian National University 2 Sparse methods A vector is S-sparse if it has at most S non-zero coefficients.

More information

Primal-dual algorithms for the sum of two and three functions 1

Primal-dual algorithms for the sum of two and three functions 1 Primal-dual algorithms for the sum of two and three functions 1 Ming Yan Michigan State University, CMSE/Mathematics 1 This works is partially supported by NSF. optimization problems for primal-dual algorithms

More information

Assorted DiffuserCam Derivations

Assorted DiffuserCam Derivations Assorted DiffuserCam Derivations Shreyas Parthasarathy and Camille Biscarrat Advisors: Nick Antipa, Grace Kuo, Laura Waller Last Modified November 27, 28 Walk-through ADMM Update Solution In our derivation

More information

An Inexact Newton Method for Optimization

An Inexact Newton Method for Optimization New York University Brown Applied Mathematics Seminar, February 10, 2009 Brief biography New York State College of William and Mary (B.S.) Northwestern University (M.S. & Ph.D.) Courant Institute (Postdoc)

More information

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches Patrick L. Combettes joint work with J.-C. Pesquet) Laboratoire Jacques-Louis Lions Faculté de Mathématiques

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in

More information

Sparse Approximation and Variable Selection

Sparse Approximation and Variable Selection Sparse Approximation and Variable Selection Lorenzo Rosasco 9.520 Class 07 February 26, 2007 About this class Goal To introduce the problem of variable selection, discuss its connection to sparse approximation

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

ADMM and Fast Gradient Methods for Distributed Optimization

ADMM and Fast Gradient Methods for Distributed Optimization ADMM and Fast Gradient Methods for Distributed Optimization João Xavier Instituto Sistemas e Robótica (ISR), Instituto Superior Técnico (IST) European Control Conference, ECC 13 July 16, 013 Joint work

More information

Variational Bayesian Inference Techniques

Variational Bayesian Inference Techniques Advanced Signal Processing 2, SE Variational Bayesian Inference Techniques Johann Steiner 1 Outline Introduction Sparse Signal Reconstruction Sparsity Priors Benefits of Sparse Bayesian Inference Variational

More information

Coordinate descent methods

Coordinate descent methods Coordinate descent methods Master Mathematics for data science and big data Olivier Fercoq November 3, 05 Contents Exact coordinate descent Coordinate gradient descent 3 3 Proximal coordinate descent 5

More information

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

CHAPTER 11. A Revision. 1. The Computers and Numbers therein CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of

More information

Key words. PDE-constrained optimization, deferred correction, time-dependent PDE, coupled system

Key words. PDE-constrained optimization, deferred correction, time-dependent PDE, coupled system A RATIONAL DEFERRED CORRECTION APPROACH TO PDE-CONSTRAINED OPTIMIZATION STEFAN GÜTTEL AND JOHN W PEARSON Abstract The accurate and efficient solution of time-dependent PDE-constrained optimization problems

More information

Composite Objective Mirror Descent

Composite Objective Mirror Descent Composite Objective Mirror Descent John C. Duchi 1,3 Shai Shalev-Shwartz 2 Yoram Singer 3 Ambuj Tewari 4 1 University of California, Berkeley 2 Hebrew University of Jerusalem, Israel 3 Google Research

More information

Optimisation in Higher Dimensions

Optimisation in Higher Dimensions CHAPTER 6 Optimisation in Higher Dimensions Beyond optimisation in 1D, we will study two directions. First, the equivalent in nth dimension, x R n such that f(x ) f(x) for all x R n. Second, constrained

More information

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices Ryota Tomioka 1, Taiji Suzuki 1, Masashi Sugiyama 2, Hisashi Kashima 1 1 The University of Tokyo 2 Tokyo Institute of Technology 2010-06-22

More information

Proceedings of the International BASP Frontiers Workshop January 29 - February 3, 2017 Villars-sur-Ollon, Switzerland

Proceedings of the International BASP Frontiers Workshop January 29 - February 3, 2017 Villars-sur-Ollon, Switzerland Proceedings of the International BASP Frontiers Workshop 2017 January 29 - February 3, 2017 Villars-sur-Ollon, Switzerland BASP Frontiers Workshop 2017 Chair Prof. Yves Wiaux, Heriot-Watt University, UK

More information

Coordinate Update Algorithm Short Course Operator Splitting

Coordinate Update Algorithm Short Course Operator Splitting Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of

More information

Adaptive Primal Dual Optimization for Image Processing and Learning

Adaptive Primal Dual Optimization for Image Processing and Learning Adaptive Primal Dual Optimization for Image Processing and Learning Tom Goldstein Rice University tag7@rice.edu Ernie Esser University of British Columbia eesser@eos.ubc.ca Richard Baraniuk Rice University

More information

Application to Hyperspectral Imaging

Application to Hyperspectral Imaging Compressed Sensing of Low Complexity High Dimensional Data Application to Hyperspectral Imaging Kévin Degraux PhD Student, ICTEAM institute Université catholique de Louvain, Belgium 6 November, 2013 Hyperspectral

More information

About Split Proximal Algorithms for the Q-Lasso

About Split Proximal Algorithms for the Q-Lasso Thai Journal of Mathematics Volume 5 (207) Number : 7 http://thaijmath.in.cmu.ac.th ISSN 686-0209 About Split Proximal Algorithms for the Q-Lasso Abdellatif Moudafi Aix Marseille Université, CNRS-L.S.I.S

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

Subgradient methods for huge-scale optimization problems

Subgradient methods for huge-scale optimization problems Subgradient methods for huge-scale optimization problems Yurii Nesterov, CORE/INMA (UCL) May 24, 2012 (Edinburgh, Scotland) Yu. Nesterov Subgradient methods for huge-scale problems 1/24 Outline 1 Problems

More information

Journal Club. A Universal Catalyst for First-Order Optimization (H. Lin, J. Mairal and Z. Harchaoui) March 8th, CMAP, Ecole Polytechnique 1/19

Journal Club. A Universal Catalyst for First-Order Optimization (H. Lin, J. Mairal and Z. Harchaoui) March 8th, CMAP, Ecole Polytechnique 1/19 Journal Club A Universal Catalyst for First-Order Optimization (H. Lin, J. Mairal and Z. Harchaoui) CMAP, Ecole Polytechnique March 8th, 2018 1/19 Plan 1 Motivations 2 Existing Acceleration Methods 3 Universal

More information

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection EUSIPCO 2015 1/19 A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection Jean-Christophe Pesquet Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est

More information

Reconstruction of Block-Sparse Signals by Using an l 2/p -Regularized Least-Squares Algorithm

Reconstruction of Block-Sparse Signals by Using an l 2/p -Regularized Least-Squares Algorithm Reconstruction of Block-Sparse Signals by Using an l 2/p -Regularized Least-Squares Algorithm Jeevan K. Pant, Wu-Sheng Lu, and Andreas Antoniou University of Victoria May 21, 2012 Compressive Sensing 1/23

More information

An Inexact Newton Method for Nonlinear Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results

More information

Electrostatic Imaging via Conformal Mapping. R. Kress. joint work with I. Akduman, Istanbul and H. Haddar, Paris

Electrostatic Imaging via Conformal Mapping. R. Kress. joint work with I. Akduman, Istanbul and H. Haddar, Paris Electrostatic Imaging via Conformal Mapping R. Kress Göttingen joint work with I. Akduman, Istanbul and H. Haddar, Paris Or: A new solution method for inverse boundary value problems for the Laplace equation

More information

Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization

Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization 1/33 Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization Xiaoqun Zhang xqzhang@sjtu.edu.cn Department of Mathematics/Institute of Natural Sciences, Shanghai Jiao Tong University

More information

Convex Optimization on Large-Scale Domains Given by Linear Minimization Oracles

Convex Optimization on Large-Scale Domains Given by Linear Minimization Oracles Convex Optimization on Large-Scale Domains Given by Linear Minimization Oracles Arkadi Nemirovski H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute of Technology Joint research

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information

At-A-Glance SIAM Events Mobile App

At-A-Glance SIAM Events Mobile App At-A-Glance SIAM Events Mobile App Scan the QR code with any QR reader and download the TripBuilder EventMobile app to your iphone, ipad, itouch or Android mobile device.you can also visit www.tripbuildermedia.com/apps/siamevents

More information

Generalized Orthogonal Matching Pursuit- A Review and Some

Generalized Orthogonal Matching Pursuit- A Review and Some Generalized Orthogonal Matching Pursuit- A Review and Some New Results Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur, INDIA Table of Contents

More information

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09 Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods

More information

Notes for System Identification: Impulse Response Functions via Wavelet

Notes for System Identification: Impulse Response Functions via Wavelet Notes for System Identification: Impulse Response Functions via Wavelet 1 Basic Wavelet Algorithm for IRF Extraction In contrast to the FFT-based extraction procedure which must process the data both in

More information

A Primal-dual Three-operator Splitting Scheme

A Primal-dual Three-operator Splitting Scheme Noname manuscript No. (will be inserted by the editor) A Primal-dual Three-operator Splitting Scheme Ming Yan Received: date / Accepted: date Abstract In this paper, we propose a new primal-dual algorithm

More information

Image restoration: edge preserving

Image restoration: edge preserving Image restoration: edge preserving Convex penalties Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP 1 / 32 Topics Image restoration,

More information

Gauge optimization and duality

Gauge optimization and duality 1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel

More information

Numerical optimization

Numerical optimization Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal

More information

Variational Image Restoration

Variational Image Restoration Variational Image Restoration Yuling Jiao yljiaostatistics@znufe.edu.cn School of and Statistics and Mathematics ZNUFE Dec 30, 2014 Outline 1 1 Classical Variational Restoration Models and Algorithms 1.1

More information

Lecture Notes 9: Constrained Optimization

Lecture Notes 9: Constrained Optimization Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form

More information

Probabilis)c image reconstruc)on, foreground removal, and power spectrum inference from radio interferometers

Probabilis)c image reconstruc)on, foreground removal, and power spectrum inference from radio interferometers Probabilis)c image reconstruc)on, foreground removal, and power spectrum inference from radio interferometers Ben Wandelt IAP, ILP, UPMC, CNRS Sorbonne University Theore>cal interest in 21cm cosmology

More information

Section Handout 5 Solutions

Section Handout 5 Solutions CS 188 Spring 2019 Section Handout 5 Solutions Approximate Q-Learning With feature vectors, we can treat values of states and q-states as linear value functions: V (s) = w 1 f 1 (s) + w 2 f 2 (s) +...

More information

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)

More information

Support Vector Machines for Regression

Support Vector Machines for Regression COMP-566 Rohan Shah (1) Support Vector Machines for Regression Provided with n training data points {(x 1, y 1 ), (x 2, y 2 ),, (x n, y n )} R s R we seek a function f for a fixed ɛ > 0 such that: f(x

More information

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems 1 Numerical optimization Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book Numerical optimization 048921 Advanced topics in vision Processing and Analysis of

More information

A Randomized Approach for Crowdsourcing in the Presence of Multiple Views

A Randomized Approach for Crowdsourcing in the Presence of Multiple Views A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion

More information

Optimization Algorithms for Compressed Sensing

Optimization Algorithms for Compressed Sensing Optimization Algorithms for Compressed Sensing Stephen Wright University of Wisconsin-Madison SIAM Gator Student Conference, Gainesville, March 2009 Stephen Wright (UW-Madison) Optimization and Compressed

More information

Robust multichannel sparse recovery

Robust multichannel sparse recovery Robust multichannel sparse recovery Esa Ollila Department of Signal Processing and Acoustics Aalto University, Finland SUPELEC, Feb 4th, 2015 1 Introduction 2 Nonparametric sparse recovery 3 Simulation

More information

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the

More information

arxiv: v1 [astro-ph.im] 16 Apr 2009

arxiv: v1 [astro-ph.im] 16 Apr 2009 Closed form solution of the maximum entropy equations with application to fast radio astronomical image formation arxiv:0904.2545v1 [astro-ph.im] 16 Apr 2009 Amir Leshem 1 School of Engineering, Bar-Ilan

More information

Sparse methods for improving gravitational wave detection. Ra Inta, The Australian National University

Sparse methods for improving gravitational wave detection. Ra Inta, The Australian National University Sparse methods for improving gravitational wave detection Ra Inta, The Australian National University ADA7, Cargèse, Corsica, May 17, 2012 Sparse methods Reminder: a vector is S-sparse (compressible) if

More information

Spectral k-support Norm Regularization

Spectral k-support Norm Regularization Spectral k-support Norm Regularization Andrew McDonald Department of Computer Science, UCL (Joint work with Massimiliano Pontil and Dimitris Stamos) 25 March, 2015 1 / 19 Problem: Matrix Completion Goal:

More information

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control ALADIN An Algorithm for Distributed Non-Convex Optimization and Control Boris Houska, Yuning Jiang, Janick Frasch, Rien Quirynen, Dimitris Kouzoupis, Moritz Diehl ShanghaiTech University, University of

More information

High-order time-splitting methods for Schrödinger equations

High-order time-splitting methods for Schrödinger equations High-order time-splitting methods for Schrödinger equations and M. Thalhammer Department of Mathematics, University of Innsbruck Conference on Scientific Computing 2009 Geneva, Switzerland Nonlinear Schrödinger

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy Caroline Chaux Joint work with X. Vu, N. Thirion-Moreau and S. Maire (LSIS, Toulon) Aix-Marseille

More information

Sparse Recovery of Streaming Signals Using. M. Salman Asif and Justin Romberg. Abstract

Sparse Recovery of Streaming Signals Using. M. Salman Asif and Justin Romberg. Abstract Sparse Recovery of Streaming Signals Using l 1 -Homotopy 1 M. Salman Asif and Justin Romberg arxiv:136.3331v1 [cs.it] 14 Jun 213 Abstract Most of the existing methods for sparse signal recovery assume

More information

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard

More information

Distributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London

Distributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London Distributed Data Fusion with Kalman Filters Simon Julier Computer Science Department University College London S.Julier@cs.ucl.ac.uk Structure of Talk Motivation Kalman Filters Double Counting Optimal

More information

Solving DC Programs that Promote Group 1-Sparsity

Solving DC Programs that Promote Group 1-Sparsity Solving DC Programs that Promote Group 1-Sparsity Ernie Esser Contains joint work with Xiaoqun Zhang, Yifei Lou and Jack Xin SIAM Conference on Imaging Science Hong Kong Baptist University May 14 2014

More information

A posteriori analysis of a discontinuous Galerkin scheme for a diffuse interface model

A posteriori analysis of a discontinuous Galerkin scheme for a diffuse interface model A posteriori analysis of a discontinuous Galerkin scheme for a diffuse interface model Jan Giesselmann joint work with Ch. Makridakis (Univ. of Sussex), T. Pryer (Univ. of Reading) 9th DFG-CNRS WORKSHOP

More information

8 A pseudo-spectral solution to the Stokes Problem

8 A pseudo-spectral solution to the Stokes Problem 8 A pseudo-spectral solution to the Stokes Problem 8.1 The Method 8.1.1 Generalities We are interested in setting up a pseudo-spectral method for the following Stokes Problem u σu p = f in Ω u = 0 in Ω,

More information

Three Generalizations of Compressed Sensing

Three Generalizations of Compressed Sensing Thomas Blumensath School of Mathematics The University of Southampton June, 2010 home prev next page Compressed Sensing and beyond y = Φx + e x R N or x C N x K is K-sparse and x x K 2 is small y R M or

More information

SCMA292 Mathematical Modeling : Machine Learning. Krikamol Muandet. Department of Mathematics Faculty of Science, Mahidol University.

SCMA292 Mathematical Modeling : Machine Learning. Krikamol Muandet. Department of Mathematics Faculty of Science, Mahidol University. SCMA292 Mathematical Modeling : Machine Learning Krikamol Muandet Department of Mathematics Faculty of Science, Mahidol University February 9, 2016 Outline Quick Recap of Least Square Ridge Regression

More information

CHAPTER 7. Regression

CHAPTER 7. Regression CHAPTER 7 Regression This chapter presents an extended example, illustrating and extending many of the concepts introduced over the past three chapters. Perhaps the best known multi-variate optimisation

More information

Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France.

Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France. Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France remi.gribonval@inria.fr Overview of the course Introduction sparsity & data compression inverse problems

More information

( nonlinear constraints)

( nonlinear constraints) Wavelet Design & Applications Basic requirements: Admissibility (single constraint) Orthogonality ( nonlinear constraints) Sparse Representation Smooth functions well approx. by Fourier High-frequency

More information

Chapter 2. Optimization. Gradients, convexity, and ALS

Chapter 2. Optimization. Gradients, convexity, and ALS Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve

More information

Reconstructing inclusions from Electrostatic Data

Reconstructing inclusions from Electrostatic Data Reconstructing inclusions from Electrostatic Data Isaac Harris Texas A&M University, Department of Mathematics College Station, Texas 77843-3368 iharris@math.tamu.edu Joint work with: W. Rundell Purdue

More information

Sparse Signal Reconstruction with Hierarchical Decomposition

Sparse Signal Reconstruction with Hierarchical Decomposition Sparse Signal Reconstruction with Hierarchical Decomposition Ming Zhong Advisor: Dr. Eitan Tadmor AMSC and CSCAMM University of Maryland College Park College Park, Maryland 20742 USA November 8, 2012 Abstract

More information

Spread spectrum magnetic resonance imaging

Spread spectrum magnetic resonance imaging 1 Spread spectrum magnetic resonance imaging Gilles Puy, Jose P. Marques, Rolf Gruetter, Jean-Philippe Thiran, Senior Member, EEE, Dimitri Van De Ville, Member, EEE, Pierre Vandergheynst, Senior Member,

More information

Lecture 23: November 21

Lecture 23: November 21 10-725/36-725: Convex Optimization Fall 2016 Lecturer: Ryan Tibshirani Lecture 23: November 21 Scribes: Yifan Sun, Ananya Kumar, Xin Lu Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Optimisation in imaging

Optimisation in imaging Optimisation in imaging Hugues Talbot ICIP 2014 Tutorial Optimisation on Hierarchies H. Talbot : Optimisation 1/59 Outline of the lecture 1 Concepts in optimization Cost function Constraints Duality 2

More information

arxiv: v1 [math.na] 23 May 2013

arxiv: v1 [math.na] 23 May 2013 ADI SPLITTING SCHEMES FOR A FOURTH-ORDER NONLINEAR PARTIAL DIFFERENTIAL EQUATION FROM IMAGE PROCESSING Luca Calatroni Cambridge Centre for Analysis, University of Cambridge Wilberforce Road, CB3 0WA, Cambridge,

More information

Dual Ascent. Ryan Tibshirani Convex Optimization

Dual Ascent. Ryan Tibshirani Convex Optimization Dual Ascent Ryan Tibshirani Conve Optimization 10-725 Last time: coordinate descent Consider the problem min f() where f() = g() + n i=1 h i( i ), with g conve and differentiable and each h i conve. Coordinate

More information

Stability of Krylov Subspace Spectral Methods

Stability of Krylov Subspace Spectral Methods Stability of Krylov Subspace Spectral Methods James V. Lambers Department of Energy Resources Engineering Stanford University includes joint work with Patrick Guidotti and Knut Sølna, UC Irvine Margot

More information

Applications of Linear Programming

Applications of Linear Programming Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 9 Non-linear programming In case of LP, the goal

More information

ARock: an algorithmic framework for asynchronous parallel coordinate updates

ARock: an algorithmic framework for asynchronous parallel coordinate updates ARock: an algorithmic framework for asynchronous parallel coordinate updates Zhimin Peng, Yangyang Xu, Ming Yan, Wotao Yin ( UCLA Math, U.Waterloo DCO) UCLA CAM Report 15-37 ShanghaiTech SSDS 15 June 25,

More information

Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 26, 2013

Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 26, 2013 Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 6, 13 SLIM University of British Columbia Motivation 3D seismic experiments - 5D

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

A Scalable and Provably Secure Hash-Based RFID Protocol

A Scalable and Provably Secure Hash-Based RFID Protocol PerSec 05 A Scalable and Provably Secure Hash-Based RFID Protocol EPFL, Lausanne, Switzerland ÉCOLE POLYTECHNIQUE FÉDÉRALE DE LAUSANNE Outline A Brief Introduction to the RFID Technology A Brief Introduction

More information

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 30 Notation f : H R { } is a closed proper convex function domf := {x R n

More information

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications Class 19: Data Representation by Design What is data representation? Let X be a data-space X M (M) F (M) X A data representation

More information

Compressive Sensing and Beyond

Compressive Sensing and Beyond Compressive Sensing and Beyond Sohail Bahmani Gerorgia Tech. Signal Processing Compressed Sensing Signal Models Classics: bandlimited The Sampling Theorem Any signal with bandwidth B can be recovered

More information