Bounds on the radius of nonsingularity

Similar documents
REGULARITY RADIUS: PROPERTIES, APPROXIMATION AND A NOT A PRIORI EXPONENTIAL ALGORITHM

A NOTE ON CHECKING REGULARITY OF INTERVAL MATRICES

arxiv: v1 [cs.na] 7 Jul 2017

NP-hardness results for linear algebraic problems with interval data

Positive Semidefiniteness and Positive Definiteness of a Linear Parametric Interval Matrix

6.854J / J Advanced Algorithms Fall 2008

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

Lecture 8: The Goemans-Williamson MAXCUT algorithm

arxiv: v1 [math.na] 28 Jun 2013

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad

Computing the Norm A,1 is NP-Hard

LINEAR INTERVAL INEQUALITIES

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Proof Complexity Meets Algebra

Total Least Squares and Chebyshev Norm

Introduction to Semidefinite Programming I: Basic properties a

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

Mustapha Ç. Pinar 1. Communicated by Jean Abadie

Componentwise perturbation analysis for matrix inversion or the solution of linear systems leads to the Bauer-Skeel condition number ([2], [13])

Optimal Preconditioning for the Interval Parametric Gauss Seidel Method

Handout 6: Some Applications of Conic Linear Programming

Separation of convex polyhedral sets with uncertain data

A DETERMINISTIC APPROXIMATION ALGORITHM FOR THE DENSEST K-SUBGRAPH PROBLEM

A Continuation Approach Using NCP Function for Solving Max-Cut Problem

Strong duality in Lasserre s hierarchy for polynomial optimization

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Checking Properties of Interval Matrices 1

For δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15])

Travelling Salesman Problem

Algebraic Proof Systems

Advanced Combinatorial Optimization Updated April 29, Lecture 20. Lecturer: Michel X. Goemans Scribe: Claudio Telha (Nov 17, 2009)

Numerical Analysis: Solving Systems of Linear Equations

Polynomiality of Linear Programming

Notice that lemma 4 has nothing to do with 3-colorability. To obtain a better result for 3-colorable graphs, we need the following observation.

Max Cut (1994; 1995; Goemans, Williamson)

Interval solutions for interval algebraic equations

Convergence rates of moment-sum-of-squares hierarchies for volume approximation of semialgebraic sets

Lecture 20: Goemans-Williamson MAXCUT Approximation Algorithm. 2 Goemans-Williamson Approximation Algorithm for MAXCUT

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

A Linear Round Lower Bound for Lovasz-Schrijver SDP Relaxations of Vertex Cover

Interval Robustness in Linear Programming

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Cheap and Tight Bounds: The Recent Result by E. Hansen Can Be Made More Efficient

Lecture Semidefinite Programming and Graph Partitioning

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Discrepancy Theory in Approximation Algorithms

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Robust optimal solutions in interval linear programming with forall-exists quantifiers

SEMIDEFINITE PROGRAM BASICS. Contents

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Advances in Convex Optimization: Theory, Algorithms, and Applications

c 2000 Society for Industrial and Applied Mathematics

Lecture 22: Hyperplane Rounding for Max-Cut SDP

Lecture 17 (Nov 3, 2011 ): Approximation via rounding SDP: Max-Cut

Lecture 9: Low Rank Approximation

Lecture: Examples of LP, SOCP and SDP

Bounds on eigenvalues of complex interval matrices

Fast Linear Iterations for Distributed Averaging 1

Lecture 3: Semidefinite Programming

Mustafa Ç. Pinar 1 and Marc Teboulle 2

LATTICE POINT COVERINGS

LINEAR PROGRAMMING III

@2 4 @ [23] n n 1. (Ribeiro, Urrutia [18]), (Nemhauser,

Lecture 16: Constraint Satisfaction Problems

Hierarchies. 1. Lovasz-Schrijver (LS), LS+ 2. Sherali Adams 3. Lasserre 4. Mixed Hierarchy (recently used) Idea: P = conv(subset S of 0,1 n )

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

Lecture 13 March 7, 2017

Optimization of Submodular Functions Tutorial - lecture I

Convex and Semidefinite Programming for Approximation

Problem 1 (Exercise 2.2, Monograph)

Applications of the Inverse Theta Number in Stable Set Problems

Quasi-Randomness and Algorithmic Regularity for Graphs with General Degree Distributions

Lift-and-Project Techniques and SDP Hierarchies

Fidelity and trace norm

CS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003

ICALP g Mingji Xia

Maximum cut and related problems

Positive Semi-definite programing and applications for approximation

A Note on Representations of Linear Inequalities in Non-Convex Mixed-Integer Quadratic Programs

Lecture 10. Semidefinite Programs and the Max-Cut Problem Max Cut

Overview. 1 Introduction. 2 Preliminary Background. 3 Unique Game. 4 Unique Games Conjecture. 5 Inapproximability Results. 6 Unique Game Algorithms

9. Iterative Methods for Large Linear Systems

MINI-TUTORIAL ON SEMI-ALGEBRAIC PROOF SYSTEMS. Albert Atserias Universitat Politècnica de Catalunya Barcelona

Approximation algorithms for nonnegative polynomial optimization problems over unit spheres

Near-Optimal Algorithms for Maximum Constraint Satisfaction Problems

Solving Global Optimization Problems by Interval Computation

Institute of Computer Science. Compact Form of the Hansen-Bliek-Rohn Enclosure

Roberto Castelli Basque Center for Applied Mathematics

arxiv: v2 [math.oc] 31 May 2010

Lecture 18. Ramanujan Graphs continued

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 8 Luca Trevisan September 19, 2017

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

On dependent randomized rounding algorithms

658 Michel X. Goemans 1 Preliminaries In this section, we collect several basic results about positive semidenite matrices and semidenite programming.

Formalizing Randomized Matching Algorithms

1 The linear algebra of linear programs (March 15 and 22, 2015)

47-831: Advanced Integer Programming Lecturer: Amitabh Basu Lecture 2 Date: 03/18/2010

The Trust Region Subproblem with Non-Intersecting Linear Constraints

Transcription:

Towards bounds on the radius of nonsingularity, Milan Hladík Department of Applied Mathematics, Charles University, Prague and Institute of Computer Science of the Academy of Czech Republic SCAN 2014, Würzburg

Radius of nonsingularity Original task For given square matrix measure the distance to the nearest singular one

Radius of nonsingularity Original task For given square matrix measure the distance to the nearest singular one Radius of nonsingularity of a square matrix Minimal distance to a singular matrix in Chebyshev norm

Radius of nonsingularity Original task For given square matrix measure the distance to the nearest singular one Radius of nonsingularity of a square matrix Minimal distance to a singular matrix in Chebyshev norm Definition (Radius of nonsingularity) Given a matrix A R n n, the radius of nonsingularity is defined by d(a) := inf {ε > 0; singular B : a ij b ij ε i, j}.

Radius of nonsingularity Original task For given square matrix measure the distance to the nearest singular one Radius of nonsingularity of a square matrix Minimal distance to a singular matrix in Chebyshev norm Definition (Radius of nonsingularity) Given a matrix A R n n, the radius of nonsingularity is defined by d(a) := inf {ε > 0; singular B : a ij b ij ε i, j}. Note In fact, a lot of results are for generalized version d(a, ) := inf {ε > 0; singular B : a ij b ij ε i,j i, j}. where R n n is a given non-negative matrix

Motivation to determine radius of nonsingularity Data uncertainty Entries of matrix A = (a i,j ) represents results of experiment Measurement device ensures some (uniform) precision δ It is thus guaranteed that the (unknown) actual values is in [a i,j δ, a i,j + δ] Question: Is A suitable for further processing? Yes: δ < d(a) No: δ d(a) Poljak, S., Rohn, J. (1993): Math.Control Signals Systems 6, 1 9

Motivation to determine radius of nonsingularity Data uncertainty Entries of matrix A = (a i,j ) represents results of experiment Measurement device ensures some (uniform) precision δ It is thus guaranteed that the (unknown) actual values is in [a i,j δ, a i,j + δ] Question: Is A suitable for further processing? Yes: δ < d(a) No: δ d(a) Rounding fixed-point arithmetic Assume A having irrational entries e.g. formal derivation of data like distance may end in 2 Rounding before numerics à A à > d(ã) problems presence of singular matrix within precision à does not reflect properties of A Poljak, S., Rohn, J. (1993): Math.Control Signals Systems 6, 1 9

Computing radius of nonsingularity Proposition (Poljak and Rohn, 1993)) For each non-singular A there hols d(a) := 1 A 1,1 where,1 is a matrix norm defined as M,1 := max { Mx 1 ; x = 1} = max { Mz 1 ; z {±1} n } Poljak, S., Rohn, J. (1993): Math.Control Signals Systems 6, 1 9

Computing radius of nonsingularity Proposition (Poljak and Rohn, 1993)) For each non-singular A there hols d(a) := 1 A 1,1 where,1 is a matrix norm defined as M,1 := max { Mx 1 ; x = 1} = max { Mz 1 ; z {±1} n } Computing exact A,1 can be reduced to max-cut of weighted graph This provides arguments for (Poljak and Rohn, 1993): Computing d(a) is NP-hard Poljak, S., Rohn, J. (1993): Math.Control Signals Systems 6, 1 9

Computing radius of nonsingularity Proposition (Poljak and Rohn, 1993)) For each non-singular A there hols d(a) := 1 A 1,1 where,1 is a matrix norm defined as M,1 := max { Mx 1 ; x = 1} = max { Mz 1 ; z {±1} n } Computing exact A,1 can be reduced to max-cut of weighted graph This provides arguments for (Poljak and Rohn, 1993): Computing d(a) is NP-hard There is no polynomial time algorithm for approximating d(a) with a relative error at most 1 4n 2 (Rohn, 1996) Poljak, S., Rohn, J. (1993): Math.Control Signals Systems 6, 1 9 Rohn, J. (1996). Technical Report 686, ICS Prague.

Bounds on radius of nonsingularity There are several bounds for d(a) Rohn provides (Rohn, 1996) 1 ρ( A 1 E) d(a) 1 max i=1,...,n (E A 1 ) ii Rump (Rump, 1997 and 1997b) developed the estimations 1 ρ( A 1 E) d(a) 6n ρ( A 1 E). Rohn, J. (1996): Technical Report 686, ICS Prague. Rump, S. M. (1997): Linear Multilinear Algebra, 42(2):93 107. Rump,S. M. (1997b): SIAM J. Matrix Anal. Appl., 18(1):83 103.

Goal and main idea Overall goal is Provide (randomized) aprroximate algorithm for d(a) with better constant relative error

Goal and main idea Overall goal is Provide (randomized) aprroximate algorithm for d(a) with better constant relative error Approach 1 Develop approximation of M,1 using semi-definite relation (This gives randomized approximation algorithm) 2 This provides also approximation for d(a) through d(a) := 1 A 1,1

SDP relaxation The problem of computing M,1 can be formulated as n max m ij x i y j subject to x, y {±1} n (1) i,j=1

SDP relaxation The problem of computing M,1 can be formulated as n max m ij x i y j subject to x, y {±1} n (1) i,j=1 The SDP relaxation replacing discrete variables x i, y j {±1}, i, j = 1,..., n by unit vectors u i, v j R n, i, j = 1,..., n as follows n max m ij ui T v j subject to u i, v i R n, (2) i,j=1 where u i 2 = v i 2 = 1, i = 1,..., n Gärtner, B. and Matoušek, J (2012): Springer

SDP relaxation The problem of computing M,1 can be formulated as n max m ij x i y j subject to x, y {±1} n (1) i,j=1 The SDP relaxation replacing discrete variables x i, y j {±1}, i, j = 1,..., n by unit vectors u i, v j R n, i, j = 1,..., n as follows n max m ij ui T v j subject to u i, v i R n, (2) i,j=1 where u i 2 = v i 2 = 1, i = 1,..., n Connection For any feasible solution x, y to (1) solution u, v to (2) as u i := (0,..., 0, x i ) and v i := (0,..., 0, y i ), i = 1,..., n Thus (2) is relaxation of (1) the OV to (2) is an upper bound on the OV to (1) Gärtner, B. and Matoušek, J (2012): Springer

Further adjustment Starting at transformed problem n max m ij ui T v j subject to u i, v i R n, i,j=1 where u i 2 = v i 2 = 1, i = 1,..., n

Further adjustment Starting at transformed problem n max m ij ui T v j subject to u i, v i R n, i,j=1 where u i 2 = v i 2 = 1, i = 1,..., n Define the matrix Z R 2n 2n as Z := U T U so that U has the columns u 1,..., u n, v 1,..., v n note also that Z is positive semidefinite (denoted by Z 0)

Further adjustment Starting at transformed problem n max m ij ui T v j subject to u i, v i R n, i,j=1 where u i 2 = v i 2 = 1, i = 1,..., n Define the matrix Z R 2n 2n as Z := U T U so that U has the columns u 1,..., u n, v 1,..., v n note also that Z is positive semidefinite (denoted by Z 0) We can thus transform above program into n max m ij z i,j+n subject to Z 0, z ii = 1, i = 1,..., 2n (3) i,j=1

Solving the problem The semidefinite problem n max m ij z i,j+n subject to Z 0, z ii = 1, i = 1,..., 2n, (3) i,j=1 Grötschel, M., Lovász, L. and Schrijver, A. (1981): Combinatorica 1, 169 197.

Solving the problem The semidefinite problem n max m ij z i,j+n subject to Z 0, z ii = 1, i = 1,..., 2n, (3) i,j=1 can be solved with arbitrary (a priori given) precision ε given by (Grötschel, Lovász, Schrijver, 1981) Grötschel, M., Lovász, L. and Schrijver, A. (1981): Combinatorica 1, 169 197.

Solving the problem The semidefinite problem n max m ij z i,j+n subject to Z 0, z ii = 1, i = 1,..., 2n, (3) i,j=1 can be solved with arbitrary (a priori given) precision ε given by (Grötschel, Lovász, Schrijver, 1981) for optimal value γ holds γ M,1 ε and thus is a bound for our norm γ + ε M,1 Grötschel, M., Lovász, L. and Schrijver, A. (1981): Combinatorica 1, 169 197.

Solving the problem The semidefinite problem n max m ij z i,j+n subject to Z 0, z ii = 1, i = 1,..., 2n, (3) i,j=1 can be solved with arbitrary (a priori given) precision ε given by (Grötschel, Lovász, Schrijver, 1981) for optimal value γ holds γ M,1 ε and thus is a bound for our norm γ + ε M,1 Handling Utilize optimal solution of (3) resp. (2) to find optimal solution to (1) Grötschel, M., Lovász, L. and Schrijver, A. (1981): Combinatorica 1, 169 197.

Optimal solution of transformed problem Rounding the vector solution Let approximate optimal solution of transformed problem be u1,..., u n, v1,..., v n Use p R n being a unit vector via following mapping { 1 if p T w 0, w (4) 1 otherwise to determine solution to original problem p +1 +1 +1 1 1

Randomized algorithm computing d(a) 1 Determine inverse matrix M = A 1 2 Compute approximation of optimal solution of n max m ij ui T v j subject to u i, v i R n, (2) i,j=1 3 Choose randomly unit vector p R n and apply { 1 if p T w 0, w 1 otherwise which provides solution to problem n M,1 = max m ij x i y j subject to x, y {±1} n (1) 4 Finally compute i,j=1 d(a) := 1 A 1,1 = 1 M,1 (4)

Expected value of the solution Lemma (Gärtner and Matoušek, 2012) Let u, v R n be unit vectors. The probability that the mapping (4) maps u and v to different values is 1 π arccos ut v.

Expected value of the solution Lemma (Gärtner and Matoušek, 2012) Let u, v R n be unit vectors. The probability that the mapping (4) maps u and v to different values is 1 π arccos ut v. Apply this lemma Let have a feasible solution to (1) x 1,..., x n, y 1,..., y n determined by images (of defined mapping) u 1,..., u n, v 1,..., v n

Expected value of the solution Lemma (Gärtner and Matoušek, 2012) Let u, v R n be unit vectors. The probability that the mapping (4) maps u and v to different values is 1 π arccos ut v. Apply this lemma Let have a feasible solution to (1) x1,..., x n, y1,..., y n determined by images (of defined mapping) u1,..., u n, v1,..., v n By the Lemma, the expected value of the solution is i,j m ij ( 1 1 π arccos u T i = i,j v j ) m ij 1 π m ij ( 1 2 π arccos u T i vj arccos u T i ) vj.

Bound of the expected value of the solution Looking for bound of expected value ( m ij 1 2 π i,j arccos u T i ) vj.

Bound of the expected value of the solution Looking for bound of expected value ( m ij 1 2 π i,j arccos u T i ) vj. Goemans-Williamson ratio α for approximating MAX-CUT It value is α 0.87856723 It represents optimal value of the problem min z [ 1,1] 2 arccos z π(1 z) Goemans, M. X. and Williamson, D. P. (1995) Journal of the ACM, 42:1115 1145.

Bound of the expected value of the solution Looking for bound of expected value ( m ij 1 2 π i,j arccos u T i ) vj. Goemans-Williamson ratio α for approximating MAX-CUT It value is α 0.87856723 It represents optimal value of the problem min z [ 1,1] 2 arccos z π(1 z) Using this constant we can bound the term from expected value Lemma For each z [ 1, 1] we have αz + α 1 1 2 arccos z αz + 1 α. π Goemans, M. X. and Williamson, D. P. (1995) Journal of the ACM, 42:1115 1145.

Applying the bound Now, we use the Lemma to find a lower bound to expected value Let i, j {1, 2,..., n}, then { ( m ij 1 2 ) π arccos z m ij (αz + α 1) if m ij 0, or m ij (αz α + 1) otherwise, ( m ij 1 2 ) π arccos z m ij αz + m ij (α 1).

Applying the bound Now, we use the Lemma to find a lower bound to expected value Let i, j {1, 2,..., n}, then { ( m ij 1 2 ) π arccos z m ij (αz + α 1) if m ij 0, or m ij (αz α + 1) otherwise, ( m ij 1 2 ) π arccos z m ij αz + m ij (α 1). Thus the lower bound to expected value is ( m ij 1 2 ) arccos u T i vj m ij αui T vj + m ij (α 1) π i,j i,j = αγ + (α 1)e T M e,

Applying the bound Now, we use the Lemma to find a lower bound to expected value Let i, j {1, 2,..., n}, then { ( m ij 1 2 ) π arccos z m ij (αz + α 1) if m ij 0, or m ij (αz α + 1) otherwise, ( m ij 1 2 ) π arccos z m ij αz + m ij (α 1). Thus the lower bound to expected value is ( m ij 1 2 ) arccos u T i vj m ij αui T vj + m ij (α 1) π i,j i,j = αγ + (α 1)e T M e, Hence we have an expected lower bound on M,1 M,1 αγ + (α 1)e T M e.

Determining approximation ratio We have an expected lower bound on M,1 M,1 αγ + (α 1)e T M e.

Determining approximation ratio We have an expected lower bound on M,1 M,1 αγ + (α 1)e T M e. The right-hand side depends on the entries of M we can employ the estimate M,1 e T M e to obtain M,1 αγ + (α 1)e T M e αγ + (α 1) M,1, whence M,1 α 2 α γ.

Determining approximation ratio We have an expected lower bound on M,1 M,1 αγ + (α 1)e T M e. The right-hand side depends on the entries of M we can employ the estimate M,1 e T M e to obtain M,1 αγ + (α 1)e T M e αγ + (α 1) M,1, whence M,1 α 2 α γ. This gives us a randomized algorithm with the approximation ratio α 2 α 0.78343281

Conclusions Computing radius of nonsingularity Randomized approximation of with a constant relative error 0.78343281 Using relaxation to SDP and principles from approximating MAX-CUT Perspectives Derandomization of the algorithm

Thank you