Equivalence constants for certain matrix norms II

Similar documents
A property concerning the Hadamard powers of inverse M-matrices

Spectrally arbitrary star sign patterns

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

Where is matrix multiplication locally open?

Interpolating the arithmetic geometric mean inequality and its operator version

Absolute value equations

Linear Algebra and its Applications

Linear Algebra and its Applications

Minimizing the Laplacian eigenvalues for trees with given domination number

Intrinsic products and factorizations of matrices

Boolean Inner-Product Spaces and Boolean Matrices

On matrix equations X ± A X 2 A = I

LOWER BOUNDS FOR HAAR PROJECTIONS: DETERMINISTIC EXAMPLES. 1. Introduction

Linear Algebra and its Applications

Subset selection for matrices

Jordan Journal of Mathematics and Statistics (JJMS) 8(3), 2015, pp THE NORM OF CERTAIN MATRIX OPERATORS ON NEW DIFFERENCE SEQUENCE SPACES

INVESTIGATING THE NUMERICAL RANGE AND Q-NUMERICAL RANGE OF NON SQUARE MATRICES. Aikaterini Aretaki, John Maroulas

RANK AND PERIMETER PRESERVER OF RANK-1 MATRICES OVER MAX ALGEBRA

Inequalities for the spectra of symmetric doubly stochastic matrices

ACI-matrices all of whose completions have the same rank

LESSON 2 5 CHAPTER 2 OBJECTIVES

Perron eigenvector of the Tsetlin matrix

Orthogonal similarity of a real matrix and its transpose

A CLASS OF SCHUR MULTIPLIERS ON SOME QUASI-BANACH SPACES OF INFINITE MATRICES

Linear Algebra and its Applications

Review of Basic Concepts in Linear Algebra

Linear Algebra and its Applications

Product distance matrix of a tree with matrix weights

The minimum rank of matrices and the equivalence class graph

Locally linearly dependent operators and reflexivity of operator spaces

Minimum number of non-zero-entries in a 7 7 stable matrix

An iterative method for the skew-symmetric solution and the optimal approximate solution of the matrix equation AXB =C

BALANCING GAUSSIAN VECTORS. 1. Introduction

Characterization of 2 n -Periodic Binary Sequences with Fixed 2-error or 3-error Linear Complexity

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Linear Algebra and its Applications

Graphs with given diameter maximizing the spectral radius van Dam, Edwin

Determinants of Partition Matrices

ICS 6N Computational Linear Algebra Matrix Algebra

7.5 Operations with Matrices. Copyright Cengage Learning. All rights reserved.

Hierarchical open Leontief models

RESEARCH ARTICLE. Linear Magic Rectangles

An almost sure invariance principle for additive functionals of Markov chains

Rate of convergence for certain families of summation integral type operators

Trace Inequalities for a Block Hadamard Product

Matrix Arithmetic. a 11 a. A + B = + a m1 a mn. + b. a 11 + b 11 a 1n + b 1n = a m1. b m1 b mn. and scalar multiplication for matrices via.

c 2010 Society for Industrial and Applied Mathematics

Yimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract

The skew-symmetric orthogonal solutions of the matrix equation AX = B

Difference Sets Corresponding to a Class of Symmetric Designs

The effect on the algebraic connectivity of a tree by grafting or collapsing of edges

Zero estimates for polynomials inspired by Hilbert s seventh problem

On the Infinitude of Twin-Primes Bao Qi Feng

The cancellable range of rings

RESEARCH ARTICLE. An extension of the polytope of doubly stochastic matrices

New lower bounds for hypergraph Ramsey numbers

The symmetric minimal rank solution of the matrix equation AX=B and the optimal approximation

Nilpotent completions of partial upper triangular matrices

MATH6081A Homework 8. In addition, when 1 < p 2 the above inequality can be refined using Lorentz spaces: f

Minimal locally linearly dependent spaces of operators

The maximal determinant and subdeterminants of ±1 matrices

On factorization of Hilbert Schmidt mappings

Self-dual interval orders and row-fishburn matrices

Mathematics 13: Lecture 10

Linear Algebra March 16, 2019

On some linear combinations of hypergeneralized projectors

MATRICES. a m,1 a m,n A =

Lecture 3. Random Fourier measurements

Knot Floer Homology and the Genera of Torus Knots

Herz (cf. [H], and also [BS]) proved that the reverse inequality is also true, that is,

Linear Algebra and its Applications

Assignment 1: From the Definition of Convexity to Helley Theorem

Decouplings and applications

Sums of diagonalizable matrices

Your first day at work MATH 806 (Fall 2015)

Hardy spaces of Dirichlet series and function theory on polydiscs

DETERMINANTAL DIVISOR RANK OF AN INTEGRAL MATRIX

SOME INEQUALITIES FOR THE EUCLIDEAN OPERATOR RADIUS OF TWO OPERATORS IN HILBERT SPACES

Matrix Algebra 2.1 MATRIX OPERATIONS Pearson Education, Inc.

Spectral radius of bipartite graphs

Ep Matrices and Its Weighted Generalized Inverse

Graphs with few total dominating sets

MATH 2030: MATRICES ,, a m1 a m2 a mn If the columns of A are the vectors a 1, a 2,...,a n ; A is represented as A 1. .

Bounds for Levinger s function of nonnegative almost skew-symmetric matrices

Linear Algebra and its Applications

Robust linear optimization under general norms

On factor width and symmetric H -matrices

Ordering trees by their largest eigenvalues

Linear Algebra and its Applications

arxiv:math/ v1 [math.ra] 3 Nov 2004

The Drazin inverses of products and differences of orthogonal projections

Laplacian Integral Graphs with Maximum Degree 3

arxiv: v1 [math.co] 13 Oct 2014

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

The Hardy Operator and Boyd Indices

Interval solutions for interval algebraic equations

Basic Concepts in Linear Algebra

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.

A Brief Introduction to Tensors

THE UNIVERSITY OF TORONTO UNDERGRADUATE MATHEMATICS COMPETITION. Sunday, March 14, 2004 Time: hours No aids or calculators permitted.

Transcription:

Linear Algebra and its Applications 420 (2007) 388 399 www.elsevier.com/locate/laa Equivalence constants for certain matrix norms II Bao Qi Feng a,, Andrew Tonge b a Department of Mathematical Sciences, Kent State University, Tuscarawas Campus, New Philadelphia, OH 44663-9403, USA b Department of Mathematical Sciences, Kent State University, Kent, OH 44240, USA Received 17 March 2006; accepted 17 July 2006 Available online 12 September 2006 Submitted by C.-K. Li Abstract We give an almost complete solution of a problem posed by Klaus and Li [A.-L. Klaus, C.-K. Li, Isometries for the vector (p, q) norm and the induced (p, q) norm, Linear and Multilinear Algebra 38 (1995) 315 332]. Klaus and Li s problem, which arose during their investigations of isometries, was to relate the Frobenius (or Hilbert Schmidt) norm of a matrix to various operator norms of that matrix. Our methods are based on earlier work of Feng [B.Q. Feng, Equivalence constants for certain matrix norms, Linear Algebra Appl. 374 (2003) 247 253] and Tonge [A. Tonge, Equivalence constants for matrix norms: a problem of Goldberg, Linear Algebra Appl. 306 (2000) 1 13], but introduce as a new ingredient some techniques developed by Hardy and Littlewood [G.H. Hardy, J.E. Littlewood, Bilinear forms bounded in space [p, q], Quart. J. Math. (Oxford) 5 (1934) 241 254]. 2006 Elsevier Inc. All rights reserved. AMS classification: 15A60; 65F35 Keywords: p-norms; (p, q)-norms; p-operator norms; (p, q)-operator norms; Essentially sharp Let C be the set of all complex numbers. Let lp n (x 1,...,x n ) C n equipped with the norm ( ) n 1/p x p = x j p (1 p< ), max 1 j n x j (p = ). be the Banach space of all n-tuples x = Corresponding author. E-mail addresses: bfeng@tusc.kent.edu (B.Q. Feng), tonge@math.kent.edu (A. Tonge). 0024-3795/$ - see front matter ( 2006 Elsevier Inc. All rights reserved. doi:10.1016/j.laa.2006.07.015

B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 389 The vector space C m n of all complex m n matrices A = (a ij ) can be given many natural norms. We focus on the (p, q)-norm, ( n m ) q/p 1/q A pq = a ij p (1 p, q ), i=1 and the (p, q)-operator norm, A pq = max{ Ax q : x p 1} (1 p, q ). In the special case that p = q, we call these norms the p-norm and p-operator norm, respectively. For fixed m and n these norms are all equivalent. Our objective is to investigate the equivalence constants among them. In [2], we asked the following problem: Problem 1. Given integers m and n and real numbers 1 p, q, r, s, determine the best constants a, b, c, d (depending on m, n, p, q, r, s) such that for all A C m n, A pq a A rs, A pq b A rs, A pq c A rs, A pq d A rs. (1) (2) (3) (4) In [2], we completely determined the values of a,b and c. In 1995, Kaus and Li [4] specifically asked how to solve the following problem: Problem 2. Determine the largest α>0such that α A 2 A rs for all A C m n. Problem 2 is the special case of (4) of Problem 1 when p = q = 2. In this paper we discuss the more general case when p = q 2. As our methods only give very incomplete and complicated results when p = q<2 or when p/= q, we will not pursue these cases here. An important ingredient in our arguments will be the following lemma, which can found in Hardy and Littlewood [3]. A modern treatment can be found in Osikiewicz and Tonge [5]. Lemma 1 [3, 4.1.5 of Theorem 5]. Suppose r>s 1and 1/p = 1/s 1/r 1/2. Then for all A C m n, A p A rs, Already, Lemma 1 answers the last part (4) of Problem 1 in the special case when p = q and (1/r,1/s) lies on the line segment in the (1/r)(1/s)-plane: 1 s = 1 r + 1 p, 0 1 r 1 p, where 1 p :=1 1 p. As we proceed, we will need to divide the unit square {0 1/r 1} {0 1/s 1} in the (1/r,1/s) plane into eight regions in order to catalog our results. These eight regions will be

390 B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 Fig. 1. m n, p 2. different in the cases m n and m n. Fig. 1 shows how these eight regions partition the unit (1/r,1/s) square in the case m n. Note that the line segment from Lemma 1 is the intersection of regions R 1 and R 2. We describe these eight regions analytically as follows: R 1 ={(1/r,1/s) : 0 1/r 1/p and 1/p + 1/r 1/s 1}, R 2 ={(1/r,1/s) : 0 1/r 1/p and 1/r 1/s 1/p + 1/r}, R 3 ={(1/r,1/s) : 0 1/r 1/p and 0 1/s 1/r}, R 4 ={(1/r,1/s) : 1/p 1/r 1, 0 1/s min(1/p, log n/(r log m)}, R 5 ={(1/r,1/s) : 1 log m/(p log n) 1/r 1 and log n/(r log m) 1/s 1/p}, R 6 ={(1/r,1/s) : 1/p 1/r 1/s 1}, R 7 ={(1/r,1/s) : 1/p 1/r 1 log m/(p log n) and 1/p 1/s 1/r}, R 8 ={(1/r,1/s) : 1 log m/(p log n) 1/r 1 and 1/p 1/s 1/r}. When m n, it is necessary to partition the unit square differently into the regions pictured in Fig. 2. These eight regions are described analytically below: R 1 ={(1/r,1/s) : 1/p 1/s 1 and 0 1/r 1/s 1/p}, R 2 ={(1/r,1/s) : 1/p 1/s 1 and 1/s 1/p 1/r 1/s}, R 3 ={(1/r,1/s) : 1/p 1/s 1/r 1}, R 4 ={(1/r,1/s) : log n/(r log m) 1/s 1/p and 1/p 1/r 1}, R 5 ={(1/r,1/s) : 0 1/s log n/(r log m) and 1/p 1/r 1}, R 6 ={(1/r,1/s) : 0 1/r 1/s 1/p},

B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 391 Fig. 2. m n, p 2. R 7 ={(1/r,1/s) : log n/(p log m) 1/s min(1/r,1/p) and 1/r 1/p }, R 8 ={(1/r,1/s) : 0 1/s min(1/r,log n/(p log m)) and 0 1/r 1/p }. Note that R 1 = R 1. All our results in the case m n will be obtained from the case m n by transposition, using the following well-known lemma: Lemma 2 [2, Lemma 3]. Let A t be the transpose matrix of A C m n. Then for every 1 r, s, A rs = A t s r, where 1/r = 1 1/r and 1/s = 1 1/s. We frequently make use of the notation: { 1, p q, λ pq (n):= n 1/p 1/q, p < q. The following lemmas are crucial to our work for the case m n. Lemma 3 [2, Lemma 1]. Let 1 p, q, and let I n : lp n ln q be the identity operator. Then I n pq = λ qp (n). In other words, for all x C n, x q λ qp (n) x p. Lemma 4 [2, Theorem 5]. Let 1 p, q, r, s. Then for all A C m n, A pq λ pr (m)λ qs (n) A rs,

392 B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 Lemma 5 [2, Theorem 6]. Let 1 p, q, r, s. Then for all A C m n, A pq λ qs (m)λ rp (n) A rs, For the discussion of sharpness, we need to introduce a class of test matrices. Define Q(m, n) C m n by { 1 (i = j), Q(m, n) ij = 0 (i /= j), where 1 i m and 1 j n. Lemma 6 (i) If m n, then Q(m, n) p = m 1/p and Q(m, n) rs = λ sr (m); (ii) If m n, then Q(m, n) p = n 1/p and Q(m, n) rs = λ sr (n). Proof. (i) It is clear that Q(m, n) p = m 1/p, so we just verify the second assertion. For each x = (x 1,...,x n ) t C n, Lemma 3 gives m Q(m, n)x s = (x 1,...,x m ) t s = x j s m λ sr (m) x j r 1/r 1/s λ sr (m) x r. Hence Q(m, n) rs λ sr (m). If r s, let x = (m 1/r,...,m 1/r, 0,...,0) t C n, where there are m copies of m 1/r. Then x r = 1, and m Q(m, n) rs Q(m, n)x s = m s/r 1/s = m 1/s 1/r = λ sr (m). So, Q(m, n) rs = λ sr (m). If r s, let x = (1, 0,...,0) t C n. Then x r = 1 and Q(m, n)x s = 1. Thus, Q(m, n) rs Q(m, n)x s = 1 = λ sr (m). Hence Q(m, n) rs = λ sr (m). (ii) This follows from (i), using Lemma 2. Simply note that Q(m, n) p = Q(m, n) t p = n 1/p and Q(m, n) rs = Q(m, n) t s r = λ r s (n) = λ sr(n). We now start to answer our problem. The first stage is to extend the validity of Lemma 1 into the whole of regions R 1 R 2 R 3 and R 1 R 2 R 3.

B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 393 Theorem 1. Suppose p 2 and 1/σ = 1/p + 1/r. (i) If m n and (1/r,1/s) R 1 R 2 R 3, then for all A C m n, A p λ σs (m)λ rs (m) 1 A rs, (5) (ii) If m n and (1/r,1/s) R 1 R 2 R 3, then for all A C m n, A p λ σs (n)λ rs (n) 1 A rs, (5 ) Proof. (i) Firstly, suppose that (1/r,1/s) R 1, then r s and λ rs (m) = 1. Since 1/σ = 1/p + 1/r, then s σ, and by Lemmas 1 and 5 we have A p A rσ λ σs (m) A rs = λ σs (m)λ rs (m) 1 A rs. Inequality (1) is sharp in R 1, since E(m, n) p = E(m, n) rs = 1, where E(m, n) is the matrix, in which all entries are 0 except 1 in the position (1, 1) (see [[2], Lemma 2]). Next, suppose that (1/r,1/s) R 2, then r s is also true and λ rs (m) = 1. Since 1/σ = 1/p + 1/r, then s σ, and by Lemmas 1 and 5 A p A rσ λ σs (m) A rs = m 1/σ 1/s A rs = m 1/p+1/r 1/s A rs = λ σs (m)λ rs (m) 1 A rs. For the sharpness of (1)inR 2, consider the matrix Q(m, n). Since m n, Lemma 6 gives Q(m, n) p = m 1/p and Q(m, n) rs = λ sr (m). Since s r, Q(m, n) rs = m 1/s 1/r. Hence inequality (1) becomes an equality when we set A = Q(m, n). Finally, suppose that (1/r,1/s) R 3. By Lemma 2, wehave n A rs = A t s r = max m r 1/r a ij x i : x s 1. i=1 Let x = (0,...,0, 1, 0,...,0) t C m, where the 1 is in the ith position. Then x s = 1. Therefore n 1/r A rs a ij r for all 1 i m. As 1/r 1/p,wehave1/p 1/r, and so λ pr (n) = 1. By Lemma 3, wehave 1/p n n 1/r a ij p λ pr (n) a ij r A rs. Considering (1/r,1/s) R 3 implies σ s and r s. Thus, 1/p ( m n m 1/p A p = a ij p A rs) p = m 1/p A rs i=1 i=1 =m 1/σ 1/r A rs = m 1/σ 1/s m 1/s 1/r A rs = λ σs (m)λ rs (m) 1 A rs.

394 B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 This establishes inequality (1)inR 3. To prove the sharpness, we begin by noting that for m n we have Q(m, n) p = m 1/p.As s r,wehave Q(m, n) rs = λ sr (m) = 1. So Q(m, n) p = m 1/p Q(m, n) rs, and inequality (1) is sharp in R 3. (ii) This follows from (i), using Lemma 2. For the next regions, R 4,R 4,R 5 and R 5, we use a probabilistic result to discuss essential sharpness. By saying that an inequality of the type we consider is essentially sharp, we mean that the only possible improvement is by a numerical factor independent of m and n, for which the inequality is true for all matrices of all dimensions. We need the following lemma which is due to Bennett. Lemma 7 [1, Proposition 3.2]. For each 1 r, s there is a constant K such that for all positive integers m and n there is an m n matrix A, all of whose entries are ±1, for which K 1 A rs max(m 1/s n max(1/2 1/r,0),n 1/r m max(1/s 1/2,0) K. ) Theorem 2. Suppose p 2. (i) Let m n. (a) If (1/r,1/s) R 4, then for all A C m n, A p m 1/p n 1/p 1/r A rs, (6) (b) If (1/r,1/s) R 5, then for all A C m n, A p m 1/p 1/s n 1/p A rs, (7) (ii) Let m n. (a ) If (1/r,1/s) R 4, then for all A C m n, A p m 1/p 1/s n 1/p A rs, (6 ) (b ) If (1/r,1/s) R 5, then for all A C m n, A p m 1/p n 1/p 1/r A rs, (7 ) Proof. (i) From the proof of Theorem 1,wehave n 1/r a ij r A rs for all 1 i m. Now, since 1/p 1/r we have 1/r 1/p. Thus, Lemma 3 gives 1/p n n 1/r a ij p λ pr (n) a ij r n 1/p 1/r A rs.

Hence m A p = B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 395 i=1 n a ij p 1/p ( m i=1 ( n 1/p 1/r A rs ) p ) 1/p = m 1/p n 1/p 1/r A rs. This is inequality (6), and it is valid in a rectangular region containing the union of R 4 and R 5, namely {(1/r,1/s) : 1/p 1/r 1 and 0 1/s 1}. We now show that inequality (7) is also valid in a rectangular region containing the union of R 4 and R 5. Recall that s m n A rs = max a ij x j i=1 1/s : x r 1. For x = (0,...,0, 1, 0,...,0) t C n, with the 1 in the jth position, we have x r = 1. Thus ( m ) 1/s a ij s A rs for all 1 j n. i=1 This implies that n m A s = a ij s i=1 1/s n 1/s A rs. Now, if 1/s 1/p, then λ ps (mn) = (mn) 1/p 1/s. Using Lemma 4, in the case q = p and r = s, we find A p λ ps (mn) A s = (mn) 1/p 1/s A s (mn) 1/p 1/s n 1/s A rs = m 1/p 1/s n 1/p A rs, and so we have established inequality (7) in the rectangular region {(1/r,1/s) : 0 1/r 1 and 0 1/s 1/p}, which clearly contains the union of R 4 and R 5. It is easy to check that m 1/p n 1/p 1/r m 1/p 1/s n 1/p m 1/s n 1/r 0 1/s log n/(r log m), m 1/p n 1/p 1/r m 1/p 1/s n 1/p m 1/s n 1/r log n/(r log m) 1/s. Thus, although inequalities (6) and (7) both hold in the rectangle R 4 R 5 ={(1/r,1/s) : 1/p 1/r 1, 0 1/s 1/p}, inequality (6) is better than inequality (7) in the region R 4, and inequality (7) is better than inequality (6) in the region R 5. Next we discuss sharpness. Let (1/r,1/s) R 4 R 5. Then, since p 2, we have 1/r 1/2 and 1/s 1/2. So, by Lemma 7, there is a constant K, independent of m and n, and a matrix A whose entries are ±1, such that

396 B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 A rs K max{m 1/s,n 1/r }. In the case that (1/r,1/s) R 4, we know that m 1/s n 1/r, and so A rs Kn 1/r. Noting that A p = m 1/p n 1/p, we obtain A p K 1 m 1/p n 1/p 1/r A rs. Hence inequality (6) is essentially sharp in region R 4. On the other hand, if (1/r,1/s) R 5, then m 1/s n 1/r, and so A rs Km 1/s. Thus A p = m 1/p n 1/p K 1 m 1/p 1/s n 1/p A rs. Hence inequality (7) is essentially sharp in region R 5. (ii) This follows from (i), using Lemma 2. We now turn to regions R 6 and R 6. The discussion of sharpness requires a new collection of test matrices. When m n, it is possible to find a pair of integers k 1 and 0 l<m, such that n = km + l. Define T (m, n) C m n by 1 1 0 0 0 0 0 0 1 1 T (m, n):=.. 0 0.. 0 0, 0 0 0 0 1 1 where matrix T (m, n) has m blocks. The first l blocks on the left have size m (k + 1), and the remaining m l blocks have size m k. Lemma 8. Let 1 p, and let m n. If (1/r,1/s) R 6, then T (m, n) p 2 1 m 1/s n 1/p 1/r T (m, n) rs. Proof. It is clear that T (m, n) p = n 1/p. For each x lr n, Hölder s inequality gives l 1 T (m, n)x s s = (k+1)(i+1) s m l 1 k(i+1) s x j + x j+(k+1)l i=0 +(k+1)i l 1 (k+1)(i+1) (k + 1) s/s i=0 (k+1)l = (k + 1) s/s (k + 1) s/s x s s. Since 1/s 1/r, Lemma 3 gives +(k+1)i i=0 +ki m l 1 x j s + x j s + k s/s n i=0 +(k+1)l T (m, n) rs = max x r 1 T (m, n)x s (k + 1) 1/s max x r 1 x s k s/s x j s k(i+1) +ki (k + 1) 1/s λ sr (n) max x r 1 x r = (k + 1) 1/s n 1/s 1/r. x j+(k+1)l s

B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 397 Since n = km + l, we find 2n = 2km + 2l 2km (k + 1)m, and so k + 1 2n/m. Hence T (m, n) rs (2n/m) 1/s n 1/s 1/r = 2 1/s m 1/s n 1/r. Thus T (m, n) p = n 1/p 2 1/s m 1/s n 1/p 1/r T (m, n) rs. Since 1/s 1, T (m, n) p 2 1 m 1/s n 1/p 1/r T (m, n) rs. We are now ready to settle our problem for regions R 6 and R 6. Theorem 3. Suppose p 2. (i) Let m n. If (1/r,1/s) R 6 R 7 R 8, then for all A C m n, A p m 1/s n 1/p 1/r A rs, (8) and this inequality is essentially sharp in R 6. (ii) Let m n. If (1/r,1/s) R 6 R 7 R 8, then for all A C m n, A p m 1/p 1/s n 1/r A rs, (8 ) and this inequality is essentially sharp in R 6. Proof. (i) First note that (1/r,1/s) R 6 implies 1/p 1/r. Now consider the composition of maps lp n I n lr n A ls m I m l1 m, and so view A as an operator from lp n to l1 m. Applying Lemma 1 in the case r = p,s = 1, and then Lemma 3, noting that 1/p 1/r, A p A p 1 = I m AI n p 1 I m s1 A rs I n p r = m 1 1/s n 1/r 1/p A rs = m 1/s n 1/p 1/r A rs. This is inequality (1). We know from Lemma 8 that inequality (1) is essentially sharp in region R 6. (ii) This follows from (i), using Lemma 2. However we are unable to determine whether these estimates are sharp or essentially sharp in R 7 R 8, when m n, orinr 7 R 8, when m n. The uncertainty in region R 7, when m n, and R 7, when m n, can be resolved when p = 2. Pulling together Theorems 1 3 we can then answer the problem of Klaus and Li for most values of r and s as follows. Theorem 4. Suppose m n, and S i = R i,i = 1,...,8, when p = 2. (i) If 1/σ = 1/2 + 1/r and (1/r,1/s) S 1 S 2 S 3, then for all A C m n, A 2 λ σs (m)λ rs (m) 1 A rs,

398 B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 (ii) If (1/r,1/s) S 4, then for all A C m n, A 2 m 1/2 n 1/2 1/r A rs, (iii) If (1/r,1/s) S 5, then for all A C m n, A 2 m 1/2 1/s n 1/2 A rs, (iv) If (1/r,1/s) S 6, then for all A C m n, A 2 m 1/s n 1/2 1/r A rs, (v) If (1/r,1/s) S 7, then for all A C m n, A 2 m 1/s n 1/2 1/r A rs, Proof. If we let p = 2 in Theorems 1 3, then we have (i) (iv) respectively. It remains to prove (v). From Theorem 3, we know that inequality (1) holds in S 7 when p = 2, and so we just need to discuss sharpness to finish the proof of (v). Assume that (1/r,1/s) S 7. Then 1/r 1 log m/(2 log n), and so 1/r log(m 1/2 )/ log n, that is n 1/r m 1/2. Thus n 1/r m 1/s 1/2 m 1/2 m 1/s 1/2 = m 1/s. Consequently, by Lemma 7, there is a constant K, independent of m and n, and a matrix A, all of whose entries are ±1, such that Hence A rs K max(m 1/s n max(1/2 1/r,0),n 1/r m max(1/s 1/2,0) ) = K max(m 1/s,n 1/r m 1/s 1/2 ) = Kn 1/r m 1/s 1/2. A 2 =m 1/2 n 1/2 K 1 m 1/2 n 1/2 n 1/r m 1/2 1/s A rs =K 1 m 1/s n 1/2 1/r A rs, and we have shown that the inequality in (v) is essentially sharp. Using Lemma 2 we can make a translation of Theorem 4 to deal with the case m n. Theorem 5. Suppose m n, and S i = R i,i= 1,...,8, when p = 2. (i) If 1/σ = 1/2 + 1/r and (1/r,1/s) S 1 S 2 S 3, then for all A C m n, A 2 λ σs (n)λ rs (n) 1 A rs,

B.Q. Feng, A. Tonge / Linear Algebra and its Applications 420 (2007) 388 399 399 (ii) If (1/r,1/s) S 4, then for all A C m n, A 2 m 1/2 1/s n 1/2 A rs, (iii) If (1/r,1/s) S 5, then for all A C m n, A 2 m 1/2 n 1/2 1/r A rs, (iv) If (1/r,1/s) S 6, then for all A C m n, A 2 m 1/2 1/s n 1/r A rs, (v) If (1/r,1/s) S 7, then for all A C m n, A 2 m 1/2 1/s n 1/r A rs, We finish this discussion with making a conjecture about the regions S 8 and S 8 of the problem of Klaus and Li as follows. Conjecture (i) Suppose m n.if(1/r,1/s) S 8, then for all A C m n, A 2 n 1/2 (m/n) (1/s 1/2)/(r /2 1) A rs, (ii) Suppose m n.if(1/r,1/s) S 8, then for all A C m n, A 2 m 1/2 (n/m) (1/2 1/r)/(s/2 1) A rs, Acknowledgments Many thanks to the referee for his very helpful suggestions. References [1] G. Bennett, Schur multipliers, Duke Math. J. 44 (3) (1977) 603 640. [2] B.Q. Feng, Equivalence constants for certain matrix norms, Linear Algebra Appl. 374 (2003) 247 253. [3] G.H. Hardy, J.E. Littlewood, Bilinear forms bounded in space [p, q], Quart. J. Math. (Oxford) 5 (1934) 241 254. [4] A.-L. Klaus, C.-K. Li, Isometries for the vector (p, q) norm and the induced (p, q) norm, Linear and Multilinear Algebra 38 (1995) 315 332. [5] B. Osikiewicz, A. Tonge, An interpolation approach to Hardy Littlewood inequalities for norms of operators on sequence spaces, Linear Algebra Appl. 331 (2001) 1 9.