A basic problem in harmonic analysis (as well as in linear algebra, random matrix theory, and high-dimensional geometry) is to estimate the operator norm of a linear map between two Hilbert spaces, which we will take to be complex for sake of discussion. Even the finite-dimensional case is of interest, as this operator norm is the same as the largest singular value of the matrix associated to .

In general, this operator norm is hard to compute precisely, except in special cases. One such special case is that of a *diagonal operator*, such as that associated to an diagonal matrix . In this case, the operator norm is simply the supremum norm of the diagonal coefficients:

A variant of (1) is Schur’s test, which for simplicity we will phrase in the setting of finite-dimensional operators given by a matrix via the usual formula

A simple version of this test is as follows: if all the absolute row sums and columns sums of are bounded by some constant , thus

(note that this generalises (the upper bound in) (1).) Indeed, to see (4), it suffices by duality and homogeneity to show that

whenever and are sequences with ; but this easily follows from the arithmetic mean-geometric mean inequality

Schur’s test (4) (and its many generalisations to weighted situations, or to Lebesgue or Lorentz spaces) is particularly useful for controlling operators in which the role of oscillation (as reflected in the *phase* of the coefficients , as opposed to just their magnitudes ) is not decisive. However, it is of limited use in situations that involve a lot of cancellation. For this, a different test, known as the Cotlar-Stein lemma, is much more flexible and powerful. It can be viewed in a sense as a non-commutative variant of Schur’s test (4) (or of (1)), in which the scalar coefficients or are replaced by operators instead.

To illustrate the basic flavour of the result, let us return to the bound (1), and now consider instead a *block-diagonal* matrix

where each is now a matrix, and so is an matrix with . Then we have

Indeed, the lower bound is trivial (as can be seen by testing on vectors which are supported on the block of coordinates), while to establish the upper bound, one can make use of the orthogonal decomposition

to decompose an arbitrary vector as

with , in which case we have

and the upper bound in (6) then follows from a simple computation.

The operator associated to the matrix in (5) can be viewed as a sum , where each corresponds to the block of , in which case (6) can also be written as

When is large, this is a significant improvement over the triangle inequality, which merely gives

The reason for this gain can ultimately be traced back to the “orthogonality” of the ; that they “occupy different columns” and “different rows” of the range and domain of . This is obvious when viewed in the matrix formalism, but can also be described in the more abstract Hilbert space operator formalism via the identities

whenever . (The first identity asserts that the ranges of the are orthogonal to each other, and the second asserts that the coranges of the (the ranges of the adjoints ) are orthogonal to each other.) By replacing (7) with a more abstract orthogonal decomposition into these ranges and coranges, one can in fact deduce (8) directly from (9) and (10).

The *Cotlar-Stein lemma* is an extension of this observation to the case where the are merely *almost orthogonal* rather than *orthogonal*, in a manner somewhat analogous to how Schur’s test (partially) extends (1) to the non-diagonal case. Specifically, we have

Lemma 1 (Cotlar-Stein lemma)Let be a finite sequence of bounded linear operators from one Hilbert space to another , obeying the boundsfor all and some (compare with (2), (3)). Then one has

that the hypothesis (11) (or (12)) already gives the bound

on each component of , which by the triangle inequality gives the inferior bound

the point of the Cotlar-Stein lemma is that the dependence on in this bound is eliminated in (13), which in particular makes the bound suitable for extension to the limit (see Remark 1 below).

The Cotlar-Stein lemma was first established by Cotlar in the special case of commuting self-adjoint operators, and then independently by Cotlar and Stein in full generality, with the proof appearing in a subsequent paper of Knapp and Stein.

The Cotlar-Stein lemma is often useful in controlling operators such as singular integral operators or pseudo-differential operators which “do not mix scales together too much”, in that operators map functions “that oscillate at a given scale ” to functions that still mostly oscillate at the same scale . In that case, one can often split into components which essentically capture the scale behaviour, and understanding boundedness properties of then reduces to establishing the boundedness of the simpler operators (and of establishing a sufficient decay in products such as or when and are separated from each other). In some cases, one can use Fourier-analytic tools such as Littlewood-Paley projections to generate the , but the true power of the Cotlar-Stein lemma comes from situations in which the Fourier transform is not suitable, such as when one has a complicated domain (e.g. a manifold or a non-abelian Lie group), or very rough coefficients (which would then have badly behaved Fourier behaviour). One can then select the decomposition in a fashion that is tailored to the particular operator , and is not necessarily dictated by Fourier-analytic considerations.

Once one is in the almost orthogonal setting, as opposed to the genuinely orthogonal setting, the previous arguments based on orthogonal projection seem to fail completely. Instead, the proof of the Cotlar-Stein lemma proceeds via an elegant application of the tensor power trick (or perhaps more accurately, the power method), in which the operator norm of is understood through the operator norm of a large power of (or more precisely, of its self-adjoint square or ). Indeed, from an iteration of (14) we see that for any natural number , one has

To estimate the right-hand side, we expand out the right-hand side and apply the triangle inequality to bound it by

Recall that when we applied the triangle inequality directly to , we lost a factor of in the final estimate; it will turn out that we will lose a similar factor here, but this factor will eventually be attenuated into nothingness by the tensor power trick.

To bound (17), we use the basic inequality in two different ways. If we group the product in pairs, we can bound the summand of (17) by

On the other hand, we can group the product by pairs in another way, to obtain the bound of

We bound and crudely by using (15). Taking the geometric mean of the above bounds, we can thus bound (17) by

If we then sum this series first in , then in , then moving back all the way to , using (11) and (12) alternately, we obtain a final bound of

for (16). Taking roots, we obtain

Sending , we obtain the claim.

Remark 1As observed in a number of places (see e.g. page 318 of Stein’s book, or this paper of Comech, the Cotlar-Stein lemma can be extended to infinite sums (with the obvious changes to the hypotheses (11), (12)). Indeed, one can show that for any , the sum is unconditionally convergent in (and furthermore has bounded -variation), and the resulting operator is a bounded linear operator with an operator norm bound on .

Remark 2If we specialise to the case where all the are equal, we see that the bound in the Cotlar-Stein lemma is sharp, at least in this case. Thus we see how the tensor power trick can convert an inefficient argument, such as that obtained using the triangle inequality or crude bounds such as (15), into an efficient one.

Remark 3One can prove Schur’s test by a similar method. Indeed, starting from the inequality(which follows easily from the singular value decomposition), we can bound by

Estimating the other two terms in the summand by , and then repeatedly summing the indices one at a time as before, we obtain

and the claim follows from the tensor power trick as before. On the other hand, in the converse direction, I do not know of any way to prove the Cotlar-Stein lemma that does not basically go through the tensor power argument.

## 10 comments

Comments feed for this article

25 May, 2011 at 9:37 pm

AnonymousProf. Tao,

Could you please write an expository note on Calderon-Zygmund decomposition?

Thanks

26 May, 2011 at 8:41 am

Terence Taohttp://www.math.ucla.edu/~tao/247a.1.06f/notes4.dvi

26 May, 2011 at 11:56 am

AnonymousThank you Prof. Tao

26 May, 2011 at 7:10 am

Joerg GrandeThe comment on the triangle inequality between (8) and (9) should use a sum, not the supremum.

Best regards

Joerg

[Corrected, thanks – T.]31 May, 2011 at 3:49 am

JanakiramanProf.Tao

I have a question regarding the norm of a block diagonal operator. Eq.8 and Eq. 9 and 10 are if and only if statements? Also i feel if the column vectors of the operator can be normalized such that T=QM where Q is the unitary matrix and M is the diagonal matrix whose diagonal entries are the norms of the column vectors. This is very similar to singular value decomposition (without the unitary operator on the right hand side) where the largest value of M is nothing but the largest norm of the column vector. So is there a way to relate the singular values to the norm of the column vectors?

31 May, 2011 at 3:59 am

JanakiramanI guess the unitary operator on the right hand side is just the identity. And now i feel it satisfies the singular value decomposition and the norm of the operator is the supremum amongst the norm of the column vectors.

I’m sorry i got the eqn 8 and 9 wrong. They are orthogonal in sense, i believe they map the vector to different subspaces of the vector space. I guess that has nothing to do with unitary. Sorry for the wrong correlation i tried to create

13 June, 2011 at 9:45 am

Seventh Linkfest[…] Tao: The Cotlar-Stein Lemma, Locally compact groups with faithful finite-dimensional representations, van Dantzig’s […]

2 October, 2011 at 5:29 am

Tony CarberyThe extension of the Cotlar–Stein lemma to infinite sums is classical, and is discussed on p. 318 of Stein’s book Harmonic Analysis. It certainly goes back as far as work by David, Journe and Semmes in the mid 1980’s

2 October, 2011 at 3:12 pm

Terence TaoThanks, Tony! I’m not sure how I managed to miss that bit of discussion; I’ve read through that particular chapter multiple times. Anyway, thanks again for pointing it out.

30 October, 2013 at 7:08 am

A spectral theory proof of the algebraic regularity lemma | What's new[…] now prove this lemma, by combining spectral theory with an (iterated) method. We may assume that is sufficiently large depending on , as the claim is trivial otherwise (just […]