You are currently browsing the tag archive for the ‘Gelfand-Tsetlin patterns’ tag.
Suppose we have an matrix
that is expressed in block-matrix form as
where is an
matrix,
is an
matrix,
is an
matrix, and
is a
matrix for some
. If
is invertible, we can use the technique of Schur complementation to express the inverse of
(if it exists) in terms of the inverse of
, and the other components
of course. Indeed, to solve the equation
where are
column vectors and
are
column vectors, we can expand this out as a system
Using the invertibility of , we can write the first equation as
and substituting this into the second equation yields
and thus (assuming that is invertible)
and then inserting this back into (1) gives
Comparing this with
we have managed to express the inverse of as
One can consider the inverse problem: given the inverse of
, does one have a nice formula for the inverse
of the minor
? Trying to recover this directly from (2) looks somewhat messy. However, one can proceed as follows. Let
denote the
matrix
(with the
identity matrix), and let
be its transpose:
Then for any scalar (which we identify with
times the identity matrix), one has
and hence by (2)
noting that the inverses here will exist for large enough. Taking limits as
, we conclude that
On the other hand, by the Woodbury matrix identity (discussed in this previous blog post), we have
and hence on taking limits and comparing with the preceding identity, one has
This achieves the aim of expressing the inverse of the minor in terms of the inverse of the full matrix. Taking traces and rearranging, we conclude in particular that
In the case, this can be simplified to
where is the
basis column vector.
We can apply this identity to understand how the spectrum of an random matrix
relates to that of its top left
minor
. Subtracting any complex multiple
of the identity from
(and hence from
), we can relate the Stieltjes transform
of
with the Stieltjes transform
of
:
At this point we begin to proceed informally. Assume for sake of argument that the random matrix is Hermitian, with distribution that is invariant under conjugation by the unitary group
; for instance,
could be drawn from the Gaussian Unitary Ensemble (GUE), or alternatively
could be of the form
for some real diagonal matrix
and
a unitary matrix drawn randomly from
using Haar measure. To fix normalisations we will assume that the eigenvalues of
are typically of size
. Then
is also Hermitian and
-invariant. Furthermore, the law of
will be the same as the law of
, where
is now drawn uniformly from the unit sphere (independently of
). Diagonalising
into eigenvalues
and eigenvectors
, we have
One can think of as a random (complex) Gaussian vector, divided by the magnitude of that vector (which, by the Chernoff inequality, will concentrate to
). Thus the coefficients
with respect to the orthonormal basis
can be thought of as independent (complex) Gaussian vectors, divided by that magnitude. Using this and the Chernoff inequality again, we see (for
distance
away from the real axis at least) that one has the concentration of measure
and thus
(that is to say, the diagonal entries of are roughly constant). Similarly we have
Inserting this into (5) and discarding terms of size , we thus conclude the approximate relationship
This can be viewed as a difference equation for the Stieltjes transform of top left minors of . Iterating this equation, and formally replacing the difference equation by a differential equation in the large
limit, we see that when
is large and
for some
, one expects the top left
minor
of
to have Stieltjes transform
where solves the Burgers-type equation
with initial data .
Example 1 If
is a constant multiple
of the identity, then
. One checks that
is a steady state solution to (7), which is unsurprising given that all minors of
are also
times the identity.
Example 2 If
is GUE normalised so that each entry has variance
, then by the semi-circular law (see previous notes) one has
(using an appropriate branch of the square root). One can then verify the self-similar solution
to (7), which is consistent with the fact that a top
minor of
also has the law of GUE, with each entry having variance
when
.
One can justify the approximation (6) given a sufficiently good well-posedness theory for the equation (7). We will not do so here, but will note that (as with the classical inviscid Burgers equation) the equation can be solved exactly (formally, at least) by the method of characteristics. For any initial position , we consider the characteristic flow
formed by solving the ODE
with initial data , ignoring for this discussion the problems of existence and uniqueness. Then from the chain rule, the equation (7) implies that
and thus . Inserting this back into (8) we see that
and thus (7) may be solved implicitly via the equation
for all and
.
Remark 3 In practice, the equation (9) may stop working when
crosses the real axis, as (7) does not necessarily hold in this region. It is a cute exercise (ultimately coming from the Cauchy-Schwarz inequality) to show that this crossing always happens, for instance if
has positive imaginary part then
necessarily has negative or zero imaginary part.
Example 4 Suppose we have
as in Example 1. Then (9) becomes
for any
, which after making the change of variables
becomes
as in Example 1.
Example 5 Suppose we have
as in Example 2. Then (9) becomes
If we write
one can calculate that
and hence
One can recover the spectral measure from the Stieltjes transform
as the weak limit of
as
; we write this informally as
In this informal notation, we have for instance that
which can be interpreted as the fact that the Cauchy distributions converge weakly to the Dirac mass at
as
. Similarly, the spectral measure associated to (10) is the semicircular measure
.
If we let be the spectral measure associated to
, then the curve
from
to the space of measures is the high-dimensional limit
of a Gelfand-Tsetlin pattern (discussed in this previous post), if the pattern is randomly generated amongst all matrices
with spectrum asymptotic to
as
. For instance, if
, then the curve is
, corresponding to a pattern that is entirely filled with
‘s. If instead
is a semicircular distribution, then the pattern is
thus at height from the top, the pattern is semicircular on the interval
. The interlacing property of Gelfand-Tsetlin patterns translates to the claim that
(resp.
) is non-decreasing (resp. non-increasing) in
for any fixed
. In principle one should be able to establish these monotonicity claims directly from the PDE (7) or from the implicit solution (9), but it was not clear to me how to do so.
An interesting example of such a limiting Gelfand-Tsetlin pattern occurs when , which corresponds to
being
, where
is an orthogonal projection to a random
-dimensional subspace of
. Here we have
and so (9) in this case becomes
A tedious calculation then gives the solution
For , there are simple poles at
, and the associated measure is
This reflects the interlacing property, which forces of the
eigenvalues of the
minor to be equal to
(resp.
). For
, the poles disappear and one just has
For , one has an inverse semicircle distribution
There is presumably a direct geometric explanation of this fact (basically describing the singular values of the product of two random orthogonal projections to half-dimensional subspaces of ), but I do not know of one off-hand.
The evolution of can also be understood using the
-transform and
-transform from free probability. Formally, letlet
be the inverse of
, thus
for all , and then define the
-transform
The equation (9) may be rewritten as
and hence
See these previous notes for a discussion of free probability topics such as the -transform.
Example 6 If
then the
transform is
.
Example 7 If
is given by (10), then the
transform is
Example 8 If
is given by (11), then the
transform is
This simple relationship (12) is essentially due to Nica and Speicher (thanks to Dima Shylakhtenko for this reference). It has the remarkable consequence that when is the reciprocal of a natural number
, then
is the free arithmetic mean of
copies of
, that is to say
is the free convolution
of
copies of
, pushed forward by the map
. In terms of random matrices, this is asserting that the top
minor of a random matrix
has spectral measure approximately equal to that of an arithmetic mean
of
independent copies of
, so that the process of taking top left minors is in some sense a continuous analogue of the process of taking freely independent arithmetic means. There ought to be a geometric proof of this assertion, but I do not know of one. In the limit
(or
), the
-transform becomes linear and the spectral measure becomes semicircular, which is of course consistent with the free central limit theorem.
In a similar vein, if one defines the function
and inverts it to obtain a function with
for all , then the
-transform
is defined by
Writing
for any ,
, we have
and so (9) becomes
which simplifies to
replacing by
we obtain
and thus
and hence
One can compute to be the
-transform of the measure
; from the link between
-transforms and free products (see e.g. these notes of Guionnet), we conclude that
is the free product of
and
. This is consistent with the random matrix theory interpretation, since
is also the spectral measure of
, where
is the orthogonal projection to the span of the first
basis elements, so in particular
has spectral measure
. If
is unitarily invariant then (by a fundamental result of Voiculescu) it is asymptotically freely independent of
, so the spectral measure of
is asymptotically the free product of that of
and of
.
Israel Gelfand, who made profound and prolific contributions to many areas of mathematics, including functional analysis, representation theory, operator algebras, and partial differential equations, died on Monday, age 96.
Gelfand’s beautiful theory of -algebras and related spaces made quite an impact on me as a graduate student in Princeton, to the point where I was seriously considering working in this area; but there was not much activity in operator algebras at the time there, and I ended up working in harmonic analysis under Eli Stein instead. (Though I am currently involved in another operator algebras project, of which I hope to be able to discuss in the near future. The commutative version of Gelfand’s theory is discussed in these lecture notes of mine.)
I met Gelfand only once, in one of the famous “Gelfand seminars” at the IHES in 2000. The speaker was Tim Gowers, on his new proof of Szemerédi’s theorem. (Endre Szemerédi, incidentally, was Gelfand’s student.) Gelfand’s introduction to the seminar, on the subject of Banach spaces which both mathematicians contributed so greatly to, was approximately as long as Gowers’ talk itself!
There are far too many contributions to mathematics by Gelfand to name here, so I will only mention two. The first are the Gelfand-Tsetlin patterns, induced by an Hermitian matrix
. Such matrices have
real eigenvalues
. If one takes the top
minor, this is another Hermitian matrix, whose
eigenvalues
intersperse the
eigenvalues of the original matrix:
for every
. This interspersing can be easily seen from the minimax characterisation
of the eigenvalues of , with the eigenvalues of the
minor being similar but with
now restricted to a subspace of
rather than
.
Similarly, the eigenvalues of the top
minor of
intersperse those of the previous minor. Repeating this procedure one eventually gets a pyramid of real numbers of height and width
, with the numbers in each row interspersing the ones in the row below. Such a pattern is known as a Gelfand-Tsetlin pattern. The space of such patterns forms a convex cone, and (if one fixes the initial eigenvalues
) becomes a compact convex polytope. If one fixes the initial eigenvalues
of
but chooses the eigenvectors randomly (using the Haar measure of the unitary group), then the resulting Gelfand-Tsetlin pattern is uniformly distributed across this polytope; the
case of this observation is essentially the classic observation of Archimedes that the cross-sectional areas of a sphere and a circumscribing cylinder are the same. (Ultimately, the reason for this is that the Gelfand-Tsetlin pattern almost turns the space of all
with a fixed spectrum (i.e. the co-adjoint orbit associated to that spectrum) into a toric variety. More precisely, there exists a mostly diffeomorphic map from the co-adjoint orbit to a (singular) toric variety, and the Gelfand-Tsetlin pattern induces a complete set of action variables on that variety.) There is also a “quantum” (or more precisely, representation-theoretic) version of this observation, in which one can decompose any irreducible representation of the unitary group
into a canonical basis (the Gelfand-Tsetlin basis), indexed by integer-valued Gelfand-Tsetlin patterns, by first decomposing this representation into irreducible representations of
, then
, and so forth.
The structure, symplectic geometry, and representation theory of Gelfand-Tsetlin patterns was enormously influential in my own work with Allen Knutson on honeycomb patterns, which control the sums of Hermitian matrices and also the structure constants of the tensor product operation for representations of ; indeed, Gelfand-Tsetlin patterns arise as the degenerate limit of honeycombs in three different ways, and we in fact discovered honeycombs by trying to glue three Gelfand-Tsetlin patterns together. (See for instance our Notices article for more discussion. The honeycomb analogue of the representation-theoretic properties of these patterns was eventually established by Henriques and Kamnitzer, using
crystals and their Kashiwara bases.)
The second contribution of Gelfand I want to discuss is the Gelfand-Levitan-Marchenko equation for solving the one-dimensional inverse scattering problem: given the scattering data of an unknown potential function , recover
. This is already interesting in and of itself, but is also instrumental in solving integrable systems such as the Korteweg-de Vries equation, because the Lax pair formulation of such equations implies that they can be linearised (and solved explicitly) by applying the scattering and inverse scattering transforms associated with the Lax operator. I discuss the derivation of this equation below the fold.

Recent Comments