You are currently browsing the monthly archive for August 2019.
William Banks, Kevin Ford, and I have just uploaded to the arXiv our paper “Large prime gaps and probabilistic models“. In this paper we introduce a random model to help understand the connection between two well known conjectures regarding the primes , the Cramér conjecture and the Hardy-Littlewood conjecture:
Conjecture 1 (Cramér conjecture) If
is a large number, then the largest prime gap
in
is of size
. (Granville refines this conjecture to
, where
. Here we use the asymptotic notation
for
,
for
,
for
, and
for
.)
Conjecture 2 (Hardy-Littlewood conjecture) If
are fixed distinct integers, then the number of numbers
with
all prime is
as
, where the singular series
is defined by the formula
(One can view these conjectures as modern versions of two of the classical Landau problems, namely Legendre’s conjecture and the twin prime conjecture respectively.)
A well known connection between the Hardy-Littlewood conjecture and prime gaps was made by Gallagher. Among other things, Gallagher showed that if the Hardy-Littlewood conjecture was true, then the prime gaps with
were asymptotically distributed according to an exponential distribution of mean
, in the sense that
as for any fixed
. Roughly speaking, the way this is established is by using the Hardy-Littlewood conjecture to control the mean values of
for fixed
, where
ranges over the primes in
. The relevance of these quantities arises from the Bonferroni inequalities (or “Brun pure sieve“), which can be formulated as the assertion that
when is even and
when is odd, for any natural number
; setting
and taking means, one then gets upper and lower bounds for the probability that the interval
is free of primes. The most difficult step is to control the mean values of the singular series
as
ranges over
-tuples in a fixed interval such as
.
Heuristically, if one extrapolates the asymptotic (1) to the regime , one is then led to Cramér’s conjecture, since the right-hand side of (1) falls below
when
is significantly larger than
. However, this is not a rigorous derivation of Cramér’s conjecture from the Hardy-Littlewood conjecture, since Gallagher’s computations only establish (1) for fixed choices of
, which is only enough to establish the far weaker bound
, which was already known (see this previous paper for a discussion of the best known unconditional lower bounds on
). An inspection of the argument shows that if one wished to extend (1) to parameter choices
that were allowed to grow with
, then one would need as input a stronger version of the Hardy-Littlewood conjecture in which the length
of the tuple
, as well as the magnitudes of the shifts
, were also allowed to grow with
. Our initial objective in this project was then to quantify exactly what strengthening of the Hardy-Littlewood conjecture would be needed to rigorously imply Cramer’s conjecture. The precise results are technical, but roughly we show results of the following form:
Theorem 3 (Large gaps from Hardy-Littlewood, rough statement)
- If the Hardy-Littlewood conjecture is uniformly true for
-tuples of length
, and with shifts
of size
, with a power savings in the error term, then
.
- If the Hardy-Littlewood conjecture is “true on average” for
-tuples of length
and shifts
of size
for all
, with a power savings in the error term, then
.
In particular, we can recover Cramer’s conjecture given a sufficiently powerful version of the Hardy-Littlewood conjecture “on the average”.
Our proof of this theorem proceeds more or less along the same lines as Gallagher’s calculation, but now with allowed to grow slowly with
. Again, the main difficulty is to accurately estimate average values of the singular series
. Here we found it useful to switch to a probabilistic interpretation of this series. For technical reasons it is convenient to work with a truncated, unnormalised version
of the singular series, for a suitable cutoff ; it turns out that when studying prime tuples of size
, the most convenient cutoff
is the “Pólya magic cutoff“, defined as the largest prime for which
(this is well defined for ); by Mertens’ theorem, we have
. One can interpret
probabilistically as
where is the randomly sifted set of integers formed by removing one residue class
uniformly at random for each prime
. The Hardy-Littlewood conjecture can be viewed as an assertion that the primes
behave in some approximate statistical sense like the random sifted set
, and one can prove the above theorem by using the Bonferroni inequalities both for the primes
and for the random sifted set, and comparing the two (using an even
for the sifted set and an odd
for the primes in order to be able to combine the two together to get a useful bound).
The proof of Theorem 3 ended up not using any properties of the set of primes other than that this set obeyed some form of the Hardy-Littlewood conjectures; the theorem remains true (with suitable notational changes) if this set were replaced by any other set. In order to convince ourselves that our theorem was not vacuous due to our version of the Hardy-Littlewood conjecture being too strong to be true, we then started exploring the question of coming up with random models of
which obeyed various versions of the Hardy-Littlewood and Cramér conjectures.
This line of inquiry was started by Cramér, who introduced what we now call the Cramér random model of the primes, in which each natural number
is selected for membership in
with an independent probability of
. This model matches the primes well in some respects; for instance, it almost surely obeys the “Riemann hypothesis”
and Cramér also showed that the largest gap was almost surely
. On the other hand, it does not obey the Hardy-Littlewood conjecture; more precisely, it obeys a simplified variant of that conjecture in which the singular series
is absent.
Granville proposed a refinement to Cramér’s random model
in which one first sieves out (in each dyadic interval
) all residue classes
for
for a certain threshold
, and then places each surviving natural number
in
with an independent probability
. One can verify that this model obeys the Hardy-Littlewood conjectures, and Granville showed that the largest gap
in this model was almost surely
, leading to his conjecture that this bound also was true for the primes. (Interestingly, this conjecture is not yet borne out by numerics; calculations of prime gaps up to
, for instance, have shown that
never exceeds
in this range. This is not necessarily a conflict, however; Granville’s analysis relies on inspecting gaps in an extremely sparse region of natural numbers that are more devoid of primes than average, and this region is not well explored by existing numerics. See this previous blog post for more discussion of Granville’s argument.)
However, Granville’s model does not produce a power savings in the error term of the Hardy-Littlewood conjectures, mostly due to the need to truncate the singular series at the logarithmic cutoff . After some experimentation, we were able to produce a tractable random model
for the primes which obeyed the Hardy-Littlewood conjectures with power savings, and which reproduced Granville’s gap prediction of
(we also get an upper bound of
for both models, though we expect the lower bound to be closer to the truth); to us, this strengthens the case for Granville’s version of Cramér’s conjecture. The model can be described as follows. We select one residue class
uniformly at random for each prime
, and as before we let
be the sifted set of integers formed by deleting the residue classes
with
. We then set
with Pólya’s magic cutoff (this is the cutoff that gives
a density consistent with the prime number theorem or the Riemann hypothesis). As stated above, we are able to show that almost surely one has
and that the Hardy-Littlewood conjectures hold with power savings for up to
for any fixed
and for shifts
of size
. This is unfortunately a tiny bit weaker than what Theorem 3 requires (which more or less corresponds to the endpoint
), although there is a variant of Theorem 3 that can use this input to produce a lower bound on gaps in the model
(but it is weaker than the one in (3)). In fact we prove a more precise almost sure asymptotic formula for
that involves the optimal bounds for the linear sieve (or interval sieve), in which one deletes one residue class modulo
from an interval
for all primes
up to a given threshold. The lower bound in (3) relates to the case of deleting the
residue classes from
; the upper bound comes from the delicate analysis of the linear sieve by Iwaniec. Improving on either of the two bounds looks to be quite a difficult problem.
The probabilistic analysis of is somewhat more complicated than of
or
as there is now non-trivial coupling between the events
as
varies, although moment methods such as the second moment method are still viable and allow one to verify the Hardy-Littlewood conjectures by a lengthy but fairly straightforward calculation. To analyse large gaps, one has to understand the statistical behaviour of a random linear sieve in which one starts with an interval
and randomly deletes a residue class
for each prime
up to a given threshold. For very small
this is handled by the deterministic theory of the linear sieve as discussed above. For medium sized
, it turns out that there is good concentration of measure thanks to tools such as Bennett’s inequality or Azuma’s inequality, as one can view the sieving process as a martingale or (approximately) as a sum of independent random variables. For larger primes
, in which only a small number of survivors are expected to be sieved out by each residue class, a direct combinatorial calculation of all possible outcomes (involving the random graph that connects interval elements
to primes
if
falls in the random residue class
) turns out to give the best results.
I’ve just uploaded to the arXiv my paper “Quantitative bounds for critically bounded solutions to the Navier-Stokes equations“, submitted to the proceedings of the Linde Hall Inaugural Math Symposium. (I unfortunately had to cancel my physical attendance at this symposium for personal reasons, but was still able to contribute to the proceedings.) In recent years I have been interested in working towards establishing the existence of classical solutions for the Navier-Stokes equations
that blow up in finite time, but this time for a change I took a look at the other side of the theory, namely the conditional regularity results for this equation. There are several such results that assert that if a certain norm of the solution stays bounded (or grows at a controlled rate), then the solution stays regular; taken in the contrapositive, they assert that if a solution blows up at a certain finite time , then certain norms of the solution must also go to infinity. Here are some examples (not an exhaustive list) of such blowup criteria:
- (Leray blowup criterion, 1934) If
blows up at a finite time
, and
, then
for an absolute constant
.
- (Prodi–Serrin–Ladyzhenskaya blowup criterion, 1959-1967) If
blows up at a finite time
, and
, then
, where
.
- (Beale-Kato-Majda blowup criterion, 1984) If
blows up at a finite time
, then
, where
is the vorticity.
- (Kato blowup criterion, 1984) If
blows up at a finite time
, then
for some absolute constant
.
- (Escauriaza-Seregin-Sverak blowup criterion, 2003) If
blows up at a finite time
, then
.
- (Seregin blowup criterion, 2012) If
blows up at a finite time
, then
.
- (Phuc blowup criterion, 2015) If
blows up at a finite time
, then
for any
.
- (Gallagher-Koch-Planchon blowup criterion, 2016) If
blows up at a finite time
, then
for any
.
- (Albritton blowup criterion, 2016) If
blows up at a finite time
, then
for any
.
My current paper is most closely related to the Escauriaza-Seregin-Sverak blowup criterion, which was the first to show a critical (i.e., scale-invariant, or dimensionless) spatial norm, namely , had to become large. This result now has many proofs; for instance, many of the subsequent blowup criterion results imply the Escauriaza-Seregin-Sverak one as a special case, and there are also additional proofs by Gallagher-Koch-Planchon (building on ideas of Kenig-Koch), and by Dong-Du. However, all of these proofs rely on some form of a compactness argument: given a finite time blowup, one extracts some suitable family of rescaled solutions that converges in some weak sense to a limiting solution that has some additional good properties (such as almost periodicity modulo symmetries), which one can then rule out using additional qualitative tools, such as unique continuation and backwards uniqueness theorems for parabolic heat equations. In particular, all known proofs use some version of the backwards uniqueness theorem of Escauriaza, Seregin, and Sverak. Because of this reliance on compactness, the existing proofs of the Escauriaza-Seregin-Sverak blowup criterion are qualitative, in that they do not provide any quantitative information on how fast the
norm will go to infinity (along a subsequence of times).
On the other hand, it is a general principle that qualitative arguments established using compactness methods ought to have quantitative analogues that replace the use of compactness by more complicated substitutes that give effective bounds; see for instance these previous blog posts for more discussion. I therefore was interested in trying to obtain a quantitative version of this blowup criterion that gave reasonably good effective bounds (in particular, my objective was to avoid truly enormous bounds such as tower-exponential or Ackermann function bounds, which often arise if one “naively” tries to make a compactness argument effective). In particular, I obtained the following triple-exponential quantitative regularity bounds:
Theorem 1 If
is a classical solution to Navier-Stokes on
with
and
for
and
.
As a corollary, one can now improve the Escauriaza-Seregin-Sverak blowup criterion to
for some absolute constant , which to my knowledge is the first (very slightly) supercritical blowup criterion for Navier-Stokes in the literature.
The proof uses many of the same quantitative inputs as previous arguments, most notably the Carleman inequalities used to establish unique continuation and backwards uniqueness theorems for backwards heat equations, but also some additional techniques that make the quantitative bounds more efficient. The proof focuses initially on points of concentration of the solution, which we define as points where there is a frequency
for which one has the bound
for a large absolute constant , where
is a Littlewood-Paley projection to frequencies
. (This can be compared with the upper bound of
for the quantity on the left-hand side that follows from (1).) The factor of
normalises the left-hand side of (2) to be dimensionless (i.e., critical). The main task is to show that the dimensionless quantity
cannot get too large; in particular, we end up establishing a bound of the form
from which the above theorem ends up following from a routine adaptation of the local well-posedness and regularity theory for Navier-Stokes.
The strategy is to show that any concentration such as (2) when is large must force a significant component of the
norm of
to also show up at many other locations than
, which eventually contradicts (1) if one can produce enough such regions of non-trivial
norm. (This can be viewed as a quantitative variant of the “rigidity” theorems in some of the previous proofs of the Escauriaza-Seregin-Sverak theorem that rule out solutions that exhibit too much “compactness” or “almost periodicity” in the
topology.) The chain of causality that leads from a concentration (2) at
to significant
norm at other regions of the time slice
is somewhat involved (though simpler than the much more convoluted schemes I initially envisaged for this argument):
- Firstly, by using Duhamel’s formula, one can show that a concentration (2) can only occur (with
large) if there was also a preceding concentration
at some slightly previous point
in spacetime, with
also close to
(more precisely, we have
,
, and
). This can be viewed as a sort of contrapositive of a “local regularity theorem”, such as the ones established by Caffarelli, Kohn, and Nirenberg. A key point here is that the lower bound
in the conclusion (3) is precisely the same as the lower bound in (2), so that this backwards propagation of concentration can be iterated.
- Iterating the previous step, one can find a sequence of concentration points
with the
propagating backwards in time; by using estimates ultimately resulting from the dissipative term in the energy identity, one can extract such a sequence in which the
increase geometrically with time, the
are comparable (up to polynomial factors in
) to the natural frequency scale
, and one has
. Using the “epochs of regularity” theory that ultimately dates back to Leray, and tweaking the
slightly, one can also place the times
in intervals
(of length comparable to a small multiple of
) in which the solution is quite regular (in particular,
enjoy good
bounds on
).
- The concentration (4) can be used to establish a lower bound for the
norm of the vorticity
near
. As is well known, the vorticity obeys the vorticity equation
In the epoch of regularity
, the coefficients
of this equation obey good
bounds, allowing the machinery of Carleman estimates to come into play. Using a Carleman estimate that is used to establish unique continuation results for backwards heat equations, one can propagate this lower bound to also give lower
bounds on the vorticity (and its first derivative) in annuli of the form
for various radii
, although the lower bounds decay at a gaussian rate with
.
- Meanwhile, using an energy pigeonholing argument of Bourgain (which, in this Navier-Stokes context, is actually an enstrophy pigeonholing argument), one can locate some annuli
where (a slightly normalised form of) the entrosphy is small at time
; using a version of the localised enstrophy estimates from a previous paper of mine, one can then propagate this sort of control forward in time, obtaining an “annulus of regularity” of the form
in which one has good estimates; in particular, one has
type bounds on
in this cylindrical annulus.
- By intersecting the previous epoch of regularity
with the above annulus of regularity, we have some lower bounds on the
norm of the vorticity (and its first derivative) in the annulus of regularity. Using a Carleman estimate first introduced by Escauriaza, Seregin, and Sverak, as well as a second application of the Carleman estimate used previously, one can then propagate this lower bound back up to time
, establishing a lower bound for the vorticity on the spatial annulus
. By some basic Littlewood-Paley theory one can parlay this lower bound to a lower bound on the
norm of the velocity
; crucially, this lower bound is uniform in
.
- If
is very large (triple exponential in
!), one can then find enough scales
with disjoint
annuli that the total lower bound on the
norm of
provided by the above arguments is inconsistent with (1), thus establishing the claim.
The chain of causality is summarised in the following image:
It seems natural to conjecture that similar triply logarithmic improvements can be made to several of the other blowup criteria listed above, but I have not attempted to pursue this question. It seems difficult to improve the triple logarithmic factor using only the techniques here; the Bourgain pigeonholing argument inevitably costs one exponential, the Carleman inequalities cost a second, and the stacking of scales at the end to contradict the upper bound costs the third.
Peter Denton, Stephen Parke, Xining Zhang, and I have just uploaded to the arXiv the short unpublished note “Eigenvectors from eigenvalues“. This note gives two proofs of a general eigenvector identity observed recently by Denton, Parke and Zhang in the course of some quantum mechanical calculations. The identity is as follows:
Theorem 1 Let
be an
Hermitian matrix, with eigenvalues
. Let
be a unit eigenvector corresponding to the eigenvalue
, and let
be the
component of
. Then
where
is the
Hermitian matrix formed by deleting the
row and column from
.
For instance, if we have
for some real number ,
-dimensional vector
, and
Hermitian matrix
, then we have
assuming that the denominator is non-zero.
Once one is aware of the identity, it is not so difficult to prove it; we give two proofs, each about half a page long, one of which is based on a variant of the Cauchy-Binet formula, and the other based on properties of the adjugate matrix. But perhaps it is surprising that such a formula exists at all; one does not normally expect to learn much information about eigenvectors purely from knowledge of eigenvalues. In the random matrix theory literature, for instance in this paper of Erdos, Schlein, and Yau, or this later paper of Van Vu and myself, a related identity has been used, namely
but it is not immediately obvious that one can derive the former identity from the latter. (I do so below the fold; we ended up not putting this proof in the note as it was longer than the two other proofs we found. I also give two other proofs below the fold, one from a more geometric perspective and one proceeding via Cramer’s rule.) It was certainly something of a surprise to me that there is no explicit appearance of the components of
in the formula (1) (though they do indirectly appear through their effect on the eigenvalues
; for instance from taking traces one sees that
).
One can get some feeling of the identity (1) by considering some special cases. Suppose for instance that is a diagonal matrix with all distinct entries. The upper left entry
of
is one of the eigenvalues of
. If it is equal to
, then the eigenvalues of
are the other
eigenvalues of
, and now the left and right-hand sides of (1) are equal to
. At the other extreme, if
is equal to a different eigenvalue of
, then
now appears as an eigenvalue of
, and both sides of (1) now vanish. More generally, if we order the eigenvalues
and
, then the Cauchy interlacing inequalities tell us that
for , and
for , so that the right-hand side of (1) lies between
and
, which is of course consistent with (1) as
is a unit vector. Thus the identity relates the coefficient sizes of an eigenvector with the extent to which the Cauchy interlacing inequalities are sharp.
Recent Comments