You are currently browsing the monthly archive for March 2015.
We have seen in previous notes that the operation of forming a Dirichlet series
or twisted Dirichlet series
is an incredibly useful tool for questions in multiplicative number theory. Such series can be viewed as a multiplicative Fourier transform, since the functions and
are multiplicative characters.
Similarly, it turns out that the operation of forming an additive Fourier series
where lies on the (additive) unit circle
and
is the standard additive character, is an incredibly useful tool for additive number theory, particularly when studying additive problems involving three or more variables taking values in sets such as the primes; the deployment of this tool is generally known as the Hardy-Littlewood circle method. (In the analytic number theory literature, the minus sign in the phase
is traditionally omitted, and what is denoted by
here would be referred to instead by
,
or just
.) We list some of the most classical problems in this area:
- (Even Goldbach conjecture) Is it true that every even natural number
greater than two can be expressed as the sum
of two primes?
- (Odd Goldbach conjecture) Is it true that every odd natural number
greater than five can be expressed as the sum
of three primes?
- (Waring problem) For each natural number
, what is the least natural number
such that every natural number
can be expressed as the sum of
or fewer
powers?
- (Asymptotic Waring problem) For each natural number
, what is the least natural number
such that every sufficiently large natural number
can be expressed as the sum of
or fewer
powers?
- (Partition function problem) For any natural number
, let
denote the number of representations of
of the form
where
and
are natural numbers. What is the asymptotic behaviour of
as
?
The Waring problem and its asymptotic version will not be discussed further here, save to note that the Vinogradov mean value theorem (Theorem 13 from Notes 5) and its variants are particularly useful for getting good bounds on ; see for instance the ICM article of Wooley for recent progress on these problems. Similarly, the partition function problem was the original motivation of Hardy and Littlewood in introducing the circle method, but we will not discuss it further here; see e.g. Chapter 20 of Iwaniec-Kowalski for a treatment.
Instead, we will focus our attention on the odd Goldbach conjecture as our model problem. (The even Goldbach conjecture, which involves only two variables instead of three, is unfortunately not amenable to a circle method approach for a variety of reasons, unless the statement is replaced with something weaker, such as an averaged statement; see this previous blog post for further discussion. On the other hand, the methods here can obtain weaker versions of the even Goldbach conjecture, such as showing that “almost all” even numbers are the sum of two primes; see Exercise 34 below.) In particular, we will establish the following celebrated theorem of Vinogradov:
Theorem 1 (Vinogradov’s theorem) Every sufficiently large odd number
is expressible as the sum of three primes.
Recently, the restriction that be sufficiently large was replaced by Helfgott with
, thus establishing the odd Goldbach conjecture in full. This argument followed the same basic approach as Vinogradov (based on the circle method), but with various estimates replaced by “log-free” versions (analogous to the log-free zero-density theorems in Notes 7), combined with careful numerical optimisation of constants and also some numerical work on the even Goldbach problem and on the generalised Riemann hypothesis. We refer the reader to Helfgott’s text for details.
We will in fact show the more precise statement:
Theorem 2 (Quantitative Vinogradov theorem) Let
be an natural number. Then
The implied constants are ineffective.
We dropped the hypothesis that is odd in Theorem 2, but note that
vanishes when
is even. For odd
, we have
Unfortunately, due to the ineffectivity of the constants in Theorem 2 (a consequence of the reliance on the Siegel-Walfisz theorem in the proof of that theorem), one cannot quantify explicitly what “sufficiently large” means in Theorem 1 directly from Theorem 2. However, there is a modification of this theorem which gives effective bounds; see Exercise 32 below.
Exercise 4 Obtain a heuristic derivation of the main term
using the modified Cramér model (Section 1 of Supplement 4).
To prove Theorem 2, we consider the more general problem of estimating sums of the form
for various integers and functions
, which we will take to be finitely supported to avoid issues of convergence.
Suppose that are supported on
; for simplicity, let us first assume the pointwise bound
for all
. (This simple case will not cover the case in Theorem 2, when
are truncated versions of the von Mangoldt function
, but will serve as a warmup to that case.) Then we have the trivial upper bound
A basic observation is that this upper bound is attainable if all “pretend” to behave like the same additive character
for some
. For instance, if
, then we have
when
, and then it is not difficult to show that
as .
The key to the success of the circle method lies in the converse of the above statement: the only way that the trivial upper bound (2) comes close to being sharp is when all correlate with the same character
, or in other words
are simultaneously large. This converse is largely captured by the following two identities:
Exercise 5 Let
be finitely supported functions. Then for any natural number
, show that
and
The traditional approach to using the circle method to compute sums such as proceeds by invoking (3) to express this sum as an integral over the unit circle, then dividing the unit circle into “major arcs” where
are large but computable with high precision, and “minor arcs” where one has estimates to ensure that
are small in both
and
senses. For functions
of number-theoretic significance, such as truncated von Mangoldt functions, the “major arcs” typically consist of those
that are close to a rational number
with
not too large, and the “minor arcs” consist of the remaining portions of the circle. One then obtains lower bounds on the contributions of the major arcs, and upper bounds on the contribution of the minor arcs, in order to get good lower bounds on
.
This traditional approach is covered in many places, such as this text of Vaughan. We will emphasise in this set of notes a slightly different perspective on the circle method, coming from recent developments in additive combinatorics; this approach does not quite give the sharpest quantitative estimates, but it allows for easier generalisation to more combinatorial contexts, for instance when replacing the primes by dense subsets of the primes, or replacing the equation with some other equation or system of equations.
From Exercise 5 and Hölder’s inequality, we immediately obtain
Corollary 6 Let
be finitely supported functions. Then for any natural number
, we have
Similarly for permutations of the
.
In the case when are supported on
and bounded by
, this corollary tells us that we have
is
whenever one has
uniformly in
, and similarly for permutations of
. From this and the triangle inequality, we obtain the following conclusion: if
is supported on
and bounded by
, and
is Fourier-approximated by another function
supported on
and bounded by
in the sense that
Thus, one possible strategy for estimating the sum is, one can effectively replace (or “model”)
by a simpler function
which Fourier-approximates
in the sense that the exponential sums
agree up to error
. For instance:
Exercise 7 Let
be a natural number, and let
be a random subset of
, chosen so that each
has an independent probability of
of lying in
.
- (i) If
and
, show that with probability
as
, one has
uniformly in
. (Hint: for any fixed
, this can be accomplished with quite a good probability (e.g.
) using a concentration of measure inequality, such as Hoeffding’s inequality. To obtain the uniformity in
, round
to the nearest multiple of (say)
and apply the union bound).
- (ii) Show that with probability
, one has
representations of the form
with
(with
treated as an ordered triple, rather than an unordered one).
In the case when is something like the truncated von Mangoldt function
, the quantity
is of size
rather than
. This costs us a logarithmic factor in the above analysis, however we can still conclude that we have the approximation (4) whenever
is another sequence with
such that one has the improved Fourier approximation
uniformly in . (Later on we will obtain a “log-free” version of this implication in which one does not need to gain a factor of
in the error term.)
This suggests a strategy for proving Vinogradov’s theorem: find an approximant to some suitable truncation
of the von Mangoldt function (e.g.
or
) which obeys the Fourier approximation property (5), and such that the expression
is easily computable. It turns out that there are a number of good options for such an approximant
. One of the quickest ways to obtain such an approximation (which is used in Chapter 19 of Iwaniec and Kowalski) is to start with the standard identity
, that is to say
and obtain an approximation by truncating to be less than some threshold
(which, in practice, would be a small power of
):
Thus, for instance, if , the approximant
would be taken to be
One could also use the slightly smoother approximation
in which case we would take
The function is somewhat similar to the continuous Selberg sieve weights studied in Notes 4, with the main difference being that we did not square the divisor sum as we will not need to take
to be non-negative. As long as
is not too large, one can use some sieve-like computations to compute expressions like
quite accurately. The approximation (5) can be justified by using a nice estimate of Davenport that exemplifies the Mobius pseudorandomness heuristic from Supplement 4:
Theorem 8 (Davenport’s estimate) For any
and
, we have
uniformly for all
. The implied constants are ineffective.
This estimate will be proven by splitting into two cases. In the “major arc” case when is close to a rational
with
small (of size
or so), this estimate will be a consequence of the Siegel-Walfisz theorem ( from Notes 2); it is the application of this theorem that is responsible for the ineffective constants. In the remaining “minor arc” case, one proceeds by using a combinatorial identity (such as Vaughan’s identity) to express the sum
in terms of bilinear sums of the form
, and use the Cauchy-Schwarz inequality and the minor arc nature of
to obtain a gain in this case. This will all be done below the fold. We will also use (a rigorous version of) the approximation (6) (or (7)) to establish Vinogradov’s theorem.
A somewhat different looking approximation for the von Mangoldt function that also turns out to be quite useful is
for some that is not too large compared to
. The methods used to establish Theorem 8 can also establish a Fourier approximation that makes (8) precise, and which can yield an alternate proof of Vinogradov’s theorem; this will be done below the fold.
The approximation (8) can be written in a way that makes it more similar to (7):
Exercise 9 Show that the right-hand side of (8) can be rewritten as
where
Then, show the inequalities
and conclude that
(Hint: for the latter estimate, use Theorem 27 of Notes 1.)
The coefficients in the above exercise are quite similar to optimised Selberg sieve coefficients (see Section 2 of Notes 4).
Another approximation to , related to the modified Cramér random model (see Model 10 of Supplement 4) is
where and
is a slowly growing function of
(e.g.
); a closely related approximation is
for as above and
coprime to
. These approximations (closely related to a device known as the “
-trick”) are not as quantitatively accurate as the previous approximations, but can still suffice to establish Vinogradov’s theorem, and also to count many other linear patterns in the primes or subsets of the primes (particularly if one injects some additional tools from additive combinatorics, and specifically the inverse conjecture for the Gowers uniformity norms); see this paper of Ben Green and myself for more discussion (and this more recent paper of Shao for an analysis of this approach in the context of Vinogradov-type theorems). The following exercise expresses the approximation (9) in a form similar to the previous approximation (8):
Exercise 10 With
as above, show that
for all natural numbers
.
Kaisa Matomaki, Maksym Radziwill, and I have just uploaded to the arXiv our paper “An averaged form of Chowla’s conjecture“. This paper concerns a weaker variant of the famous conjecture of Chowla (discussed for instance in this previous post) that
as for any distinct natural numbers
, where
denotes the Liouville function. (One could also replace the Liouville function here by the Möbius function
and obtain a morally equivalent conjecture.) This conjecture remains open for any
; for instance the assertion
is a variant of the twin prime conjecture (though possibly a tiny bit easier to prove), and is subject to the notorious parity barrier (as discussed in this previous post).
Our main result asserts, roughly speaking, that Chowla’s conjecture can be established unconditionally provided one has non-trivial averaging in the parameters. More precisely, one has
Theorem 1 (Chowla on the average) Suppose
is a quantity that goes to infinity as
(but it can go to infinity arbitrarily slowly). Then for any fixed
, we have
In fact, we can remove one of the averaging parameters and obtain
Actually we can make the decay rate a bit more quantitative, gaining about over the trivial bound. The key case is
; while the unaveraged Chowla conjecture becomes more difficult as
increases, the averaged Chowla conjecture does not increase in difficulty due to the increasing amount of averaging for larger
, and we end up deducing the higher
case of the conjecture from the
case by an elementary argument.
The proof of the theorem proceeds as follows. By exploiting the Fourier-analytic identity
(related to a standard Fourier-analytic identity for the Gowers norm) it turns out that the
case of the above theorem can basically be derived from an estimate of the form
uniformly for all . For “major arc”
, close to a rational
for small
, we can establish this bound from a generalisation of a recent result of Matomaki and Radziwill (discussed in this previous post) on averages of multiplicative functions in short intervals. For “minor arc”
, we can proceed instead from an argument of Katai and Bourgain-Sarnak-Ziegler (discussed in this previous post).
The argument also extends to other bounded multiplicative functions than the Liouville function. Chowla’s conjecture was generalised by Elliott, who roughly speaking conjectured that the copies of
in Chowla’s conjecture could be replaced by arbitrary bounded multiplicative functions
as long as these functions were far from a twisted Dirichlet character
in the sense that
(This type of distance is incidentally now a fundamental notion in the Granville-Soundararajan “pretentious” approach to multiplicative number theory.) During our work on this project, we found that Elliott’s conjecture is not quite true as stated due to a technicality: one can cook up a bounded multiplicative function which behaves like
on scales
for some
going to infinity and some slowly varying
, and such a function will be far from any fixed Dirichlet character whilst still having many large correlations (e.g. the pair correlations
will be large). In our paper we propose a technical “fix” to Elliott’s conjecture (replacing (1) by a truncated variant), and show that this repaired version of Elliott’s conjecture is true on the average in much the same way that Chowla’s conjecture is. (If one restricts attention to real-valued multiplicative functions, then this technical issue does not show up, basically because one can assume without loss of generality that
in this case; we discuss this fact in an appendix to the paper.)
A major topic of interest of analytic number theory is the asymptotic behaviour of the Riemann zeta function in the critical strip
in the limit
. For the purposes of this set of notes, it is a little simpler technically to work with the log-magnitude
of the zeta function. (In principle, one can reconstruct a branch of
, and hence
itself, from
using the Cauchy-Riemann equations, or tools such as the Borel-Carathéodory theorem, see Exercise 40 of Supplement 2.)
One has the classical estimate
(See e.g. Exercise 37 from Supplement 3.) In view of this, let us define the normalised log-magnitudes for any
by the formula
informally, this is a normalised window into near
. One can rephrase several assertions about the zeta function in terms of the asymptotic behaviour of
. For instance:
- (i) The bound (1) implies that
is asymptotically locally bounded from above in the limit
, thus for any compact set
we have
for
and
sufficiently large. In fact the implied constant in
only depends on the projection of
to the real axis.
- (ii) For
, we have the bounds
which implies that
converges locally uniformly as
to zero in the region
.
- (iii) The functional equation, together with the symmetry
, implies that
which by Exercise 17 of Supplement 3 shows that
as
, locally uniformly in
. In particular, when combined with the previous item, we see that
converges locally uniformly as
to
in the region
.
- (iv) From Jensen’s formula (Theorem 16 of Supplement 2) we see that
is a subharmonic function, and thus
is subharmonic as well. In particular we have the mean value inequality
for any disk
, where the integral is with respect to area measure. From this and (ii) we conclude that
for any disk with
and sufficiently large
; combining this with (i) we conclude that
is asymptotically locally bounded in
in the limit
, thus for any compact set
we have
for sufficiently large
.
From (iv) and the usual Arzela-Ascoli diagonalisation argument, we see that the are asymptotically compact in the topology of distributions: given any sequence
tending to
, one can extract a subsequence such that the
converge in the sense of distributions. Let us then define a normalised limit profile of
to be a distributional limit
of a sequence of
; they are analogous to limiting profiles in PDE, and also to the more recent introduction of “graphons” in the theory of graph limits. Then by taking limits in (i)-(iv) we can say a lot about such normalised limit profiles
(up to almost everywhere equivalence, which is an issue we will address shortly):
- (i)
is bounded from above in the critical strip
.
- (ii)
vanishes on
.
- (iii) We have the functional equation
for all
. In particular
for
.
- (iv)
is subharmonic.
Unfortunately, (i)-(iv) fail to characterise completely. For instance, one could have
for any convex function
of
that equals
for
,
for
, and obeys the functional equation
, and this would be consistent with (i)-(iv). One can also perturb such examples in a region where
is strictly convex to create further examples of functions obeying (i)-(iv). Note from subharmonicity that the function
is always going to be convex in
; this can be seen as a limiting case of the Hadamard three-lines theorem (Exercise 41 of Supplement 2).
We pause to address one minor technicality. We have defined as a distributional limit, and as such it is a priori only defined up to almost everywhere equivalence. However, due to subharmonicity, there is a unique upper semi-continuous representative of
(taking values in
), defined by the formula
for any (note from subharmonicity that the expression in the limit is monotone nonincreasing as
, and is also continuous in
). We will now view this upper semi-continuous representative of
as the canonical representative of
, so that
is now defined everywhere, rather than up to almost everywhere equivalence.
By a classical theorem of Riesz, a function is subharmonic if and only if the distribution
is a non-negative measure, where
is the Laplacian in the
coordinates. Jensen’s formula (or Greens’ theorem), when interpreted distributionally, tells us that
away from the real axis, where ranges over the non-trivial zeroes of
. Thus, if
is a normalised limit profile for
that is the distributional limit of
, then we have
where is a non-negative measure which is the limit in the vague topology of the measures
Thus is a normalised limit profile of the zeroes of the Riemann zeta function.
Using this machinery, we can recover many classical theorems about the Riemann zeta function by “soft” arguments that do not require extensive calculation. Here are some examples:
Theorem 1 The Riemann hypothesis implies the Lindelöf hypothesis.
Proof: It suffices to show that any limiting profile (arising as the limit of some
) vanishes on the critical line
. But if the Riemann hypothesis holds, then the measures
are supported on the critical line
, so the normalised limit profile
is also supported on this line. This implies that
is harmonic outside of the critical line. By (ii) and unique continuation for harmonic functions, this implies that
vanishes on the half-space
(and equals
on the complementary half-space, by (iii)), giving the claim.
In fact, we have the following sharper statement:
Theorem 2 (Backlund) The Lindelöf hypothesis is equivalent to the assertion that for any fixed
, the number of zeroes in the region
is
as
.
Proof: If the latter claim holds, then for any , the measures
assign a mass of
to any region of the form
as
for any fixed
and
. Thus the normalised limiting profile measure
is supported on the critical line, and we can repeat the previous argument.
Conversely, suppose the claim fails, then we can find a sequence and
such that
assigns a mass of
to the region
. Extracting a normalised limiting profile, we conclude that the normalised limiting profile measure
is non-trivial somewhere to the right of the critical line, so the associated subharmonic function
is not harmonic everywhere to the right of the critical line. From the maximum principle and (ii) this implies that
has to be positive somewhere on the critical line, but this contradicts the Lindelöf hypothesis. (One has to take a bit of care in the last step since
only converges to
in the sense of distributions, but it turns out that the subharmonicity of all the functions involved gives enough regularity to justify the argument; we omit the details here.)
Theorem 3 (Littlewood) Assume the Lindelöf hypothesis. Then for any fixed
, the number of zeroes in the region
is
as
.
Proof: By the previous arguments, the only possible normalised limiting profile for is
. Taking distributional Laplacians, we see that the only possible normalised limiting profile for the zeroes is Lebesgue measure on the critical line. Thus,
can only converge to
as
, and the claim follows.
Even without the Lindelöf hypothesis, we have the following result:
Theorem 4 (Titchmarsh) For any fixed
, there are
zeroes in the region
for sufficiently large
.
Among other things, this theorem recovers a classical result of Littlewood that the gaps between the imaginary parts of the zeroes goes to zero, even without assuming unproven conjectures such as the Riemann or Lindelöf hypotheses.
Proof: Suppose for contradiction that this were not the case, then we can find and a sequence
such that
contains
zeroes. Passing to a subsequence to extract a limit profile, we conclude that the normalised limit profile measure
assigns no mass to the horizontal strip
. Thus the associated subharmonic function
is actually harmonic on this strip. But by (ii) and unique continuation this forces
to vanish on this strip, contradicting the functional equation (iii).
Exercise 5 Use limiting profiles to obtain the matching upper bound of
for the number of zeroes in
for sufficiently large
.
Remark 6 One can remove the need to take limiting profiles in the above arguments if one can come up with quantitative (or “hard”) substitutes for qualitative (or “soft”) results such as the unique continuation property for harmonic functions. This would also allow one to replace the qualitative decay rates
with more quantitative decay rates such as
or
. Indeed, the classical proofs of the above theorems come with quantitative bounds that are typically of this form (see e.g. the text of Titchmarsh for details).
Exercise 7 Let
denote the quantity
, where the branch of the argument is taken by using a line segment connecting
to (say)
, and then to
. If we have a sequence
producing normalised limit profiles
for
and the zeroes respectively, show that
converges in the sense of distributions to the function
, or equivalently
Conclude in particular that if the Lindelöf hypothesis holds, then
as
.
A little bit more about the normalised limit profiles are known unconditionally, beyond (i)-(iv). For instance, from Exercise 3 of Notes 5 we have
as
, which implies that any normalised limit profile
for
is bounded by
on the critical line, beating the bound of
coming from convexity and (ii), (iii), and then convexity can be used to further bound
away from the critical line also. Some further small improvements of this type are known (coming from various methods for estimating exponential sums), though they fall well short of determining
completely at our current level of understanding. Of course, given that we believe the Riemann hypothesis (and hence the Lindelöf hypothesis) to be true, the only actual limit profile that should exist is
(in fact this assertion is equivalent to the Lindelöf hypothesis, by the arguments above).
Better control on limiting profiles is available if we do not insist on controlling for all values of the height parameter
, but only for most such values, thanks to the existence of several mean value theorems for the zeta function, as discussed in Notes 6; we discuss this below the fold.
Recent Comments