You are currently browsing the category archive for the ‘math.NT’ category.
Tamar Ziegler and I have just uploaded to the arXiv our paper “Infinite partial sumsets in the primes“. This is a short paper inspired by a recent result of Kra, Moreira, Richter, and Robertson (discussed for instance in this Quanta article from last December) showing that for any set of natural numbers of positive upper density, there exists a sequence
of natural numbers and a shift
such that
for all
this answers a question of Erdős). In view of the “transference principle“, it is then plausible to ask whether the same result holds if
is replaced by the primes. We can show the following results:
Theorem 1
- (i) If the Hardy-Littlewood prime tuples conjecture (or the weaker conjecture of Dickson) is true, then there exists an increasing sequence
of primes such that
is prime for all
.
- (ii) Unconditionally, there exist increasing sequences
and
of natural numbers such that
is prime for all
.
- (iii) These conclusions fail if “prime” is replaced by “positive (relative) density subset of the primes” (even if the density is equal to 1).
We remark that it was shown by Balog that there (unconditionally) exist arbitrarily long but finite sequences of primes such that
is prime for all
. (This result can also be recovered from the later results of Ben Green, myself, and Tamar Ziegler.) Also, it had previously been shown by Granville that on the Hardy-Littlewood prime tuples conjecture, there existed increasing sequences
and
of natural numbers such that
is prime for all
.
The conclusion of (i) is stronger than that of (ii) (which is of course consistent with the former being conditional and the latter unconditional). The conclusion (ii) also implies the well-known theorem of Maynard that for any given , there exist infinitely many
-tuples of primes of bounded diameter, and indeed our proof of (ii) uses the same “Maynard sieve” that powers the proof of that theorem (though we use a formulation of that sieve closer to that in this blog post of mine). Indeed, the failure of (iii) basically arises from the failure of Maynard’s theorem for dense subsets of primes, simply by removing those clusters of primes that are unusually closely spaced.
Our proof of (i) was initially inspired by the topological dynamics methods used by Kra, Moreira, Richter, and Robertson, but we managed to condense it to a purely elementary argument (taking up only half a page) that makes no reference to topological dynamics and builds up the sequence recursively by repeated application of the prime tuples conjecture.
The proof of (ii) takes up the majority of the paper. It is easiest to phrase the argument in terms of “prime-producing tuples” – tuples for which there are infinitely many
with
all prime. Maynard’s theorem is equivalent to the existence of arbitrarily long prime-producing tuples; our theorem is equivalent to the stronger assertion that there exist an infinite sequence
such that every initial segment
is prime-producing. The main new tool for achieving this is the following cute measure-theoretic lemma of Bergelson:
Lemma 2 (Bergelson intersectivity lemma) Letbe subsets of a probability space
of measure uniformly bounded away from zero, thus
. Then there exists a subsequence
such that
for all
.
This lemma has a short proof, though not an entirely obvious one. Firstly, by deleting a null set from , one can assume that all finite intersections
are either positive measure or empty. Secondly, a routine application of Fatou’s lemma shows that the maximal function
has a positive integral, hence must be positive at some point
. Thus there is a subsequence
whose finite intersections all contain
, thus have positive measure as desired by the previous reduction.
It turns out that one cannot quite combine the standard Maynard sieve with the intersectivity lemma because the events that show up (which roughly correspond to the event that
is prime for some random number
(with a well-chosen probability distribution) and some shift
) have their probability going to zero, rather than being uniformly bounded from below. To get around this, we borrow an idea from a paper of Banks, Freiberg, and Maynard, and group the shifts
into various clusters
, chosen in such a way that the probability that at least one of
is prime is bounded uniformly from below. One then applies the Bergelson intersectivity lemma to those events and uses many applications of the pigeonhole principle to conclude.
Kaisa Matomäki, Xuancheng Shao, Joni Teräväinen, and myself have just uploaded to the arXiv our preprint “Higher uniformity of arithmetic functions in short intervals I. All intervals“. This paper investigates the higher order (Gowers) uniformity of standard arithmetic functions in analytic number theory (and specifically, the Möbius function , the von Mangoldt function
, and the generalised divisor functions
) in short intervals
, where
is large and
lies in the range
for a fixed constant
(that one would like to be as small as possible). If we let
denote one of the functions
, then there is extensive literature on the estimation of short sums
Traditionally, asymptotics for such sums are expressed in terms of a “main term” of some arithmetic nature, plus an error term that is estimated in magnitude. For instance, a sum such as would be approximated in terms of a main term that vanished (or is negligible) if
is “minor arc”, but would be expressible in terms of something like a Ramanujan sum if
was “major arc”, together with an error term. We found it convenient to cancel off such main terms by subtracting an approximant
from each of the arithmetic functions
and then getting upper bounds on remainder correlations such as
- For the Möbius function
, we simply set
, as per the Möbius pseudorandomness conjecture. (One could choose a more sophisticated approximant in the presence of a Siegel zero, as I did with Joni in this recent paper, but we do not do so here.)
- For the von Mangoldt function
, we eventually went with the Cramér-Granville approximant
, where
and
.
- For the divisor functions
, we used a somewhat complicated-looking approximant
for some explicit polynomials
, chosen so that
and
have almost exactly the same sums along arithmetic progressions (see the paper for details).
The objective is then to obtain bounds on sums such as (1) that improve upon the “trivial bound” that one can get with the triangle inequality and standard number theory bounds such as the Brun-Titchmarsh inequality. For and
, the Siegel-Walfisz theorem suggests that it is reasonable to expect error terms that have “strongly logarithmic savings” in the sense that they gain a factor of
over the trivial bound for any
; for
, the Dirichlet hyperbola method suggests instead that one has “power savings” in that one should gain a factor of
over the trivial bound for some
. In the case of the Möbius function
, there is an additional trick (introduced by Matomäki and Teräväinen) that allows one to lower the exponent
somewhat at the cost of only obtaining “weakly logarithmic savings” of shape
for some small
.
Our main estimates on sums of the form (1) work in the following ranges:
- For
, one can obtain strongly logarithmic savings on (1) for
, and power savings for
.
- For
, one can obtain weakly logarithmic savings for
.
- For
, one can obtain power savings for
.
- For
, one can obtain power savings for
.
Conjecturally, one should be able to obtain power savings in all cases, and lower down to zero, but the ranges of exponents and savings given here seem to be the limit of current methods unless one assumes additional hypotheses, such as GRH. The
result for correlation against Fourier phases
was established previously by Zhan, and the
result for such phases and
was established previously by by Matomäki and Teräväinen.
By combining these results with tools from additive combinatorics, one can obtain a number of applications:
- Direct insertion of our bounds in the recent work of Kanigowski, Lemanczyk, and Radziwill on the prime number theorem on dynamical systems that are analytic skew products gives some improvements in the exponents there.
- We can obtain a “short interval” version of a multiple ergodic theorem along primes established by Frantzikinakis-Host-Kra and Wooley-Ziegler, in which we average over intervals of the form
rather than
.
- We can obtain a “short interval” version of the “linear equations in primes” asymptotics obtained by Ben Green, Tamar Ziegler, and myself in this sequence of papers, where the variables in these equations lie in short intervals
rather than long intervals such as
.
We now briefly discuss some of the ingredients of proof of our main results. The first step is standard, using combinatorial decompositions (based on the Heath-Brown identity and (for the result) the Ramaré identity) to decompose
into more tractable sums of the following types:
- Type
sums, which are basically of the form
for some weights
of controlled size and some cutoff
that is not too large;
- Type
sums, which are basically of the form
for some weights
,
of controlled size and some cutoffs
that are not too close to
or to
;
- Type
sums, which are basically of the form
for some weights
of controlled size and some cutoff
that is not too large.
The precise ranges of the cutoffs depend on the choice of
; our methods fail once these cutoffs pass a certain threshold, and this is the reason for the exponents
being what they are in our main results.
The Type sums involving nilsequences can be treated by methods similar to those in this previous paper of Ben Green and myself; the main innovations are in the treatment of the Type
and Type
sums.
For the Type sums, one can split into the “abelian” case in which (after some Fourier decomposition) the nilsequence
is basically of the form
, and the “non-abelian” case in which
is non-abelian and
exhibits non-trivial oscillation in a central direction. In the abelian case we can adapt arguments of Matomaki and Shao, which uses Cauchy-Schwarz and the equidistribution properties of polynomials to obtain good bounds unless
is “major arc” in the sense that it resembles (or “pretends to be”)
for some Dirichlet character
and some frequency
, but in this case one can use classical multiplicative methods to control the correlation. It turns out that the non-abelian case can be treated similarly. After applying Cauchy-Schwarz, one ends up analyzing the equidistribution of the four-variable polynomial sequence
For the type sum, a model sum to study is
In a sequel to this paper (currently in preparation), we will obtain analogous results for almost all intervals with
in the range
, in which we will be able to lower
all the way to
.
Joni Teräväinen and I have just uploaded to the arXiv our preprint “The Hardy–Littlewood–Chowla conjecture in the presence of a Siegel zero“. This paper is a development of the theme that certain conjectures in analytic number theory become easier if one makes the hypothesis that Siegel zeroes exist; this places one in a presumably “illusory” universe, since the widely believed Generalised Riemann Hypothesis (GRH) precludes the existence of such zeroes, yet this illusory universe seems remarkably self-consistent and notoriously impossible to eliminate from one’s analysis.
For the purposes of this paper, a Siegel zero is a zero of a Dirichlet
-function
corresponding to a primitive quadratic character
of some conductor
, which is close to
in the sense that
One of the early influential results in this area was the following result of Heath-Brown, which I previously blogged about here:
Theorem 1 (Hardy-Littlewood assuming Siegel zero) Letbe a fixed natural number. Suppose one has a Siegel zero
associated to some conductor
. Then we have
for all
, where
is the von Mangoldt function and
is the singular series
In particular, Heath-Brown showed that if there are infinitely many Siegel zeroes, then there are also infinitely many twin primes, with the correct asymptotic predicted by the Hardy-Littlewood prime tuple conjecture at infinitely many scales.
Very recently, Chinis established an analogous result for the Chowla conjecture (building upon earlier work of Germán and Katai):
Theorem 2 (Chowla assuming Siegel zero) Letbe distinct fixed natural numbers. Suppose one has a Siegel zero
associated to some conductor
. Then one has
in the range
, where
is the Liouville function.
In our paper we unify these results and also improve the quantitative estimates and range of :
Theorem 3 (Hardy-Littlewood-Chowla assuming Siegel zero) Letbe distinct fixed natural numbers with
. Suppose one has a Siegel zero
associated to some conductor
. Then one has
for
for any fixed
.
Our argument proceeds by a series of steps in which we replace and
by more complicated looking, but also more tractable, approximations, until the correlation is one that can be computed in a tedious but straightforward fashion by known techniques. More precisely, the steps are as follows:
- (i) Replace the Liouville function
with an approximant
, which is a completely multiplicative function that agrees with
at small primes and agrees with
at large primes.
- (ii) Replace the von Mangoldt function
with an approximant
, which is the Dirichlet convolution
multiplied by a Selberg sieve weight
to essentially restrict that convolution to almost primes.
- (iii) Replace
with a more complicated truncation
which has the structure of a “Type I sum”, and which agrees with
on numbers that have a “typical” factorization.
- (iv) Replace the approximant
with a more complicated approximant
which has the structure of a “Type I sum”.
- (v) Now that all terms in the correlation have been replaced with tractable Type I sums, use standard Euler product calculations and Fourier analysis, similar in spirit to the proof of the pseudorandomness of the Selberg sieve majorant for the primes in this paper of Ben Green and myself, to evaluate the correlation to high accuracy.
Steps (i), (ii) proceed mainly through estimates such as (1) and standard sieve theory bounds. Step (iii) is based primarily on estimates on the number of smooth numbers of a certain size.
The restriction in our main theorem is needed only to execute step (iv) of this step. Roughly speaking, the Siegel approximant
to
is a twisted, sieved version of the divisor function
, and the types of correlation one is faced with at the start of step (iv) are a more complicated version of the divisor correlation sum
Step (v) is a tedious but straightforward sieve theoretic computation, similar in many ways to the correlation estimates of Goldston and Yildirim used in their work on small gaps between primes (as discussed for instance here), and then also used by Ben Green and myself to locate arithmetic progressions in primes.
Joni Teräväinen and myself have just uploaded to the arXiv our preprint “Quantitative bounds for Gowers uniformity of the Möbius and von Mangoldt functions“. This paper makes quantitative the Gowers uniformity estimates on the Möbius function and the von Mangoldt function
.
To discuss the results we first discuss the situation of the Möbius function, which is technically simpler in some (though not all) ways. We assume familiarity with Gowers norms and standard notations around these norms, such as the averaging notation and the exponential notation
. The prime number theorem in qualitative form asserts that
Once one restricts to arithmetic progressions, the situation gets worse: the Siegel-Walfisz theorem gives the bound
for any residue classIn 1937, Davenport was able to show the discorrelation estimate
For the situation with the norm the previously known results were much weaker. Ben Green and I showed that
For higher norms , the situation is even worse, because the quantitative inverse theory for these norms is poorer, and indeed it was only with the recent work of Manners that any such bound is available at all (at least for
). Basically, Manners establishes if
Our first result gives an effective decay bound:
Theorem 1 For any, we have
for some
. The implied constants are effective.
This is off by a logarithm from the best effective bound (2) in the case. In the
case there is some hope to remove this logarithm based on the improved quantitative inverse theory currently available in this case, but there is a technical obstruction to doing so which we will discuss later in this post. For
the above bound is the best one could hope to achieve purely using the quantitative inverse theory of Manners.
We have analogues of all the above results for the von Mangoldt function . Here a complication arises that
does not have mean close to zero, and one has to subtract off some suitable approximant
to
before one would expect good Gowers norms bounds. For the prime number theorem one can just use the approximant
, giving
Theorem 2 For any, we have
for some
. The implied constants are effective.
By standard methods, this result also gives quantitative asymptotics for counting solutions to various systems of linear equations in primes, with error terms that gain a factor of with respect to the main term.
We now discuss the methods of proof, focusing first on the case of the Möbius function. Suppose first that there is no “Siegel zero”, by which we mean a quadratic character of some conductor
with a zero
with
for some small absolute constant
. In this case the Siegel-Walfisz bound (1) improves to a quasipolynomial bound
Now suppose we have a Siegel zero . In this case the bound (5) will not hold in general, and hence also (6) will not hold either. Here, the usual way out (while still maintaining effective estimates) is to approximate
not by
, but rather by a more complicated approximant
that takes the Siegel zero into account, and in particular is such that one has the (effective) pseudopolynomial bound
For the analogous problem with the von Mangoldt function (assuming a Siegel zero for sake of discussion), the approximant is simpler; we ended up using
In principle, the above results can be improved for due to the stronger quantitative inverse theorems in the
setting. However, there is a bottleneck that prevents us from achieving this, namely that the equidistribution theory of two-step nilmanifolds has exponents which are exponential in the dimension rather than polynomial in the dimension, and as a consequence we were unable to improve upon the doubly logarithmic results. Specifically, if one is given a sequence of bracket quadratics such as
that fails to be
-equidistributed, one would need to establish a nontrivial linear relationship modulo 1 between the
(up to errors of
), where the coefficients are of size
; current methods only give coefficient bounds of the form
. An old result of Schmidt demonstrates proof of concept that these sorts of polynomial dependencies on exponents is possible in principle, but actually implementing Schmidt’s methods here seems to be a quite non-trivial task. There is also another possible route to removing a logarithm, which is to strengthen the inverse
theorem to make the dimension of the nilmanifold logarithmic in the uniformity parameter
rather than polynomial. Again, the Freiman-Bilu theorem (see for instance this paper of Ben and myself) demonstrates proof of concept that such an improvement in dimension is possible, but some work would be needed to implement it.
Kaisa Matomäki, Maksym Radziwill, Xuancheng Shao, Joni Teräväinen, and myself have just uploaded to the arXiv our preprint “Singmaster’s conjecture in the interior of Pascal’s triangle“. This paper leverages the theory of exponential sums over primes to make progress on a well known conjecture of Singmaster which asserts that any natural number larger than appears at most a bounded number of times in Pascal’s triangle. That is to say, for any integer
, there are at most
solutions to the equation
Our main result settles this conjecture in the “interior” region of the triangle:
Theorem 1 (Singmaster’s conjecture in the interior of the triangle) Ifand
is sufficiently large depending on
, there are at most two solutions to (1) in the region
and hence at most four in the region
Also, there is at most one solution in the region
To verify Singmaster’s conjecture in full, it thus suffices in view of this result to verify the conjecture in the boundary region
(or equivalentlyThe upper bound of two here for the number of solutions in the region (2) is best possible, due to the infinite family of solutions to the equation
coming from
The appearance of the quantity in Theorem 1 may be familiar to readers that are acquainted with Vinogradov’s bounds on exponential sums, which ends up being the main new ingredient in our arguments. In principle this threshold could be lowered if we had stronger bounds on exponential sums.
To try to control solutions to (1) we use a combination of “Archimedean” and “non-Archimedean” approaches. In the “Archimedean” approach (following earlier work of Kane on this problem) we view primarily as real numbers rather than integers, and express (1) in terms of the Gamma function as
Proposition 2 Let, and suppose
is sufficiently large depending on
. If
is a solution to (1) in the left half
of Pascal’s triangle, then there is at most one other solution
to this equation in the left half with
Again, the example of (4) shows that a cluster of two solutions is certainly possible; the convexity argument only kicks in once one has a cluster of three or more solutions.
To finish the proof of Theorem 1, one has to show that any two solutions to (1) in the region of interest must be close enough for the above proposition to apply. Here we switch to the “non-Archimedean” approach, in which we look at the
-adic valuations
of the binomial coefficients, defined as the number of times a prime
divides
. From the fundamental theorem of arithmetic, a collision
A key idea in our approach is to view this condition (6) statistically, for instance by viewing as a prime drawn randomly from an interval such as
for some suitably chosen scale parameter
, so that the two sides of (6) now become random variables. It then becomes advantageous to compare correlations between these two random variables and some additional test random variable. For instance, if
and
are far apart from each other, then one would expect the left-hand side of (6) to have a higher correlation with the fractional part
, since this term shows up in the summation on the left-hand side but not the right. Similarly if
and
are far apart from each other (although there are some annoying cases one has to treat separately when there is some “unexpected commensurability”, for instance if
is a rational multiple of
where the rational has bounded numerator and denominator). In order to execute this strategy, it turns out (after some standard Fourier expansion) that one needs to get good control on exponential sums such as
A modification of the arguments also gives similar results for the equation
where
Theorem 3 Ifand
is sufficiently large depending on
, there are at most two solutions to (7) in the region
Again the upper bound of two is best possible, thanks to identities such as
Previous set of notes: Notes 3. Next set of notes: 246C Notes 1.
One of the great classical triumphs of complex analysis was in providing the first complete proof (by Hadamard and de la Vallée Poussin in 1896) of arguably the most important theorem in analytic number theory, the prime number theorem:
Theorem 1 (Prime number theorem) Letdenote the number of primes less than a given real number
. Then
(or in asymptotic notation,
as
).
(Actually, it turns out to be slightly more natural to replace the approximation in the prime number theorem by the logarithmic integral
, which happens to be a more precise approximation, but we will not stress this point here.)
The complex-analytic proof of this theorem hinges on the study of a key meromorphic function related to the prime numbers, the Riemann zeta function . Initially, it is only defined on the half-plane
:
Definition 2 (Riemann zeta function, preliminary definition) Letbe such that
. Then we define
Note that the series is locally uniformly convergent in the half-plane , so in particular
is holomorphic on this region. In previous notes we have already evaluated some special values of this function:
The Riemann zeta function has several remarkable properties, some of which we summarise here:
Theorem 3 (Basic properties of the Riemann zeta function)
- (i) (Euler product formula) For any
with
, we have
where the product is absolutely convergent (and locally uniform in
) and is over the prime numbers
.
- (ii) (Trivial zero-free region)
has no zeroes in the region
.
- (iii) (Meromorphic continuation)
has a unique meromorphic continuation to the complex plane (which by abuse of notation we also call
), with a simple pole at
and no other poles. Furthermore, the Riemann xi function
is an entire function of order
(after removing all singularities). The function
is an entire function of order one after removing the singularity at
.
- (iv) (Functional equation) After applying the meromorphic continuation from (iii), we have
for all
(excluding poles). Equivalently, we have
for all
. (The equivalence between the (5) and (6) is a routine consequence of the Euler reflection formula and the Legendre duplication formula, see Exercises 26 and 31 of Notes 1.)
Proof: We just prove (i) and (ii) for now, leaving (iii) and (iv) for later sections.
The claim (i) is an encoding of the fundamental theorem of arithmetic, which asserts that every natural number is uniquely representable as a product
over primes, where the
are natural numbers, all but finitely many of which are zero. Writing this representation as
, we see that
The claim (ii) is immediate from (i) since the Euler product is absolutely convergent and all terms are non-zero.
We remark that by sending to
in Theorem 3(i) we conclude that
The meromorphic continuation (iii) of the zeta function is initially surprising, but can be interpreted either as a manifestation of the extremely regular spacing of the natural numbers occurring in the sum (1), or as a consequence of various integral representations of
(or slight modifications thereof). We will focus in this set of notes on a particular representation of
as essentially the Mellin transform of the theta function
that briefly appeared in previous notes, and the functional equation (iv) can then be viewed as a consequence of the modularity of that theta function. This in turn was established using the Poisson summation formula, so one can view the functional equation as ultimately being a manifestation of Poisson summation. (For a direct proof of the functional equation via Poisson summation, see these notes.)
Henceforth we work with the meromorphic continuation of . The functional equation (iv), when combined with special values of
such as (2), gives some additional values of
outside of its initial domain
, most famously
From Theorem 3 and the non-vanishing nature of , we see that
has simple zeroes (known as trivial zeroes) at the negative even integers
, and all other zeroes (the non-trivial zeroes) inside the critical strip
. (The non-trivial zeroes are conjectured to all be simple, but this is hopelessly far from being proven at present.) As we shall see shortly, these latter zeroes turn out to be closely related to the distribution of the primes. The functional equation tells us that if
is a non-trivial zero then so is
; also, we have the identity
Conjecture 4 (Riemann hypothesis) All the non-trivial zeroes oflie on the critical line
.
This conjecture would have many implications in analytic number theory, particularly with regard to the distribution of the primes. Of course, it is far from proven at present, but the partial results we have towards this conjecture are still sufficient to establish results such as the prime number theorem.
Return now to the original region where . To take more advantage of the Euler product formula (3), we take complex logarithms to conclude that
The series and
that show up in the above formulae are examples of Dirichlet series, which are a convenient device to transform various sequences of arithmetic interest into holomorphic or meromorphic functions. Here are some more examples:
Exercise 5 (Standard Dirichlet series) Letbe a complex number with
.
- (i) Show that
.
- (ii) Show that
, where
is the divisor function of
(the number of divisors of
).
- (iii) Show that
, where
is the Möbius function, defined to equal
when
is the product of
distinct primes for some
, and
otherwise.
- (iv) Show that
, where
is the Liouville function, defined to equal
when
is the product of
(not necessarily distinct) primes for some
.
- (v) Show that
, where
is the holomorphic branch of the logarithm that is real for
, and with the convention that
vanishes for
.
- (vi) Use the fundamental theorem of arithmetic to show that the von Mangoldt function is the unique function
such that
for every positive integer
. Use this and (i) to provide an alternate proof of the identity (8). Thus we see that (8) is really just another encoding of the fundamental theorem of arithmetic.
Given the appearance of the von Mangoldt function , it is natural to reformulate the prime number theorem in terms of this function:
Theorem 6 (Prime number theorem, von Mangoldt form) One has(or in asymptotic notation,
as
).
Let us see how Theorem 6 implies Theorem 1. Firstly, for any , we can write
Exercise 7 Show that Theorem 1 conversely implies Theorem 6.
The alternate form (8) of the Euler product identity connects the primes (represented here via proxy by the von Mangoldt function) with the logarithmic derivative of the zeta function, and can be used as a starting point for describing further relationships between and the primes. Most famously, we shall see later in these notes that it leads to the remarkably precise Riemann-von Mangoldt explicit formula:
Theorem 8 (Riemann-von Mangoldt explicit formula) For any non-integer, we have
where
ranges over the non-trivial zeroes of
with imaginary part in
. Furthermore, the convergence of the limit is locally uniform in
.
Actually, it turns out that this formula is in some sense too precise; in applications it is often more convenient to work with smoothed variants of this formula in which the sum on the left-hand side is smoothed out, but the contribution of zeroes with large imaginary part is damped; see Exercise 22. Nevertheless, this formula clearly illustrates how the non-trivial zeroes of the zeta function influence the primes. Indeed, if one formally differentiates the above formula in
, one is led to the (quite nonrigorous) approximation
Comparing Theorem 8 with Theorem 6, it is natural to suspect that the key step in the proof of the latter is to establish the following slight but important extension of Theorem 3(ii), which can be viewed as a very small step towards the Riemann hypothesis:
Theorem 9 (Slight enlargement of zero-free region) There are no zeroes ofon the line
.
It is not quite immediate to see how Theorem 6 follows from Theorem 8 and Theorem 9, but we will demonstrate it below the fold.
Although Theorem 9 only seems like a slight improvement of Theorem 3(ii), proving it is surprisingly non-trivial. The basic idea is the following: if there was a zero at , then there would also be a different zero at
(note
cannot vanish due to the pole at
), and then the approximation (9) becomes
In fact, Theorem 9 is basically equivalent to the prime number theorem:
Exercise 10 For the purposes of this exercise, assume Theorem 6, but do not assume Theorem 9. For any non-zero real, show that
as
, where
denotes a quantity that goes to zero as
after being multiplied by
. Use this to derive Theorem 9.
This equivalence can help explain why the prime number theorem is remarkably non-trivial to prove, and why the Riemann zeta function has to be either explicitly or implicitly involved in the proof.
This post is only intended as the briefest of introduction to complex-analytic methods in analytic number theory; also, we have not chosen the shortest route to the prime number theorem, electing instead to travel in directions that particularly showcase the complex-analytic results introduced in this course. For some further discussion see this previous set of lecture notes, particularly Notes 2 and Supplement 3 (with much of the material in this post drawn from the latter).
Previous set of notes: Notes 2. Next set of notes: Notes 4.
On the real line, the quintessential examples of a periodic function are the (normalised) sine and cosine functions ,
, which are
-periodic in the sense that
What about periodic functions on the complex plane? We can start with singly periodic functions which obey a periodicity relationship
for all
in the domain and some period
; such functions can also be viewed as functions on the “additive cylinder”
(or equivalently
). We can rescale
as before. For holomorphic functions, we have the following characterisations:
Proposition 1 (Description of singly periodic holomorphic functions)In both cases, the coefficients
- (i) Every
-periodic entire function
has an absolutely convergent expansion
where
is the nome
, and the
are complex coefficients such that
Conversely, every doubly infinite sequence
of coefficients obeying (2) gives rise to a
-periodic entire function
via the formula (1).
- (ii) Every bounded
-periodic holomorphic function
on the upper half-plane
has an expansion
where the
are complex coefficients such that
Conversely, every infinite sequence
obeying (4) gives rise to a
-periodic holomorphic function
which is bounded away from the real axis (i.e., bounded on
for every
).
can be recovered from
by the Fourier inversion formula
for any
in
(in case (i)) or
(in case (ii)).
Proof: If is
-periodic, then it can be expressed as
for some function
on the “multiplicative cylinder”
, since the fibres of the map
are cosets of the integers
, on which
is constant by hypothesis. As the map
is a covering map from
to
, we see that
will be holomorphic if and only if
is. Thus
must have a Laurent series expansion
with coefficients
obeying (2), which gives (1), and the inversion formula (5) follows from the usual contour integration formula for Laurent series coefficients. The converse direction to (i) also follows by reversing the above arguments.
For part (ii), we observe that the map is also a covering map from
to the punctured disk
, so we can argue as before except that now
is a bounded holomorphic function on the punctured disk. By the Riemann singularity removal theorem (Exercise 35 of 246A Notes 3)
extends to be holomorphic on all of
, and thus has a Taylor expansion
for some coefficients
obeying (4). The argument now proceeds as with part (i).
The additive cylinder and the multiplicative cylinder
can both be identified (on the level of smooth manifolds, at least) with the geometric cylinder
, but we will not use this identification here.
Now let us turn attention to doubly periodic functions of a complex variable , that is to say functions
that obey two periodicity relations
Within the world of holomorphic functions, the collection of doubly periodic functions is boring:
Proposition 2 Letbe an entire doubly periodic function (with periods
linearly independent over
). Then
is constant.
In the language of Riemann surfaces, this proposition asserts that the torus is a non-hyperbolic Riemann surface; it cannot be holomorphically mapped non-trivially into a bounded subset of the complex plane.
Proof: The fundamental domain (up to boundary) enclosed by is compact, hence
is bounded on this domain, hence bounded on all of
by double periodicity. The claim now follows from Liouville’s theorem. (One could alternatively have argued here using the compactness of the torus
.
To obtain more interesting examples of doubly periodic functions, one must therefore turn to the world of meromorphic functions – or equivalently, holomorphic functions into the Riemann sphere . As it turns out, a particularly fundamental example of such a function is the Weierstrass elliptic function
I’ve just uploaded to the arXiv my paper The Ionescu-Wainger multiplier theorem and the adeles“. This paper revisits a useful multiplier theorem of Ionescu and Wainger on “major arc” Fourier multiplier operators on the integers (or lattices
), and strengthens the bounds while also interpreting it from the viewpoint of the adelic integers
(which were also used in my recent paper with Krause and Mirek).
For simplicity let us just work in one dimension. Any smooth function then defines a discrete Fourier multiplier operator
for any
by the formula
We will be interested in discrete Fourier multiplier operators whose symbols are supported on a finite union of arcs. One way to construct such operators is by “folding” continuous Fourier multiplier operators into various target frequencies. To make this folding operation precise, given any continuous Fourier multiplier operator , and any frequency
, we define the discrete Fourier multiplier operator
for any frequency shift
by the formula
There are a body of results relating the theory of discrete Fourier multiplier operators such as
or
with the
theory of their continuous counterparts. For instance we have the basic result of Magyar, Stein, and Wainger:
Proposition 1 (Magyar-Stein-Wainger sampling principle) Letand
.
- (i) If
is a smooth function supported in
, then
, where
denotes the operator norm of an operator
.
- (ii) More generally, if
is a smooth function supported in
for some natural number
, then
.
When the implied constant in these bounds can be set to equal
. In the paper of Magyar, Stein, and Wainger it was posed as an open problem as to whether this is the case for other
; in an appendix to this paper I show that the answer is negative if
is sufficiently close to
or
, but I do not know the full answer to this question.
This proposition allows one to get a good multiplier theory for symbols supported near cyclic groups ; for instance it shows that a discrete Fourier multiplier with symbol
for a fixed test function
is bounded on
, uniformly in
and
. For many applications in discrete harmonic analysis, one would similarly like a good multiplier theory for symbols supported in “major arc” sets such as
In the regime where is fixed and
is small, there is a good theory:
Theorem 2 (Ionescu-Wainger theorem, rough version) Ifis an even integer or the dual of an even integer, and
is supported on
for a sufficiently small
, then
There is a more explicit description of how small needs to be for this theorem to work (roughly speaking, it is not much more than what is needed for all the arcs
in (2) to be disjoint), but we will not give it here. The logarithmic loss of
was reduced to
by Mirek. In this paper we refine the bound further to
The proof of (3) follows a similar strategy as to previous proofs of Ionescu-Wainger type. By duality we may assume . We use the following standard sequence of steps:
- (i) (Denominator orthogonality) First one splits
into various pieces depending on the denominator
appearing in the element of
, and exploits “superorthogonality” in
to estimate the
norm by the
norm of an appropriate square function.
- (ii) (Nonconcentration) One expands out the
power of the square function and estimates it by a “nonconcentrated” version in which various factors that arise in the expansion are “disjoint”.
- (iii) (Numerator orthogonality) We now decompose based on the numerators
appearing in the relevant elements of
, and exploit some residual orthogonality in this parameter to reduce to estimating a square-function type expression involving sums over various cosets
.
- (iv) (Marcinkiewicz-Zygmund) One uses the Marcinkiewicz-Zygmund theorem relating scalar and vector valued operator norms to eliminate the role of the multiplier
.
- (v) (Rubio de Francia) Use a reverse square function estimate of Rubio de Francia type to conclude.
The main innovations are that of using the probabilistic decoupling method to remove some logarithmic losses in (i), and recent progress on the Erdos-Rado sunflower conjecture (as discussed in this recent post) to improve the bounds in (ii). For (i), the key point is that one can express a sum such as
In this paper we interpret the Ionescu-Wainger multiplier theorem as being essentially a consequence of various quantitative versions of the Shannon sampling theorem. Recall that this theorem asserts that if a (Schwartz) function has its Fourier transform supported on
, then
can be recovered uniquely from its restriction
. In fact, as can be shown from a little bit of routine Fourier analysis, if we narrow the support of the Fourier transform slightly to
for some
, then the restriction
has the same
behaviour as the original function, in the sense that
The quantitative sampling theorem (4) can be used to give an alternate proof of Proposition 1(i), basically thanks to the identity
The locally compact abelian groups and
can all be viewed as projections of the adelic integers
(the product of the reals and the profinite integers
). By using the Ionescu-Wainger multiplier theorem, we are able to obtain an adelic version of the quantitative sampling estimate (5), namely
Kaisa Matomäki, Maksym Radziwill, Joni Teräväinen, Tamar Ziegler and I have uploaded to the arXiv our paper Higher uniformity of bounded multiplicative functions in short intervals on average. This paper (which originated from a working group at an AIM workshop on Sarnak’s conjecture) focuses on the local Fourier uniformity conjecture for bounded multiplicative functions such as the Liouville function . One form of this conjecture is the assertion that
The conjecture gets more difficult as increases, and also becomes more difficult the more slowly
grows with
. The
conjecture is equivalent to the assertion
For , the conjecture is equivalent to the assertion
Now we apply the same strategy to (4). For abelian the claim follows easily from (3), so we focus on the non-abelian case. One now has a polynomial sequence
attached to many
, and after a somewhat complicated adaptation of the above arguments one again ends up with an approximate functional equation
We give two applications of this higher order Fourier uniformity. One regards the growth of the number
The second application is to obtain cancellation for various polynomial averages involving the Liouville function or von Mangoldt function
, such as
Define the Collatz map on the natural numbers
by setting
to equal
when
is odd and
when
is even, and let
denote the forward Collatz orbit of
. The notorious Collatz conjecture asserts that
for all
. Equivalently, if we define the backwards Collatz orbit
to be all the natural numbers
that encounter
in their forward Collatz orbit, then the Collatz conjecture asserts that
. As a partial result towards this latter statement, Krasikov and Lagarias in 2003 established the bound
for all and
. (This improved upon previous values of
obtained by Applegate and Lagarias in 1995,
by Applegate and Lagarias in 1995 by a different method,
by Wirsching in 1993,
by Krasikov in 1989,
by Sander in 1990, and some
by Crandall in 1978.) This is still the largest value of
for which (1) has been established. Of course, the Collatz conjecture would imply that we can take
equal to
, which is the assertion that a positive density set of natural numbers obeys the Collatz conjecture. This is not yet established, although the results in my previous paper do at least imply that a positive density set of natural numbers iterates to an (explicitly computable) bounded set, so in principle the
case of (1) could now be verified by an (enormous) finite computation in which one verifies that every number in this explicit bounded set iterates to
. In this post I would like to record a possible alternate route to this problem that depends on the distribution of a certain family of random variables that appeared in my previous paper, that I called Syracuse random variables.
Definition 1 (Syracuse random variables) For any natural number
, a Syracuse random variable
on the cyclic group
is defined as a random variable of the form
where
are independent copies of a geometric random variable
on the natural numbers with mean
, thus
} for
. In (2) the arithmetic is performed in the ring
.
Thus for instance
and so forth. After reversing the labeling of the , one could also view
as the mod
reduction of a
-adic random variable
The probability density function of the Syracuse random variable can be explicitly computed by a recursive formula (see Lemma 1.12 of my previous paper). For instance, when
,
is equal to
for
respectively, while when
,
is equal to
when respectively.
The relationship of these random variables to the Collatz problem can be explained as follows. Let denote the odd natural numbers, and define the Syracuse map
by
where the –valuation
is the number of times
divides
. We can define the forward orbit
and backward orbit
of the Syracuse map as before. It is not difficult to then see that the Collatz conjecture is equivalent to the assertion
, and that the assertion (1) for a given
is equivalent to the assertion
for all , where
is now understood to range over odd natural numbers. A brief calculation then shows that for any odd natural number
and natural number
, one has
where the natural numbers are defined by the formula
so in particular
Heuristically, one expects the -valuation
of a typical odd number
to be approximately distributed according to the geometric distribution
, so one therefore expects the residue class
to be distributed approximately according to the random variable
.
The Syracuse random variables will always avoid multiples of three (this reflects the fact that
is never a multiple of three), but attains any non-multiple of three in
with positive probability. For any natural number
, set
Equivalently, is the greatest quantity for which we have the inequality
for all integers not divisible by three, where
is the set of all tuples
for which
Thus for instance ,
, and
. On the other hand, since all the probabilities
sum to
as
ranges over the non-multiples of
, we have the trivial upper bound
There is also an easy submultiplicativity result:
Lemma 2 For any natural numbers
, we have
Proof: Let be an integer not divisible by
, then by (4) we have
If we let denote the set of tuples
that can be formed from the tuples in
by deleting the final component
from each tuple, then we have
with an integer not divisible by three. By definition of
and a relabeling, we then have
for all . For such tuples we then have
so that . Since
for each , the claim follows.
From this lemma we see that for some absolute constant
. Heuristically, we expect the Syracuse random variables to be somewhat approximately equidistributed amongst the multiples of
(in Proposition 1.4 of my previous paper I prove a fine scale mixing result that supports this heuristic). As a consequence it is natural to conjecture that
. I cannot prove this, but I can show that this conjecture would imply that we can take the exponent
in (1), (3) arbitrarily close to one:
Proposition 3 Suppose that
(that is to say,
as
). Then
as
, or equivalently
I prove this proposition below the fold. A variant of the argument shows that for any value of , (1), (3) holds whenever
, where
is an explicitly computable function with
as
. In principle, one could then improve the Krasikov-Lagarias result
by getting a sufficiently good upper bound on
, which is in principle achievable numerically (note for instance that Lemma 2 implies the bound
for any
, since
for any
).
Recent Comments