You are currently browsing the category archive for the ‘question’ category.

This is the fourth “research” thread of the Polymath15 project to upper bound the de Bruijn-Newman constant , continuing https://terrytao.wordpress.com/2018/01/24/polymath-proposal-upper-bounding-the-de-bruijn-newman-constant/. Progress will be summarised at this Polymath wiki page.

We are getting closer to finishing off the following test problem: can one show that whenever , ? This would morally show that . A wiki page for this problem has now been created here. We have obtained a number of approximations to (see wiki page), though numeric evidence indicates that the approximations are all very close to each other. (Many of these approximations come with a correction term , but thus far it seems that we may be able to avoid having to use this refinement to the approximations.) The effective approximation also comes with an effective error bound

for some explicit (but somewhat messy) error terms : see this wiki page for details. The original approximations can be considered deprecated at this point in favour of the (slightly more complicated) approximation ; the approximation is a simplified version of which is not quite as accurate but might be useful for testing purposes.

It is convenient to normalise everything by an explicit non-zero factor . Asymptotically, converges to 1; numerically, it appears that its magnitude (and also its real part) stays roughly between 0.4 and 3 in the range , and we seem to be able to keep it (or at least the toy counterpart ) away from zero starting from about (here it seems that there is a useful trick of multiplying by Euler-type factors like to cancel off some of the oscillation). Also, the bounds on the error seem to be of size about 0.1 or better in these ranges also. So we seem to be on track to be able to rigorously eliminate zeroes starting from about or so. We have not discussed too much what to do with the small values of ; at some point our effective error bounds will become unusable, and we may have to find some more faster ways to compute .

In addition to this main direction of inquiry, there have been additional discussions on the dynamics of zeroes, and some numerical investigations of the behaviour Lehmer pairs under heat flow. Contributors are welcome to summarise any findings from these discussions from previous threads (or on any other related topic, e.g. improvements in the code) in the comments below.

This is the third “research” thread of the Polymath15 project to upper bound the de Bruijn-Newman constant , continuing this previous thread. Discussion of the project of a non-research nature can continue for now in the existing proposal thread. Progress will be summarised at this Polymath wiki page.

We are making progress on the following test problem: can one show that whenever , , and ? This would imply that

which would be the first quantitative improvement over the de Bruijn bound of (or the Ki-Kim-Lee refinement of ). Of course we can try to lower the two parameters of later on in the project, but this seems as good a place to start as any. One could also potentially try to use finer analysis of dynamics of zeroes to improve the bound further, but this seems to be a less urgent task.

Probably the hardest case is , as there is a good chance that one can then recover the case by a suitable use of the argument principle. Here we appear to have a workable Riemann-Siegel type formula that gives a tractable approximation for . To describe this formula, first note that in the case we have

and the Riemann-Siegel formula gives

for any natural numbers , where is a contour from to that winds once anticlockwise around the zeroes of but does not wind around any other zeroes. A good choice of to use here is

In this case, a classical steepest descent computation (see wiki) yields the approximation

where

Thus we have

where

with and given by (1).

Heuristically, we have derived (see wiki) the more general approximation

for (and in particular for ), where

In practice it seems that the term is negligible once the real part of is moderately large, so one also has the approximation

For large , and for fixed , e.g. , the sums converge fairly quickly (in fact the situation seems to be significantly better here than the much more intensively studied case), and we expect the first term

of the series to dominate. Indeed, analytically we know that (or ) as (holding fixed), and it should also be provable that as well. Numerically with , it seems in fact that (or ) stay within a distance of about of once is moderately large (e.g. ). This raises the hope that one can solve the toy problem of showing for by numerically controlling for small (e.g. ), numerically controlling and analytically bounding the error for medium (e.g. ), and analytically bounding both and for large (e.g. ). (These numbers and are arbitrarily chosen here and may end up being optimised to something else as the computations become clearer.)

Thus, we now have four largely independent tasks (for suitable ranges of “small”, “medium”, and “large” ):

- Numerically computing for small (with enough accuracy to verify that there are no zeroes)
- Numerically computing for medium (with enough accuracy to keep it away from zero)
- Analytically bounding for large (with enough accuracy to keep it away from zero); and
- Analytically bounding for medium and large (with a bound that is better than the bound away from zero in the previous two tasks).

Note that tasks 2 and 3 do not directly require any further understanding of the function .

Below we will give a progress report on the numeric and analytic sides of these tasks.

** — 1. Numerics report (contributed by Sujit Nair) — **

There is some progress on the code side but not at the pace I was hoping. Here are a few things which happened (rather, mistakes which were taken care of).

- We got rid of code which wasn’t being used. For example, @dhjpolymath computed based on an old version but only realized it after the fact.
- We implemented tests to catch human/numerical bugs before a computation starts. Again, we lost some numerical cycles but moving forward these can be avoided.
- David got set up on GitHub and he is able to compare his output (in C) with the Python code. That is helping a lot.

Two areas which were worked on were

- Computing and zeroes for around
- Computing quantities like , , , etc. with the goal of understanding the zero free regions.

Some observations for , , include:

- does seem to avoid the negative real axis
- (based on the oscillations and trends in the plots)
- seems to be settling around range.

See the figure below. The top plot is on the complex plane and the bottom plot is the absolute value. The code to play with this is here.

** — 2. Analysis report — **

The Riemann-Siegel formula and some manipulations (see wiki) give , where

where is a contour that goes from to staying a bounded distance away from the upper imaginary and right real axes, and is the complex conjugate of . (In each of these sums, it is the first term that should dominate, with the second one being about as large.) One can then evolve by the heat flow to obtain , where

Steepest descent heuristics then predict that , , and . For the purposes of this project, we will need effective error estimates here, with explicit error terms.

A start has been made towards this goal at this wiki page. Firstly there is a “effective Laplace method” lemma that gives effective bounds on integrals of the form if the real part of is either monotone with large derivative, or has a critical point and is decreasing on both sides of that critical point. In principle, all one has to do is manipulate expressions such as , , by change of variables, contour shifting and integration by parts until it is of the form to which the above lemma can be profitably applied. As one may imagine though the computations are messy, particularly for the term. As a warm up, I have begun by trying to estimate integrals of the form

for smallish complex numbers , as these sorts of integrals appear in the form of . As of this time of writing, there are effective bounds for the case, and I am currently working on extending them to the case, which should give enough control to approximate and . The most complicated task will be that of upper bounding , but it also looks eventually doable.

This is the first official “research” thread of the Polymath15 project to upper bound the de Bruijn-Newman constant . Discussion of the project of a non-research nature can continue for now in the existing proposal thread. Progress will be summarised at this Polymath wiki page.

The proposal naturally splits into at least three separate (but loosely related) topics:

- Numerical computation of the entire functions , with the ultimate aim of establishing zero-free regions of the form for various .
- Improved understanding of the dynamics of the zeroes of .
- Establishing the zero-free nature of when and is sufficiently large depending on and .

Below the fold, I will present each of these topics in turn, to initiate further discussion in each of them. (I thought about splitting this post into three to have three separate discussions, but given the current volume of comments, I think we should be able to manage for now having all the comments in a single post. If this changes then of course we can split up some of the discussion later.)

To begin with, let me present some formulae for computing (inspired by similar computations in the Ki-Kim-Lee paper) which may be useful. The initial definition of is

where

is a variant of the Jacobi theta function. We observe that in fact extends analytically to the strip

as has positive real part on this strip. One can use the Poisson summation formula to verify that is even, (see this previous post for details). This lets us obtain a number of other formulae for . Most obviously, one can unfold the integral to obtain

In my previous paper with Brad, we used this representation, combined with Fubini’s theorem to swap the sum and integral, to obtain a useful series representation for in the case. Unfortunately this is not possible in the case because expressions such as diverge as approaches . Nevertheless we can still perform the following contour integration manipulation. Let be fixed. The function decays super-exponentially fast (much faster than , in particular) as with ; as is even, we also have this decay as with (this is despite each of the summands in having much slower decay in this direction – there is considerable cancellation!). Hence by the Cauchy integral formula we have

Splitting the horizontal line from to at and using the even nature of , we thus have

Using the functional equation , we thus have the representation

where is the oscillatory integral

The formula (2) is valid for any . Naively one would think that it would be simplest to take ; however, when and is large (with bounded), it seems asymptotically better to take closer to , in particular something like seems to be a reasonably good choice. This is because the integrand in (3) becomes significantly less oscillatory and also much lower in amplitude; the term in (3) now generates a factor roughly comparable to (which, as we will see below, is the main term in the decay asymptotics for ), while the term still exhibits a reasonable amount of decay as . We will use the representation (2) in the asymptotic analysis of below, but it may also be a useful representation to use for numerical purposes.

Building on the interest expressed in the comments to this previous post, I am now formally proposing to initiate a “Polymath project” on the topic of obtaining new upper bounds on the de Bruijn-Newman constant . The purpose of this post is to describe the proposal and discuss the scope and parameters of the project.

De Bruijn introduced a family of entire functions for each real number , defined by the formula

where is the super-exponentially decaying function

As discussed in this previous post, the Riemann hypothesis is equivalent to the assertion that all the zeroes of are real.

De Bruijn and Newman showed that there existed a real constant – the de Bruijn-Newman constant – such that has all zeroes real whenever , and at least one non-real zero when . In particular, the Riemann hypothesis is equivalent to the upper bound . In the opposite direction, several lower bounds on have been obtained over the years, most recently in my paper with Brad Rodgers where we showed that , a conjecture of Newman.

As for upper bounds, de Bruijn showed back in 1950 that . The only progress since then has been the work of Ki, Kim and Lee in 2009, who improved this slightly to . The primary proposed aim of this Polymath project is to obtain further explicit improvements to the upper bound of . Of course, if we could lower the upper bound all the way to zero, this would solve the Riemann hypothesis, but I do not view this as a realistic outcome of this project; rather, the upper bounds that one could plausibly obtain by known methods and numerics would be comparable in achievement to the various numerical verifications of the Riemann hypothesis that exist in the literature (e.g., that the first non-trivial zeroes of the zeta function lie on the critical line, for various large explicit values of ).

In addition to the primary goal, one could envisage some related secondary goals of the project, such as a better understanding (both analytic and numerical) of the functions (or of similar functions), and of the dynamics of the zeroes of these functions. Perhaps further potential goals could emerge in the discussion to this post.

I think there is a plausible plan of attack on this project that proceeds as follows. Firstly, there are results going back to the original work of de Bruijn that demonstrate that the zeroes of become attracted to the real line as increases; in particular, if one defines to be the supremum of the imaginary parts of all the zeroes of , then it is known that this quantity obeys the differential inequality

whenever is positive; furthermore, once for some , then for all . I hope to explain this in a future post (it is basically due to the attraction that a zero off the real axis has to its complex conjugate). As a corollary of this inequality, we have the upper bound

for any real number . For instance, because all the non-trivial zeroes of the Riemann zeta function lie in the critical strip , one has , which when inserted into (2) gives . The inequality (1) also gives for all . If we could find some explicit between and where we can improve this upper bound on by an explicit constant, this would lead to a new upper bound on .

Secondly, the work of Ki, Kim and Lee (based on an analysis of the various terms appearing in the expression for ) shows that for any positive , all but finitely many of the zeroes of are real (in contrast with the situation, where it is still an open question as to whether the proportion of non-trivial zeroes of the zeta function on the critical line is asymptotically equal to ). As a key step in this analysis, Ki, Kim, and Lee show that for any and , there exists a such that all the zeroes of with real part at least , have imaginary part at most . Ki, Kim and Lee do not explicitly compute how depends on and , but it looks like this bound could be made effective.

If so, this suggests a possible strategy to get a new upper bound on :

- Select a good choice of parameters .
- By refining the Ki-Kim-Lee analysis, find an explicit such that all zeroes of with real part at least have imaginary part at most .
- By a numerical computation (e.g. using the argument principle), also verify that zeroes of with real part between and have imaginary part at most .
- Combining these facts, we obtain that ; hopefully, one can insert this into (2) and get a new upper bound for .

Of course, there may also be alternate strategies to upper bound , and I would imagine this would also be a legitimate topic of discussion for this project.

One appealing thing about the above strategy for the purposes of a polymath project is that it naturally splits the project into several interacting but reasonably independent parts: an analytic part in which one tries to refine the Ki-Kim-Lee analysis (based on explicitly upper and lower bounding various terms in a certain series expansion for – I may detail this later in a subsequent post); a numerical part in which one controls the zeroes of in a certain finite range; and perhaps also a dynamical part where one sees if there is any way to improve the inequality (2). For instance, the numerical “team” might, over time, be able to produce zero-free regions for with an increasingly large value of , while in parallel the analytic “team” might produce increasingly smaller values of beyond which they can control zeroes, and eventually the two bounds would meet up and we obtain a new bound on . This factoring of the problem into smaller parts was also a feature of the successful Polymath8 project on bounded gaps between primes.

The project also resembles Polymath8 in another aspect: that there is an obvious way to numerically measure progress, by seeing how the upper bound for decreases over time (and presumably there will also be another metric of progress regarding how well we can control in terms of and ). However, in Polymath8 the final measure of progress (the upper bound on gaps between primes) was a natural number, and thus could not decrease indefinitely. Here, the bound will be a real number, and there is a possibility that one may end up having an infinite descent in which progress slows down over time, with refinements to increasingly less significant digits of the bound as the project progresses. Because of this, I think it makes sense to follow recent Polymath projects and place an expiration date for the project, for instance one year after the launch date, in which we will agree to end the project and (if the project was successful enough) write up the results, unless there is consensus at that time to extend the project. (In retrospect, we should probably have imposed similar sunset dates on older Polymath projects, some of which have now been inactive for years, but that is perhaps a discussion for another time.)

Some Polymath projects have been known for a breakneck pace, making it hard for some participants to keep up. It’s hard to control these things, but I am envisaging a relatively leisurely project here, perhaps taking the full year mentioned above. It may well be that as the project matures we will largely be waiting for the results of lengthy numerical calculations to come in, for instance. Of course, as with previous projects, we would maintain some wiki pages (and possibly some other resources, such as a code repository) to keep track of progress and also to summarise what we have learned so far. For instance, as was done with some previous Polymath projects, we could begin with some “online reading seminars” where we go through some relevant piece of literature (most obviously the Ki-Kim-Lee paper, but there may be other resources that become relevant, e.g. one could imagine the literature on numerical verification of RH to be of value).

One could also imagine some incidental outcomes of this project, such as a more efficient way to numerically establish zero free regions for various analytic functions of interest; in particular, the project may well end up focusing on some other aspect of mathematics than the specific questions posed here.

Anyway, I would be interested to hear in the comments below from others who might be interested in participating, or at least observing, this project, particularly if they have suggestions regarding the scope and direction of the project, and on organisational structure (e.g. if one should start with reading seminars, or some initial numerical exploration of the functions , etc..) One could also begin some preliminary discussion of the actual mathematics of the project itself, though (in line with the leisurely pace I was hoping for), I expect that the main burst of mathematical activity would happen later, once the project is formally launched (with wiki page resources, blog posts dedicated to specific aspects of the project, etc.).

The Polymath14 online collaboration has uploaded to the arXiv its paper “Homogeneous length functions on groups“, submitted to Algebra & Number Theory. The paper completely classifies *homogeneous length functions* on an arbitrary group , that is to say non-negative functions that obey the symmetry condition , the non-degeneracy condition , the triangle inequality , and the homogeneity condition . It turns out that these norms can only arise from pulling back the norm of a Banach space by an isometric embedding of the group. Among other things, this shows that can only support a homogeneous length function if and only if it is abelian and torsion free, thus giving a metric description of this property.

The proof is based on repeated use of the homogeneous length function axioms, combined with elementary identities of commutators, to obtain increasingly good bounds on quantities such as , until one can show that such norms have to vanish. See the previous post for a full proof. The result is robust in that it allows for some loss in the triangle inequality and homogeneity condition, allowing for some new results on “quasinorms” on groups that relate to quasihomomorphisms.

As there are now a large number of comments on the previous post on this project, this post will also serve as the new thread for any final discussion of this project as it winds down.

In the tradition of “Polymath projects“, the problem posed in the previous two blog posts has now been solved, thanks to the cumulative effect of many small contributions by many participants (including, but not limited to, Sean Eberhard, Tobias Fritz, Siddharta Gadgil, Tobias Hartnick, Chris Jerdonek, Apoorva Khare, Antonio Machiavelo, Pace Nielsen, Andy Putman, Will Sawin, Alexander Shamov, Lior Silberman, and David Speyer). In this post I’ll write down a streamlined resolution, eliding a number of important but ultimately removable partial steps and insights made by the above contributors en route to the solution.

Theorem 1Let be a group. Suppose one has a “seminorm” function which obeys the triangle inequalityfor all , with equality whenever . Then the seminorm factors through the abelianisation map .

*Proof:* By the triangle inequality, it suffices to show that for all , where is the commutator.

We first establish some basic facts. Firstly, by hypothesis we have , and hence whenever is a power of two. On the other hand, by the triangle inequality we have for all positive , and hence by the triangle inequality again we also have the matching lower bound, thus

for all . The claim is also true for (apply the preceding bound with and ). By replacing with if necessary we may now also assume without loss of generality that , thus

Next, for any , and any natural number , we have

so on taking limits as we have . Replacing by gives the matching lower bound, thus we have the conjugation invariance

Next, we observe that if are such that is conjugate to both and , then one has the inequality

Indeed, if we write for some , then for any natural number one has

where the and terms each appear times. From (2) we see that conjugation by does not affect the norm. Using this and the triangle inequality several times, we conclude that

and the claim (3) follows by sending .

The following special case of (3) will be of particular interest. Let , and for any integers , define the quantity

Observe that is conjugate to both and to , hence by (3) one has

which by (2) leads to the recursive inequality

We can write this in probabilistic notation as

where is a random vector that takes the values and with probability each. Iterating this, we conclude in particular that for any large natural number , one has

where and are iid copies of . We can write where are iid signs. By the triangle inequality, we thus have

noting that is an even integer. On the other hand, has mean zero and variance , hence by Cauchy-Schwarz

But by (1), the left-hand side is equal to . Dividing by and then sending , we obtain the claim.

The above theorem reduces such seminorms to abelian groups. It is easy to see from (1) that any torsion element of such groups has zero seminorm, so we can in fact restrict to torsion-free groups, which we now write using additive notation , thus for instance for . We think of as a -module. One can then extend the seminorm to the associated -vector space by the formula , and then to the associated -vector space by continuity, at which point it becomes a genuine seminorm (provided we have ensured the symmetry condition ). Conversely, any seminorm on induces a seminorm on . (These arguments also appear in this paper of Khare and Rajaratnam.)

This post is a continuation of the previous post, which has attracted a large number of comments. I’m recording here some calculations that arose from those comments (particularly those of Pace Nielsen, Lior Silberman, Tobias Fritz, and Apoorva Khare). Please feel free to either continue these calculations or to discuss other approaches to the problem, such as those mentioned in the remaining comments to the previous post.

Let be the free group on two generators , and let be a quantity obeying the triangle inequality

and the linear growth property

for all and integers ; this implies the conjugation invariance

or equivalently

We consider inequalities of the form

for various real numbers . For instance, since

we have (1) for . We also have the following further relations:

Proposition 1

*Proof:* For (i) we simply observe that

For (ii), we calculate

giving the claim.

For (iii), we calculate

giving the claim.

For (iv), we calculate

giving the claim.

Here is a typical application of the above estimates. If (1) holds for , then by part (i) it holds for , then by (ii) (2) holds for , then by (iv) (1) holds for . The map has fixed point , thus

For instance, if , then .

Here is a curious question posed to me by Apoorva Khare that I do not know the answer to. Let be the free group on two generators . Does there exist a metric on this group which is

- bi-invariant, thus for all ; and
- linear growth in the sense that for all and all natural numbers ?

By defining the “norm” of an element to be , an equivalent formulation of the problem asks if there exists a non-negative norm function that obeys the conjugation invariance

for all , the triangle inequality

for all , and the linear growth

for all and , and such that for all non-identity . Indeed, if such a norm exists then one can just take to give the desired metric.

One can normalise the norm of the generators to be at most , thus

This can then be used to upper bound the norm of other words in . For instance, from (1), (3) one has

A bit less trivially, from (3), (2), (1) one can bound commutators as

In a similar spirit one has

What is not clear to me is if one can keep arguing like this to continually improve the upper bounds on the norm of a given non-trivial group element to the point where this norm must in fact vanish, which would demonstrate that no metric with the above properties on would exist (and in fact would impose strong constraints on similar metrics existing on other groups as well). It is also tempting to use some ideas from geometric group theory (e.g. asymptotic cones) to try to understand these metrics further, though I wasn’t able to get very far with this approach. Anyway, this feels like a problem that might be somewhat receptive to a more crowdsourced attack, so I am posing it here in case any readers wish to try to make progress on it.

By an odd coincidence, I stumbled upon a second question in as many weeks about power series, and once again the only way I know how to prove the result is by complex methods; once again, I am leaving it here as a challenge to any interested readers, and I would be particularly interested in knowing of a proof that was not based on complex analysis (or thinly disguised versions thereof), or for a reference to previous literature where something like this identity has occured. (I suspect for instance that something like this may have shown up before in free probability, based on the answer to part (ii) of the problem.)

Here is a purely algebraic form of the problem:

Problem 1Let be a formal function of one variable . Suppose that is the formal function defined bywhere we use to denote the -fold derivative of with respect to the variable .

- (i) Show that can be formally recovered from by the formula
- (ii) There is a remarkable further formal identity relating with that does not explicitly involve any infinite summation. What is this identity?

To rigorously formulate part (i) of this problem, one could work in the commutative differential ring of formal infinite series generated by polynomial combinations of and its derivatives (with no constant term). Part (ii) is a bit trickier to formulate in this abstract ring; the identity in question is easier to state if are formal power series, or (even better) convergent power series, as it involves operations such as composition or inversion that can be more easily defined in those latter settings.

To illustrate Problem 1(i), let us compute up to third order in , using to denote any quantity involving four or more factors of and its derivatives, and similarly for other exponents than . Then we have

and hence

multiplying, we have

and

and hence after a lot of canceling

Thus Problem 1(i) holds up to errors of at least. In principle one can continue verifying Problem 1(i) to increasingly high order in , but the computations rapidly become quite lengthy, and I do not know of a direct way to ensure that one always obtains the required cancellation at the end of the computation.

Problem 1(i) can also be posed in formal power series: if

is a formal power series with no constant term with complex coefficients with , then one can verify that the series

makes sense as a formal power series with no constant term, thus

For instance it is not difficult to show that . If one further has , then it turns out that

as formal power series. Currently the only way I know how to show this is by first proving the claim for power series with a positive radius of convergence using the Cauchy integral formula, but even this is a bit tricky unless one has managed to guess the identity in (ii) first. (In fact, the way I discovered this problem was by first trying to solve (a variant of) the identity in (ii) by Taylor expansion in the course of attacking another problem, and obtaining the transform in Problem 1 as a consequence.)

The transform that takes to resembles both the exponential function

and Taylor’s formula

but does not seem to be directly connected to either (this is more apparent once one knows the identity in (ii)).

My colleague Tom Liggett recently posed to me the following problem about power series in one real variable . Observe that the power series

has very rapidly decaying coefficients (of order ), leading to an infinite radius of convergence; also, as the series converges to , the series decays very rapidly as approaches . The problem is whether this is essentially the only example of this type. More precisely:

Problem 1Let be a bounded sequence of real numbers, and suppose that the power series(which has an infinite radius of convergence) decays like as , in the sense that the function remains bounded as . Must the sequence be of the form for some constant ?

As it turns out, the problem has a very nice solution using complex analysis methods, which by coincidence I happen to be teaching right now. I am therefore posing as a challenge to my complex analysis students and to other readers of this blog to answer the above problem by complex methods; feel free to post solutions in the comments below (and in particular, if you don’t want to be spoiled, you should probably refrain from reading the comments). In fact, the *only* way I know how to solve this problem currently is by complex methods; I would be interested in seeing a purely real-variable solution that is not simply a thinly disguised version of a complex-variable argument.

(To be fair to my students, the complex variable argument does require one additional tool that is not directly covered in my notes. That tool can be found here.)

## Recent Comments