We consider the incompressible Euler equations on the (Eulerian) torus , which we write in divergence form as

where is the (inverse) Euclidean metric. Here we use the summation conventions for indices such as (reserving the symbol for other purposes), and are retaining the convention from Notes 1 of denoting vector fields using superscripted indices rather than subscripted indices, as we will eventually need to change variables to Lagrangian coordinates at some point. In principle, much of the discussion in this set of notes (particularly regarding the positive direction of Onsager’s conjecture) could also be modified to also treat non-periodic solutions that decay at infinity if desired, but some non-trivial technical issues do arise non-periodic settings for the negative direction.

As noted previously, the kinetic energy

is formally conserved by the flow, where is the Euclidean metric. Indeed, if one assumes that are continuously differentiable in both space and time on , then one can multiply the equation (1) by and contract against to obtain

which rearranges using (2) and the product rule to

and then if one integrates this identity on and uses Stokes’ theorem, one obtains the required energy conservation law

It is then natural to ask whether the energy conservation law continues to hold for lower regularity solutions, in particular weak solutions that only obey (1), (2) in a distributional sense. The above argument no longer works as stated, because is not a test function and so one cannot immediately integrate (1) against . And indeed, as we shall soon see, it is now known that once the regularity of is low enough, energy can “escape to frequency infinity”, leading to failure of the energy conservation law, a phenomenon known in physics as *anomalous energy dissipation*.

But what is the precise level of regularity needed in order to for this anomalous energy dissipation to occur? To make this question precise, we need a quantitative notion of regularity. One such measure is given by the Hölder space for , defined as the space of continuous functions whose norm

is finite. The space lies between the space of continuous functions and the space of continuously differentiable functions, and informally describes a space of functions that is “ times differentiable” in some sense. The above derivation of the energy conservation law involved the integral

that roughly speaking measures the fluctuation in energy. Informally, if we could take the derivative in this integrand and somehow “integrate by parts” to split the derivative “equally” amongst the three factors, one would morally arrive at an expression that resembles

which suggests that the integral can be made sense of for once . More precisely, one can make

Conjecture 1 (Onsager’s conjecture)Let and , and let .

- (i) If , then any weak solution to the Euler equations (in the Leray form ) obeys the energy conservation law (3).
- (ii) If , then there exist weak solutions to the Euler equations (in Leray form) which do not obey energy conservation.

This conjecture was originally arrived at by Onsager by a somewhat different heuristic derivation; see Remark 7. The numerology is also compatible with that arising from the Kolmogorov theory of turbulence (discussed in this previous post), but we will not discuss this interesting connection further here.

The positive part (i) of Onsager conjecture was established by Constantin, E, and Titi, building upon earlier partial results by Eyink; the proof is a relatively straightforward application of Littlewood-Paley theory, and they were also able to work in larger function spaces than (using -based Besov spaces instead of Hölder spaces, see Exercise 3 below). The negative part (ii) is harder. Discontinuous weak solutions to the Euler equations that did not conserve energy were first constructed by Sheffer, with an alternate construction later given by Shnirelman. De Lellis and Szekelyhidi noticed the resemblance of this problem to that of the Nash-Kuiper theorem in the isometric embedding problem, and began adapting the *convex integration* technique used in that theorem to construct weak solutions of the Euler equations. This began a long series of papers in which increasingly regular weak solutions that failed to conserve energy were constructed, culminating in a recent paper of Isett establishing part (ii) of the Onsager conjecture in the non-endpoint case in three and higher dimensions ; the endpoint remains open. (In two dimensions it may be the case that the positive results extend to a larger range than Onsager’s conjecture predicts; see this paper of Cheskidov, Lopes Filho, Nussenzveig Lopes, and Shvydkoy for more discussion.) Further work continues into several variations of the Onsager conjecture, in which one looks at other differential equations, other function spaces, or other criteria for bad behavior than breakdown of energy conservation. See this recent survey of de Lellis and Szekelyhidi for more discussion.

In these notes we will first establish (i), then discuss the convex integration method in the original context of the Nash-Kuiper embedding theorem. Before tackling the Onsager conjecture (ii) directly, we discuss a related construction of high-dimensional weak solutions in the Sobolev space for close to , which is slightly easier to establish, though still rather intricate. Finally, we discuss the modifications of that construction needed to establish (ii), though we shall stop short of a full proof of that part of the conjecture.

We thank Phil Isett for some comments and corrections.

** — 1. Energy conservation for sufficiently regular weak solutions — **

We now prove the positive part (i) of Onsager’s conjecture, which turns out to be a straightforward application of Littlewood-Paley theory. We need the following relation between Hölder spaces and Littlewood-Paley projections:

Let be a weak solution to the Euler equations for some , thus

To show (3), it will suffice from dominated convergence to show that

as . Applying to (4), we have

From Bernstein’s inequality we conclude that , and thus . Thus solves the PDE

in the classical sense. We can then apply the fundamental theorem of calculus to write

and so it will suffice to show that

We can integrate by parts to place the Leray projection onto the divergence-free factor , at which point it may be removed. Moving the derivative over there as well, we now reduce to showing that

On the other hand, the expression is a total derivative (as is divergence-free), and thus has vanishing integral. Thus it will remain to show that

From Bernstein’s inequality, Exercise 2, and the triangle inequality one has for any time that

as is finite, it thus suffices to establish the pointwise bound

We split the left-hand side into the sum of

To treat (7), we use Exercise 2 to conclude that

and so the quantity (7) is , which is acceptable since . Now we turn to (5). This is a commutator of the form

where . Observe that this commutator wouuld vanish if were replaced by , thus we may write this commutator as

where . If we write

for a suitable Schwartz function of total mass one, we have

writing , we thus have the bound

But from Bernstein’s inequality and Exercise 2 we have

and

and so we see that (5) is also of size , which is acceptable since . A similar argument gives (6), and the claim follows.

As shown by Constantin, E, and Titi, the Hölder regularity in the above result can be relaxed to Besov regularity, at least in non-endpoint cases:

Exercise 3Let . Define the Besov space to be the space of functions such that the Besov space normShow that if is a weak solution to the Euler equations, the energy is conserved in time.

The endpoint case of the above exercise is still open; however energy conservation in the slightly smaller space is known thanks to the work of Cheskidov, Constantin, Friedlander, and Shvydkoy (see also this paper of Isett for further discussion, and this paper of Isett and Oh for an alternate argument that also works on Riemannian manifolds).

As observed by Isett (see also the recent paper of Colombo and de Rosa), the above arguments also give some partial information about the energy in the low regularity regime:

Exercise 4Let be a weak solution to the Euler equations for .

- (i) If , show that the energy is a function of time. (
Hint:express the energy as the uniform limit of .)- (ii) If , show that the energy is a function of time.

Exercise 5Let be a weak solution to the Navier-Stokes equations for some with initial data . Establish the energy identity

Remark 6An alternate heuristic derivation of the threshold for the Onsager conjecture is as follows. If , then from Exercise 2 we see that the portion of that fluctuates at frequency has amplitude at most ; in particular, the amount of energy at frequencies is at most . On the other hand, by the heuristics in Remark 11 of 254A Notes 3, the time needed for the portion of the solution at frequency to evolve to a higher frequency scale such as is of order . Thus the rate ofenergy fluxat frequency should be . For , the energy flux goes to zero as , and so energy cannot escape to frequency infinity in finite time.

Remark 7Yet another alternate heuristic derivation of the threshold arises by considering the dynamics of individual Fourier coefficients. Using a Fourier expansionthe Euler equations may be written as

In particular, the energy at a single Fourier mode evolves according to the equation

If , then we have for any , hence by Plancherel’s theorem

which suggests that (up to logarithmic factors) one would expect to be of magnitude about . Onsager posited that for typically “turbulent” or “chaotic” flows, the main contributions to (8) come when have magnitude roughly comparable to that of , and that the various summands should not be correlated strongly to each other. For , one might expect about significant terms in the sum, which according to standard “square root cancellation heuristics” (cf. the central limit theorem) suggests that the sum is about as large as times the main term. Thus the total flux of energy in or out of a single mode would be expected to be of size , and so the total flux in or out of the frequency range (which consists of modes ) should be about . As such, for the energy flux should decay in and so there is no escape of energy to frequency infinity, whereas for such an escape should be possible. Related heuristics can also support Kolmogorov’s 1941 model of the distribution of energy in the vanishing viscosity limit; see this blog post for more discussion. On the other hand, while Onsager did discuss the dynamics of individual Fourier coefficients in his paper, it appears that he arrived at the threshold by a more physical space based approach, a rigorous version of which was eventually established by Duchon and Robert; see this survey of Eyink and Sreenivasan for more discussion.

** — 2. The (local) isometric embedding problem — **

Before we develop the convex integration method for fluid equations, we first discuss the simpler (and historically earlier) instance of this technique for the isometric embedding problem for Riemannian manifolds. To avoid some topological technicalities that are not the focus of the present set of notes, we only consider the *local problem* of embedding a small neighbourhood of the origin in into Euclidean space .

Let be an open neighbourhood of in . A (smooth) Riemannian metric on , when expressed in coordinates, is a family of smooth maps for , such that for each point , the matrix is symmetric and positive definite. Any such metric gives the structure of an (incomplete) Riemannian manifold . An *isometric embedding* of this manifold into a Euclidean space is a map which is continuously differentiable, injective, and obeys the equation

pointwise on , where is the usual inner product (or dot product) on . In the differential geometry language from Notes 1, we are looking for an injective map such that the Euclidean metric on pulls back to via : .

The *isometric embedding problem* asks, given a Riemannian manifold such as , whether there is an isometric embedding from this manifold to a Euclidean space ; for simplicity we only discuss the simpler *local isometric embedding problem* of constructing an isometric immersion of into for some sufficiently small neighbourhood of the origin. In particular for the local problem we do not need to worry about injectivity since (9) ensures that the derivative map is injective at the origin, and hence is injective near the origin by the inverse function theorem (indeed it is an immersion near the origin).

It is a celebrated theorem of Nash (discussed in this previous blog post) that the isometric embedding problem is possible in the smooth category if the dimension is large enough. For sake of discussion we just present the local version:

Theorem 8 (Nash embedding theorem)Suppose that is sufficiently large depending on . Then for any smooth metric on a neighbourhood of the origin, there is a smooth local isometric embedding on some smaller neighbourhood of the origin.

The optimal value of depending on is not completely known, but it grows roughly quadratically in . Indeed, in this paper of Günther it is shown that one can take

In the other direction, one cannot take below :

Proposition 9Suppose that . Then there exists a smooth Riemannian metric on an open neighbourhood of the origin in such that there is no smooth embedding from any smaller neighbourhood of the origin to .

*Proof:* Informally, the reason for this is that the given field has degrees of freedom (which is the number of independent fields after accounting for the symmetry ), but there are only degrees of freedom for the unknown . To make this rigorous, we perform a Taylor expansion of both and around the origin up to some large order , valid for a sufficiently small neighbourhood :

Equating coefficients, we see that the coefficients

are a polynomial function of the coefficients

this polynomial can be written down explicitly if desired, but its precise form will not be relevant for the argument. Observe that the space of possible coefficients contains an open ball, as can be seen by considering arbitrary perturbations of the Euclidean metric on (here it is important to restrict to in order to avoid the symmetry constraint ; also, the positive definiteness of the metric will be automatic as long as one restricts to sufficiently small perturbations). Comparing dimensions, we conclude that if every smooth metric had a smooth embedding , one must have the inequality

Dividing by and sending , we conclude that . Taking contrapositives, the claim follows.

Remark 10If one replaces “smooth” with “analytic”, one can reverse the arguments here using the Cauchy-Kowaleski theorem and show that any analytic metric on can be locally analytically embedded into ; this is a classical result of Cartan and Janet.

Apart from the slight gap in dimensions, this would seem to settle the question of when a -dimensional metric may be locally isometrically embedded in . However, all of the above arguments required the immersion map to be smooth (i.e., ), whereas the definition of an isometric embedding only required the regularity of .

It is a remarkable (and somewhat counter-intuitive) result of Nash and Kuiper that if one only requires the embedding to be in , then one can embed into a much lower dimensional space:

Theorem 11 (Nash-Kuiper embedding theorem)Let . Then for any smooth metric on a neighbourhood of the origin, there is a local isometric embedding on some smaller neighbourhood of the origin.

Nash originally proved this theorem with the slightly weaker condition ; Kuiper then obtained the optimal condition . The case fails due to curvature obstructions; for instance, if the Riemannian metric has positive scalar curvature, then small Riemannian balls of radius will have (Riemannian) volume slightly less than their Euclidean counterparts, whereas any embedding into will preserve both Riemannian length and volume, preventing such an isometric embedding from existing.

Remark 12One striking illustration of the distinction between the and smooth categories comes when considering isometric embeddings of the round sphere (with the usual metric) into Euclidean space . It is a classical result (see e.g., Spivak’s book) that the only isometric embeddings of in are the obvious ones coming from composing the inclusion map with an isometry of the Euclidean space; however, the Nash-Kuiper construction allows one to create an embedding of into an arbitrarily small ball! Thus the embedding problem lacks the “rigidity” of the embedding problem. This is an instance of a more general principle that nonlinear differential equations such as (10) can become much less rigid when one weakens the regularity hypotheses demanded on the solution.

To prove this theorem we work with a relaxation of the isometric embedding problem. We say that is a *short isometric embedding* on if , solve the equation

on with the matrix symmetric and positive definite for all . With the additional unknown field it is much easier to solve the short isometric problem than the true problem. For instance:

Proposition 13Let , and let be a smooth Riemannian metric on a neighbourhood of the origin in . There is at least one short isometric embedding .

*Proof:* Set and for a sufficiently small , where is the standard embedding, and the Euclidean metric on ; this will be a short isometric embedding on some neighbourhood of the origin.

To create a true isometric embedding , we will first construct a sequence of short embeddings with converging to zero in a suitable sense, and then pass to a limit. The key observation is then that by using the fact that the positive matrices lie in the convex hull of the rank one matrices, one can add a high frequency perturbation to the first component of a short embedding to largely erase the error term , replacing it instead with a much higher frequency error.

We now prove Theorem 11. The key iterative step is

Theorem 14 (Iterative step)Let , let be a closed ball in , let be a smooth Riemannian metric on , and let be a short isometric embedding on . Then for any , one can find a short isometric embedding to (11) on with

Suppose for the moment that we had Theorem 14. Starting with the short isometric embedding on a ball provided by Proposition 13, we can iteratively apply the above theorem to obtain a sequence of short isometric embeddings on with

for . From this we see that converges uniformly to zero, while converges in norm to a limit , which then solves (10) on , giving Theorem 11. (Indeed, this shows that the space of isometric embeddings is dense in the space of short maps in the topology.)

We prove Theorem 14 through a sequence of reductions. Firstly, we can rearrange it slightly:

Theorem 15 (Iterative step, again)Let , let be a closed ball in , let be a smooth immersion, and let be a smooth Riemannian metric on . Then there exists a sequence of smooth immersions for obeying the bounds

uniformly on for , where denotes a quantity that goes to zero as (for fixed choices of ).

Let us see how Theorem 15 implies Theorem 14. Let the notation and hypotheses be as in Theorem 14. We may assume to be small. Applying Theorem 15 with replaced by (which will be positive definite for small enough), we obtain a sequence of smooth immersions obeying the estimates

If we set

then is smooth, symmetric, and (from (13)) will be positive definite for large enough. By construction, we thus have solving (11), and Theorem 14 follows.

To prove Theorem 15, it is convenient to break up the metric into more “primitive” pieces that are rank one matrices:

Lemma 16 (Rank one decomposition)Let be a closed ball in , and let be a smooth Riemannian metric on . Then there exists a finite collection of unit vectors in , and smooth functions , such thatfor all . Furthermore, for each , at most of the are non-zero.

Remark 17Strictly speaking, the unit vectors should belong to the dual space of rather than itself, in order to have the index appear as subscripts instead of superscripts. (A similar consideration occurs for the frequency vectors from Remark 7. However, we will not bother to distinguish between and here (since they are identified using the Euclidean metric).

*Proof:* Fix a point in . Then the matrix is symmetric and positive definite; one can thus write , where is an orthonormal basis of (column) eigenvectors of and are the eigenvalues (we suppress for now the dependence of these objects on ). Using the parallelogram identity

for some positive real numbers , where are the unit vectors of the form for , enumerated in an arbitrary order. From the further parallelogram identity

we see that every sufficiently small symmetric perturbation of also has a representation of the form (14) with slightly different coefficients that depend smoothly on the perturbation. As is smooth, we thus see that for sufficiently close to we have the decomposition

for some positive quantities varying smoothly with . This gives the lemma in a small ball centred at ; the claim then follows by covering by a finite number balls of the form (say), covering these balls by balls of a fixed radius smaller than all the in the finite cover, in such a way that any point lies in at most of the balls , constructing a smooth partition of unity adapted to the , multiplying each of the decompositions of previously obtained on (which each lie in one of the ) by , and summing to obtain the required decomposition on .

Remark 18Informally, Lemma 16 lets one synthesize a metric as a “convex integral” of rank one pieces, so that if the problem at hand has the freedom to “move” in the direction of each of these rank one pieces, then it also has the freedom to move in the direction , at least if one is working in low enough regularities that one can afford to rapidly change direction from one rank one perturbation to another. This convex integration technique was formalised by Gromov in his interpretation of the Nash-Kuiper method as part of his “-principle“, which we will not discuss further here.

One can now deduce Theorem 15 from

Theorem 19 (Iterative step, rank one version)Let , let be a closed ball in , let be a smooth immersion, let be a unit vector, and let be smooth. Then there exists a sequence of smooth immersions for obeying the bounds

uniformly on for . Furthermore, the support of is contained in the support of .

Indeed, suppose that Theorem 19 holds, and we are in the situation of Theorem 15. We apply Lemma 16 to obtain the decomposition

with the stated properties. On taking traces we see that

for all . Applying Theorem 19 times (and diagonalising the sequences as necessary), we obtain sequences of smooth immersions for such that and one has

an such that the support of is contained in that of . The claim then follows from the triangle inequality, noting that the implied constant in (12) will not depend on because of the bounded overlap in the supports of the .

It remains to prove Theorem 19. We note that the requirement that be an immersion will be automatic from (15) for large enough since was already an immersion, making the matrix positive definite uniformly for , and this being unaffected by the addition of the perturbation and the positive semi-definite rank one matrix .

Writing , it will suffice to find a sequence of smooth maps supported in the support of and obeying the approximate difference equation

To locate these functions , we use the *method of slow and fast variables*. First we observe by applying a rotation that we may assume without loss of generality that is the unit vector , thus . We then use the ansatz

where is a smooth function independent of to be chosen later; thus is a function both of the “slow” variable and the “fast” variable taking values in the“fast torus” . (We adopt the convention here of using boldface symbols to denote functions of both the fast and slow variables. The fast torus is isomorphic to the Eulerian torus from the introduction, but we denote them by slightly different notation as they play different roles.) Thus is a low amplitude but highly oscillating perturbation to . The fast variable oscillation means that will not be bounded in regularity norms higher than (and so this ansatz is not available for use in the smooth embedding problem), but because we only wish to control and type quantities, we will still be able to get adequate bounds for the purposes of embedding. Now that we have twice as many variables, the problem becomes more “underdetermined” and we can arrive at a simpler PDE by decoupling the role of the various variables (in particular, we will often work with PDE where the derivatives of the main terms are in the fast variables, but the coefficients only depend on the slow variables, and are thus effectively constant coefficient with respect to the fast variables).

Remark 20Informally, one should think of functions that are independent of the fast variable as being of “low frequency”, and conversely functions that have mean zero in the fast variable (thus for all ) as being of “high frequency”. Thus for instance any smooth function on can be uniquely decomposed into a “low frequency” component and a “high frequency” component, with the two components orthogonal to each other. In later sections we will start inverting “fast derivatives” on “high frequency” functions, which will effectively gain important factors of in the analysis. See also the table below for the dictionary between ordinary physical coordinates and fast-slow coordinates.

Position | Slow variable |

Fast variable | Fast variable |

Function | Function |

Low-frequency function | Function independent of |

High-frequency function | Function mean zero in |

N/A | Slow derivative |

N/A | Fast derivative |

If we expand out using the chain rule, using and to denote partial differentiation in the coordinates of the slow and fast variables respectively, and noting that all terms with at least one power of can be absorbed into the error, we see that we will be done as long as we can construct to obey the bounds

where are viewed as functions of the slow variable only. The original approach of Nash to solve this equation was to use a function that was orthogonal to the entire gradient of , thus

for . Taking derivatives in one would conclude that

and similarly

and one now just had to solve the equation

For this, Nash used a “spiral” construction

where were unit vectors varying smoothly with respect to the slow variable; this obeys (22) and (19), and would also obey (21) if the vectors and were both always orthogonal to the entire gradient of . This is not possible in (as cannot then support linearly independent vectors), but there is no obstruction for :

Lemma 21 (Constructing an orthogonal frame)Let be an immersion. If , then there exist smooth vector fields such that at every point , are unit vectors orthogonal to each other and to for .

*Proof:* Applying the Gram-Schmidt process to the linearly independent vectors for , we can find an orthonormal system of vectors , depending smoothly on , whose span is the same as the span of the . Our task is now to find smooth functions solving the system of equations

For this is possible at the origin from the Gram-Schmidt process. Now we extend in the direction to the line segment . To do this we evolve the fields by the parallel transport ODE

on this line segment. From the Picard existence and uniqueness theorem we can uniquely extend smoothly to this segment with the specified initial data at , and a simple calculation using Gronwall’s inequality shows that the system of equations (23), (24), (25) is preserved by this evolution. Then, one can extend to the disk by using the previous extension to the segment as initial data and solving the parallel transport ODE

Iterating this procedure we obtain the claim.

This concludes Nash’s proof of Theorem 11 when . Now suppose that . In this case we cannot locate two unit vector fields orthogonal to each other and to the entire gradient of ; however, we may still obtain one such vector field by repeating the above arguments. By Gram-Schmidt, we can then locate a smooth unit vector field which is orthogonal to and to for , but has a positive inner product with . If we use the “corrugation” ansatz

for some smooth functions , one is reduced to locating such functions that obey the bounds

and the ODE

This can be done by an explicit construction:

Exercise 22 (One-dimensional corrugation)For any positive , show that there exist smooth functions solving the ODEand which vary smoothly with , and obey the bounds

(

Hint:one can renormalise . The problem is basically to locate a periodic function taking values in the circle of mean zero and Lipschitz norm that varies smoothly with .)

This exercise supplies the required functions , completing Kuiper’s proof of Theorem 11 when .

Remark 23For sake of discussion let us restrict attention to the surface case . For the local isometric embedding problem, we have seen that we have rigidity at regularities at or above , but lack of regularity at . The precise threshold at which rigidity occurs is not completely known at present: a result of Borisov (also reproven here) gives rigidity at the level for , while a result of de Lellis, Inauen, and Szekelyhidi (building upon a series pf previous results) establishes non-rigidity when . For recent results in higher dimensions, see this paper of Cao and Szekelyhidi.

** — 3. Low regularity weak solutions to Navier-Stokes in high dimensions — **

We now turn to constructing solutions (or near-solutions) to the Euler and Navier-Stokes equations. For minor technical reasons it is convenient to work with solutions that are periodic in both space and time, and normalised to have zero mean at every time (although the latter restriction is not essential for our arguments, since one can always reduce to this case after a Galilean transformation as in 254A Notes 1). Accordingly, let denote the periodic spacetime

and let denote the space of smooth periodic functions that have mean zero and are divergence-free at every time , thus

and

We use as an abbreviation for for various vector spaces (the choice of which will be clear from context).

Let (for now, our discussion will apply both to the Navier-Stokes equations and the Euler equations ). Smooth solutions to Navier-Stokes equations then take the form

for some and smooth . Here of course denotes the spatial Laplacian.

Much as we replaced the equation (10) in the previous section with (11), we will consider the relaxed version

of the Navier-Stokes equations, where we have now introduced an additional field , known as the *Reynolds stress* (cf. the Cauchy stress tensor from 254A Notes 0). If , , are smooth solutions to (26), (27), (28), with having mean zero at every time, then we call a *Navier-Stokes-Reynolds flow* (or *Euler-Reynolds flow*, if ). Note that if then we recover a solution to the true Navier-Stokes equations. Thus, heuristically, the smaller is, the closer and should become to a solution to the true Navier-Stokes equations. (The Reynolds stress tensor here is a rank tensor, as opposed to the rank tensor used in the previous section to measure the failure of isometric embedding, but this will not be a particularly significant distinction.)

Note that if is a Navier-Stokes-Reynolds flow, and , , are smooth functions, then will also be a Navier-Stokes-Reynolds flow if and only if has mean zero at every time, and obeys the difference equation

When this occurs, we say that is a *difference Navier-Stokes-Reynolds flow* at .

It will be thus of interest to find, for a given , difference Navier-Stokes-Reynolds flows at with small, as one could hopefully iterate this procedure and take a limit to construct weak solutions to the true Euler equations. The main strategy here will be to choose a highly oscillating (and divergence-free) correction velocity field such that approximates up to an error which is also highly oscillating (and somewhat divergence-free). The effect of this error can then eventually be absorbed efficiently into the new Reynolds stress tensor . Of course, one also has to manage the other terms , , , appearing in (29). In high dimensions it turns out that these terms can be made very small in norm, and can thus be easily disposed of. In three dimensions the situation is considerably more delicate, particularly with regards to the and terms; in particular, the transport term term is best handled by using a local version of Lagrangian coordinates. We will discuss these subtleties in later sections.

To execute above strategy, it will be convenient to have an even more flexible notion of solution, in which is no longer required to be perfectly divergence-free and mean zero, and is also allowed to be slightly inaccurate in solving (29). We say that is an *approximate difference Navier-Stokes-Reynolds flow* at if , , are smooth functions obeying the system

If the error terms , as well as the mean of , are all small, one can correct an approximate difference Navier-Stokes-Reynolds flow to a true difference Navier-Stokes-Reynolds flow with only small adjustments:

Exercise 24 (Removing the error terms)Let be a Navier-Stokes-Reynolds flow, and let be an approximate difference Navier-Stokes-Reynolds flow at . Show that is an approximate difference Navier-Stokes-Reynolds flow at , whereand is the mean of , thus

(

Hint:one will need at some point to show that has mean zero in space at every time; this can be achieved by integrating (32) in space.)

Because of this exercise we will be able to tolerate the error terms if they (and the mean ) are sufficiently small.

As a simple corollary of Exercise 24, we have the following analogue of Proposition 13:

Proposition 25Let . Then there exist smooth fields , such that is a Navier-Stokes-Reynolds flow. Furthermore, if is supported in for some compact time interval , then can be chosen to also be supported in this region.

*Proof:* Clearly is an approximate difference Navier-Stokes-Reynolds flow at , where

Applying Exercise 24, we can construct an difference Navier-Stokes-Reynolds flow at , which then verifies the claimed properties.

Now, we show that, in sufficiently high dimension, a Navier-Stokes-Reynolds flow can be approximated (in an sense) as the limit of Navier-Stokes-Reynolds flows , with the Reynolds stress going to zero.

Proposition 26 (Weak improvement of Navier-Stokes-Reynolds flows)Let , and let be sufficiently large depending on . Let be a Navier-Stokes-Reynolds flow. Then for sufficiently large , there exists a Navier-Stokes-Reynolds flow obeying the estimates

Furthermore, if is supported in for some interval , then one can arrange for to be supported on for any interval containing in its interior (at the cost of allowing the implied constants in the above to depend also on ).

This proposition can be viewed as an analogue of Theorem 14. For an application at the end of this section it is important that the implied constant in (36) is uniform in the choice of initial flow . The estimate (35) can be viewed as asserting that the new velocity field is oscillating at frequencies , at least in an sense. In the next section, we obtain a stronger version of this proposition with more quantitative estimates that can be iterated to obtain higher regularity weak solutions.

To simplify the notation we adopt the following conventions. Given an -dimensional vector of differential operators, we use to denote the -tuple of differential operators with . We use to denote the -tuple formed by concatenating for . Thus for instance the estimate (35) can be abbreviated as

for all . Informally, one should read the above estimate as asserting that is bounded in with norm , and oscillates with frequency in time and in space (or equivalently, with a temporal wavelength of and a spatial wavelength of ).

*Proof:* We can assume that is non-zero, since if we can just take . We may assume that is supported in for some interval (which may be all of ), and let be an interval containing in its interior. To abbreviate notation, we allow all implied constants to depend on .

Assume is sufficiently large. Using the ansatz

and the triangle inequality, it suffices to construct a difference Navier-Stokes-Reynolds flow at supported on and obeying the bounds

for all .

It will in fact suffice to construct an *approximate* difference Navier-Stokes-Reynolds flow at supported on and obeying the bounds

for , since an application of Exercise 24 and some simple estimation will then give a difference Navier-Stokes-Reynolds flow obeying the desired estimates (using in particular the fact that is bounded on and , as can be seen from Littlewood-Paley theory; also note that (39) can be used to ensure that the mean of is very small).

To construct this approximate solution, we again use the method of fast and slow variables. Set , and introduce the fast-slow spacetime , which we coordinatise as ; we use to denote partial differentiation in the coordinates of the slow variable , and to denote partial differentiation in the coordinates of the fast variable . We also use as shorthand for . Define an *approximate fast-slow solution* to the difference Navier-Stokes-Reynolds equation at (at the frequency scale ) to be a tuple of smooth functions , , that obey the system of equations

Here we think of as a “low-frequency” function (in the sense of Remark 20) that only depends on and the slow variable , but not on the fast variable .

Let denote the tuple . Suppose that for any sufficiently large , we can construct an approximate fast-slow solution to the difference equations at supported on supported on that obeys the bounds

for all . (Informally, the presence of the derivatives means that the fields involved are allowed to oscillate in time at wavelength , in the slow variable at wavelength , and in the fast variable at wavelength .) From (46) and the choice of we then have

for all , and similarly for (48), (49). (Note here that there was considerable room in the estimates with regards to regularity in the variable; this room will be exploited more in the next section.) For any shift , we see from the chain rule that is an approximate difference Navier-Stokes-Reynolds flow at supported on , where

Also from \eqerf{bd-1-fast} and Fubini’s theorem we have

for all . By Markov’s inequality and (52), we see that for each , we have

for all outside of an exceptional set of measure (say) . Similarly for the other equations above. Applying the union bound, we can then find a such that obeys all the required bounds (37)-\eqref[bd-5} simultaneously for all . (This is an example of the probabilistic method, originally developed in combinatorics; one can think of probabilistically as a shift drawn uniformly at random from the torus , in order to relate the fast-slow Lebesgue norms to the original Lebesgue norms .)

It remains to construct an approximate fast-slow solution supported on with the required bounds (46)–(51). Actually, in this high-dimensional setting we can afford to simplify the situation here by removing some of the terms (and in particular eliminating the role of the reference velocity field ). Define a *simplified fast-slow solution* at to be a tuple of smooth functions on obeying the simplified equations

If we can find a simplified fast-slow solution of smooth functions on supported on obeying the bounds

for all , then the will be an approximate fast-slow solution supported on obeying the required bounds (46)–(51), where

Now we need to construct a simplified fast-slow solution supported on obeying the bounds (56)–(60). We do this in stages, first finding a solution that cancels off the highest order terms and , and also such that has mean zero in the fast variable (so that it is “high frequency” in the sense of Remark 20). This still leads to fairly large values of and , but we will then apply a “divergence corrector” to almost completely eliminate , followed by a “stress corrector” that almost completely eliminates , at which point we will be done.

We turn to the details. Our preliminary construction of the velocity field will be a “Mikado flow”, consisting of flows along narrow tubes. (Earlier literature used other flows, such as Beltrami flows; however, Mikado flows have the advantage of being localisable to small subsets of spacetime, which is particularly useful in high dimensions.) We need the following modification of Lemma 16:

Exercise 27Let be a compact subset of the space of positive definite matrices . Show that there exist non-zero lattice vectors and smooth functions for some such that

for all . (This decomposition is essentially due to de Lellis and Szekelyhidi. The subscripting and superscripting here is reversed from that in Lemma 16; this is because we are now trying to decompose a rank tensor rather than a rank tensor.)

We would like to apply this exercise to the matrix with entries . We thus need to select the pressure so that this matrix is positive definite. There are many choices available for this pressure; we will take

where is the Frobenius norm of . Then is smooth and is positive definite on all of the compact spacetime (recall that we can assume to not be identically zero), and in particular ranges in a compact subset of positive definite matrices. Applying the previous exercise and composing with the function , we conclude that there exist non-zero lattice vectors and smooth functions for some such that

for all . As depend only on , and is a component of , all norms of these quantities are bounded by ; they are independent of . Furthermore, on taking traces and integrating on , we obtain the estimate

(note here that the implied constant is uniform in , ). By applying a smooth cutoff in time that equals on and vanishes outside of , we may assume that the are supported in .

Now for each , the closed subgroup is a one-dimensional subset of , so the -neighbourhood of this subgroup has measure ; crucially, this will be a large negative power of when is very large. let be a translate of this -neighbourhood such that all the are disjoint; this is easily accomplished by the probabilistic method for large enough, translating each of the by an independent random shift and noting that the probability of a collision goes to zero as (here we need the fact that we are in at least three dimensions).

Let be a large integer (depending on ) to be chosen later. For any , let be a scalar test function supported on that is constant in the direction, thus

and is not identically zero, which implies that the iterated Laplacian of is also not identically zero (thanks to the unique continuation property of harmonic functions). We can normalise so that

and we can also arrange to have the bounds

for all (basically by first constructing a version of on a standard cylinder and the applying an affine transformation to map onto ).

Let denote the function

intuitively this represents the velocity field of a fluid traveling along the tube , with the presence of the Laplacian ensuring that this function is extremely well balanced (for instance, it will have mean zero, and thus “high frequency” in the sense of Remark 20). Clearly is divergence free, and one also has the steady-state Euler equation

for all and . If we then set

then one easily checks that is a simplified fast-slow solution supported in . Direct calculation using the Leibniz rule then gives the bounds

for all , while from (64) one has

(note here that the implied constant is uniform in ).

This looks worse than (56)–(60). However, observe that is supported on the set , which has measure , which for large enough can be taken to be (say) . Thus by Cauchy-Schwarz one has

for all . Also, from construction (particularly (106)) we see that is of mean zero in the variable (thus it is “high frequency” in the sense of Remark 20).

We are now a bit closer to (56)–(60), but our bounds on are not yet strong enough. We now apply a “divergence corrector” to make much smaller. Observe from construction that where

and is supported on and obeys the estimates

We abbreviate the differential operator as . Iterating the above identity times, we obtain

where

and

In particular, is supported in . Observe that is a simplified fast-slow solution supported in , where

From (72) we have

so in particular for large enough

for any . Meanwhile, another appeal to (72) yields

for any , and hence by (67) and the triangle inequality

Similarly one has

Since continues to be supported on the thin set , we can apply Hölder as before to conclude that

for any . Also, from (73) and Hölder we have

We have now achieved the bound (59); the remaining estimate that needs to be corrected for is (60). This we can do by a modification of the previous argument, where we now work to reduce the size of rather than . Observe that as is “high frequency” (mean zero in the variable), one can write

where is the linear operator on smooth vector-valued functions on of mean zero defined by the formula

Note that also has mean zero. We can thus iterate and obtain

where

and is a smooth function whose exact form is explicit but irrelevant for our argument. We then see that is a simplified fast-slow solution supported in . Since is bounded in , we see from (69) that

and

if is large enough. Thus obeys the required bounds (56)–(60), concluding the proof.

As an application of this proposition we construct a low-regularity weak solution to high-dimensional Navier-Stokes that does not obey energy conservation. More precisely, for any , let be the Banach space of periodic functions which are divergence free, and of mean zero at every time. For , define a *time-periodic weak solution* of the Navier-Stokes (or Euler, if ) equations to be a function that solves the equation

in the sense of distributions. (Note that one may easily define on functions in a distributional sense, basically because the adjoint operator maps test functions to bounded functions.)

Corollary 28 (Low regularity non-trivial weak solutions)Assume that the dimension is sufficiently large. Then for any , there exists a periodic weak solution to Navier-Stokes which equals zero at time , but is not identically zero. In particular, periodic weak solutions are not uniquely determined by their initial data, and do not necessarily obey the energy inequality

*Proof:* Let be an element of that is supported on and is not identically zero (it is easy to show that such an element exists). By Proposition 25, we may then find a Navier-Stokes-Reynolds flow also supported on . Let be sufficiently large. By applying Proposition 26 repeatedly (with say ) and with a sufficiently rapidly increasing sequence , we can find a sequence of Navier-Stokes-Reynolds flows supported on (say) obeying the bounds

(say) for . For sufficiently rapidly growing, this implies that converges strongly in to zero, while converges strongly in to some limit supported in . From the triangle inequality we have

(if is sufficiently rapidly growing) and hence is not identically zero if is chosen large enough. Applying Leray projections to the Navier-Stokes-Reynolds equation we have

in the sense of distributions (where is the vector field with components for ); taking distributional limits as , we conclude that is a periodic weak solution to the Navier-Stokes equations, as required.

** — 4. High regularity weak solutions to Navier-Stokes in high dimensions — **

Now we refine the above arguments to give a higher regularity version of Corollary 28, in which we can give the weak solutions almost half a derivative of regularity in the Sobolev scale:

Theorem 29 (Non-trivial weak solutions)Let , and assume that the dimension is sufficiently large depending on . Then for any , there exists a periodic weak solution which equals zero at time , but is not identically zero. In particular, periodic weak solutions are not uniquely determined by their initial data, and do not necessarily obey the energy inequality (74).

This result is inspired by a three-dimensional result of Buckmaster and Vicol (with a small value of ) and a higher dimensional result of Luo (taking , and restricting attention to time-independent solutions). In high dimensions one can create fairly regular functions which are large in type norms but tiny in type norms; when using the Sobolev scale to control the solution (and type norms to measure an associated stress tensor), this has the effect of allowing one to treat as negligible the linear terms in (variants of) the Navier-Stokes equation, as well as interaction terms between low and high frequencies. As such, the analysis here is simpler than that required to establish the Onsager conjecture. The construction used to prove this theorem shows in fact that periodic weak solutions are in some sense “dense” in , but we will not attempt to quantify this fact here.

In the proof of Corollary 28, we took the frequency scales to be extremely rapidly growing in . This will no longer be good enough for proving Theorem 29, and in fact we need to take a fairly dense set of frequency scales in which for a small . In order to do so, we have to replace Proposition 26 with a more quantitative version in which the dependence of bounds on the size of the original Navier-Stokes-Reynolds flow is made much more explicit.

We turn to the details. We select the following parameters:

- A regularity ;
- A quantity , assumed to be sufficiently small depending on ;
- An integer , assumed to be sufficiently large depending on ; and
- A dimension , assumed to be sufficiently large depending on .

Then we let . To simplify the notation we allow all implied constants to depend on unless otherwise specified. We recall from the previous section the notion of a Navier-Stokes-Reynolds flow . The basic strategy is to start with a Navier-Stokes-Reynolds flow and repeatedly adjust by increasingly high frequency corrections in order to significantly reduce the size of the stress (at the cost of making both of these expressions higher in frequency).

As before, we abbreviate as . We write for the spatial gradient to distinguish it from the time derivative .

The main iterative statement (analogous to Theorem 14) starts with a Navier-Stokes-Reynolds flow oscillating at spatial scales up to some small wavelength , and modifies it to a Navier-Stokes-Reynolds flow oscillating at a slightly smaller wavelength , with a smaller Reynolds stress. It can be viewed as a more quantitative version of Proposition 26.

Theorem 30 (Iterative step)Let be sufficiently large depending on the parameters . Set . Suppose that one has a Navier-Stokes-Reynolds flow obeying the estimates

for some . Set . Then there exists a Navier-Stokes-Reynolds flow obeying the estimates

Furthermore, if is supported on for some interval , then one can ensure that is supported in , where is the -neighbourhood of .

Let us assume Theorem 30 for the moment and establish Theorem 29. Let be chosen to be supported on (say) and not be identically zero. By Proposition 25, we can then find a Navier-Stokes-Reynolds flow supported on . Let be a sufficiently large parameter, and set , then the hypotheses (75), (76) will be obeyed for large enough. Set for all . By iteratively applying Theorem 30, we may find a sequence of Navier-Stokes-Reynolds flows, all supported on (say) , obeying the bounds

for . In particular, the converge weakly to zero on , and we have the bound

from Plancherel’s theorem, and hence by Sobolev embedding in time

Thus converges strongly in (and in particular also in for some ) to some limit ; as the are all divergence-free, is also. From applying Leray projections to (26) one has

Taking weak limits we conclude that is a weak solution to Navier-Stokes. Also, from construction one has

(say), and so for large enough is not identically zero. This proves Theorem 29.

It remains to establish Theorem 30. It will be convenient to introduce the intermediate frequency scales

where

is slightly larger than , and

is slightly smaller than (and constrained to be integer).

Before we begin the rigorous argument, we first give a heuristic explanation of the numerology. The initial solution has about degrees of regularity controlled at . For technical reasons we will upgrade this to an infinite amount of regularity, at the cost of worsening the frequency bound slightly from to . Next, to cancel the Reynolds stress up to a smaller error , we will perturb by some high frequency correction , basically oscillating at spatial frequency (and temporal frequency ), so that is approximately equal to (minus a pressure term) after averaging at spatial scales . Given the size bound (76), one expects to achieve this with of norm about . By exploiting the small gap between and , we can make concentrate on a fairly small measure set (of density something like ), which in high dimension allows us to make linear terms such as and (as well as the correlation terms and ) negligible in size (as measured using type norms) when compared against quadratic terms such as (cf. the proof of Proposition 26). The defect will then oscillate at frequency , but can be selected to be of size about in norm, because can choose to cancel off all the high-frequency (by which we mean or greater) contributions to this term, leaving only low frequency contributions (at frequencies or below). Using the ellipticity of the Laplacian, we can then express this defect as where the norm of is of order

When , this is slightly less than , allowing one to close the argument.

We now turn to the rigorous details. In a later part of the argument we will encounter a *loss of derivatives*, in that the new Navier-Stokes-Reynolds flow has lower amounts of controlled regularity (in both space and time) than the Navier-Stokes-Reynolds flow used to construct it. To counteract this loss of derivatives we need to perform an initial *mollification step*, which improves the amount of regularity from derivatives in space and one in time to an unlimited number of derivatives in space and two derivatives in time, at the cost of worsening the estimates on slightly (basically by replacing with ).

Proposition 31 (Mollification)Let the notation and hypotheses be as above. Then we can find a Navier-Stokes-Reynolds flow obeying the estimates

Furthermore, if is supported on for some interval , then one can ensure that is supported in , where is the -neighbourhood of .

We remark that this sort of mollification step is now a standard technique in any iteration scheme that involves loss of derivatives, including the Nash-Moser iteration scheme that was first used to prove Theorem 8.

*Proof:* Let be a bump function (depending only on ) supported on the region of total mass , and define the averaging operator on smooth functions by the formula

From the fundamental theorem of calculus we have

where is the identity operator and

The operators and will behave like low and high frequency Littlewood-Paley projections. (We cannot directly use these projections here because their convolution kernels are not localised in time.)

Observe that are convolution operators and thus commute with each other and with the partial derivatives . If we apply the operator to (26), (27), (28), we see that is Navier-Stokes-Reynolds flow, where

Since , is a linear combination of the operators . In particular, we see that is supported on .

We abbreviate . For any , we have

and therefore deduce the bounds

for any , thanks to Young’s inequality. A similar application of Young’s inequality gives

From (81) and decomposing as linear combinations of , we have

for any , and hence (77) follows from (75). In a similar spirit, from (82), (75) one has

if is large enough, and this gives (79), (80).

Finally we prove (78). By the triangle inequality it suffices to show that

for any . The claim (83) follows from (81), (76), after writing as a linear combination of and noting that . For (84), if we again write as a linear combination of and uses (81) and the Leibniz rule, one can bound the left-hand side of (84) by

and hence by (75) (bounding by ) this is bounded by

This gives (84) when . For , we rewrite the expression

as

The contribution of the first term to (84) can be bounded using (82), (81) (splitting ) by

which by the Leibniz rule, bounding by , and (75) is bounded by

which is again an acceptable contribution to (84) since is large. The other terms are treated similarly.

We return to the proof of Theorem 30. We abbreviate . Let be the Navier-Stokes-Reynolds flow constructed by Proposition 31. By using the ansatz

and the triangle inequality, it will suffice to locate a difference Navier-Stokes-Reynolds flow at supported on , obeying the estimates

From (77) we have

so by the triangle inequality we can replace (85) by

and then (85), (87) may then be replaced by the single estimate

(say). By using Exercise 24 as in the previous section, it then suffices to construct an *approximate* difference Navier-Stokes-Reynolds flow to the difference equation at supported on supported in obeying the bounds (86), (89),

Now, we pass to fast and slow variables. Let denote the tuple

informally, the use of is consistent with oscillations in time of wavelength , in the slow variable of wavelength , and in the fast variable of wavelength .

Exercise 32By using the method of fast and slow variables as in the previous section, show that to construct the approximate Navier-Stokes-Reynolds flow at obeying the bounds (86), (89), (90), (91), it suffices to locate an approximate fast-slow solution to the difference Navier-Stokes-Reynolds equation at (at frequency scale rather than ) and supported in that obey the bounds

As in the previous section, we can then pass to simplified fast-slow soutions:

Exercise 33Show that to construct the approximate fast-slow solution to the difference equation at obeying the estimates of the previous exercise, it will in fact suffice to locate a simplified fast-slow solution at (again at frequency scale ) supported on , obeying the bounds (92), (93), (95), (96) and

(

Hint:one will need the estimatefrom Proposition 31.)

Now we need to construct a simplified fast-slow solution at supported on obeying the bounds (92), (93), (95), (96), (97). As in the previous section, we do this in stages, first finding a solution that cancels off the top order terms and , and also such that is “ high frequency” (mean zero in ). Then we apply a divergence corrector to completely eliminate , followed by a stress corrector that almost completely eliminates .

As before, we need to select so that is positive definite. In the previous section we essentially took to be a large multiple of , but now we will need good control on the derivatives of , which requires a little more care. Namely, we will need the following technical lemma:

Lemma 34 (Smooth polar-type decomposition)There exists a factorisation , where , are smooth, supported on , and obey the estimates

*Proof:* We may assume that is not identically zero, since otherwise the claim is trivial. For brevity we write and . From (78) we have

Let denote the spacetime cylinder , and let denote the maximal function

From the fundamental theorem of calculus (or Sobolev embedding) one has the pointwise estimate

thus by Fubini’s theorem and (101)

We do not have good control on the derivatives of , so we apply a smoothing operator. Let denote the function

where , then by Fubini’s theorem (or Young’s inequality)

Also, is smooth and strictly positive everywhere, and from differentation under the integral sign and integration by parts we have

for any . Also, from construction one has

Write . From many applications of the chain rule (or the Faá di Bruno formula), we see that for any , is a linear combination of terms of the form

where sum up to (more precisely, each component of is a linear combination of expressions of the above form in which one works with individual components of each factor rather than the full tuple ). From (103) we thus have the pointwise estimate

for any , and (98) now follows from (102). A similar argument gives

for any , hence if we set , then by the product rule

and (99) now follows from (104).

Strictly speaking we are not quite done because is not supported in , but if one applies a smooth cutoff function in time that equals on (where is supported in time) and vanishes outside of , we obtain the required support property without significantly affecting the estimates.

Let be the factorisation given by the above lemma. If we set for a sufficiently large constant depending only on , then

For large enough, we see from (99) that the matrix with entries takes values in a compact subset of positive definite matrices) that depends only on . Applying Exercise 27, we conclude that there exist non-zero lattice vectors and smooth functions for some such that

for all , and furthermore (from (99) the chain rule) we have the derivative estimates

for . Setting , we thus have

and from the Leibniz rule and (98) we have

Let be the disjoint tubes in from the previous section, with width rather than . Construct the functions as in the previous section, and again set

Then as before, each is divergence free, and obeys the identities (65), (106) and the counds

for all and . As in the preceding section, we then set

and one easily checks that is a simplified fast-slow solution supported in . Direct calculation using the Leibniz rule and (105), (107) then gives the bounds

As before, is “high frequency” (mean zero in the variable). Also, is supported on the set , and for large enough the latter set has measure (say) . Thus by Cauchy-Schwarz (in just the variable) one has

The divergence corrector can be applied without difficulty:

Exercise 35Show that there is a simplified fast-slow solution supported in obeying the estimates

The crucial thing here is the tiny gain in the third estimate, with the first factor coming from a “slow” derivative and the second factor coming from essentially inverting a “fast” derivative .

Finally, we apply a stress corrector:

Exercise 36Show that there is a simplified fast-slow solution supported in obeying the estimates

Again, we have a crucial gain of coming from applying a slow derivative and inverting a fast one.

Since

(with implied constant in the exponent uniform in ) and , we see (for small enough) that

and the desired estimates (92), (93), (95), (96), (97) now follow.

** — 5. Constructing low regularity weak solutions to Euler — **

Throughout this section, we specialise to the Euler equations in the three-dimensional case (although all of the arguments here also apply without much modification to as well). In this section we establish an analogue of Corollary 28:

Proposition 37 (Low regularity non-trivial weak solutions)There exists a periodic weak solution to Euler which equals zero at time , but is not identically zero.

This result was first established by de Lellis and Szekelyhidi. Our approach will deviate from the one in that paper in a number of technical respects (for instance, we use Mikado flows in place of Beltrami flows, and we place more emphasis on the method of fast and slow variables). A key new feature, which was not present in the high-dimensional Sobolev-scale setting, is that the material derivative term in the difference Euler-Reynolds equations is no longer negligible, and needs to be treated by working with an ansatz in Lagrangian coordinates (or equivalently, an ansatz transported by the flow). (This use of Lagrangian coordinates is implicit in the thesis of Isett, this paper of de Lellis and Szekelyhidi, and in the later work of Isett.)

Just as Corollary 28 was derived from Proposition 26, the above proposition may be derived from

Proposition 38 (Weak improvement of Euler-Reynolds flows)Let be an Euler-Reynolds flow supported on a strict compact subinterval . Let be another interval in containing in its interior. Then for sufficiently large , there exists a Euler-Reynolds flow supported in obeying the estimates

where are smooth functions obeying the bounds

The point of the decomposition (115) is that it (together with the smallness bounds (116)) asserts that the velocity correction is mostly “high frequency” in nature, in that its low frequency components are small. Together with (112), the bounds roughly speaking assert that it is only the frequency components of that can be large in norm. Unlike the previous estimates, it will be important for our arguments that is supported in a strict subinterval of , because we will not be able to extend Lagrangian coordinates periodically around the circle. Actually the long-time instability of Lagrangian coordinates causes significant technical difficulties to overcome when one wants to construct solutions in higher regularity Hölder spaces , and in particular for close to ; we discuss this in the next section.

Exercise 39Deduce Proposition 37 from Proposition 38. (The decomposition (116) is needed to keep close to in a very weak topology – basically the topology – but one which is still sufficent to ensure that the limiting solution constructed is not identically zero.)

We now begin the proof of Proposition 38, repeating many of the steps used to prove Proposition 26. As before we may assume that is non-zero, and that is supported in . We can assume that is also a strict subinterval of .

Assume is sufficiently large; by rounding we may assume that is a natural number. Using the ansatz

and the triangle inequality, it suffices to construct a difference Euler-Reynolds flow at supported on and obeying the bounds

for all , and for which we have a decomposition obeying (116).

As before, we permit ourselves some error:

Exercise 40Show that it suffices to construct anapproximatedifference Euler-Reynolds flow at supported on and obeying the bounds

for , and for which we have a decomposition obeying (116).

It still remains to construct the approximate difference Euler-Reynolds flow obeying the claimed estimates. By definition, has to obey the system of equations

As is divergence-free, the first equation (122) may be rewritten as

where is the material Lie derivative of , thus

The lower order terms in (126) will turn out to be rather harmless; the main new difficulty is dealing with the material Lie derivative term . We will therefore invoke Lagrangian coordinates in order to convert the material Lie derivative into the more tractable time derivative (at the cost of mildly complicating all the other terms in the system).

We introduce a “Lagrangian torus” that is an isomorphic copy of the Eulerian torus ; as in the previous section, we paramterise this torus by , and adopt the usual summation conventions for the indices . Let be a trajectory map for , that is to say a smooth map such that for every time , the map is a diffeomorphism and one obeys the ODE

for all . The existence of such a trajectory map is guaranteed by the Picard existence theorem (it is important here that is not all of the torus ); see also Exercise 1 from Notes 1. From (the periodic version of) Lemma 3 of Notes 1, we can ensure that the map is volume-preserving, thus

Recall from Notes 1 that

- (i) Any Eulerian scalar field on can be pulled back to a Lagrangian scalar field on by the formula
- (ii) Any Eulerian vector field on can be pulled back to a Lagrangian vector field on by the formula
where is the inverse of the matrix , defined by

and

- (iii) Any Eulerian rank tensor on , can be pulled back to a Lagrangian rank tensor on by the formula

(One can pull back other tensors also, but these are the only ones we will need here.) Each of these pullback operations may be inverted by the corresponding pullback operation for the labels map (also known as *pushforward* by ). One can compute how these pullbacks interact with divergences:

Exercise 41 (Pullback and divergence)

- (i) If is a smooth Eulerian vector field, show that the pullback of the divergence of equals the divergence of the pullback of :
(

Hint:you will need to use the fact that is volume-preserving. Similarly to Lemma 3 and Exercise 4 of Notes 1, one can establish this either using the laws of integration or the laws of differentiation.).- (ii) Show that there exist smooth functions for with the following property: for any smooth Eulerian rank tensor on , with divergence , one has
(In fact, can be given explicitly as , where is the Kronecker delta and is the Christoffel symbol associated with the pullback of the Euclidean metric – but we will not need this precise formula. The right-hand side may also be written (in Penrose abstract index notation) as , where is the covariant derivative associated to .)

As remarked upon in the exercise, these calculations can be streamlined using the theory of the covariant derivative in Riemannian geometry; we will not develop this theory further here, but see for instance these two blog posts.

If one now applies the pullback operation to the system (126), (123), (124), (125) (and uses Exercise 16 from Notes 1 to convert the material Lie derivative into the ordinary time derivative) one obtain the equivalent system

where denotes the rank tensor

Thus, if one introduces the Lagrangian fields

and also

then (from many applications of the chain rule) we see that our task has now transformed to that of obtaining a supported on obeying the equations

for , where denotes the supremum on the Lagrangian spacetime . (In (137) we have to move back to Eulerian coordinates because the coefficients in the pushforward depend on , and we want an estimate here uniform in .)

This problem looks complicated, but the net effect of moving to the Lagrangian formulation is to arrive at a problem that is nearly identical to the Eulerian one, but in which the material Lie derivative has been replaced by the ordinary time derivative , and several lower order terms with smooth variable coefficients have been added to the system.

Now that the dangerous transport term in the material Lie derivative has been eliminated, it is now safe to use the method of fast-slow variables, but now on the Lagrangian torus rather than the Eulerian torus . We now parameterise the fast torus by (thus we think of now as a “Lagrangian fast torus” rather than a “Eulerian fast torus”) and use the ansatz

so that the equations of motion now become

where we think of as “low frequency” functions of time and the slow Lagrangian variable only (thus they are independent of the fast Lagrangian variable ). Set

It will now suffice to find a smooth solution to the above system supported on obeying the estimates

and obeying the pointwise estimate

If is chosen to be “high frequency” (mean zero in the fast variable ), then we can automatically obtain the estimate (148), as one may obtain the decomposition (142) with

and

at which point the estimates (148) follow from (144). Thus we may drop (148) and (142) from our requirements as long as we instead require to be high frequency.

As in previous sections, we can relax the conditions on and :

Exercise 42

- (i) (Stress corrector) Show that we may replace the condition (147) with the condition that
for all , and that is of high frequency (mean zero in ). (

Hint:add a corrector of size to , so that the main term now cancels off , and the other terms created by the correction are of size and of mean zero in . Then iterate.)- (ii) (Divergence corrector) After performing the modifications in (i), show that we may replace the condition (146) with the condition that
for all , and that has mean zero in . (Note that correcting for will modify by , but this will make a negligible contribution to (149) for large enough.)

We can now repeat the Mikado flow construction from previous sections:

Exercise 43Set . Construct a smooth vector field supported on , with of mean zero in , obeying the equationsand such that

is of mean zero in , obeying the bounds (144) for all . Also show that for a sufficiently large absolute constant not depending on , one can ensure that the matrix with entries

If we now set

one can verify that the equations (139), (140), (141) hold, and that have mean zero in . Furthermore, by pushing forward (150) by , we conclude that the matrix with entries

is positive semi-definite for all ; taking traces one concludes (149). Thus we have obtained all the properties required in Exercise 43, concluding the proof of Proposition 38.

** — 6. Constructing high regularity weak solutions to Euler — **

We now informally discuss how to modify the arguments above to establish the negative direction (ii) of Onsager’s conjecture. The full details are rather complicated (and arranged slightly differently from the presentation here), and we refer to Isett’s original paper for details. See also a subsequent paper of Buckmaster, De Lellis, Szekelyhidi, and Vicol for a simplified argument establishing this statement (as well as some additional strengthenings of it).

Let , let be a small quantity, and let be a large integer. The main iterative step, analogous to Theorem 30, roughly speaking (ignoring some technical logarithmic factors) takes an Euler-Reynolds flow obeying the estimates that look like

for some sufficiently large , and obtains a new Euler-Reynolds flow close to that obeys the estimates

where ; see Lemma 2.1 of Isett’s original paper for a more precise statement (in a slightly different notation), which also includes various estimates on the difference that we omit here. In contrast to previous arguments, it is useful for technical reasons to not impose time regularity in the estimates. Once this claim is formalised and proved, conclusions such as Onsager’s conjecture follow from the usual iteration arguments.

To achieve this iteration step, the first step is a mollification step analogous to Proposition 31 in which one perturbs the initial flow to obtain additional spatial regularity on the solution. Roughly speaking, this mollification allows one to replace the purely spatial differential operator appearing in (151), (153) with for much larger than (in practice there are some slight additional losses, which we will ignore here).

Now one has to solve the difference equation. We focus on the equation (126) and omit the small errors involving . Suppose for the time being that we could magically replace the material Lie derivative here by an ordinary time derivative, thus we would be trying to construct solving an equation such as

As before, we can use the method of fast and slow variables to construct a of amplitude roughly , oscillating at frequency , such that is high frequency (mean zero in the fast variable) and has amplitude about . We can also arrange matters (using something like (65)) so that the fast derivative component of vanishes, leaving only a slower derivative of size about . This makes the expression of magnitude about and oscillating at frequency about , which can be cancelled by a stress corrector in of magnitude about . Such a term would be acceptable (smaller than ) for as large as , in the spirit of Theorem 29.

However, one also has the terms and , which (in contrast to the high-dimensional Sobolev scale setting) cannot be ignored in this low-dimensional Hölder scale problem. The natural time scale of oscillation here is , coming from the usual heuristics concerning the Euler equation (see Remark 11 from 254A Notes 3). With this heuristic, and should both behave like , these expressions would be expected to have amplitude . They still oscillate at the high frequency , though, and lead to a stress corrector of magnitude about . This however remains acceptable for up to , which in principle resolves Onsager’s conjecture.

Now we have to remove the “cheat” of replacing the material Lie derivative by the ordinary time derivative. As we saw in the previous section, the natural way to fix this is to work in Lagrangian coordinates. However, we encounter a new problem: if one initialises the trajectory flow map to be the identity at some given time , then it turns out that one only gets good control on the flow map and its derivatives for times within the natural time scale of that initial time ; beyond this, the Gronwall-type arguments used to obtain bounds start to deteriorate exponentially. Because of this, one cannot rely on a “global” Lagrangian coordinate system as in the previous section. To get around this, one needs to partition the time domain into intervals of length about , and construct a separate trajectory map adapted to each such interval. One can then use these “local Lagrangian coordinates” to construct local components of the velocity perturbation that obey the required properties on each such interval. This construction is essentially the content of the “convex integration lemma” in Lemma 3.3 of Isett’s paper.

However, a new problem arises when trying to “glue” these local corrections together: two consecutive time intervals will overlap, and their corresponding local corrections will also overlap. This leads to some highly undesirable interaction terms between and (such as the fast derivative of ) which are very difficult to make small (for instance, one cannot simply ensure that have disjoint spatial supports as they are constructed using different local Lagrangian coordinate systems). On the other hand, if the original Reynolds stress had a special structure, namely that it was only supported on every other interval (i.e., on the for all even, or the for all odd), then these interactions no longer occur and the iteration step can proceed.

One could try to then resolve the problem by correcting the odd and even interval components of the stress in separate stages (cf. how Theorem 19 can be iterated to establish Theorem 15), but this is inefficient with regards to the parameter, in particular this makes the argument stop well short of the optimal threshold. To attain this threshold one needs the final ingredient of Isett’s argument, namely a “gluing approximation” (see Lemma 3.2 of Isett’s paper), in which one tales the (mollified) initial Euler-Reynolds flow and replaces it with a nearby flow in which the new Reynolds stress is only supported on every other interval . Combining this gluing approximation lemma with the mollification lemma and convex integration lemma gives the required iteration step. (One technical point is that this gluing has to create some useful additional regularity along the material derivative, in the spirit of Remark 38 of Notes 1, as such regularity will be needed in order to justify the convex integration step.)

To obtain this gluing approximation, one takes an -separated sequence of times , and for each such time , one solves the *true* Euler equations with initial data at to obtain smooth Euler solutions on a time interval centred at of lifespan , that agree with at time . (It is non-trivial to check that these solutions even exist on such an interval, let alone obey good estimates, but this can be done if the initial data was suitably mollified, as is consistent with the heuristics in Remark 11 from 254A Notes 3.) One can then glue these solutions together around the reference solution by defining

for a suitable partition of unity in time. This gives fields that solve the Euler equation near each time . To finish the proof of the gluing approximation lemma, one needs to then find a matching Reynolds stress for the intervals at which the Euler equation is not solved exactly. Isett’s original construction of this stress was rather intricate; see Sections 7-10 of Isett’s paper for the technical details. However, with improved estimates, a simpler construction was used in a subsequent paper of Buckmaster, De Lellis, Szekelyhidi, and Vicol, leading to a simplified proof of (the non-endpoint version of) this direction of Onsager’s conjecture.

## 14 comments

Comments feed for this article

10 January, 2019 at 3:23 am

AnonymousThe formatting is messed up (equations showing as raw TeX instead of being rendered) in several places in the article. Btw if you don’t mind my asking, how do you compose these entries? Do you write them out on paper before transcribing to TeX? Or just type TeX directly? Thanks.

10 January, 2019 at 9:50 am

Terence TaoI didn’t see any raw TeX in the latest version (though there were some overflowing latex displays which I reformatted). It could have been a transient WordPress LaTeX rendering issue.

I use Luca Trevisan’s LateX to WordPress converter, see the bottom of https://terrytao.wordpress.com/about/ as well as http://lucatrevisan.wordpress.com/latex-to-wordpress/ .

10 January, 2019 at 11:18 am

AnonymousWhat a wonderful post. Thank you very much Terry

10 January, 2019 at 11:48 am

Auf AchseThank you very much for this very interesting post.

It is also worth checking out a beautiful recent survey on the developments towards Onsager‘s conjecture by De Lellis and Székelyhidi:

https://arxiv.org/pdf/1901.02318.pdf

10 January, 2019 at 4:20 pm

Camillo De LellisDear Terry,

thanks for the beautiful post!

I am not sure whether I have understood correctly your last paragraph that naive attempts to invert the divergence operator lead to poor estimates. But it might be inaccurate, because in this note we do invert the divergence in the most “obvious way”, through an elliptic operator, after the glueing (cf. Section 4):

http://www.math.uzh.ch/fileadmin/user/delellis/publikation/Onsager96.pdf

Best,

Camillo

10 January, 2019 at 7:39 pm

Terence TaoThanks for this reference, which I had somehow missed!

My discussion was based on that in the start of Section 7 of Isett’s paper https://arxiv.org/pdf/1608.08301.pdf where he says the best estimate he can get from naively inverting the divergence is estimate (50) of that paper which is unsatisfactory. I’m not sure how to reconcile this statement with the estimates in Proposition 4.3 of your paper. Is the point that the commutator estimates with the advective derivative provide a better estimate on the “obvious” choice of antidivergence than what is described in Section 7 of Isett’s paper?

11 January, 2019 at 1:46 am

Camillo De LellisYes, the point is precisely that commuting the antidivergence with the advective derivative gives a better estimate. In a first draft of our proof we did have a lemma where we commuted the antidivergence and the advective derivative. Then we discovered that the proof becomes very elementary and simple if one instead commutes the Biot-Savart operator. So, in the version I linked above, the key is actually Proposition 3.4.

If I remember correctly (I have a bad connection at the moment and I don’t find a copy of Phil’s paper in my laptop), the improvement compared to Phil’s estimate (50) is, in our notation, the factor in our estimate (3.16). The relevant computations on the commutator are between eq. (3.19) and eq. (3.21).

11 January, 2019 at 9:11 am

Terence TaoThanks for the clarification! The construction and estimation of the antidivergence was the one part of Isett’s paper that I didn’t delve into in detail as it seemed particularly technical; it’s good to know that in fact one can avoid much of this technicality. I have updated the lecture notes accordingly.

16 January, 2019 at 10:16 pm

Phil IsettAbout estimating the anti-divergence, one can say a little more about why it works. Commuting with the advective derivative produces a suitable bound for the anti-divergence only if one assumes a priori that the anti-divergence itself already obeys a suitable bound. To get this argument to close depends on the structure of the evolution equation obeyed by the anti-divergence; it is a dynamical estimate that is stronger than the bound one would get by trying to estimate the anti-divergence directly. A similar statement applies to the linearized Euler term that appears in this evolution equation; the direct bound is not sharp enough and the term has to be treated by a dynamical argument.

About improving the timescale: It is possible to improve the timescale of the gluing approximation by using a combination of the approach of Camillo and coauthors (Buckmaster et al.) and my own. Doing so helps to obtain the best known estimate towards the endpoint case of Onsager’s conjecture. The way this improvement works is explained in the last section of my endpoint paper:

https://arxiv.org/abs/1706.01549

Note that obtaining any result of endpoint type requires avoiding power losses in frequency during the iteration, so the strategy I developed in my original proof of Onsager’s conjecture is important for this application because it avoids such power losses.

14 January, 2019 at 9:55 am

ChungachangaI’m an alien from nearby galaxy, and I wonder how stupid Terence Tao is. The problems he is trying to solve are actualy our elementary school homework.

17 January, 2019 at 1:34 am

terminology: equation collapse | Physics Forums[…] happens. Another term are regular and abnormal solutions, I think. At least Tao used them here: https://terrytao.wordpress.com/2019/01/08/255b-notes-2-onsagers-conjecture/ fresh_42, Jan 17, 2019 at 3:34 […]

17 January, 2019 at 5:10 am

JavohirMr. Terence Tao

What kind of email can I send you?

18 January, 2019 at 3:47 am

JavohirDear prof. Tao

I have a simple question in math. How can I send you an email?

18 January, 2019 at 6:39 am

MohdHi I learnt from your math thank you for your teaching moreover I am always looking forward for more mathematics journal entries.thank you sir