You are currently browsing the tag archive for the ‘pseudodifferential operators’ tag.

In contrast to previous notes, in this set of notes we shall focus exclusively on Fourier analysis in the one-dimensional setting ${d=1}$ for simplicity of notation, although all of the results here have natural extensions to higher dimensions. Depending on the physical context, one can view the physical domain ${{\bf R}}$ as representing either space or time; we will mostly think in terms of the former interpretation, even though the standard terminology of “time-frequency analysis”, which we will make more prominent use of in later notes, clearly originates from the latter.

In previous notes we have often performed various localisations in either physical space or Fourier space ${{\bf R}}$, for instance in order to take advantage of the uncertainty principle. One can formalise these operations in terms of the functional calculus of two basic operations on Schwartz functions ${{\mathcal S}({\bf R})}$, the position operator ${X: {\mathcal S}({\bf R}) \rightarrow {\mathcal S}({\bf R})}$ defined by

$\displaystyle (Xf)(x) := x f(x)$

and the momentum operator ${D: {\mathcal S}({\bf R}) \rightarrow {\mathcal S}({\bf R})}$, defined by

$\displaystyle (Df)(x) := \frac{1}{2\pi i} \frac{d}{dx} f(x). \ \ \ \ \ (1)$

(The terminology comes from quantum mechanics, where it is customary to also insert a small constant ${h}$ on the right-hand side of (1) in accordance with de Broglie’s law. Such a normalisation is also used in several branches of mathematics, most notably semiclassical analysis and microlocal analysis, where it becomes profitable to consider the semiclassical limit ${h \rightarrow 0}$, but we will not emphasise this perspective here.) The momentum operator can be viewed as the counterpart to the position operator, but in frequency space instead of physical space, since we have the standard identity

$\displaystyle \widehat{Df}(\xi) = \xi \hat f(\xi)$

for any ${\xi \in {\bf R}}$ and ${f \in {\mathcal S}({\bf R})}$. We observe that both operators ${X,D}$ are formally self-adjoint in the sense that

$\displaystyle \langle Xf, g \rangle = \langle f, Xg \rangle; \quad \langle Df, g \rangle = \langle f, Dg \rangle$

for all ${f,g \in {\mathcal S}({\bf R})}$, where we use the ${L^2({\bf R})}$ Hermitian inner product

$\displaystyle \langle f, g\rangle := \int_{\bf R} f(x) \overline{g(x)}\ dx.$

Clearly, for any polynomial ${P(x)}$ of one real variable ${x}$ (with complex coefficients), the operator ${P(X): {\mathcal S}({\bf R}) \rightarrow {\mathcal S}({\bf R})}$ is given by the spatial multiplier operator

$\displaystyle (P(X) f)(x) = P(x) f(x)$

and similarly the operator ${P(D): {\mathcal S}({\bf R}) \rightarrow {\mathcal S}({\bf R})}$ is given by the Fourier multiplier operator

$\displaystyle \widehat{P(D) f}(\xi) = P(\xi) \hat f(\xi).$

Inspired by this, if ${m: {\bf R} \rightarrow {\bf C}}$ is any smooth function that obeys the derivative bounds

$\displaystyle \frac{d^j}{dx^j} m(x) \lesssim_{m,j} \langle x \rangle^{O_{m,j}(1)} \ \ \ \ \ (2)$

for all ${j \geq 0}$ and ${x \in {\bf R}}$ (that is to say, all derivatives of ${m}$ grow at most polynomially), then we can define the spatial multiplier operator ${m(X): {\mathcal S}({\bf R}) \rightarrow {\mathcal S}({\bf R})}$ by the formula

$\displaystyle (m(X) f)(x) := m(x) f(x);$

one can easily verify from several applications of the Leibniz rule that ${m(X)}$ maps Schwartz functions to Schwartz functions. We refer to ${m(x)}$ as the symbol of this spatial multiplier operator. In a similar fashion, we define the Fourier multiplier operator ${m(D)}$ associated to the symbol ${m(\xi)}$ by the formula

$\displaystyle \widehat{m(D) f}(\xi) := m(\xi) \hat f(\xi).$

For instance, any constant coefficient linear differential operators ${\sum_{k=0}^n c_k \frac{d^k}{dx^k}}$ can be written in this notation as

$\displaystyle \sum_{k=0}^n c_k \frac{d^k}{dx^k} =\sum_{k=0}^n c_k (2\pi i D)^k;$

however there are many Fourier multiplier operators that are not of this form, such as fractional derivative operators ${\langle D \rangle^s = (1- \frac{1}{4\pi^2} \frac{d^2}{dx^2})^{s/2}}$ for non-integer values of ${s}$, which is a Fourier multiplier operator with symbol ${\langle \xi \rangle^s}$. It is also very common to use spatial cutoffs ${\psi(X)}$ and Fourier cutoffs ${\psi(D)}$ for various bump functions ${\psi}$ to localise functions in either space or frequency; we have seen several examples of such cutoffs in action in previous notes (often in the higher dimensional setting ${d>1}$).

We observe that the maps ${m \mapsto m(X)}$ and ${m \mapsto m(D)}$ are ring homomorphisms, thus for instance

$\displaystyle (m_1 + m_2)(D) = m_1(D) + m_2(D)$

and

$\displaystyle (m_1 m_2)(D) = m_1(D) m_2(D)$

for any ${m_1,m_2}$ obeying the derivative bounds (2); also ${m(D)}$ is formally adjoint to ${\overline{m(D)}}$ in the sense that

$\displaystyle \langle m(D) f, g \rangle = \langle f, \overline{m}(D) g \rangle$

for ${f,g \in {\mathcal S}({\bf R})}$, and similarly for ${m(X)}$ and ${\overline{m}(X)}$. One can interpret these facts as part of the functional calculus of the operators ${X,D}$, which can be interpreted as densely defined self-adjoint operators on ${L^2({\bf R})}$. However, in this set of notes we will not develop the spectral theory necessary in order to fully set out this functional calculus rigorously.

In the field of PDE and ODE, it is also very common to study variable coefficient linear differential operators

$\displaystyle \sum_{k=0}^n c_k(x) \frac{d^k}{dx^k} \ \ \ \ \ (3)$

where the ${c_0,\dots,c_n}$ are now functions of the spatial variable ${x}$ obeying the derivative bounds (2). A simple example is the quantum harmonic oscillator Hamiltonian ${-\frac{d^2}{dx^2} + x^2}$. One can rewrite this operator in our notation as

$\displaystyle \sum_{k=0}^n c_k(X) (2\pi i D)^k$

and so it is natural to interpret this operator as a combination ${a(X,D)}$ of both the position operator ${X}$ and the momentum operator ${D}$, where the symbol ${a: {\bf R} \times {\bf R} \rightarrow {\bf C}}$ this operator is the function

$\displaystyle a(x,\xi) := \sum_{k=0}^n c_k(x) (2\pi i \xi)^k. \ \ \ \ \ (4)$

Indeed, from the Fourier inversion formula

$\displaystyle f(x) = \int_{\bf R} \hat f(\xi) e^{2\pi i x \xi}\ d\xi$

for any ${f \in {\mathcal S}({\bf R})}$ we have

$\displaystyle (2\pi i D)^k f(x) = \int_{\bf R} (2\pi i \xi)^k \hat f(\xi) e^{2\pi i x \xi}\ d\xi$

and hence on multiplying by ${c_k(x)}$ and summing we have

$\displaystyle (\sum_{k=0}^n c_k(X) (2\pi i D)^k) f(x) = \int_{\bf R} a(x,\xi) \hat f(\xi) e^{2\pi i x \xi}\ d\xi.$

Inspired by this, we can introduce the Kohn-Nirenberg quantisation by defining the operator ${a(X,D) = a_{KN}(X,D): {\mathcal S}({\bf R}) \rightarrow {\mathcal S}({\bf R})}$ by the formula

$\displaystyle a(X,D) f(x) = \int_{\bf R} a(x,\xi) \hat f(\xi) e^{2\pi i x \xi}\ d\xi \ \ \ \ \ (5)$

whenever ${f \in {\mathcal S}({\bf R})}$ and ${a: {\bf R} \times {\bf R} \rightarrow {\bf C}}$ is any smooth function obeying the derivative bounds

$\displaystyle \frac{\partial^j}{\partial x^j} \frac{\partial^l}{\partial \xi^l} a(x,\xi) \lesssim_{a,j,l} \langle x \rangle^{O_{a,j}(1)} \langle \xi \rangle^{O_{a,j,l}(1)} \ \ \ \ \ (6)$

for all ${j,l \geq 0}$ and ${x \in {\bf R}}$ (note carefully that the exponent in ${x}$ on the right-hand side is required to be uniform in ${l}$). This quantisation clearly generalises both the spatial multiplier operators ${m(X)}$ and the Fourier multiplier operators ${m(D)}$ defined earlier, which correspond to the cases when the symbol ${a(x,\xi)}$ is a function of ${x}$ only or ${\xi}$ only respectively. Thus we have combined the physical space ${{\bf R} = \{ x: x \in {\bf R}\}}$ and the frequency space ${{\bf R} = \{ \xi: \xi \in {\bf R}\}}$ into a single domain, known as phase space ${{\bf R} \times {\bf R} = \{ (x,\xi): x,\xi \in {\bf R} \}}$. The term “time-frequency analysis” encompasses analysis based on decompositions and other manipulations of phase space, in much the same way that “Fourier analysis” encompasses analysis based on decompositions and other manipulations of frequency space. We remark that the Kohn-Nirenberg quantization is not the only choice of quantization one could use; see Remark 19 below.

Exercise 1

• (i) Show that for ${a}$ obeying (6), that ${a(X,D)}$ does indeed map ${{\mathcal S}({\bf R})}$ to ${{\mathcal S}({\bf R})}$.
• (ii) Show that the symbol ${a}$ is uniquely determined by the operator ${a(X,D)}$. That is to say, if ${a,b}$ are two functions obeying (6) with ${a(X,D) f = b(X,D) f}$ for all ${f \in {\mathcal S}({\bf R})}$, then ${a=b}$. (Hint: apply ${a(X,D)-b(X,D)}$ to a suitable truncation of a plane wave ${x \mapsto e^{2\pi i x \xi}}$ and then take limits.)

In principle, the quantisations ${a(X,D)}$ are potentially very useful for such tasks as inverting variable coefficient linear operators, or to localize a function simultaneously in physical and Fourier space. However, a fundamental difficulty arises: map from symbols ${a}$ to operators ${a(X,D)}$ is now no longer a ring homomorphism, in particular

$\displaystyle (a_1 a_2)(X,D) \neq a_1(X,D) a_2(X,D) \ \ \ \ \ (7)$

in general. Fundamentally, this is due to the fact that pointwise multiplication of symbols is a commutative operation, whereas the composition of operators such as ${X}$ and ${D}$ does not necessarily commute. This lack of commutativity can be measured by introducing the commutator

$\displaystyle [A,B] := AB - BA$

of two operators ${A,B}$, and noting from the product rule that

$\displaystyle [X,D] = -\frac{1}{2\pi i} \neq 0.$

(In the language of Lie groups and Lie algebras, this tells us that ${X,D}$ are (up to complex constants) the standard Lie algebra generators of the Heisenberg group.) From a quantum mechanical perspective, this lack of commutativity is the root cause of the uncertainty principle that prevents one from simultaneously localizing in both position and momentum past a certain point. Here is one basic way of formalising this principle:

Exercise 2 (Heisenberg uncertainty principle) For any ${x_0, \xi_0 \in {\bf R}}$ and ${f \in \mathcal{S}({\bf R})}$, show that

$\displaystyle \| (X-x_0) f \|_{L^2({\bf R})} \| (D-\xi_0) f\|_{L^2({\bf R})} \geq \frac{1}{4\pi} \|f\|_{L^2({\bf R})}^2.$

(Hint: evaluate the expression ${\langle [X-x_0, D - \xi_0] f, f \rangle}$ in two different ways and apply the Cauchy-Schwarz inequality.) Informally, this exercise asserts that the spatial uncertainty ${\Delta x}$ and the frequency uncertainty ${\Delta \xi}$ of a function obey the Heisenberg uncertainty relation ${\Delta x \Delta \xi \gtrsim 1}$.

Nevertheless, one still has the correspondence principle, which asserts that in certain regimes (which, with our choice of normalisations, corresponds to the high-frequency regime), quantum mechanics continues to behave like a commutative theory, and one can sometimes proceed as if the operators ${X,D}$ (and the various operators ${a(X,D)}$ constructed from them) commute up to “lower order” errors. This can be formalised using the pseudodifferential calculus, which we give below the fold, in which we restrict the symbol ${a}$ to certain “symbol classes” of various orders (which then restricts ${a(X,D)}$ to be pseudodifferential operators of various orders), and obtains approximate identities such as

$\displaystyle (a_1 a_2)(X,D) \approx a_1(X,D) a_2(X,D)$

where the error between the left and right-hand sides is of “lower order” and can in fact enjoys a useful asymptotic expansion. As a first approximation to this calculus, one can think of functions ${f \in {\mathcal S}({\bf R})}$ as having some sort of “phase space portrait${\tilde f(x,\xi)}$ which somehow combines the physical space representation ${x \mapsto f(x)}$ with its Fourier representation ${\xi \mapsto f(\xi)}$, and pseudodifferential operators ${a(X,D)}$ behave approximately like “phase space multiplier operators” in this representation in the sense that

$\displaystyle \widetilde{a(X,D) f}(x,\xi) \approx a(x,\xi) \tilde f(x,\xi).$

Unfortunately the uncertainty principle (or the non-commutativity of ${X}$ and ${D}$) prevents us from making these approximations perfectly precise, and it is not always clear how to even define a phase space portrait ${\tilde f}$ of a function ${f}$ precisely (although there are certain popular candidates for such a portrait, such as the FBI transform (also known as the Gabor transform in signal processing literature), or the Wigner quasiprobability distribution, each of which have some advantages and disadvantages). Nevertheless even if the concept of a phase space portrait is somewhat fuzzy, it is of great conceptual benefit both within mathematics and outside of it. For instance, the musical score one assigns a piece of music can be viewed as a phase space portrait of the sound waves generated by that music.

To complement the pseudodifferential calculus we have the basic Calderón-Vaillancourt theorem, which asserts that pseudodifferential operators of order zero are Calderón-Zygmund operators and thus bounded on ${L^p({\bf R})}$ for ${1 < p < \infty}$. The standard proof of this theorem is a classic application of one of the basic techniques in harmonic analysis, namely the exploitation of almost orthogonality; the proof we will give here will achieve this through the elegant device of the Cotlar-Stein lemma.

Pseudodifferential operators (especially when generalised to higher dimensions ${d \geq 1}$) are a fundamental tool in the theory of linear PDE, as well as related fields such as semiclassical analysis, microlocal analysis, and geometric quantisation. There is an even wider class of operators that is also of interest, namely the Fourier integral operators, which roughly speaking not only approximately multiply the phase space portrait ${\tilde f(x,\xi)}$ of a function by some multiplier ${a(x,\xi)}$, but also move the portrait around by a canonical transformation. However, the development of theory of these operators is beyond the scope of these notes; see for instance the texts of Hormander or Eskin.

This set of notes is only the briefest introduction to the theory of pseudodifferential operators. Many texts are available that cover the theory in more detail, for instance this text of Taylor.

Lars Hörmander, who made fundamental contributions to all areas of partial differential equations, but particularly in developing the analysis of variable-coefficient linear PDE, died last Sunday, aged 81.

I unfortunately never met Hörmander personally, but of course I encountered his work all the time while working in PDE. One of his major contributions to the subject was to systematically develop the calculus of Fourier integral operators (FIOs), which are a substantial generalisation of pseudodifferential operators and which can be used to (approximately) solve linear partial differential equations, or to transform such equations into a more convenient form. Roughly speaking, Fourier integral operators are to linear PDE as canonical transformations are to Hamiltonian mechanics (and one can in fact view FIOs as a quantisation of a canonical transformation). They are a large class of transformations, for instance the Fourier transform, pseudodifferential operators, and smooth changes of the spatial variable are all examples of FIOs, and (as long as certain singular situations are avoided) the composition of two FIOs is again an FIO.

The full theory of FIOs is quite extensive, occupying the entire final volume of Hormander’s famous four-volume series “The Analysis of Linear Partial Differential Operators”. I am certainly not going to try to attempt to summarise it here, but I thought I would try to motivate how these operators arise when trying to transform functions. For simplicity we will work with functions ${f \in L^2({\bf R}^n)}$ on a Euclidean domain ${{\bf R}^n}$ (although FIOs can certainly be defined on more general smooth manifolds, and there is an extension of the theory that also works on manifolds with boundary). As this will be a heuristic discussion, we will ignore all the (technical, but important) issues of smoothness or convergence with regards to the functions, integrals and limits that appear below, and be rather vague with terms such as “decaying” or “concentrated”.

A function ${f \in L^2({\bf R}^n)}$ can be viewed from many different perspectives (reflecting the variety of bases, or approximate bases, that the Hilbert space ${L^2({\bf R}^n)}$ offers). Most directly, we have the physical space perspective, viewing ${f}$ as a function ${x \mapsto f(x)}$ of the physical variable ${x \in {\bf R}^n}$. In many cases, this function will be concentrated in some subregion ${\Omega}$ of physical space. For instance, a gaussian wave packet

$\displaystyle f(x) = A e^{-(x-x_0)^2/\hbar} e^{i \xi_0 \cdot x/\hbar}, \ \ \ \ \ (1)$

where ${\hbar > 0}$, ${A \in {\bf C}}$ and ${x_0, \xi_0 \in {\bf R}^n}$ are parameters, would be physically concentrated in the ball ${B(x_0,\sqrt{\hbar})}$. Then we have the frequency space (or momentum space) perspective, viewing ${f}$ now as a function ${\xi \mapsto \hat f(\xi)}$ of the frequency variable ${\xi \in {\bf R}^n}$. For this discussion, it will be convenient to normalise the Fourier transform using a small constant ${\hbar > 0}$ (which has the physical interpretation of Planck’s constant if one is doing quantum mechanics), thus

$\displaystyle \hat f(\xi) := \frac{1}{(2\pi \hbar)^{n/2}} \int_{\bf R} e^{-i\xi \cdot x/\hbar} f(x)\ dx.$

For instance, for the gaussian wave packet (1), one has

$\displaystyle \hat f(\xi) = A e^{i\xi_0 \cdot x_0/\hbar} e^{-(\xi-\xi_0)^2/\hbar} e^{-i \xi \cdot x_0/\hbar},$

and so we see that ${f}$ is concentrated in frequency space in the ball ${B(\xi_0,\sqrt{\hbar})}$.

However, there is a third (but less rigorous) way to view a function ${f}$ in ${L^2({\bf R}^n)}$, which is the phase space perspective in which one tries to view ${f}$ as distributed simultaneously in physical space and in frequency space, thus being something like a measure on the phase space ${T^* {\bf R}^n := \{ (x,\xi): x, \xi \in {\bf R}^n\}}$. Thus, for instance, the function (1) should heuristically be concentrated on the region ${B(x_0,\sqrt{\hbar}) \times B(\xi_0,\sqrt{\hbar})}$ in phase space. Unfortunately, due to the uncertainty principle, there is no completely satisfactory way to canonically and rigorously define what the “phase space portrait” of a function ${f}$ should be. (For instance, the Wigner transform of ${f}$ can be viewed as an attempt to describe the distribution of the ${L^2}$ energy of ${f}$ in phase space, except that this transform can take negative or even complex values; see Folland’s book for further discussion.) Still, it is a very useful heuristic to think of functions has having a phase space portrait, which is something like a non-negative measure on phase space that captures the distribution of functions in both space and frequency, albeit with some “quantum fuzziness” that shows up whenever one tries to inspect this measure at scales of physical space and frequency space that together violate the uncertainty principle. (The score of a piece of music is a good everyday example of a phase space portrait of a function, in this case a sound wave; here, the physical space is the time axis (the horizontal dimension of the score) and the frequency space is the vertical dimension. Here, the time and frequency scales involved are well above the uncertainty principle limit (a typical note lasts many hundreds of cycles, whereas the uncertainty principle kicks in at ${O(1)}$ cycles) and so there is no obstruction here to musical notation being unambiguous.) Furthermore, if one takes certain asymptotic limits, one can recover a precise notion of a phase space portrait; for instance if one takes the semiclassical limit ${\hbar \rightarrow 0}$ then, under certain circumstances, the phase space portrait converges to a well-defined classical probability measure on phase space; closely related to this is the high frequency limit of a fixed function, which among other things defines the wave front set of that function, which can be viewed as another asymptotic realisation of the phase space portrait concept.

If functions in ${L^2({\bf R}^n)}$ can be viewed as a sort of distribution in phase space, then linear operators ${T: L^2({\bf R}^n) \rightarrow L^2({\bf R}^n)}$ should be viewed as various transformations on such distributions on phase space. For instance, a pseudodifferential operator ${a(X,D)}$ should correspond (as a zeroth approximation) to multiplying a phase space distribution by the symbol ${a(x,\xi)}$ of that operator, as discussed in this previous blog post. Note that such operators only change the amplitude of the phase space distribution, but not the support of that distribution.

Now we turn to operators that alter the support of a phase space distribution, rather than the amplitude; we will focus on unitary operators to emphasise the amplitude preservation aspect. These will eventually be key examples of Fourier integral operators. A physical translation ${Tf(x) := f(x-x_0)}$ should correspond to pushing forward the distribution by the transformation ${(x,\xi) \mapsto (x+x_0,\xi)}$, as can be seen by comparing the physical and frequency space supports of ${Tf}$ with that of ${f}$. Similarly, a frequency modulation ${Tf(x) := e^{i \xi_0 \cdot x/\hbar} f(x)}$ should correspond to the transformation ${(x,\xi) \mapsto (x,\xi+\xi_0)}$; a linear change of variables ${Tf(x) := |\hbox{det} L|^{-1/2} f(L^{-1} x)}$, where ${L: {\bf R}^n \rightarrow {\bf R}^n}$ is an invertible linear transformation, should correspond to ${(x,\xi) \mapsto (Lx, (L^*)^{-1} \xi)}$; and finally, the Fourier transform ${Tf(x) := \hat f(x)}$ should correspond to the transformation ${(x,\xi) \mapsto (\xi,-x)}$.

Based on these examples, one may hope that given any diffeomorphism ${\Phi: T^* {\bf R}^n \rightarrow T^* {\bf R}^n}$ of phase space, one could associate some sort of unitary (or approximately unitary) operator ${T_\Phi: L^2({\bf R}^n) \rightarrow L^2({\bf R}^n)}$, which (heuristically, at least) pushes the phase space portrait of a function forward by ${\Phi}$. However, there is an obstruction to doing so, which can be explained as follows. If ${T_\Phi}$ pushes phase space portraits by ${\Phi}$, and pseudodifferential operators ${a(X,D)}$ multiply phase space portraits by ${a}$, then this suggests the intertwining relationship

$\displaystyle a(X,D) T_\Phi \approx T_\Phi (a \circ \Phi)(X,D),$

and thus ${(a \circ \Phi)(X,D)}$ is approximately conjugate to ${a(X,D)}$:

$\displaystyle (a \circ \Phi)(X,D) \approx T_\Phi^{-1} a(X,D) T_\Phi. \ \ \ \ \ (2)$

The formalisation of this fact in the theory of Fourier integral operators is known as Egorov’s theorem, due to Yu Egorov (and not to be confused with the more widely known theorem of Dmitri Egorov in measure theory).

Applying commutators, we conclude the approximate conjugacy relationship

$\displaystyle \frac{1}{i\hbar} [(a \circ \Phi)(X,D), (b \circ \Phi)(X,D)] \approx T_\Phi^{-1} \frac{1}{i\hbar} [a(X,D), b(X,D)] T_\Phi.$

Now, the pseudodifferential calculus (as discussed in this previous post) tells us (heuristically, at least) that

$\displaystyle \frac{1}{i\hbar} [a(X,D), b(X,D)] \approx \{ a, b \}(X,D)$

and

$\displaystyle \frac{1}{i\hbar} [(a \circ \Phi)(X,D), (b \circ \Phi)(X,D)] \approx \{ a \circ \Phi, b \circ \Phi \}(X,D)$

where ${\{,\}}$ is the Poisson bracket. Comparing this with (2), we are then led to the compatibility condition

$\displaystyle \{ a \circ \Phi, b \circ \Phi \} \approx \{ a, b \} \circ \Phi,$

thus ${\Phi}$ needs to preserve (approximately, at least) the Poisson bracket, or equivalently ${\Phi}$ needs to be a symplectomorphism (again, approximately at least).

Now suppose that ${\Phi: T^* {\bf R}^n \rightarrow T^* {\bf R}^n}$ is a symplectomorphism. This is morally equivalent to the graph ${\Sigma := \{ (z, \Phi(z)): z \in T^* {\bf R}^n \}}$ being a Lagrangian submanifold of ${T^* {\bf R}^n \times T^* {\bf R}^n}$ (where we give the second copy of phase space the negative ${-\omega}$ of the usual symplectic form ${\omega}$, thus yielding ${\omega \oplus -\omega}$ as the full symplectic form on ${T^* {\bf R}^n \times T^* {\bf R}^n}$; this is another instantiation of the closed graph theorem, as mentioned in this previous post. This graph is known as the canonical relation for the (putative) FIO that is associated to ${\Phi}$. To understand what it means for this graph to be Lagrangian, we coordinatise ${T^* {\bf R}^n \times T^* {\bf R}^n}$ as ${(x,\xi,y,\eta)}$ suppose temporarily that this graph was (locally, at least) a smooth graph in the ${x}$ and ${y}$ variables, thus

$\displaystyle \Sigma = \{ (x, F(x,y), y, G(x,y)): x, y \in {\bf R}^n \}$

for some smooth functions ${F, G: {\bf R}^n \rightarrow {\bf R}^n}$. A brief computation shows that the Lagrangian property of ${\Sigma}$ is then equivalent to the compatibility conditions

$\displaystyle \frac{\partial F_i}{\partial x_j} = \frac{\partial F_j}{\partial x_i}$

$\displaystyle \frac{\partial G_i}{\partial y_j} = \frac{\partial G_j}{\partial y_i}$

$\displaystyle \frac{\partial F_i}{\partial y_j} = - \frac{\partial G_j}{\partial x_i}$

for ${i,j=1,\ldots,n}$, where ${F_1,\ldots,F_n, G_1,\ldots,G_n}$ denote the components of ${F,G}$. Some Fourier analysis (or Hodge theory) lets us solve these equations as

$\displaystyle F_i = -\frac{\partial \phi}{\partial x_i}; \quad G_j = \frac{\partial \phi}{\partial y_j}$

for some smooth potential function ${\phi: {\bf R}^n \times {\bf R}^n \rightarrow {\bf R}}$. Thus, we have parameterised our graph ${\Sigma}$ as

$\displaystyle \Sigma = \{ (x, -\nabla_x \phi(x,y), y, \nabla_y \phi(x,y)): x,y \in {\bf R}^n \} \ \ \ \ \ (3)$

so that ${\Phi}$ maps ${(x, -\nabla_x \phi(x,y))}$ to ${(y, \nabla_y \phi(x,y))}$.

A reasonable candidate for an operator associated to ${\Phi}$ and ${\Sigma}$ in this fashion is the oscillatory integral operator

$\displaystyle Tf(y) := \frac{1}{(2\pi \hbar)^{n/2}} \int_{{\bf R}^n} e^{i \phi(x,y)/\hbar} a(x,y) f(x)\ dx \ \ \ \ \ (4)$

for some smooth amplitude function ${a}$ (note that the Fourier transform is the special case when ${a=1}$ and ${\phi(x,y)=xy}$, which helps explain the genesis of the term “Fourier integral operator”). Indeed, if one computes an inner product ${\int_{{\bf R}^n} Tf(y) \overline{g(y)}\ dy}$ for gaussian wave packets ${f, g}$ of the form (1) and localised in phase space near ${(x_0,\xi_0), (y_0,\eta_0)}$ respectively, then a Taylor expansion of ${\phi}$ around ${(x_0,y_0)}$, followed by a stationary phase computation, shows (again heuristically, and assuming ${\phi}$ is suitably non-degenerate) that ${T}$ has (3) as its canonical relation. (Furthermore, a refinement of this stationary phase calculation suggests that if ${a}$ is normalised to be the half-density ${|\det \nabla_x \nabla_y \phi|^{1/2}}$, then ${T}$ should be approximately unitary.) As such, we view (4) as an example of a Fourier integral operator (assuming various smoothness and non-degeneracy hypotheses on the phase ${\phi}$ and amplitude ${a}$ which we do not detail here).

Of course, it may be the case that ${\Sigma}$ is not a graph in the ${x,y}$ coordinates (for instance, the key examples of translation, modulation, and dilation are not of this form), but then it is often a graph in some other pair of coordinates, such as ${\xi,y}$. In that case one can compose the oscillatory integral construction given above with a Fourier transform, giving another class of FIOs of the form

$\displaystyle Tf(y) := \frac{1}{(2\pi \hbar)^{n/2}} \int_{{\bf R}^n} e^{i \phi(\xi,y)/\hbar} a(\xi,y) \hat f(\xi)\ d\xi. \ \ \ \ \ (5)$

This class of FIOs covers many important cases; for instance, the translation, modulation, and dilation operators considered earlier can be written in this form after some Fourier analysis. Another typical example is the half-wave propagator ${T := e^{it \sqrt{-\Delta}}}$ for some time ${t \in {\bf R}}$, which can be written in the form

$\displaystyle Tf(y) = \frac{1}{(2\pi \hbar)^{n/2}} \int_{{\bf R}^n} e^{i (\xi \cdot y + t |\xi|)/\hbar} a(\xi,y) \hat f(\xi)\ d\xi.$

This corresponds to the phase space transformation ${(x,\xi) \mapsto (x+t\xi/|\xi|, \xi)}$, which can be viewed as the classical propagator associated to the “quantum” propagator ${e^{it\sqrt{-\Delta}}}$. More generally, propagators for linear Hamiltonian partial differential equations can often be expressed (at least approximately) by Fourier integral operators corresponding to the propagator of the associated classical Hamiltonian flow associated to the symbol of the Hamiltonian operator ${H}$; this leads to an important mathematical formalisation of the correspondence principle between quantum mechanics and classical mechanics, that is one of the foundations of microlocal analysis and which was extensively developed in Hörmander’s work. (More recently, numerically stable versions of this theory have been developed to allow for rapid and accurate numerical solutions to various linear PDE, for instance through Emmanuel Candés’ theory of curvelets, so the theory that Hörmander built now has some quite significant practical applications in areas such as geology.)

In some cases, the canonical relation ${\Sigma}$ may have some singularities (such as fold singularities) which prevent it from being written as graphs in the previous senses, but the theory for defining FIOs even in these cases, and in developing their calculus, is now well established, in large part due to the foundational work of Hörmander.