This is going to be a somewhat experimental post. In class, I mentioned that when solving the type of homework problems encountered in a graduate real analysis course, there are really only about a dozen or so basic tricks and techniques that are used over and over again. But I had not thought to actually try to make these tricks explicit, so I am going to try to compile here a list of some of these techniques here. But this list is going to be far from exhaustive; perhaps if other recent students of real analysis would like to share their own methods, then I encourage you to do so in the comments (even – or especially – if the techniques are somewhat vague and general in nature).

(See also the Tricki for some general mathematical problem solving tips. Once this page matures somewhat, I might migrate it to the Tricki.)

Note: the tricks occur here in no particular order, reflecting the stream-of-consciousness way in which they were arrived at. Indeed, this list will be extended on occasion whenever I find another trick that can be added to this list.

**1. Split up equalities into inequalities.**

If one has to show that two numerical quantities X and Y are equal, try proving that and separately. (Often one of these will be very easy, and the other one harder; but the easy direction may still provide some clue as to what needs to be done to establish the other direction.)

In a similar spirit, to show that two sets E and F are equal, try proving that and .

**2. Give yourself an epsilon of room.**

If one has to show that , try proving that for any . (This trick combines well with Trick 1.)

In a similar spirit, if one needs to show that a quantity vanishes, try showing that for every .

Or: if one wishes to show that two functions agree almost everywhere, try showing first that holds for almost every x, or even just outside of a set of measure at most , for any given .

Or: if one wants to show that a sequence of real numbers converges to zero, try showing that for every .

Don’t be too focused on getting all your error terms adding up to exactly – usually, as long as the final error bound consists of terms that can all be made as small as one wishes by choosing parameters in a suitable way, that is enough. For instance, an error term such as is certainly OK, or even more complicated expressions such as if one has the ability to choose as small as one wishes, and then after is chosen, one can then also set as small as one wishes (in a manner that can depend on ).

One caveat: for finite , and any , it is true that and , but this statement is not true when is equal to (or ). So remember to exercise some care with the epsilon of room trick when some quantities are infinite.

See also the Tricki article on this topic.

**3. Decompose or approximate a rough or general object by a smoother or simpler one.**

If one has to prove something about an unbounded (or infinite measure) set, consider proving it for bounded (or finite measure) sets first if this looks easier.

Similarly:

- if one has to prove something about a measurable set, try proving it for open, closed, compact, bounded, or elementary sets first.
- if one has to prove something about a measurable function, try proving it for functions that are continuous, bounded, compactly supported, simple, absolutely integrable, etc.
- if one has to prove something about an infinite sum or sequence, try proving it first for finite truncations of that sum or sequence (but try to get all the bounds independent of the number of terms in that truncation, so that you can still pass to the limit!).
- if one has to prove something about a complex-valued function, try it for real-valued functions first.
- If one has to prove something about a real-valued function, try it for unsigned functions first.
- If one has to prove something about a simple function, try it for indicator functions first.

In order to pass back to the general case from these special cases, one will have to somehow decompose the general object into a combination of special ones, or approximate general objects by special ones (or as a limit of a sequence of special objects). In the latter case, one may need an epsilon of room (trick 2), and some sort of limiting analysis may be needed to deal with the errors in the approximation (it is not always enough to just “pass to the limit”, as one has to justify that the desirable properties of the approximating object are preserved in the limit).

*Note*: one should not do this blindly, as one might then be loading on a bunch of distracting but ultimately useless hypotheses that end up being a lot less help than one might hope. But they should be kept in mind as something to try if one starts having thoughts such as “Gee, it would be nice at this point if I could assume that is continuous / real-valued / simple / unsigned / etc. “.

In the more quantitative areas of analysis and PDE, one sees a common variant of the above technique, namely the method of *a priori* estimates. Here, one needs to prove an estimate or inequality for all functions in a large, rough class (e.g. all rough solutions to a PDE). One can often then first prove this inequality in a much smaller (but still “dense”) class of “nice” functions, so that there is little difficulty justifying the various manipulations (e.g. exchanging integrals, sums, or limits, or integrating by parts) that one wishes to perform. Once one obtains these *a priori* estimates, one can then often take some sort of limiting argument to recover the general case.

**4. If one needs to flip an upper bound to a lower bound or vice versa, look for a way to take reflections or complements.**

Sometimes one needs a lower bound for some quantity, but only has techniques that give upper bounds. In some cases, though, one can “reflect” an upper bound into a lower bound (or vice versa) by replacing a set contained in some space with its complement , or a function with its negation (or perhaps subtracting from some dominating function to obtain ). This trick works best when the objects being reflected are contained in some sort of “bounded”, “finite measure”, or “absolutely integrable” container, so that one avoids having the dangerous situation of having to subtract infinite quantities from each other.

The triangle inequality

can be used to flip an upper bound on to a lower bound on , provided of course one has a lower bound on . Similarly, the Cauchy-Schwarz inequality

can flip a upper bound on to a lower bound on , provided that one has a lower bound on . Holder’s inequality can also be used in a similar fashion.

**5. Uncountable unions can sometimes be replaced by countable or finite unions.**

Uncountable unions are not well-behaved in measure theory; for instance, an uncountable union of null sets need not be a null set (or even a measurable set). (On the other hand, the uncountable union of *open* sets remains open; this can often be important to know.) However, in many cases one can replace an uncountable union by a countable one. For instance, if one needs to prove a statement for all , then there are an uncountable number of ‘s one needs to check, which may threaten measurability; but in many cases it is enough to only work with a countable sequence of s, such as the numbers for .

In a similar spirit, given a real parameter , this parameter initially ranges over uncountably many values, but in some cases one can get away with only working with a countable set of such values, such as the rationals. In a similar spirit, rather than work with all boxes (of which there are uncountably many), one might work with the dyadic boxes (of which there are only countably many; also, they obey nicer nesting properties than general boxes and so are often desirable to work with in any event).

If you are working on a compact set, then one can often replace even uncountable unions with finite ones, so long as one is working with open sets. When this option is available, it is often worth spending an epsilon of measure (or whatever other resource is available to spend) to make one’s sets open, just so that one can take advantage of compactness.

**6. If it is difficult to work globally, work locally instead.**

A domain such as Euclidean space has infinite measure, and this creates a number of technical difficulties when trying to do measure theory directly on such spaces. Sometimes it is best to work more locally, for instance working on a large ball or even a small ball such as first, and then figuring out how to patch things together later. Compactness (or the closely related property of total boundedness) is often useful for patching together small balls to cover a large ball. Patching together large balls into the whole space tends to work well when the properties one are trying to establish are local in nature (such as continuity, or pointwise convergence) or behave well with respect to countable unions. For instance, to prove that a sequence of functions converges pointwise almost everywhere to on , it suffices to verify this pointwise almost everywhere convergence on the ball for every (which one can take to be an integer to get countability, see Trick 5).

**7. Be willing to throw away an exceptional set.**

The “Lebesgue philosophy” to measure theory is that null sets are often “irrelevant”, and so one should be very willing to cut out a set of measure zero on which bad things are happening (e.g. a function is undefined or infinite, a sequence of functions is not converging, etc.). One should also be only slightly less willing to throw away sets of positive but small measure, e.g. sets of measure at most . If such sets can be made arbitrarily small in measure, this is often almost as good as just throwing away a null set.

Many things in measure theory improve after throwing away a small set. The most notable examples of this are Egorov’s theorem (pointwise a.e. convergence becomes locally uniform convergence after throwing away a small set) and Lusin’s theorem (measurable functions become continuous after throwing away a small set). A related (and simpler) principle is that a measurable function that is finite almost everywhere can become locally bounded after throwing away a small set (this can be seen by applying downward monotone convergence to exceptional sets such as as ). From Markov’s inequality, one also sees that a function that is small in norm becomes small uniformly after throwing away a small set. (The same is true for functions that are small in other norms, or (by definition) for functions that are converging to zero in measure.)

A little later in this course, we will also start seeing a similar trick of throwing away most of a sequence and working with a subsequence instead. (This trick can also be interpreted as “throwing away a small set”, but to understand what “small” means in this context, one needs the language of ultrafilters, which will not be discussed here.) See Trick 17 below.

**8. Draw pictures and try to build counterexamples.**

Measure theory, particularly on Euclidean spaces, has a significant geometric aspect to it, and you should be exploiting your geometric intuition. Drawing pictures and graphs of all the objects being studied is a good way to start. These pictures need not be completely realistic; they should be just complicated enough to hint at the complexities of the problem, but not more. (For instance, usually one- or two-dimensional pictures suffice for understanding problems in ; drawing intricate 3D (or 4D, etc.) pictures does not often make things simpler. To indicate that a function is not continuous, one or two discontinuities or oscillations might suffice; make it too ornate and it becomes less clear what to do about that function.)

A common mistake is to try to draw a picture in which both the hypotheses and conclusion of the problem hold. This is actually not all that useful, as it often does not reveal the causal relationship between the former and the latter. One should try instead to draw a picture in which the hypotheses hold but for which the conclusion does not – in other words, a *counterexample* to the problem. Of course, you should be expected to fail at this task, given that the statement of the problem is presumably true. However, the *way* in which your picture fails to accomplish this task is often very instructive, and can reveal vital clues as to how the solution to the problem is supposed to proceed.

**9. Try simpler cases first.**

This advice of course extends well beyond measure theory, but if one is completely stuck on a problem, try making the problem simpler (while still capturing at least one of the difficulties of the problem that you cannot currently resolve). For instance, if faced with a problem in , try the one-dimensional case first. Faced with a problem about a general measurable function , try a much simpler case first, such as an indicator function . Faced with a problem about a general measurable set, try an elementary set first. Faced with a problem about a sequence of functions, try a monotone sequence of functions first. And so forth. (Note that this trick overlaps quite a bit with Trick 3.)

The problem should not be made so simple that it becomes trivial, as this doesn’t really gain you any new insight about the original problem; instead, one should try to keep the “essential” difficulties of the problem while throwing away those aspects that you think are less important (but are still serving to add to the overall difficulty level).

On the other hand, if the simplified problem is unexpectedly easy, but one cannot extend the methods to the general case (or somehow leverage the simplified case to the general case, as in Trick 3), this is an indication that the true difficulty lies elsewhere. For instance, if a problem involving general functions could be solved easily for monotone functions, but one cannot then extend that argument to the general case, this suggests that the true enemy is oscillation, and perhaps one should try another simple case in which the function is allowed to be highly oscillatory (but perhaps simple in other ways, e.g. bounded with compact support).

**10. Abstract away any information that you believe or suspect to be irrelevant.**

Sometimes one is faced with an embarrassment of riches when it comes to what choice of technique to use on a problem; there are so many different facts that one knows about the problem, and so many different pieces of theory that one could apply, that one doesn’t quite know where to begin.

When this happens, *abstraction* can be a vital tool to clear away some of the conceptual clutter. Here, one wants to “forget” part of the setting that the problem is phrased in, and only keep the part that seems to be most relevant to the hypotheses and conclusions of the problem (and thus, presumably, to the solution as well).

For instance, if one is working in a problem that is set in Euclidean space , but the hypotheses and conclusions only involve measure-theoretic concepts (e.g. measurability, integrability, measure, etc.) rather than topological structure, metric structure, etc., then it may be worthwhile to try abstracting the problem to the more general setting of an abstract measure space, thus forgetting that one was initially working in . The point of doing this is that it cuts down on the number of possible ways to start attacking the problem. For instance, facts such as outer regularity (every measurable set can be approximated from above by an open set) do not hold in abstract measure spaces (which do not even have a meaningful notion of an open set), and so presumably will not play a role in the solution; similarly for any facts involving boxes. Instead, one should be trying to use general facts about measure, such as countable additivity, which are not specific to .

[It is worth noting that sometimes this abstraction method does not always work; for instance, when viewed as a measure space, is not completely arbitrary, but does have one or two features that distinguish it from a generic measure space, most notably the fact that it is -finite. So, even if the hypotheses and conclusion of a problem about is purely measure-theoretic in nature, one may still need to use some measure-theoretic facts specific to . Here, it becomes useful to know a little bit about the classification of measure spaces to have some intuition as to how “generic” a measure space such as really is. This intuition is hard to convey at this level of the subject, but in general, measure spaces form a very “non-rigid” category, with very few invariants, and so it is largely true that one measure space usually behaves much the same as any other.]

Another example of abstraction: suppose that a problem involves a large number of sets (e.g. and ) and their measures, but that the conclusion of the problem only involves the measures of the sets, rather than the sets themselves. Then one can try to abstract the sets out of the problem, by trying to write down every single relationship between the numerical quantities that one can easily deduce from the given hypotheses (together with basic properties of measure, such as monotonicity or countable additivity). One can then rename these quantities (e.g. and ) to “forget” that these quantities arose from a measure-theoretic context, and then work with a purely numerical problem, in which one is starting with hypotheses on some sequences of numbers and trying to deduce a conclusion about such sequences. Such a problem is often easier to solve than the original problem due to the simpler context. Sometimes, this simplified problem will end up being false, but the counterexample will often be instructive, either in indicating the need to add an additional hypothesis connecting the , or to indicate that one cannot work at this level of abstraction but must introduce some additional concrete ingredient.

A real-life example: a student asked how to justify the claim that

uniformly for and , where is a fixed linear operator and is a fixed natural number. This expression looks very complicated, but if one observes that the expression is common to both sides, and that this expression is of magnitude , one sees that actually one can abstract this claim to the simpler claim

At this point it is clear what to do: use Taylor’s theorem with remainder.

Note that this trick is in many ways the antithesis of Trick 9, because by passing to a special case, one often makes the problem more concrete, with more things that one is now able to start trying. However, the two tricks can work together. One particularly useful “advanced move” in mathematical problem solving is to first abstract the problem to a more general one, and *then* consider a special case of that more abstract problem which is not directly related to the original one, but is somehow simpler than the original while still capturing some of the essence of the difficulty. Attacking this alternate problem can then lead to some indirect but important ways to make progress on the original problem.

**11. Exploit Zeno’s paradox: a single epsilon can be cut up into countably many sub-epsilons.**

A particularly useful fact in measure theory is that one can cut up a single epsilon into countably many pieces, for instance by using the geometric series identity

this observation arguably goes all the way back to Zeno. As such, even if one only has an epsilon of room budgeted for a problem, one can still use this budget to do a countably infinite number of things. This fact underlies many of the countable additivity and subadditivity properties in measure theory, and also makes the ability to approximate rough objects by smoother ones to be useful even when countably many rough objects need to be approximated.

In general, one should be alert to this sort of trick when one has to spend an epsilon or so on an infinite number of objects. If one was forced to spend the same epsilon on each object, one would soon end up with an unacceptable loss; but if one can get away with using a different epsilon each time, then Zeno’s trick comes in very handy.

Needless to say, this trick combines well with Trick 2.

**12. If you expand your way to a double sum, a double integral, a sum of an integral, or an integral of a sum, try interchanging the two operations.**

Or, to put it another way: “The Fubini-Tonelli theorem is your friend”. Provided that one is either in the unsigned or absolutely convergent worlds, this theorem allows you to interchange sums and integrals with each other. In many cases, a double sum or integral that is difficult to sum in one direction can become easier to sum (or at least to upper bound, which is often all that one needs in analysis). In fact, if in the course of expanding an expression, you encounter such a double sum or integral, you should reflexively try interchanging the operations to see if the resulting expression looks any simpler.

Note that in some cases the parameters in the summation may be constrained, and one may have to take a little care to sum it properly. For instance,

(1)

interchanges (assuming that the are either unsigned or absolutely convergent) to

(why? try plotting the set of pairs that appear in both). If one is having trouble interchanging constrained sums or integrals, one solution is to re-express the constraint using indicator functions. For instance, one can rewrite the constrained sum (1) as the unconstrained sum

(extending the domain of if necessary), at which point interchanging the summations is easily accomplished.

The following point is obvious, but bears mentioning explicitly: while the interchanging sums/integrals trick can be very powerful, one should not apply it twice in a row to the same double sum or double operation, unless one is doing something genuinely non-trivial in between those two applications. So, after one exchanges a sum or integral, the next move should be something other than another exchange (unless one is dealing with a triple or higher sum or integral).

A related move (not so commonly used in measure theory, but occurring in other areas of analysis, particularly those involving the geometry of Euclidean spaces) is to merge two sums or integrals into a single sum or integral over the product space, in order to use some additional feature of the product space (e.g. rotation symmetry) that is not readily visible in the factor spaces. The classic example of this trick is the evaluation of the gaussian integral by squaring it, rewriting that square as the two-dimensional gaussian integral , and then switching to polar coordinates.

**13. Pointwise control, uniform control, and integrated (average) control are all partially convertible to each other.**

There are three main ways to control functions (or sequences of functions, etc.) in analysis. Firstly, there is pointwise control, in which one can control the function at every point (or almost every point), but in a non-uniform way. Then there is uniform control, where one can control the function in the same way at most points (possibly throwing out a set of zero measure, or small measure). Finally, there is integrated control (or control “on the average”), in which one controls the integral of a function, rather than the pointwise values of that function.

It is important to realise that control of one type can often be partially converted to another type. Simple examples include the deduction of pointwise convergence from uniform convergence, or integrating a pointwise bound to obtain an integrated bound . Of course, these conversions are not reversible and thus lose some information; not every pointwise convergent sequence is uniformly convergent, and an integral bound does not imply a pointwise bound. However, one can partially reverse such implications if one is willing to throw away an exceptional set (Trick 7). For instance, Egorov’s theorem lets one convert pointwise convergence to (local) uniform convergence after throwing away an exceptional set, and Markov’s inequality lets one convert integral bounds to pointwise bounds, again after throwing away an exceptional set.

**14. If the conclusion and hypotheses look particularly close to each other, just expand out all the definitions and follow your nose.**

This trick is particularly useful when building the most basic foundations of a theory. Here, one may not need to experiment too much with generalisations, abstractions, or special cases, or try to marshall a lot of possibly relevant facts about the objects being studied: sometimes, all one has to do is go back to first principles, write out all the definitions with their epsilons and deltas, and start plugging away at the problem.

Knowing when to just follow one’s nose, and when to instead look for a more high-level approach to a problem, can require some judgement or experience. A direct approach tends to work best when the conclusion and hypothesis already look quite similar to each other (e.g. they both state that a certain set or family of sets is measurable, or they both state that a certain function or family of functions is continuous, etc.). But when the conclusion looks quite different from the hypotheses (e.g. the conclusion is some sort of integral identity, and the hypotheses involve measurability or convergence properties), then one may need to use more sophisticated tools than what one can easily get from using first principles.

**15. Don’t worry too much about exactly what (or , or , etc.) needs to be. It can usually be chosen or tweaked later if necessary.**

Often in the middle of an argument, you will want to use some fact that involves a parameter, such as , that you are completely free to choose (subject of course to reasonable constraints, such as requiring to be positive). For instance, you may have a measurable set and decide to approximate it from above by an open set of at most more measure. But it may not be obvious exactly what value to give this parameter ; you have so many choices available that you don’t know which one to pick!

In many cases, one can postpone thinking about this problem by leaving undetermined for now, and continuing on with one’s argument, which will gradually start being decorated with ‘s all over the place. At some point, one will need to do something (and, in the particular case of , “doing something” almost always means “being small enough”), e.g. one may need to be less than , where are some other positive quantities in one’s problem that do not depend on . At this point, one could now set to be whatever is needed to get past this step in the argument, e.g. one could set to equal . But perhaps one still wishes to retain the freedom to set because it might come in handy later. In that case, one sets aside the requirement “” and keeps going. Perhaps a bit later on, one might need to do something else; for instance, one might also need . Once one has compiled the complete “wish list” of everything one wishes one’s parameters to do, then one can finally make the decision of what value to set those parameters equal to. For instance, if the above two inequalities are the only inequalities required of , one can choose equal to . This may be a choice of which was not obvious at the start of the argument, but becomes so as the argument progresses.

There is however one big caveat when adopting this “choose parameters later” approach, which is that one needs to avoid a circular dependence of constants. For instance, it is perfectly fine to have two arbitrary parameters and floating around unspecified for most of the argument, until at some point you realise that you need to be smaller than , and so one chooses accordingly (e.g. one sets it to equal ). Or, perhaps instead one needs to be smaller than , and so sets equal to . One can execute either of these two choices separately, but of course one cannot perform them simultaneously; this sets up an inconsistent circular dependency in which needs to be defined after is chosen, and can only be chosen after is fixed. So, if one is going to delay choosing a parameter such as until later, it becomes important to mentally keep track of what objects in one’s argument depend on , and which ones are independent of . One can choose in terms of the latter quantities, but one usually cannot do so in terms of the former quantities (unless one takes the care to show that the interlinked constraints between the quantities are still consistent, and thus simultaneously satisfiable).

See also the Tricki article “Keep parameters unspecified until it is clear how to optimize them“.

**16. Once one has started to lose some constants (or some epsilons), don’t be hesitant to lose some more.**

Many techniques in analysis end up giving inequalities that are inefficient by a constant factor. For instance, any argument involving dyadic decomposition and powers of two tends to involve losses of factors of 2. When arguing using balls in Euclidean space, one sometimes loses factors involving the volume of the unit ball (although this factor often cancels itself out if one tracks it more carefully). And so forth. However, in many cases these constant factors end up being of little importance: an upper bound of or is often just as good as an upper bound of for the purposes of analysis (cf. Trick 15). So it is often best not to invest too much energy in carefully computing and optimising these constants; giving these constants a symbol such as , and not worrying about their exact value, is often the simplest approach. (One can also use asymptotic notation, such as , which is very convenient to use once you know how it works.)

Now there are some cases in which one really does not want to lose any constants at all. For instance, if one is using Trick 1 to prove that , it is not enough to show that and ; one really needs to show and without losing any constants. (But proving and for every is OK, by Trick 2.) But once one has already performed one step that loses a constant, there is little further to be lost by losing more; there can be a big difference between and , but there is little difference in practice between and , at least for the purposes of mathematical analysis. At that stage, one should put oneself in the mental mode of thought where “constants don’t matter”, which can lead to some simplifications. For instance, if one has to estimate a sum of two positive quantities, one can start using such estimates as

which says that, up to afactor of , is the same thing as . In some cases the latter is easier to work with (e.g. is equal to , whereas the formula for is messier).

A related principle: if one is in the common situation where one can afford to have some error terms, then often approximate formulae can be more useful than exact formulae. For instance, a Taylor expansion with remainder such as can often be more useful than the full Taylor series . As a general rule of thumb, once one has a formula that is accurate up to the order of magnitude of error one is willing to tolerate, it is usually not very productive to try to refine the formula much further to improve the accuracy.

In a similar spirit, if one wants to optimise some expression (such as ) in a parameter such as (where in this example is a small quantity), and one is willing to lose a constant factor, then it is often not necessary to perform the undergraduate calculus steps of differentiating in the parameter and setting the derivative equal to zero; one can instead just test a few values of the parameter and see what produces a near-optimal result. For instance, setting any larger than will give a terrible first term, but setting gives a first term that is only of size , and a second term that is the smaller size . To balance this one can instead set , and now one sees that both terms are of size . One can also easily show that it is not possible for both terms to simultaneously be significantly smaller than . so the optimal bound here is indeed . (Now try to arrive at the same conclusion using undergraduate calculus.)

**17. One can often pass to a subsequence to improve the convergence properties.**

In real analysis, one often ends up possessing a sequence of objects, such as a sequence of functions , which may converge in some rather slow or weak fashion to a limit . Often, one can improve the convergence of this sequence by passing to a subsequence. For instance:

- In a metric space, if a sequence converges to a limit , then one can find a subsequence which converges quickly to the same limit , for instance one can ensure that (or one can replace with any other positive expression depending on ). In particular, one can make and absolutely convergent, which is sometimes useful.
- A sequence of functions that converges in norm or in measure can be refined to a subsequence that converges pointwise almost everywhere as well.
- A sequence in a (sequentially) compact space may not converge at all, but some subsequence of it will always converge.
- The pigeonhole principle: A sequence which takes only finitely many values has a subsequence that is constant. More generally, a sequence which lives in the union of finitely many sets has a subsequence that lives in just one of these sets.

Often, the subsequence is good enough for one’s applications, and there are also a number of ways to get back from a subsequence to the original sequence:

- In a metric space, if you know that is a Cauchy sequence, and some subsequence of already converges to , then this drags the entire sequence with it, i.e. converges to also.
- The Urysohn subsequence principle: in a topological space, if every subsequence of a sequence itself has a subsequence that converges to a limit , then the entire sequence converges to .

**18. A real limit can be viewed as a meeting of the limit superior and limit inferior.**

A sequence of real numbers does not necessarily have a limit , but the limit superior and the limit inferior always exist (though they may be infinite), and can be easily defined in terms of infima and suprema. Because of this, it is often convenient to work with the lim sup and lim inf instead of a limit. For instance, to show that a limit exists, it suffices to show that

for all . In a similar spirit, to show that a sequence of real numbers converges to zero, it suffices to show that

for all . It can be more convenient to work with lim sups and lim infs instead of limits because one does not need to worry about the issue of whether the limit exists or not, and many tools (notably Fatou’s lemma and its relatives) still work in this setting. One should however be cautious that lim sups and lim infs tend to have only one half of the linearity properties that limits do; for instance, lim sups are subadditive but not necessarily additive, while lim infs are superadditive but not necessarily additive.

**19. Be aware of and exploit the symmetries of the problem, for instance to normalise a key mathematical object, or to locate instructive limiting cases.**

In many problems in analysis enjoy some sort of symmetry (such as translation symmetry or homogeneity) in which the truth of the statement to be proven is clearly unchanged (or nearly unchanged) if one applies this symmetry to the parameters of the problem. For instance, if one wishes to establish an estimate of the form

for all , then one can notice that this estimate enjoys a homogeneity symmetry, in that one can multiply by an arbitrary non-zero constant without changing the truth value of the above estimate, since the inequality

is clearly equivalent to the original inequality.

Being aware of such symmetries can have many advantages. Most obviously, one can *spend* such symmetries to achieve a convenient normalisation; for instance, in the above problem, one can spend the homogeneity symmetry to obtain the normalisation (in some rare cases, would be a more convenient normalisation). Once one enforces such a normalisation, though, the symmetry is “spent” and cannot be simultaneously used to achieve a different normalisation that is incompatible with the previous one (for instance one cannot simultaneously normalise and ). It is thus desirable to collect as many symmetries of the problem as one can, so that one can normalise as much as possible (but one has to check that spending one symmetry to achieve a normalisation does not destroy the other symmetries one is planning to spend). For instance, to prove Holder’s inequality

one can use the homogeneity in and separately to simultaneously achieve the normalisations ; but one can also use the symmetry to make non-negative, and also the symmetry to normalise one of the exponent, say , to be equal to 1. Finally there is a symmetry , , arising from reweighting the measure, which can be used (for instance) to achieve a normalisation , which can be convenient for some proofs of this inequality (see this previous blog post for more discussion).

A more advanced move is to *arbitrage* an imbalance in symmetry in a statement to strengthen it, or to locate interesting asymptotic special cases; I discuss these sorts of tricks in this post. This is particularly powerful with respect to the symmetry of taking tensor products, where it is known as the “tensor power trick”. As another example, if one wishes to prove a nonlinear inequality of the form

for some given and all in some vector space, one can exploit the imbalance of symmetry with respect to dilations to rewrite this as

and then by sending one obtains some useful linearisations of the problem which can then be studied first (e.g., if is differentiable, one has the sub-problem of checking that , and if is twice differentiable, there is also the sub-problem of checking that is negative semidefinite).

Symmetries can also be used to detect errors (particularly with regards to exponents), much as dimensional analysis is used in physics; see also this related Tricki article.

Because symmetries are so useful, it is often worth abstracting or generalising the problem to make these symmetries more evident; thus this trick combines well with Trick 10.

**20. To show that a “completed” mathematical object exists, one can first build an “incomplete” mathematical object, find a way to make it more complete, and appeal to Zorn’s lemma to conclude.**

This trick is so standard in analysis that one sometimes sees lemmas in the literature whose entire proof is basically “Apply Zorn’s lemma in the obvious fashion.” The Hahn-Banach theorem is a well known example of this, but there are many others (e.g. the proof of existence of non-principal ultrafilters, discussed in this previous post). See my blog post on Zorn’s lemma for more discussion. One subtlety when applying this lemma is that one has to check that the class of incomplete objects one is applying Zorn’s lemma to is actually a set rather than a proper class; see this previous blog post for more discussion.

Zorn’s lemma of course relies on the axiom of choice, and the objects produced by this lemma are generally considered to be “nonconstructive”, but this should not deter one from using it; in practical applications (not involving pathological or hugely infinite objects), one can often rework an argument using Zorn’s lemma into a lengthier but more constructive one that does not require the axiom of choice.

**21. To demonstrate that all objects in a certain class obey a certain property, first show that this property is closed under various basic operations of the class, and then establish the property for “generating” elements of the class.**

This is a variant of Trick 3. It can be illustrated by some basic examples:

- (Induction) To show that a property holds for all natural numbers , first show that it is preserved under incrementation (if holds, then holds), then check the base case arising from the generator .
- (Backwards induction) To show that a property holds for all numbers , first show that it is preserved under decrementation (if holds for , then holds), then check the base case arising from the generator .
- (Strong induction) To show that a property holds for all natural numbers , show that it is preserved under least strict upper bounds (if holds for all , then holds). In this case an empty set of generators suffices, so one can omit the base case step. (This also works more generally if the natural numbers are replaced by a well-ordered set.)
- (Transfinite induction) To show that a property holds for all in a well-ordered set, show that it is preserved under least upper bounds and incrementation, and verify the base case.
- (Continuous induction) To show that a property holds for all in a metric space, show that the property is preserved under limits (if and holds, then holds), is open (if holds, then holds for all sufficiently close to ), and verify a base case for at least one point . (Continuous induction also works in more general topological spaces than metric spaces, though if the space is not first countable one may need to work with limits of nets instead of limits of sequences.)
- (Dense subclass argument) To show that a property holds for all in a topological space , show that it is preserved under limits, then check it for a dense subclass. (This argument is also known as the “density argument” or “limiting argument”.) Again, for spaces that are not first countable, one has to work with limits of nets rather than limits of sequences.
- (Linear algebra generation) To show that a property holds for all vectors in a vector space , show that the property is preserved under linear combinations (thus if hold, then holds, as well as for any scalar ), and also verify for an algebraic basis (or Hamel basis) of . This sort of argument can also be adapted to other bases of than Hamel bases, but then one often needs to also verify that the property is preserved under limits.
- (Sigma algebra generation) To show that a property holds for all in a -algebra , show that the property is preserved under countable unions, countable intersections, and complements, and then verify for the empty set and for a generating set of the -algebra (for instance, if is a Borel -algebra, one can take open sets, closed sets, or compact sets as the generating sets; in some cases one can also take balls, cubes, or rectangles); see Remark 4 of Notes 3. One can relax the set of properties required, for instance countable intersections is redundant by de Morgan’s laws if one already has countable unions and complements, and by the monotone class lemma one only has to verify monotone countable intersections and unions (and drop complements) if the generating set is already a Boolean algebra. (In probability theory, the theorem of Dynkin plays a similar role to the monotone class lemma.)
- (Convex generation) To show that a property holds for all in a finite-dimensional compact convex set , show that it holds for convex linear combinations, then test it on all extreme points. The same argument also works for infinite dimensions if one allows for convex integrals rather than convex combinations, thanks to Choquet’s theorem; alternatively, by the Krein-Milman theorem, one can keep working with finite convex combinations as long the property is closed under limits.
- (Stone-Weierstrass generation) To show that a property holds for all continuous functions on a compact space , show that the property is closed under algebraic operations (linear operations, pointwise multiplication, and also complex conjugation, if one is working with complex-valued functions), and also is true for a class of functions that is large enough to separate points in the space, so that the Stone-Weierstrass theorem may be applied.
- (Symmetry reduction) To show that a property holds for all in a space , which has a group acting on it, it suffices to show that the property is invariant under the group action (if is true, then is true for any , and then verify for one representative from each orbit of . (This reduction overlaps significantly with Trick 19.)

One caveat with applying these sorts of reductions: it is necessary that the predicate be precisely defined to avoid vagueness paradoxes. In particular, the predicate should not involve asymptotic notation such as or .

## 42 comments

Comments feed for this article

21 October, 2010 at 10:09 pm

tksmAnother trick: see whether it is possible to extend a given function to a larger domain in order to access more tools. For example, a continuous function defined on an open interval may extend continuously to the closed interval, a smooth real function may extend to a complex analytic function.

21 October, 2010 at 11:24 pm

TKI am new in research field. This is one of the most useful and appropriate article that i have read till date. Thanking you for posting.

22 October, 2010 at 1:17 am

GerardAlthough “How to solve it” is a good reading about the philosophy and general procedures of problem’ solving, it’s conforting to see the common tricks made explicit. Thanks!

22 October, 2010 at 1:20 am

Colin McQuillanTo show inequalities between norms, scaling to 1 helps with the algebra. For example in exercise 11 of

https://terrytao.wordpress.com/2009/03/30/245c-notes-1-interpolation-of-lp-spaces/

to show that the bounds C and C’ are equivalent up to a multiplicative constant, scale C to 1 and bound C’, and vice versa.

22 October, 2010 at 1:51 am

LeiI think the Remark 4 in notes 3 is also a very useful trick.

22 October, 2010 at 2:24 am

Thanks!I’m not in your class (unfortunately) but your post helped me a lot in my class!…

22 October, 2010 at 7:12 am

cyDear Prof Tao,

This is a great post to help students (like me) who are struggling with graduate analysis.

Do you have any advice to tackle problems of the type: Prove or disprove the following statements … So instead of attacking these problems directly, we have to think twice whether we are heading in the right direction. And usually it’s hard to think of suitable counterexamples during exams!

Thanks!

22 October, 2010 at 4:20 pm

Top Posts — WordPress.com[…] 245A: Problem solving strategies This is going to be a somewhat experimental post. In class, I mentioned that when solving the type of homework problems […] […]

23 October, 2010 at 11:25 am

David SpeyerHere are some more basic points that I remember being useful, related to yours:

Suppose you have some quantity which you are trying to bound by , as in strategy . Suppose depends on several parameters: maybe there is a grid spacing which is going to zero and a length of some sequence which is going to infinity. The hard way to do this is to start out your proof: “Let and …” The easy way is to start out with “Let and .” Run the whole proof and get a bound for in terms of and ; maybe something like . Only once you have those bounds, do you go back and figure out how small to take and how large to take in terms of .

Regarding Terry’s advice: “A common mistake is to try to draw a picture in which both the hypotheses and conclusion of the problem hold.” I would say that this holds in every field of math. The attempt to build a counter-example is much more informative than the attempt to find an example. I like to take my statement and strip out all implications using the equivalence of “p implies q” and “not(p) or q”. So what I am trying to prove is simply “A or B or C or D”. Then a counterexample is “not(A) and not(B) and not(C) and not(D)”, and I can search for a counterexample without being biased as to which parts of the result to focus on.

Regarding 12: In analysis, you’ll learn lots of subtle results about when sums and integrals can be exchanged. The one that I have found most useful is “If your sum/integral is absolutely convergent, anything goes.” I am willing to sacrifice a lot in an analysis problem in order to work with an absolutely convergent summand/integrand, so that I can apply Terry’s advice 12 without worrying.

24 October, 2010 at 7:15 am

V.L.A trick for proving things about Borel sets: if one wants to prove that the Borel sets have a certain property, one can prove that the collection of all sets that have that property is a sigma algebra that contains the open sets (or closed sets or intervals). Since the Borel sets are the smallest such sigma algebra, they must be contained in this collection. This avoids having to make arguments that use the way in which the Borel sets are generated.

26 October, 2010 at 8:27 am

anonymousAll these fall under the general philosophy ( I think I first saw it in Polya’s How to Solve it):

If you cannot solve a problem, find an easier one that you still cannot solve.

26 October, 2010 at 2:10 pm

edenh“Or: if one wishes to show that two functions f, g agree almost everywhere, try showing first that |f(x)-g(x)| \leq \varepsilon holds for almost every x.”

Is it obvious that this second condition implies the first?

26 October, 2010 at 4:41 pm

Terence TaoBy the Archimedean principle, the set is the union of the sets for all . Thus, if all the latter sets have measure zero, the former set does also.

26 October, 2010 at 4:12 pm

Linkage « An Ergodic Walk[…] October 26, 2010 Linkage Posted by Anand Sarwate under Uncategorized | Tags: mathematics, news, politics | Leave a Comment Terry Tao gives general tips on solving analysis problems. […]

4 November, 2010 at 3:54 pm

AHThanks for this very helpful post!

The first display of item 18 contains a slight typo: the subscript on the liminf should be {n \to \infty}, not {n \to infty}.

[Corrected, thanks – T.]17 November, 2010 at 10:19 am

-IamLittle-Dr Tao, is it good to learn some tricks to solve mathematic problems?

I am not used to learning tricks, I always solve problems without seeing any examples, tricks, etc. I know it wastes my time, so I would like to ask for your suggestion

18 November, 2010 at 3:46 am

AnonymousIn the first paragraph fir trick 2 shouldn’t it be for ‘some’ instead of for ‘any’ ?

[No – T.]18 November, 2010 at 10:09 am

AnonymousCould you please explain? (Sorry to bother with such a basic question but I am not a mathematician) As I see it if some quantity then .

Thanks.

18 November, 2010 at 10:12 am

Terence TaoThe issue is the reverse implication. does not imply , but does.

28 August, 2019 at 9:23 am

AnonymousI’m no mathematician but why is there a ‘0’ in /leq$ Y+/epsilon]$ I didn’t understand it .

29 August, 2019 at 4:26 am

MikeRQuantifiers [there exists] and [for all] have very different meanings.

“There exists a pitch that I will hit for a home run” is different from “for all pitches I will hit a home run.”

10 December, 2010 at 5:23 pm

Universality and Characteristic Polynomials « Graduated Understanding[…] you feel are irrelevant to the current problem. Terry tao discusses a similar trick in analysis (trick 10) in his “problem solving strategies,” […]

2 January, 2011 at 9:09 pm

science and mathNice tips.

Thanks for sharing the tips with us.

5 January, 2011 at 5:18 am

AnonymousWhen you call “Give yourself an epsilon of room” a standard trick, do you mean it is the ONLY way to handle such kind of problems? I was wondering whether it is right to say that most of the problems in analysis is about dealing with “epsilon”. (maybe according to Halmos’s experiences?)

1 May, 2011 at 9:01 pm

Some outside help… | Analysis Quals[…] Posted on May 1, 2011 by eekelly2388 Jeremy just sent me this link to Terry Tao’s blog about common problem solving strategies for analysis. I can especially relate to number 9, try simpler cases first. Also, number 15 is a great trick; […]

28 December, 2012 at 12:53 pm

Parth KohliProf. Terrence, please let me point out a mistake.

“In a similar spirit, to show that two sets E and F are equal, try proving that and $E \subset B$ and $B \subset E$.”

I believe that it’d be the following instead:

“In a similar spirit, to show that two sets E and F are equal, try proving that $E \subseteq B$ and $B \subseteq E$.

Regards.

28 December, 2012 at 12:54 pm

Parth KohliSorry, I do not know how to render MathJax here.

28 February, 2013 at 12:36 am

Clark ZinzowAlthough I am not fluent in MathJax, I do not think that your correction is correct. His use of the normal subset is correct, since it is stating that E is a proper subset of F or E is equal to F, and vice versa, allowing for the potential of equality. Whereas your notation (assuming that “\subseteq” is referring to the subset symbol with a horizontal line underneath it) refers to a proper subset exclusively, therefore not allowing for the potential of equality since E being a proper subset of F implies that there exists some element in F such that the same element is not in E, and vice versa.

Please let me know if I misunderstood! I know that some mathematicians occasionally use the proper subset symbol to mean “a subset of or equal to”, much to the chagrin of those raised by baby and/or adolescent Rudin (the text, not the man).

13 June, 2014 at 6:20 am

SpencerIn my experience the symbol for a proper subset is $\subset$. See for instance: http://mathworld.wolfram.com/Subset.html .

3 June, 2015 at 4:45 pm

Clark ZinzowOh man I forgot about this post. My professor used what is widely considered incorrect notation, with subset being $\subset$ and proper subset being $\subseteq$, which is quite a terrible convention to use. I didn’t realize that this was incorrect until I took a topology class the next semester. Thanks for correcting me, I wouldn’t want to plant that incorrect seed into anyone else’s brain!

1 June, 2014 at 6:18 am

AnonymousWould it be much more useful that if this list contains concrete examples (proofs/problems) that illustrate each strategy?

[The book version of this post at https://terrytao.wordpress.com/books/an-introduction-to-measure-theory/ contains links to such examples in other sections of the book. -T.]6 February, 2015 at 5:27 pm

Career Advice by Prof Terence Tao, Mozart of Mathematics | MScMathematics[…] See also Eric Schechter’s “Common errors in undergraduate mathematics“. I also have a post on problem solving strategies in real analysis. […]

11 July, 2015 at 6:04 am

译：解决数学问题 by 陶哲轩 | 万里风云[…] 请读Eric Schechter’s的《本科数学学习的常见错误》。除此，我也发表过一篇关于实分析解题策略的博客。 […]

18 July, 2016 at 8:29 am

Solving mathematical problems | nguyen Huynh Huy's Blog[…] See also Eric Schechter’s “Common errors in undergraduate mathematics“. I also have a post on problem solving strategies in real analysis. […]

30 December, 2018 at 6:55 am

Máté WierdlHere is a simple example for exploiting symmetries to make proofs go simpler: when a weak inequality is to be proved for a homogeneous operator , one can assume that . All books I have seen don’t make this observation and carry the around throughout the proof. The same happens when discussing the Calderón-Zygmund decomposition. For example, keeping around, the estimates , valid for , looks lucky, while with it looks like business as usual, since “of course” for .

30 December, 2018 at 11:03 am

Terence TaoThis is a great example! I used to agree with you on immediately spending the homogeneity symmetry to normalise , but my thoughts on this are mixed now. Certainly it does simplify the calculations as you say. But once one spends the symmetry to buy the normalisation, one can’t spend it again to achieve any other normalisation (for instance, one might wish to normalise instead ), unless one undoes the normalisation to return to a homogeneous statement. If the normalisation is not immediately needed to achieve a simplification, it may actually be better to save the symmetry in reserve, though perhaps noting near the beginning of the argument that the reader could, if he or she wished, spend the symmetry right away to buy a normalisation.

A related point (which Jim Colliander shared with me some years ago) is that sometimes seemingly disposable parameters such as can be useful markers of information. For instance consider the Navier-Stokes equations , where is a constant viscosity. It is a common practice to spend a homogeneity symmetry to normalise , which certainly does clean up many calculations a little bit. But if one instead retains when doing, say, an energy estimate, one can see quite visibly what terms in the estimate are “viscosity terms” and which ones have nothing to do with viscosity, merely by the presence of factors. Also, the presence of positive or negative powers of in a term conveys a lot of information about what happens to that term in the inviscid limit . (One can think of these factors as being analogous to the dimensional units of physics such as length, mass, and time; . Indeed, much as dimensional analysis can be used to detect an incorrect physical calculation, retaining parameters such as and can also help guard against error, since if an equation or inequality suddenly stops being “lucky” in that the exponents of these parameters are no longer canceling out as expected, this immediately signals that a mistake has been made near this point.)

13 January, 2019 at 8:46 pm

Anonymous… but if one observes that the expression

is common to both sides

Since is used later, I think there is a redundant in the definition above.

[Corrected – T.]16 January, 2019 at 10:58 am

Terence TaoI’ve recently extended this list of problem solving strategies by adding two or three more tricks, some of which were based on previous suggestions in the comments.

25 August, 2019 at 5:57 pm

James WrightReblogged this on >Numberery_.

30 August, 2021 at 9:10 am

Resumen de lecturas compartidas durante agosto de 2020 | Vestigium[…] Problem solving strategies. ~ Terence Tao (2010). #Math […]

28 February, 2022 at 9:32 am

AnonymousMost of the above strategies are for simplifying the mathematical structure of the problem by reducing its complexity (wrt some standard complexity measures)

2 March, 2022 at 10:44 am

Victor PortonWon’t it be counted as spam if I share my real real analysis problem? This problem is specifically interesting, because it is 1. open-ended; 2. but I tried to formalize it; 3. uses mix of both non-standard and standard analysis (limit on a filter); 4. because of being a mix, calls to reformulate itself better. Solving this problem will help me to become rich and famous and so have money and media links to spread scientific research.

From https://math.stackexchange.com/q/4394253/4876

There are nodes, each has response time (here and further I use as an index).

I want probabilistically choose a node with low response time; if there are several nodes with low response time, alternate between them.

The first idea that came to my mind was to make probability of chosing -th node proportional to . But it is not a solution, because if there are many nodes with high response times, then may overweight few good (with low response times) nodes, so that with a big probability there will be selected a bad node.

Formally (using both non-standard analysis and limit on a filter):

I define as a decreasing in each argument positive-valued continuous function of positive numbers as follows: There are numbers . I denote the (the sought for) (standard) probability of the node (that is is a function of and ).

Let extend the above function of by making to be non-standard (positive) reals in such a way that of them a “good” (infinitely smaller than the rest) and not infinitely bigger than each other.

Let by adding more “not good” nodes. Find a function of from such that we can warrant that even when (with the restriction that we don’t change existing nodes and add only bad nodes) we have and each .

Help to find such a function of from .

Also, I used both non-standard analysis and limit. Isn’t it a poor way to formulate things? How to formulate the problem better (concisely and clearly)?