At the core of almost any undergraduate real analysis course are the concepts of differentiation and integration, with these two basic operations being tied together by the fundamental theorem of calculus (and its higher dimensional generalisations, such as Stokes’ theorem). Similarly, the notion of the complex derivative and the complex line integral (that is to say, the contour integral) lie at the core of any introductory complex analysis course. Once again, they are tied to each other by the fundamental theorem of calculus; but in the complex case there is a further variant of the fundamental theorem, namely Cauchy’s theorem, which endows complex differentiable functions with many important and surprising properties that are often not shared by their real differentiable counterparts. We will give complex differentiable functions another name to emphasise this extra structure, by referring to such functions as holomorphic functions. (This term is also useful to distinguish these functions from the slightly less well-behaved meromorphic functions, which we will discuss in later notes.)
In this set of notes we will focus solely on the concept of complex differentiation, deferring the discussion of contour integration to the next set of notes. To begin with, the theory of complex differentiation will greatly resemble the theory of real differentiation; the definitions look almost identical, and well known laws of differential calculus such as the product rule, quotient rule, and chain rule carry over verbatim to the complex setting, and the theory of complex power series is similarly almost identical to the theory of real power series. However, when one compares the “one-dimensional” differentiation theory of the complex numbers with the “two-dimensional” differentiation theory of two real variables, we find that the dimensional discrepancy forces complex differentiable functions to obey a real-variable constraint, namely the Cauchy-Riemann equations. These equations make complex differentiable functions substantially more “rigid” than their real-variable counterparts; they imply for instance that the imaginary part of a complex differentiable function is essentially determined (up to constants) by the real part, and vice versa. Furthermore, even when considered separately, the real and imaginary components of complex differentiable functions are forced to obey the strong constraint of being harmonic. In later notes we will see these constraints manifest themselves in integral form, particularly through Cauchy’s theorem and the closely related Cauchy integral formula.
Despite all the constraints that holomorphic functions have to obey, a surprisingly large number of the functions of a complex variable that one actually encounters in applications turn out to be holomorphic. For instance, any polynomial with complex coefficients will be holomorphic, as will the complex exponential . From this and the laws of differential calculus one can then generate many further holomorphic functions. Also, as we will show presently, complex power series will automatically be holomorphic inside their disk of convergence. On the other hand, there are certainly basic complex functions of interest that are not holomorphic, such as the complex conjugation function , the absolute value function , or the real and imaginary part functions . We will also encounter functions that are only holomorphic at some portions of the complex plane, but not on others; for instance, rational functions will be holomorphic except at those few points where the denominator vanishes, and are prime examples of the meromorphic functions mentioned previously. Later on we will also consider functions such as branches of the logarithm or square root, which will be holomorphic outside of a branch cut corresponding to the choice of branch. It is a basic but important skill in complex analysis to be able to quickly recognise which functions are holomorphic and which ones are not, as many of useful theorems available to the former (such as Cauchy’s theorem) break down spectacularly for the latter. Indeed, in my experience, one of the most common “rookie errors” that beginning complex analysis students make is the error of attempting to apply a theorem about holomorphic functions to a function that is not at all holomorphic. This stands in contrast to the situation in real analysis, in which one can often obtain correct conclusions by formally applying the laws of differential or integral calculus to functions that might not actually be differentiable or integrable in a classical sense. (This latter phenomenon, by the way, can be largely explained using the theory of distributions, as covered for instance in this previous post, but this is beyond the scope of the current course.)
Remark 1 In this set of notes it will be convenient to impose some unnecessarily generous regularity hypotheses (e.g. continuous second differentiability) on the holomorphic functions one is studying in order to make the proofs simpler. In later notes, we will discover that these hypotheses are in fact redundant, due to the phenomenon of elliptic regularity that ensures that holomorphic functions are automatically smooth.
— 1. Complex differentiation and power series —
Recall in real analysis that if is a function defined on some subset of the real line , and is an interior point of (that is to say, contains an interval of the form for some ), then we say that is differentiable at if the limit
exists (note we have to exclude from the possible values of to avoid division by zero. If is differentiable at , we denote the above limit as or , and refer to this as the derivative of at . If is open (that is to say, every element of is an interior point), and is differentiable at every point of , then we say that is differentiable on , and call the derivative of . (One can also define differentiability at non-interior points if they are not isolated, but for simplicity we will restrict attention to interior derivatives only.)
We can adapt this definition to the complex setting without any difficulty:
Definition 2 (Complex differentiability) Let be a subset of the complex numbers , and let be a function. If is an interior point of (that is to say, contains a disk for some ), we say that is complex differentiable at if the limit
exists, in which case we denote this limit as , , or , and refer to this as the complex derivative of at . If is open (that is to say, every point in is an interior point), and is complex differentiable at every point at , we say that is complex differentiable on , or holomorphic on .
In terms of epsilons and deltas: is complex differentiable at with derivative if and only if, for every , there exists such that whenever is such that . Another way of writing this is that we have an approximate linearisation
as approaches , where denotes a quantity of the form for in a neighbourhood of , where goes to zero as goes to . Making the change of variables , one can also write the derivative in the familiar form
where is the translation of by .
If is differentiable at , then from the limit laws we see that
that is to say that is continuous at . In particular, holomorphic functions are automatically continuous. (Later on we will see that they are in fact far more regular than this, being smooth and even analytic.)
It is usually quite tedious to verify complex differentiability of a function, and to compute its derivative, from first principles. We will give just one example of this:
Proof: This is clear for , so suppose . We need to show that for any complex number , that
But we have the geometric series identity
which is valid (in any field) whenever , as can be seen either by induction or by multiplying both sides by and cancelling the telescoping series on the right-hand side. The claim then follows from the usual limit laws.
Fortunately, we have the familiar laws of differential calculus, that allow us to more quickly establish the differentiability of functions if they arise as various combinations of functions that are already known to be differentiable, and to compute the derivative:
- (i) (Linearity) Show that is complex differentiable at , with derivative . For any constant , show that is differentiable at , with derivative .
- (ii) (Product rule) Show that is complex differentiable at , with derivative .
- (iii) (Quotient rule) If is non-zero, show that (which is defined in a neighbourhood of , by continuity) is complex differentiable at , with derivative .
- (iv) (Chain rule) If is a neighbourhood of , and is a function that is complex differentiable at , show that the composition (which is defined in a neighbourhood of ) is complex differentiable at , with derivative
(Hint: take your favourite proof of the real-variable version of these facts and adapt them to the complex setting.)
One could also state and prove a complex-variable form of the inverse function theorem here, but the proof of that statement is a bit more complicated than the ones in the above exercise, so we defer it until later in the course when it becomes needed.
If a function is holomorphic on the entire complex plane, we call it an entire function; clearly such functions remain holomorphic when restricted to any open subset of the complex plane. Thus for instance Proposition 3 tells us that the functions are entire, and from linearity we then see that any complex polynomial
A function of the form , where are polynomials with not identically zero, is called a rational function, being to polynomials as rational numbers are to integers. Such a rational function is well defined as long as is not zero. From the factor theorem (which works over any field, and in particular over the complex numbers) we know that the number of zeroes of is finite, being bounded by the degree of (of course we will be able to say something stronger once we have the fundamental theorem of algebra). Because of these singularities, rational functions are rarely entire; but from the quotient rule we do at least see that is complex differentiable wherever the denominator is non-zero. Such functions are prime examples of meromorphic functions, which we will discuss later in the course.
Exercise 5 (Gauss-Lucas theorem) Let be a complex polynomial that is factored as
for some non-zero constant and roots (not necessarily distinct) with .
- (i) Suppose that all lie in the upper half-plane . Show that any root of the derivative also lies in the upper half-plane. (Hint: use the product rule to decompose the log-derivative into partial fractions, and then investigate the sign of the imaginary part of this log-derivative for outside the upper half-plane.)
- (ii) Show that all the roots of lie in the convex hull of the set of roots of , that is to say the smallest convex polygon that contains .
Now we discuss power series, which are infinite degree variants of polynomials, and which turn out to inherit many of the algebraic and analytic properties of such polynomials, at least if one stays within the disk of convergence.
Definition 6 (Power series) Let be a complex number. A formal power series with complex coefficients around the point is a formal series of the form
for some complex numbers , with an indeterminate.
One can attempt to evaluate a formal power series at a given complex number by replacing the formal indeterminate with the complex number . This may or may not produce a convergent (or absolutely convergent) series, depending on where is; for instance, the power series is always absolutely convergent at , but the geometric power series fails to be even conditionally convergent whenever (since the summands do not go to zero). As it turns out, the region of convergence is always essentially a disk, the size of which depends on how rapidly the coefficients decay (or how slowly they grow):
Proposition 7 (Convergence of power series) Let be a formal power series, and define the radius of convergence of the series to be the quantity
with the convention that is infinite if . (Note that is allowed to be zero or infinite.) Then the formal power series is absolutely convergent for any in the disk (known as the disk of convergence), and is divergent (i.e., not convergent) for any in the exterior region .
Proof: The proof is nearly identical to the analogous result for real power series. First suppose that is a complex number with (this of course implies that is finite). Then by (3), we have for infinitely many , which after some rearranging implies that for infinitely many . In particular, the sequence does not go to zero as , which implies that is divergent.
Now suppose that is a complex number with (this of course implies that is non-zero). Choose a real number with , then by (3), we have for all sufficiently large , which after some rearranging implies that
for all sufficiently large . Since the geometric series is absolutely convergent, this implies that is absolutely convergent also, as required.
Remark 8 Note that this proposition does not say what happens on the boundary of this disk (assuming for sake of discussion that the radius of convergence is finite and non-zero). The behaviour of power series on and near the boundary of the disk of convergence is in fact remarkably subtle; see for instance Example 11 below.
The above proposition gives a “root test” formula for the radius of convergence. The following “ratio test” variant gives convenient lower and upper bounds for the radius of convergence which suffices in many applications:
In particular, if the limit exists, then it is equal to . Give example to show that strict inequality can hold in both bounds in (4).
If a formal power series has a positive radius of convergence, then it defines a function in the disk of convergence by setting
We refer to such a function as a power series, and refer to as the radius of convergence of that power series. (Strictly speaking, a formal power series and a power series are different concepts, but there is little actual harm in conflating them together in practice, because of the uniqueness property established in Exercise 17 below.)
Example 10 The formal power series has a zero radius of convergence, thanks to the ratio test, and so only converges at . Conversely, the exponential formal power series has an infinite radius of convergence (thanks to the ratio test), and converges of course to when evaluated at any complex number .
and thus after some algebra we obtain the geometric series formula
as long as is inside the disk . The function does not extend continuously to the boundary point of the disk, but does extend continuously (and even smoothly) to the rest of the boundary, and is in fact holomorphic on the remainder of the complex plane. However, the geometric series diverges at every single point of this boundary (when , the coefficients of the series do not converge to zero), and of course definitely diverge outside of the disk as well. Thus we see that the function that a power series converges to can extend well beyond the disk of convergence, which thus may only capture a portion of the domain of definition of that function. For instance, if one formally applies (5) with, say, , one ends up with the apparent identity
This identity does not make sense if one interprets infinite series in the classical fashion, as the series is definitely divergent. However, by formally extending identities such as (5) beyond their disk of convergence, we can generalise the notion of summation of infinite series to assign meaningful values to such series even if they do not converge in the classical sense. This leads to generalised summation methods such as zeta function regularisation, which are discussed in this previous blog post. However, we will not use such generalised interpretations of summation very much in this course.
Exercise 12 For any complex numbers , show that the formal power series has radius of convergence (with the convention that this is infinite for ), and is equal to the function inside the disk of convergence.
Exercise 13 For any positive integer , show that the formal power series
has radius of convergence , and converges to the function in the disk . Here of course is the usual binomial coefficient.
We have seen above that power series can be well behaved as one approaches the boundary of the disk of convergence, while being divergent at the boundary. However, the converse scenario, in which the power series converges at the boundary but does not behave well as one approaches the boundary, does not occur:
- (i) (Summation by parts formula) Let be a finite sequence of complex numbers, and let be the partial sums for . Show that for any complex numbers , that
- (ii) Let be a sequence of complex numbers such that is convergent (not necessarily absolutely) to zero. Show that for any , the series is absolutely convergent, and
(Hint: use summation by parts and a limiting argument to express in terms of the partial sums .)
- (iii) (Abel’s theorem) Let be a power series with a finite positive radius of convergence , and let be a point on the boundary of the disk of convergence at which the series converges (not necessarily absolutely). Show that . (Hint: use various translations and rotations to reduce to the case considered in (ii).)
As a general rule of thumb, as long as one is inside the disk of convergence, power series behave very similarly to polynomials. In particular, we can generalise the differentiation formula (2) to such power series:
that has the same radius of convergence as .
Proof: From (3), the standard limit and the usual limit laws, it is easy to see that the power series has the same radius of convergence as . To show that this series is actually the derivative of , we use first principles. If lies in the disk of convergence, we consider the Newton quotient
for . Expanding out the absolutely convergent series and , we can write
The ratio vanishes for , and for it is equal to as in the proof of Proposition 3. Thus
As approaches , each summand converges to . This almost proves the desired limiting formula
but we need to justify the interchange of a sum and limit. Fortunately we have a standard tool for this, namely the Weierstrass -test (which works for complex-valued functions exactly as it does for real-valued functions; one could also use the dominated convergence theorem here). It will be convenient to select two real numbers with . Clearly, for close enough to , we have . By the triangle inequality we then have
On the other hand, from (3) we know that for sufficiently large , hence for sufficiently large . From the ratio test we know that the series is absolutely convergent, hence the series is also. Thus, for sufficiently close to , the summands are uniformly dominated by an absolutely summable sequence of numbers . Applying the Weierstrass -test (or dominated convergence theorem), we obtain the claim.
Exercise 16 Prove the above theorem directly using epsilon and delta type arguments, rather than invoking the -test or the dominated convergence theorem.
We remark that the above theorem is a little easier to prove once we have the complex version of the fundamental theorem of calculus, but this will have to wait until the next set of notes, where we will also prove a remarkable converse to the above theorem, in that any holomorphic function can be expanded as a power series around any point in its domain.
A convenient feature of power series is the ability to equate coefficients: if two power series around the same point agree, then their coefficients must also agree. More precisely, we have:
Exercise 17 (Taylor expansion and uniqueness of power series) Let be a power series with a positive radius of convergence. Show that , where denotes the complex derivative of . In particular, if is another power series around with a positive radius of convergence which agrees with on some neighbourhood of (thus, for all ), show that the coefficients of and are identical, that is to say that for all .
Of course, one can no longer compare coefficients so easily if the power series are based around two different points. For instance, from Exercise 11 we see that the geometric series and both converge to the same function on the unit disk , but have differing coefficients. The precise relation between the coefficients of power series of the same function is given as follows:
Exercise 18 (Changing the origin of a power series) Let be a power series with a positive radius of convergence . Let be an element of the disk of convergence . Show that the formal power series , where
has radius of convergence at least , and converges to on the disk . Here of course is the usual binomial coefficient.
has an infinite radius of convergence, and so is entire, and is its own derivative:
This makes the complex trigonometric functions
entire as well, and from the chain rule we recover the familiar formulae
Of course, one can combine these functions together in many ways to create countless other complex differentiable functions with explicitly computable derivatives, e.g. is an entire function with derivative , the tangent function is holomorphic outside of the discrete set with derivative , and so forth.
Exercise 19 (Multiplication of power series) Let and be power series that both have radius of convergence at least . Show that on the disk , we have
where the right-hand side is another power series of radius of convergence at least , with coefficients given as the convolution
of the sequences and .
— 2. The Cauchy-Riemann equations —
Thus far, the theory of complex differentiation closely resembles the analogous theory of real differentiation that one sees in an introductory real analysis class. But now we take advantage of the Argand plane representation of to view a function of one complex variable as a function of two real variables. This gives rise to some further notions of differentiation. Indeed, if is a function defined on an open subset of , and is a point in , then in addition to the complex derivative
already discussed, we can also define (if they exist) the partial derivatives
where is the Euclidean norm of .
These notions of derivative are of course closely related to each other. If a function is Fréchet differentiable at , in the sense that the gradient exists, then on specialising the limit in (7) to vectors of the form or we see that
for the gradient of a function that is Fréchet differentiable at . We caution however that it is possible for the partial derivatives of a function to exist without the function being Fréchet differentiable, in which case the formula (8) is of course not valid. (A typical example is the function defined by setting for , with ; this function has both partial derivatives existing at , but is not differentiable here.) On the other hand, if the partial derivatives exist everywhere on and are additionally known to be continuous, then the fundamental theorem of calculus gives the identity
for and sufficiently small (with the convention that if ), and from this it is not difficult to see that is then Fréchet differentiable everywhere on .
Similarly, if is complex differentiable at , then by specialising the limit (6) to variables of the form or for some non-zero real near zero, we see that
that must be satisfied in order for to be complex differentiable. More generally, from (6) we see that if is complex differentiable at , then
which on comparison with (7) shows that is also Fréchet differentiable with
which after making the substitution gives
which on comparison with (6) shows that is complex differentiable with
We summarise the above discussion as follows:
- (i) If is complex differentiable at , then it is also Fréchet differentiable at , with
In particular, the Cauchy-Riemann equations (9) hold at .
- (ii) Conversely, if is Fréchet differentiable at and obeys the Cauchy-Riemann equations at , then is complex differentiable at .
Remark 21 From part (ii) of the above proposition we see that if is Fréchet differentiable on and obeys the Cauchy-Riemann equations on , then it is holomorphic on . One can ask whether the requirement of Fréchet differentiability can be weakened. It cannot be omitted entirely; one can show, for instance, that the function defined by for non-zero and obeys the Cauchy-Riemann equations at every point , but is not complex differentiable (or even continuous) at the origin. But there is a somewhat difficult theorem of Looman and Menchoff that asserts that if is continuous on and obeys the Cauchy-Riemann equations on , then it is holomorphic. We will not prove or use this theorem in this course; generally in modern applications, when one wants to weaken the regularity hypotheses of a theorem involving classical differentiation, the best way to do so is to replace the notion of a classical derivative with that of a weak derivative, rather than insist on computing derivatives in the classical pointwise sense. See this blog post for more discussion.
Combining part (i) of the above proposition with Theorem 15, we also conclude as a corollary that any power series will be smooth inside its disk of convergence, in the sense all partial derivatives to all orders of this power series exist.
Remark 22 From the geometric perspective, one can interpret complex differentiability at a point as a requirement that a map is conformal and orientation-preserving at , at least in the non-degenerate case when is non-zero. In more detail: suppose that is a map that is complex differentiable at some point with . Let be a differentiable curve with ; we view this as the trajectory of some particle which passes through at time . The derivative (defined in the usual manner by limits of Newton quotients) can then be viewed as the velocity of the particle as it passes through . The map takes this particle to a new particle parameterised by the curve ; at time , this new particle passes through , and by the chain rule we see that the velocity of the new particle at this time is given by
Thus, if we write in polar coordinates as , the map transforms the velocity of the particle by multiplying the speed by a factor of and rotating the direction of travel counter-clockwise by . In particular, we consider two differentiable trajectories both passing through at time (with non-zero speeds), then the map preserves the angle between the two velocity vectors , as well as their orientation (e.g. if is counterclockwise to , then is counterclockwise to . This is in contrast to, for instance, shear transformations such as , which preserve orientation but not angle, or the complex conjugation map , which preserve angle but not orientation. The same preservation of angle is present for real differentiable functions , but is much less impressive in that case since the only angles possible between two vectors on the real line are and ; it is the geometric two-dimensionality of the complex plane that makes conformality a much stronger and more “rigid” property for complex differentiable functions.
One consequence of the first component of Proposition 20(i) is that the complex notion of differentiation and the real notion of differentiation are compatible with each other. More precisely, if is holomorphic and is the restriction of to the real line, then the real derivative of exists and is equal to the restriction of the complex derivative to the real line. For instance, since for a complex variable , we also have for a real variable . For similar reasons, it will be a safe “abuse of notation” to use the notation both to refer to the complex derivative of a function , and also the real derivative of the restriction of that function to the real line.
of two partial differential equations for two functions . This gives a quick way to test if various functions are differentiable. Consider for instance the conjugation function . In this case, and . These functions, being polynomial in , is certainly Fréchet differentiable everywhere; the equation (11) is always satisfied, but the equation (10) is never satisfied. As such, the conjugation function is never complex differentiable. Similarly for the real part function , the imaginary part function , or the absolute value function . The function has real part and imaginary part ; one easily checks that the system (10), (11) is only satisfied when , so this function is only complex differentiable at the origin. In particular, it is not holomorphic on any non-empty open set.
The general rule of thumb that one should take away from these examples is that complex functions that are constructed purely out of “good” functions such as polynomials, the complex exponential, complex trigonometric functions, or other convergent power series are likely to be holomorphic, whereas functions that involve “bad” functions such as complex conjugation, the real and imaginary part, or the absolute value, are unlikely to be holomorphic.
Exercise 23 (Wirtinger derivatives) Let be an open subset of , and let be a Fréchet differentiable function. Define the Wirtinger derivatives , by the formulae
- (i) Show that is holomorphic on if and only if the Wirtinger derivative vanishes identically on .
- (ii) If is given by a polynomial
in both and for some complex coefficients and some natural number , show that
(Hint: first establish a Leibniz rule for Wirtinger derivatives.) Conclude in particular that is holomorphic if and only if vanishes whenever (i.e. does not contain any terms that involve ).
- (iii) If is a point in , show that one has the Taylor expansion
as , where denotes a quantity of the form , where goes to zero as goes to (compare with (1)). Conversely, show that this property determines the numbers and uniquely (and thus can be used as an alternate definition of the Wirtinger derivatives).
Remark 24 Any polynomial
in the real and imaginary parts of can be rewritten as a polynomial in and as per (17), using the usual identities
for . Thus such a non-holomorphic polynomial of one complex variable can be viewed as the restriction of a holomorphic polynomial
of two complex variables to the anti-diagonal , and the Wirtinger derivatives can then be interpreted as genuine (complex) partial derivatives in these two complex variables. More generally, Wirtinger derivatives are convenient tools in the subject of several complex variables, which we will not cover in this course.
The Cauchy-Riemann equations couple the real and imaginary parts of a holomorphic function to each other. But it is also possible to eliminate one of these components from the equations and obtain a constraint on just the real part , or just the imaginary part . Suppose for the moment that is a holomorphic function which is twice continuously differentiable (thus the second partial derivatives , , , all exist and are continuous on ); we will show in the next set of notes that this extra hypothesis is in fact redundant. Assuming continuous second differentiability for now, we have Clairaut’s theorem
everywhere on . Similarly for the real and imaginary parts . If we then differentiate (10) in the direction, (11) in the direction, and then sum, the derivatives of cancel thanks to Clairaut’s theorem, and we obtain Laplace’s equation
where is the Laplacian operator
A similar argument gives ; by linearity we then also have .
Functions for which is continuously twice differentiable and are known as harmonic functions: thus we have shown that (continuously twice differentiable) holomorphic functions are automatically harmonic, as are the real and imaginary parts of such functions. The converse is not true: not every harmonic function is holomorphic. For instance, the conjugate function is clearly harmonic on , but not holomorphic. We will return to the precise relationship between harmonic and holomorphic functions shortly.
Harmonic functions have many remarkable properties. Since the second derivative in a given direction is a local measure of “convexity” of a function, we see from (13) that any convex behaviour of a harmonic function in one direction has to be balanced by an equal and opposite amount of concave behaviour in the orthogonal direction. A good example of a harmonic function to keep in mind is the function
which exhibits convex behavior in and concave behavior in in exactly opposite amounts. This function is the real part of the holomorphic function , which is of course consistent with the previous observation that the real parts of holomorphic functions are harmonic.
We will discuss harmonic functions more in later notes. For now, we record just one important property of these functions, namely the maximum principle:
Theorem 25 (Maximum principle) Let be an open subset of , and let be a harmonic function. Let be a compact subset of , and let be the boundary of . Then
Informally, the maximum principle asserts that the maximum of a real-valued harmonic function on a compact set is always attained on the boundary, and similarly for the minimum. In particular, any bound on the harmonic function that one can obtain on the boundary is automatically inherited by the interior. Compare this with a non-harmonic function such as , which is bounded by on the boundary of the compact unit disk , but is not bounded by on the interior of this disk.
Proof: We begin with an “almost proof” of this principle, and then repair this attempted proof so that it is an actual proof.
Suppose that this is the case. As is continuous and is compact, must attain its maximum at some point in ; from (16) we see that must be an interior point. Since is a local maximum of , and is twice differentiable, we must have
This almost, but does not quite, contradict the harmonicity of , since it is still possible that both of these partial derivatives vanish. To get around this problem we use the trick of creating an epsilon of room, adding a tiny bit of convexity to . Let be a small number to be chosen later, and let be the modified function
Since is compact, the function is bounded on . Thus, from (16), we see that if is small enough we have
Arguing as before, must attain its maximum at some interior point of , and so we again have
On the other hand, since is harmonic, we have
on . These facts contradict each other, and we are done.
Exercise 26 (Maximum principle for holomorphic functions) If is a continuously twice differentiable holomorphic function on an open set , and is a compact subset of , show that
(Hint: use Theorem 25 and the fact that for any complex number .) What happens if we replace the suprema on both sides by infima?
Exercise 27 Recall the Wirtinger derivatives defined in Exercise 23(i).
- (i) If is twice continuously differentiable on an open subset of , show that
Use this to give an alternate proof that holomorphic functions are harmonic.
- (ii) If is given by a polynomial
in both and for some complex coefficients and some natural number , show that is harmonic on if and only if vanishes whenever and are both positive (i.e. only contains terms or that only involve one of or ).
- (iii) If is a real polynomial
in and for some real coefficients and some natural number , show that is harmonic if and only if it is the real part of a polynomial of one complex variable .
We have seen that the real and imaginary parts of any holomorphic function are harmonic functions. Conversely, let us call a harmonic function a harmonic conjugate of another harmonic function if is holomorphic on ; this is equivalent by Proposition 20 to satisfying the Cauchy-Riemann equations (10), (11). Here is a short table giving some examples of harmonic conjugates:
(for the last example one of course has to exclude the origin from the domain ).
From Exercise 27(ii) we know that every harmonic polynomial has at least one harmonic conjugate; it is natural to ask whether the same fact is true for more general harmonic functions than polynomials. In the case that the domain is the entire complex plane, the answer is affirmative:
Proposition 28 Let be a harmonic function. Then there exists a harmonic conjugate of . Furthermore, this harmonic conjugate is unique up to constants: if are two harmonic conjugates of , then is a constant function.
Proof: We first prove uniqueness. If is a harmonic conjugate of , then from the fundamental theorem of calculus, we have
Similarly for any other harmonic conjugate of . It is now clear that and differ by a constant.
This is one of the two Cauchy-Riemann equations needed. To obtain the other one, we differentiate (18) in the variable. The fact that is continuously twice differentiable allows one to differentiate under the integral sign (exercise!) and conclude that
As is harmonic, we have , so by the fundamental theorem of calculus we conclude that
Thus we now have both of the Cauchy-Riemann equations (10), (11) in . Differentiating these equations again, we conclude that is twice continuously differentiable, and hence by Proposition 20 we have holomorphic on , giving the claim.
The same argument would also work for some other domains than , such as rectangles . To handle the general case, though, it becomes convenient to introduce the notion of contour integration, which we will do in the next set of notes. In some cases (specifically, when the underlying domain fails to be simply connected), it will turn out that some harmonic functions do not have conjugates!
Exercise 29 Show that an entire function can real-valued on only if it is constant.
Exercise 30 Let be a complex number. Show that if is an entire function such that for all , then for all .