Decay of waves IIb: Minkowski space, with right-hand side
by Willie Wong
In the first half of this second part of the series, we considered solutions to the linear, homogeneous wave equation on flat Minkowski space, and showed that for compactly supported initial data, we have strong Huygens’ principle. We further made references to the fact that this behaviour is expected to be unstable. In this post, we will further illustrate this instability by looking at Equation 1 first with a fixed source , and then with a nonlinearity
.
Duhamel’s Principle
To study how one can incorporate inhomogeneous terms into a linear equation, and to get a qualitative grasp of how the source term contributes to the solution, we need to discuss the abstract method known as Duhamel’s Principle. We start by illustrating this for a very simple ordinary differential equation.
Consider the ODE satisfied by a scalar function :
Equation 13
![]()
when , we can easily solve the equation with integration factors
Using this as a sort of an ansatz, we can solve the inhomogeneous equation as follows. For convenience we denote by the anti-derivative of
. Then multiplying Equation 13 through by
, we have that
Equation 14
![]()
which we solve by integrating
Equation 15
![]()
If we write , then we can rewrite Equation 15 as given by an integral operator
Equation 15′
![]()
So what does Equations 15 and 15′ mean? The term by itself would indicate a solution to the homogeneous version of ODE 13, evaluated at time
, with initial data prescribed at time
as
. So in other words, the total solution of the ODE 13 at time
corresponds to the “free evolution” of the initial data prescribed at time
, plus an infinitesimal part for each time instant between
corresponding to the "free evolution" of initial data
from time
to
.
Qualitatively, what is going on is that at every time instant, the "source term" gives a small kick to the solution. Because of the linearity of the corresponding homogeneous problem, each small kick evolves independently of the free evolution, and also independently of other small kicks coming from other instants in time. At the end to evaluate the solution at time , all one needs to do is to add up the contribution from the sum total of these small kicks, and add to it the free evolution of the homogeneous problem given by the initial data.
Now, recall how many statements about scalar ODEs can be transferred to vector valued ODEs: we can treat to be vector valued in Equation 13, with
to be matrix valued. The defining
the usual way for matrices, we get that Equations 15 and 15' equally well apply in this context.
Similarly, with a bit more care, we can generalise this to infinite dimensional systems.
For the wave equation, observe that we can write it as a first order ODE on some Hilbert space: let , then the wave equation can be written as
Equation 16
![]()
so at least on a formal level (which can in fact be made rigorous) we can obtain the solution for the inhomogeneous wave equation by considering the solutions for the homogeneous version.
To recapitulate, and to introduce some terminologies, equations like Equation 13 above, where are “time dependent” functions that are scalar-, vector-, or Hilbert-space-valued, and where
is a “time dependent” family of scalar-multipliers, linear transformations, or linear operators (in the three cases mentioned) are first order evolution equations. The corresponding object
(which is now is a two-parameter, depending on initial and final times, family of multipliers, linear transformations, or linear operators) goes under the names of free propagator, Green’s function, or fundamental solution, as it denotes the (linear) change for solutions of the homogeneous problem between initial time
and final time
.
As an aside: Interesting to note is that each evolution equation corresponds to a category. The objects are sets representing the time-dependent spaces of data. The morphisms are precisely that maps between the different times. Well-posedness of the initial value problem relative to one element
in the class of objects is then the statement that
is an initial object (existence) and that for any two objects
the class of morphisms
consists of at most one arrow (uniqueness). Time reversibility will then be the statement our category is in fact a groupoid, and hence that
is also a terminal object. Linearity amounts to requiring the objects be vector spaces and the arrows be vector space homomorphisms. Time translation symmetry would be the existence of a functor collapsing our category to a monoid. (At least, if I am not mistaken….)
Example: 1D wave equation
Let us illustrate this by considering the wave equation in 1 spatial dimension. (Which, again, using the reflection trick described in the last post, gives us a solution to the case with Dirichlet boundary conditions, and also then to the spherically symmetric wave equation in 3 spatial dimensions.) We first try to write down the free propagator in terms of the first-order formalism above. Recall Equation 7 for the solution to the homogeneous problem with initial data
Equation 17a
![]()
Now observe that also solves a homogeneous wave equation by linearity, and using that for a solution
, we have that the initial data for
is
. Hence we also have the expression
Equation 17b
![]()
The time translation symmetry tells us that the propagator . And Equations 17a,b tell us what the map
looks like. Now, going back up to Equation 16, we see that the vector representation of
only contributes to the second slot. So we have
Equation 18
![]()
and thus, the source part of the full solution to the inhomogeneous problem, would be
Equation 19
![]()
which perhaps is easier seen with a picture.
In Figure 2, we assume that we want to consider the initial value problem for Equation 16 (of equivalently, the inhomogeneous wave equation) in spatial dimension 1. Let us assume that the initial data is prescribed on the axis, and the initial data is identically zero. (In other words, we just want to consider purely the contribution coming from the source term, by setting the first term on the right hand side of Equation 15′ to be 0. In the picture, the green region represents the support of the source function
. Suppose we want to evaluate the solution
at the red dot, then Equation 19 tells us that
is given by the total integral of
in the blue, crosshatched region, which is bounded below by the initial time slice, and above by the null lines emanating from the red dot.
(In general, given a space-time point where now space can be arbitrary number of dimensions, we will write the two sets
which represents the future and past light rays emanating from our point. We see that they form two cones, intersection only at the vertex, in space-time. We will also denote the sets
to be respectively the future of
and past of
.)
Another thing to read off from the picture, and from Equations 18 and 19, is that the solution will not “settle down to a constant value” until the support set of the inhomogeneity
is completely contained in
. So in particular, if the support set for
has infinite spatial (or temporal) extent, then the solution will see “tail behaviour”. This supports the interpretation, using Duhamel’s Principle, of the right-hand side of the wave equation as a “source term”.
A quick detour to Penrose diagram
To better illustrate the global behaviour of wave equations, we will use the so-called (Carter-)Penrose diagrams for space-time. To study late time behaviour, we will need to think about points infinitely far into the future. And so we want a convenient way to represent the space-time and its infinite extent on a finite piece of paper. This requires some creative rescaling of space-time coordinates. Now, there are many ways to map the infinite to the finte. For example, consider the tangent function. It is a one-to-one mapping between the open interval and the entire real line. The Penrose diagram is a particular type of mapping. Now, if you look at any mapping from
, representing the 1+1 dimensional space-time, to a finite region, it would be necessary that some of the straight-lines be mapped to lines that are no longer straight. (If all straight lines are preserved, the mapping must be affine by the fundamental theorem of projective geometry.) So the question is what, if any, curves we want to keep “straight”. As we have already seen, what is most important in the study of wave equations is the null lines
; they, being characteristics of the partial differential equation, encode most of the phenomenon that we want to study. The Penrose diagram is a way of representing the 1+1 dimensional Minkowski space-time as a set with finite extent, such that the null lines (which are also the level sets of the
coordinate functions) are preserved.
Figure 3 shows some examples of Penrose diagrams. (a) is the Penrose diagram for the 1+1 dimensional Minkowski space, the boundaries of the diamond are precisely where the coordinate functions attain
values; we include the diagonal lines to indicate the level sets of
. (b) is the Penrose diagram for the
Minkowski half space; this also corresponds to the Penrose diagram for the 1+3 dimensional Minkowski space after we quotient out by the spherical symmetry. Notice that the central axis
is drawn slightly curved to illustrate the fact that non-null straight lines don’t necessarily map to straight lines in the compactified diagram. In (a) and (b) we see that the open circles and the dashed lines represent “idealized boundaries”, points that are not in fact in the space-time; in (b) the solid line for the central axis indicates the fact that boundary is in fact a physical region of the space-time. (c) is a representation of Figure 2 in compactified form; observe that the bottom boundary of the blue triangle is drawn slightly curved also. (d) shows the “domain of influence” of a source. In the picture the cyan ball indicates the space-time support set of the inhomogeneity. Under the assumption of either Dirichlet boundary condition (if we consider this as a 1+1 problem on the Minkowski half space) or symmetric reduction from a 1+3 dimensional problem, the pink region is the region in space-time where a priori, the contribution from the inhomogeneity to the solution of the wave equation can be non-zero. The larger the space-time extent of the inhomogeneity, the larger the domain of influence.
Nonlinear tails: an example
Be start by giving the motivation for the construction. Consider the following Penrose diagram, in which we study the future evolution of some wave equation in 1+1 dimensional space with Dirichlet boundary (which could be the symmetry reduction of a spherically symmetric wave equation in 1+3 dimensional Minkowski space).
The initial data is prescribed to be only non-vanishing on the blue line. So if we were to solve the purely linear homogeneous problem, using what we discussed in the previous post in the series, we see that the initial data splits into an incoming part, which approaches the central axis, bounces off, and then radiates to infinity; and an outgoing part which immediately radiates outward. And as we described before, following any future directed time like curve (which can be represented by a curve converging to the top open circle, the “future timelike infinity”) we will eventually escape the region influenced by this data.
Equivalently, we can consider the value of the solution at a point in the space time (here the red dot). Using the representation given by Duhamel’s principle, we see that for the inhomogeneous wave equation with possible source term, the solution can only be non-vanishing at the red dot if the initial data and any inhomogeneities have space-time support that intersects the crosshatched region. So for our linear homogeneous problem, where the initial data (blue curve) does not intersect the crosshatched set determined by strong Huygens’ principle, at the red dot the solution would vanish.
Now suppose that instead of an outside source, we impose a nonlinear self-interaction on the right hand side. That is, the source term is assumed to be generated from non-linear effects of the wave itself. To be slightly more precise, assume our equation takes the form
Equation 20
![]()
where is now only a function of the value of the displacement
, with
(these conditions ensure that the interaction is non-linear, and that waves won’t be spontaneously generated from 0 initial data; in other words, 0 is still a solution to the equation).
Now, because the interaction is nonlinear, if is small, we can expect, by Taylor expanding
, that the source term contribution would be
, which is even smaller. Thus intuitively we expect to be able to solve the equation using successive approximation, namely: we start with the given initial data
, we solve the homogeneous wave equation for it, and call the solution
. Then we run through the following iterative procedure:
- We are fed the solution from the previous iterate,
.
- We solve the inhomogeneous wave equation with prescribed source:
subject to the initial data
. Notice that on the right hand side, the source is given by
, with on the left hand side, we are solving for
. So we can directly obtain an integral expression using Duhamel’s principle.
- Feed
into this algorithm and repeat.
If this approximation scheme converges, then we will have that , the limiting function, will solve the nonlinear equation. (See Appendix for a bit more details.)
Let us consider for now the first iterate, which is enough to give a heuristic demonstration of why nonlinear tails are generic, and that the complete vanishing of waves within finite time that is possible in the linear system is very unstable.
Going back to Figure 4, suppose the initial data is precisely supported on the blue line. Then the 0th iterate, , just solves the linear homogeneous wave equation, and so its support set is represented by the cyan region, and the red dot lies outside of it. But now let us consider the first iterate. It solves the wave equation with source term
! So in particular, while it is true that the crosshatched region does not intersection the initial data (so there is no contribution to
at the red dot coming from the prescribed initial value
), the region does intersect the cyan set, on which our inhomogeneity does not identically vanish. And if the homogeneity is assumed to take only non-negative values (for example
), then it is necessary that
is nonvanishing at the red dot. Since the position of the red dot is more or less arbitrary, we see that at the level of the first iterate, we already expect some nonlinear tails.
Appendix: successive approximation and local existence of solutions for nonlinear problems
Observe that using Duhamel’s principle we have a very precise control of the rate of convergence. Consider the difference . It solves the wave equation
with 0 initial condition. The right hand side we again use Taylor expansion (assuming that is small), to replace by
But since is also small, we can approximate
by
. So by the assumption that
, we see that the source term is small. In particular, applying Duhamel’s principle again, using that the integral used to factor in the inhomogeneity is taken over a finite space-time region, we have that
where the constant depends only on the maximum time
. Using that
, we see that
is in fact controlled by a geometric series, and the series converges if
is bounded by an
such that
is less than, say, 1/100.



Thanks for the nice post! Wouldn’t time translation invariance make the groupoid a group?
Yes, if your process is reversible. Some evolution equations (the heat equation being the classic example) only admits forward time-evolution (in suitable Banach spaces yada yada yada), in which case the morphisms are not necessarily invertible. For these types of equations with time independent coefficients, the natural notion is the evolution semigroup (not group!), which since it has an identity (evolving by zero time) is just a monoid. In the case of hyperbolic/wave equations where the evolution is time-reversible, you are of course correct, that a monoid that is also a groupoid is in fact a group.
Very interesting! Is the categorical point of view on the solutions of a PDE laid out in some standard (or at least known among specialists) references on PDE theory?
Not that I know of. But if you back-track a bit, the semi-group point of view for time-independent evolution partial differential equations (linear and non-linear) are rather well known. A standard reference is Pazy’s Semigroups of linear operators and applications to partial differential equations for the linear case, and the Hille–Yoshida theorem is something most students of PDEs would recognize. A nonlinear analogue is the Crandall–Liggett theorem. You may want to look at some of the references in this Springer EOM article. Once you have semigroups, going from there to categories isn’t that huge a leap conceptually. For the most part, as far as I know, this type of structures is not very interesting to PDE specialists: the hard work is usually in showing that such a structure does exist (local wellposedness theory) for a given equation, and it seems difficult to actually make use of the more abstract notions from category theory. There is, however, probably a good functorial description of integrable PDEs, in view of the Darboux and Bäcklund transformations.
[…] So in particular will solve a 1+1 dimensional wave equation with source term on manifold (which we can partially complete by adding the axis and imposing the Dirichlet boundary condition there). This will be the starting point of our construction (recall the support propagation properties of the wave equation from Part IIb of this series). […]
Would the approach in Appendix work in 3D as well?
Yes, more or less. In the last step you will have to replace using an appropriate version of Duhamel for the 3D wave equation (and that in dimensions greater than 1, the correct norm to control is not that supremum norm of the function, but some version of L^2 Sobolev norm). The approach (a generic iteration method) works for many abstract semilinear evolution equations too.
Sorry for not saying it in the first place but I was wondering if some supremum-based norm such as C^k or Holder norm can be used in the iteration in higher dimensions. If not this is one more reason for necessity of Sobolev spaces.
Depends on what you mean, yes and no: Littman showed that you cannot get L^p -> L^p estimates for the wave equation unless either p = 2 or the spatial dimension is 1. If you try to do iterations in other spaces, you will generally have to compensate by giving up some derivatives. For example, if you look at fundamental solution of the wave equation in 1+3 dimensions, you see that to control the supremum norm of the solution at time t requires controlling the supremum norm of the solution and its first derivatives at time 0. As a side effect, that means if you naively iterate, to control, say, the supremum of the final solution at a time t, for the nonlinear problem, you need to control infinitely many derivatives for the data (and you also need that the sizes of the derivatives grow controllably), which immediately puts you in the situation of analytic function classes (C^\omega or Gevrey classes). In those cases, however, you might as well do the local existence theory using analogues of Cauchy-Kovalevskaya, and avoid dealing with fundamental solutions altogether.
But there are techniques in which you can use other L^p spaces and such in proofs, but not necessarily in the iteration step. For example, there is the well-known break-down criterion for solutions to nonlinear wave equations which states that as long as a certain supremum norm of the function stays bounded, the solution can be continued. But the local existence aspect for that proof is still tackled using Sobolev spaces. Another example where alternative L^p spaces are used in iteration is the case of subcritical semilinear wave equations, where one estimate the nonlinearity using Strichartz estimates, exploiting the fact that the time integral in the wave operator can be used to recover some of the lost regularity. You may want to consult the last section(s) of Keel and Tao’s “Endpoint Strichartz Estimates” for a discussion in that direction.