p154022_cumullect
p154022_cumullect
p154022_cumullect
Appendices 26
Appendices 52
Appendices 81
Appendices 107
4
Lecture Notes on Mathematical Methods 2022
Definition 1.1. A vector space V over a field F is a (possibly infinite) set of objects on which an
operation called addition and another called s-multiplication (multiplication by a scalar) are defined,
and that is closed under these operations. Therefore, any two elements u and v of V satisfy:
(a + b)(u + v) = (a u + a v + b u + b v) ∈ V
∀ a, b ∈ F; in what follows, F = R. Both the addition and s-multiplication operations are commuta-
tive and associative.. We will call elements of a vector space vectors.
Example 1.1. Rn , the set of all ordered n-tuples of real numbers, with addition defined as adding
entries with the same place in the n-tuple, and s-multiplication by λ defined as multiplying each entry
by λ, is perhaps the best-known and most important vector space.
of a set {eα ∈ V}, then that set is said to span, or to be a set of generators of, the vector space V.
If, furthermore, this set is linearly independent, in the sense that v = 0 =⇒ v α = 0, then it is a
basis of V. The number n of vectors in a basis defines the dimension of V, and we often write V n .
The (real, and unique!) coefficients v α are called the contravariant components of the vector v in
this basis. This one-to-one correspondence between V n and Rn can be represented by a n × 1 matrix:
1 Warning ! v and its components are different beasts and
v
v2 should never be confused. Byron and Fuller (BF) do not
v 7−→ . make this distinction clear enough. Also, the index on eα
..
identifies the vector, not a component of the vector.
vn
Example 1.2. The standard, or natural, basis Rn is the set {eα } (α = 1, 2, . . . , n), where each
n-tuple labelled by a value of α has 1 in the αth position and 0 in all other positions.
†
These notes generally follow the conventions set by the ISO (International Standards Organisation) for mathematical typography, with
one important exception: as in Byron and Fuller, vectors and tensors are in bold upright (u) instead of bold italic font (u). Sans-serif fonts
denote matrices, eg. M.
‡
Infinite linear combinations (series) require extra topological structure on V so as to allow the notion of convergence.
1
Lecture Notes on Mathematical Methods 2022
where the operations on the right are defined on W. One can define linear mappings on L (V, W), ie., one can
compose linear mappings into a linear map.
An important subset of the set of linear mappings is L(V, R) that contains all linear, real-valued functions on
a vector space. We say that it forms a space V ∗ dual to V. Since L(V m , W n ) has dimension m × n, V ∗ and V
have the same† dimension. The elements of V ∗ are called covectors, or linear functionals (in linear algebra), or
1-forms. One example would be definite integrals on the vector space of polynomials.
There comes following important definition:
Definition 1.3. If {eα } is a basis of a space V n , its unique dual basis (cobasis) in V ∗ , {ω α }, satisfies:
ω α (eβ ) = δαβ (α, β) = 1, . . . , n (1.2)
where δαβ is the Kronecker delta. The left-hand side is just classic multiplication of row-vectors with
column-vectors. Then a covector σ ∈ V ∗ is written σ = σα ω α , where the σα are the covariant
components of σ in this dual basis.
From this we derive the action of an element ω α of the cobasis of V ∗ on a vector v ∈ V:
2
Lecture Notes on Mathematical Methods 2022
Note that σ(v) = v(σ) = σα v α is basis-independent, but only if σ is referred to the cobasis of the basis in
which v is written. At this stage, there is no unique connection between σ and a vector in V. So, tempting as it is
to identify σα v α with the scalar product of two vectors, let us resist that urge.
For a given n-dimensional vector v, there exists a unique set of parallel (n-1)-dimensional hyperplanes that
can provide a pictorial representation of 1-forms. This is easiest when n = 2. Then a = σ1 v 1 + σ2 v 2 determines a
perpendicular to v with equation σ2 = a/v 2 − σ1 v 1 /v 2 . The lines generated by different a all have slope −v 1 /v 2 .
Definition 1.4. A differentiable manifold, or just manifold, M is a set of elements (“points”), all of
which have an open ball (or neighbourhood) around them in M , such that M can be entirely covered
by a union of possibly overlapping open (without boundary) subsets Ui , each mapped in a one-to-one
way to an open subset of Rn by a non-unique, differentiable coordinate map: x : Ui −→ Rn .
Each (Ui , x) forms a coordinate chart (local coordinate system), and an atlas is any collection of
charts that covers the whole M . Also, on any overlap Ui ∩ Uj ⊂ M , only charts (Ui , x) and (Uj , y)
for which the coordinate transformation y ◦ x−1 : Rn −→ Ui ∩ Uj −→ n
y R between them is (once)
x−1
differentiable are allowed.
The minimum number n of parameters—each a map xi : U −→ R (1 ≤ k ≤ n)—that uniquely
specify every point in U is its dimension.
Example 1.3. • Rn can be promoted to a manifold; it can be covered with just one coordinate chart, Cartesian
(standard) coordinates. Other charts are possible, eg. polar coordinates. to cover the manifold.
• A conical surface, even a semi-infinite one, can never be a manifold because of its tip.
• A vector space V can be made into a manifold that can be covered with one chart (V, Φ), where Φ maps
elements of V to their components in Rn in that basis. Conversely, however, a manifold is not in general a
vector space! On Earth’s surface, there is no meaning to adding the position of Toronto to that of London.
upper/lower half-plane. Then each point of any of the two submanifolds is in one-to-one correspondence
with some x ∈ R, with x < 1. To cover all of S 1 , we repeat with two submanifolds in correspondence with
x > 0 and x < 0, and an atlas with four charts has been constructed.
p
S 1 also has a local coordinate, θ, related to x by: θ = tan−1 (y/x) = tan−1 1/x2 − 1 . To avoid any
point being mapped to more than one value, θ must map to [0, 2π) in R .
An atlas can also be constructed for S 2 out of patches similar to those for S 1 for each of the Cartesian
±x > 0, ±y > 0 , and ±z > 0. Each point in each patch unambiguously maps to R2 .
On S 2 we could also use spherical coordinates θ and φ that map to he region of R2 : θ ∈ (0, π), φ ∈ [0, 2π),
with the poles removed since φ is undetermined there. More patches are needed to cover S 2 .
Notice that we have looked at S 1 and S 2 as being embedded in a higher-dimensional manifold, R2 and R3 .
Whitney’s embedding theorems guarantee that any smooth M n is a submanifold of Rm>2n , with stronger
results in restricted cases. Embedding curves and surfaces in, eg., R3 is great for visualisation, but we are
more interested in their intrinsic properties which should be independent of the embedding manifold.
Less technically, it is usually enough to view a manifold as a set that can be parametrised in a smooth way.
We can also view the curve Γ as embedded in a region U of a manifold M parametrised by coordinate functions
denoted collectively by x : U −→ Rn , with x ◦ Γ describing what the curve “looks like” in M . Then an alternate
expression for the velocity is, from definition 1.7:
n h
X i
−1
v(Γ,P) (f ) = dλ (f ◦ x ) ◦ (x ◦ Γ) λ0 = dλ (xν ◦ Γ) λ0
∂ν (f ◦ x −1
)
xν (P)
ν
X
=⇒ v(Γ,P) = d λ xν 0
(∂ν )P (1.6)
ν
where the index ν in the multidimensional chain rule runs over the number of local coordinates that specify each
point in M . The xν (λ) parametrise the curve Γ in M . Since f ◦ x−1 maps Rn to R, its derivatives behave like the
standard ∂xν f (xµ ), that is: ∂xν f P := ∂ν (f ◦ x−1 ) P .
We will interpret this important result a little later, after we have constructed the space where v(Γ,P) lives.
4
Lecture Notes on Mathematical Methods 2022
This definition rests on a very bold assertion, namely, that the velocities as defined above (definition 1.7) are in
fact vectors. For those who are interested, a proof that TP can be made into a vector space is in Appendix A.
Since the ∂ν in eq. (1.6) span the tangent space of M at P, it would be natural to think of them as basis vectors.
But are they linearly independent? Take f = xν , the coordinates on M ; then aµ ∂xµ xν P = aµ ∂µ (xν ◦ x−1 ) P =
aµ δµν = aν , where aν ∈ R. If aµ ∂xµ xν P = 0, aν = 0, which shows that the ∂ν do form a basis of the tangent
space. Thus:
To find the coordinate-basis tangent vectors, we freeze all the variables that parametrise the manifold, except
one that is varied to generate a coordinate curve whose tangent vector at a point P is the partial derivative with
respect to the parameter λ on which the coordinates xµ depend. Then the components of ∂µ at P are the derivatives
of x ∈ Rn with respect to the parameters of the manifold at P. An example should make this procedure clearer:
Example 1.4. On S 2 (embedded in R3 ), a point P is mapped into the spherical coordinates (θ, ϕ),
with θ 6= (0, π); P can also be described by the R3 coordinates (sin θ cos ϕ, sin θ sin ϕ, cos θ).
Freezing say, θ, generates a great circle on the sphere. Then these coordinates describe a circle of
radius sin θ at “colatitude” θ, and ∂ϕ is a coordinate-basis vector visualised in R3 by the vector with
components:
At each point on S 2 parametrised by (θ, ϕ), this is a vector tangent to the circle at colatitude θ.
Similarly, there is a spherical-coordinate-basis vector, ∂θ , tangent to a meridian going through that
same P, with components (cos θ cos ϕ, cos θ sin ϕ, − sin θ). ∂θ and ∂ϕ together form a basis for
vectors in the plane tangent to S 2 at P. These vectors do not live in S 2 ! Instead, any vector at a point
on S 2 lives in the R2 plane tangent to the point. Each point on S 2 has its own tangent plane.
Also, notice that ∂ϕ is nor normalised to 1, except at θ = π/2. In general, coordinate bases and
cobases are not normalised. But ∂ϕ̂ := sin1 θ ∂ϕ has components (− sin ϕ, cos ϕ, 0), which are the
components of the unit basis vector ϕ̂ in the standard basis; it is an element of a non-coordinate basis.
5
Lecture Notes on Mathematical Methods 2022
To find the components of df in the coordinate cobasis dual to the coordinate basis {∂µ } at a point P ∈ TP , recall
that the action of a 1-form (covector) on a coordinate-basis vector ∂ν outputs the corresponding component of the
1-form in that (as yet undetermined) cobasis: [df ]ν = df (∂ν ), which, from eq. (1.8), is the ordinary derivative of
f in the direction of the basis vector ∂ν , so ∂ν f . Now, taking f = xµ , the same argument immediately leads to:
which we recognise as the defining equation (1.2) for a cobasis, with eµ = ∂µ and ω µ = dxµ . Then, choosing
{∂µ } as basis for TP , we conclude that {dxµ } is the basis, dual to {∂µ }, of the cotangent space, TP∗ , dual to TP .
When written in a coordinate cobasis, σ = σα dxα is called a differential form.
If we think of f as a 0-form, the differential of f is the gradient 1-form df :
df = ∂µ f dxµ (1.10)
We recognise the well-known expression for the differential of a function in calculus, where it is taken to be a
scalar, a number. But df , interpreted as the infinitesimal change of f does not know in which direction this change
should be evaluated. Only when a vector is inserted in its input slot, as in eq. (1.8), can it output a number, the
change of f in the direction of the vector.
As for the usual calculus interpretation of dxµ as the difference between the components of two coordinate
vectors at infinitesimally close points, this is not valid on an arbitrary manifold, since dxµ , like all 1-forms at
a point, lives in the cotangent space, not the manifold, Only in Rn can one ignore with impunity this crucial
distinction between a base manifold and its tangent and cotangent spaces at a point.
vyν = ∂xµ y ν xP
vxµ (1.11)
What is remarkable about this transformation is that it is linear and homogeneous, even though the transforma-
tions beween (U1 , x) and (U2 , y) can be non-linear. Thus, in coordinate bases, the coefficients, ∂xµ y ν , in the
transformation law are the entries of the Jacobian matrix of the transformation evaluated at P. Conversely, if
vxν = vyµ ∂yµ xν , one shows, using the chain rule on partial derivatives, that v is unchanged by the transformation.
In general bases, transformations on components must be assumed homogeneous and linear, and take the form:
′ ′ ′
v α = v µ Lα µ = Lα µ v µ (1.12)
where the prime refers to the y coordinates in (1.11). This is the more traditional definition of a vector still in use
′
in physics. These two ways of writing v α are equivalent, but the second one is a matrix product. Bases transform
as:
′ ′
eµ = Lαµ eα′ = eα′ Lαµ (1.13)
The second expression in the equation, however, is not matrix multiplication, because the subscript of the basis
vector is a label for the vector, not for a component of this vector.
6
Lecture Notes on Mathematical Methods 2022
L being non-singular, and therefore invertible, the action of the inverse transformation L−1 is represented by:
′
v = L−1 v′ ⇐⇒ v µ = (L−1 )µν ′ v ν (1.14)
Do not confuse matrix and index notation! Whereas matrix notation is readily translated into index notation, the
reverse generally requires some rearrangement. This is because index notation does not care about ordering—one
of its virtues—but matrix notation most certainly does.
′ ′
Let {eµ } and {eµ′ } be two bases in V n , connected by eµ = eν ′ Lν µ , where the Lν µ are the coefficients of
′
the matrix L representing a linear transformation L. Let {ω α } and {ω α } be their two respective cobases in V ∗ .
′
Then, writing ω α = M αβ ′ ω β where the M α β ′ are the matrix coefficients of the corresponding transformation
′
M between the cobases, it can be shown (EXERCISE) that M is the inverse of L, ie. M αν ′ Lν β = δαβ in index
′ ′
notation and M = L−1 in matrix notation. This means that: ω α = Lα β ω β .
In the same way as for vector components, we can then obtain (EXERCISE) the transformation law of the
′
components σα of a 1-form σ. Since σ must be cobasis-independent, σα ω α = σβ ′ ω β yields:
σα′ = σµ (L−1 )µ α′ (1.15)
while the inverse matrix, M−1 = L, takes the components in the opposite direction.
The following table summarises all the possible transformations, both in general and in coordinate bases:
Care should be exercised when com-
′ ′
paring this table to the expressions
eα′ = eβ (L−1 )β α′ = eβ ∂α′ xβ eα = eβ ′ Lβ α = eβ ′ ∂α xβ given in §2.9 and in Box 8.4 of MTW
α′ α′ β α′ β α −1 α β ′ α β ′
v = L β v = ∂β x v v = (L ) β ′ v = ∂β x v ′ which refer to Lorentz transforma-
′ ′ ′ ′ ′
tions. In their potentially confus-
ω α = Lα β ω β = ∂β xα v β ω α = (L−1 )α β ′ ω β = ∂β ′ xα ω β ing but standard notation, the matrix
−1
σα′ = σβ (L ) α′β β ′
σα = σβ ′ L α = σβ ′ ∂α x β ′
with elements Lαβ ′ is actually the
′ inverse of the matrix with elements
σα v α = σβ ′ v β ′
Lβ α ; we prefer making this explicit
by writing (L−1 )αβ ′ .
Another word of caution: transformations in coordinate bases may well produce components in non-normalised
bases, even if one starts from a basis that happens to be normalised. This does not occur in the case of rotations
and Lorentz boosts, but it will when we transform from Cartesian to curvilinear coordinates.
Also, in a coordinate basis, we cannot call ∂µ f the components of the gradient vector, ∇f . They do not trans-
form as vector components, as can be seen by calculating ∂µ′ f in terms of ∂ν f using the chain rule (EXERCISE).
Definition 1.12. Contravariant tensors T ∈ T r of type (r, 0) are multilinear functions of r 1-forms:
T(σ1 , . . . , σr ) = σµ1 . . . σµr T(dxµ1 , . . . , dxµr ) = T µ1 ...µr σµ1 . . . σµr ∈ R (1.16)
Covariant tensors S ∈ Ts of type (0, s)s are real multilinear functions of s vectors:
S(u1 , . . . , us ) = uν1 . . . uνs S(∂ν1 , . . . , ∂νs ) = Sν1 ...νs uν1 . . . uνs (1.17)
Mixed tansors Q ∈ Tsr of type (r, s) are real multilinear functions of r covectors and s vectors:
Q(σ1 , . . . , σr , u1 , . . . , us ) = σµ1 . . . σµr uν1 . . . uνs Q(dxµ1 , . . . , dxµr , ∂ν1 , . . . , ∂νs )
= Qµ1 ...µr ν1 ...νs σµ1 . . . σµr uν1 . . . uνs (1.18)
7
Lecture Notes on Mathematical Methods 2022
T µ1 ...µr , Sν1 ...νs , and Qµ1 ...µr ν1 ...νs are the contravariant components of T, the covariant compo-
nents of S, and the mixed components of Q, respectively, in the chosen basis and cobasis.
Following the metaphor of tensors as machines, to output a number from a (r, s) tensor, one must
supply r 1-forms and s vectors as inputs, one for each slot.
Among important tensors are those whose components are completely symmetric in all their covariant (or
contravariant) indices, and those which are completely antisymmetric (skew-symmetric, alternating) in all their
covariant (or contravariant) indices.
A completely symmetric tensor of rank r in n dimensions has n+r−1 r = (n + r − 1)!/(n − 1)!r! independent
components. A skew-symmetric tensor has nr = n!/(n − r)!r! independent non-zero components.
In three dimensions, many physically relevant tensors are symmetric, eg. examples 1.6, 1.7 and 1.8 (moment
of inertia, electrical polarisation, multipole moment) in B&F, as well as the Maxwell stress tensor. Antisymmetric
3-d rank-2 tensors are not usual, although I will argue toward the end of the chapter that in three dimensions a
magnetic field is more naturally described by an antisymmetric (0, 2) tensor than by a vector.
In four dimensions, we also have symmetric tensors, such as the important energy-momentum tensor which
carries all the information about the energy and momentum density at a point, plus the flux of these quantities at
that point. And there is the famous antisymmetric (0, 2) Faraday electromagnetic field tensor F.
It can be useful to symmetrise or skew-symmetrise a general (r, 0) or (0, s) tensor. To symmetrise the compo-
nents of a (0, s) tensor T, construct:
1 X
T(µ1 ...µs ) = T (1.24)
s! π π(µ1 ...µs )
with round brackets around symmetric indices, and where the sum runs over all permutations π of µ1 . . . µs .
Contravariant components are symmetrised in the same way.
To antisymmetrise the components of a (0, s) or (r, 0) tensor T, construct:
1 ν1 ...νs
T[µ1 ...µs ] = δ Tν ...ν (1.25)
s! µ1 ...µs 1 s
j ... j
with square brackets around antisymmetric indices, and the general permutation symbol, δi11 ... iss , defined as:
+1 j1 . . . js an even permutation of i1 . . . is
−1 j . . . j an odd permutation of i . . . i
j ... j 1 s 1 s
δi11 ... iss := (1.26)
0 j1 . . . js not a permutation of i1 . . . is
0 j = j or i = i for some k, l
k l k l
where even/odd means an even/odd number of transpositions (switches) of two indices. The permutation symbol
is seen to be antisymmetric in its upper and lower indices.
s! is the number of terms in all these summations, ie. the number of permutations of the indices of the tensor.
The normalisation factor 1/s! ensures consistency in the event that the Tµ1 ...µs should already be symmetric or
skew-symmetric. A simple example is that of a (2, 0) tensor:
1 µν 1
T µν = (T + T νµ ) + (T µν − T νµ ) ≡ T (µν) + T [µν]
2 2
9
Lecture Notes on Mathematical Methods 2022
In traditional treatments, this transformation law actually defines a tensor. Scalars ((0, 0) tensors) remain invariant;
and we know how the components of vectors and 1-forms transform. What about, say, those of a (2, 0) tensor?
′ ′ ′ ′ ′
T α β = T µν Lα µ Lβ ν = Lα µ T µν L e β′ ⇐⇒ T′ = L T L e
ν
where Le is the transpose of L. Sometimes, as with 3-dim rotations, Le = L−1 ; sometimes, as with Lorentz boosts,
e = L. Tensors of rank 2 (2, 0), (0, 2), (1, 1) can be represented by n × n-dim matrices T, where n is the
L
dimension of the spaces V and V ∗ on which they are defined.
An immediate consequence of eq. (1.27) is that a tensor that is zero in a basis will remain zero in any other
basis. Thus, any equation made of tensors (or components) that is valid in one basis must hold in any other basis.
In the older view of tensors defined by transformations, an object may have tensor character under certain
transformations, but not others. For instance, 4-dim tensors might owe their tensor character to how they transform
under Lorentz transformations, while 3-dim tensors might be tensors only under rotations.
The transformation rules can always be used to establish whether an object is a tensor. For instance, on a space
of dimension n, the Kronecker delta, with components δνµ , is represented by the n × n identity matrix. It is a
mixed rank-2 tensor. Indeed, from the transformation law, eq. (1.27):
′ ′ ′ ′
δµ ν ′ = Lµ λ (L−1 )ρ ν ′ δλρ = Lµ λ (L−1 )λ ν ′ = Iµ ν ′
which are the components of the identity matrix. Here we learn that there is something more to δµν than just being
a tensor: its components remain the same under changes of basis!
It is skew-symmetric in its n indices, with ǫ1...n = +1, where the indices are in ascending order. In
pseudo-Riemannian manifolds, it is traditional to use ǫ0...n−1 , the 0 index corresponding to time.
The determinant of a n × n matrix L is a product of its elements antisymmetrised with respect to rows (or
columns):
det L = ǫν1 ...νn Lν11 · · · Lνnn (1.28)
If the Levi-Civita symbol is to be a tensor, the transformation laws on its components demand that:
′ ′
1 = ǫ1...n = ǫν1′ ...νn′ Lν1 1 · · · Lνn n = det L
This is the case when L is a 3-dim rotation or a Lorentz-boost matrix, under which ǫµ1 ...µn , like δµ ν , is invariant.
We shall discover a little later how the general Levi-Civita tensor can be constructed.
Fortunately, often we can avoid using the transformation law (1.27) if we build tensors from other objects
known to be tensors. The following section presents some important examples.
10
Lecture Notes on Mathematical Methods 2022
T(. . . , dxγ , . . . , ∂γ , . . .) = T ... γ ... ... γ ... . . .⊗∂µi−1 ⊗∂µi+1 ⊗. . . . . .⊗dxνj−1 ⊗dxνj+1 ⊗. . . (1.29)
In terms of components, one just makes a contravariant index µ the same as a covariant index, ν, by multiplying
the component by δνµ , thus forcing a summation over these indices.
For instance, consider T ∈ T11 . The contraction of T = T α β ∂α ⊗ dxβ is a scalar, called its trace:
When contracting tensors of type higher than 2, it is important to specify which indices are being contracted. Thus,
the tensor T µν λ ∂µ ⊗ ∂ν ⊗ dxλ has two possible contractions: the vectors T µν µ ∂ν and T µν ν ∂µ .
The correspondence will be unique if we demand that g be symmetric, because then g(u, ) = g( , u). Inserting
vectors in the two input slots yields the number: g(u, v) = gµν uµ v ν . In effect, g may be thought of as a map
from V to its dual space! Once defined, it establishes a unique correspondence between a vector u ∈ V and a
1-form ũ ∈ V ∗ .
Definition 1.17. The inner product of two vectors u and v, < u, v >, can now be defined as:
In a coordinate basis, gµν = < ∂µ , ∂ν >, which is just the naı̈ve scalar product of the two basis vectors
∂µ and ∂ν . In a general basis, gµν = < eµ , eν >.
g(u, u) = gµν uµ uν is called the norm of u. If it is positive (negative) ∀ u, we say that g is positive
(negative) definite. But if g(u, u) = 0 for some non-zero vector (null vector) u, then g is indefinite.
g must be invertible (ie. det g 6= 0) , and we can always write: u = g−1 g(u) = g−1 (ũ). Now g−1 must take
1-forms to vectors, which means it must be a (2, 0) tensor g−1 = (g−1 )µν ∂µ ⊗ ∂ν . Then:
u = uµ (g−1 )αβ ∂α ∂β (dxµ ) = uµ (g−1 )αβ ∂α δµ β = uµ (g−1 )αµ ∂α
As will be justified soon, we identify (g−1 )µν with the contravariant components of g, gµν , and, comparing with
u = uα ∂α , we conclude that uµ = gµν uν , uµ being thought now as the contravariant components of the 1-form.
In that sense u and ũ can both have covariant and conravariant components.
These mappings between V and V ∗ can be applied to any tensor T ∈ T ; in other words, g may be used to
convert any contravariant index of a given tensor into a covariant one (“lowering the index”), while g−1 may be
used to convert any covariant index of a given tensor into a contravariant one (“raising the index”). Thus, we say
that the inner product sets up an isomorphism between a vector space and its dual. Because of this connection, we
also have: ∂µ = gµν dxν . One tensor can have all-contravariant, all-covariant, or mixed components! In particular:
Since, as we have seen, δµν is basis-independent, so is g µ ν , unlike gµν and g µν . But g µλ gλν = g µ ν = δµν , which
justifies our earlier assertion that gµν are the components of g−1 . On a n-dim space, gµ µ = δµµ = n.
If δµν are the components of the identity matrix, I, δµν = gµρ δρν = gµν will not in general be the entries of I.
A final word of caution: we always wrote our matrices as Lµν , with the left index a row index. Why do we
not write the matrix of g’s components the same way? Because Lµν is a transformation between two bases in V n ,
whereas gµν transforms a basis in V n to its dual basis. For instance, in R3 , the basis dual to {e1 , e2 , e3 } cannot be
reached by any combination of rotations and translations. Also, Lµν is not a tensor component, but gµν is.
The result, (∆s)2 = (x1 − x2 )2 + (y1 − y2 )2 + (z1 − z2 )2 = (∆x)2 + (∆y 2 ) + (∆z)2 , is recognised
to be the “Pythagorean” distance squared between two points: |x1 − x2 |2 .
Example 1.8. In R4 , let xi (i = 1, 2) be two position vectors with (cti , xi , yi , zi ) as contravariant and
(−cti , xi , yi , zi ) as covariant components. Then take the indefinite η ≡ g with matrix representation:
−1 0 0 0
0 1 0 0
ηµν =
0 0 1
0 = diag (−1, 1, 1, 1)
0 0 0 1
ROn
bp R b pthe distance between two points λ = a and λ = b on a curve
general manifolds,
µ
parametrised by λ is given
µ ν
by a g(v, v)dλ = a gµν dλ x dλ x dλ, where v is the velocity vector and x are the coordinates describing
the curve on the manifold. The metric, or line element, is said to define the geometry of a manifold. Two manifolds
of the same dimension can have different geometries, eg. R4 with a positive-definite (∆s2 > 0) metric is not the
metric of 4-dim “flat” spacetime of special Relativity.
Quite often, we will wish to work in bases other than coordinate bases. The formal properties of g that we have
reviewed still hold, but its covariant and contravariant components can be different, even in the same coordinates.
12
Lecture Notes on Mathematical Methods 2022
Definition 1.19. A basis {eµ } such that g(eµ , eν ) = ±1 when µ = ν and 0 otherwise is said to be
orthonormal. A useful notation to distinguish it from a coordinate basis is {eµ̂ }, extending the usual
definition of orthonormality which admits only +1 and 0, and useful in the case of indefinite metrics.
Let n+ (n− ) denote the number of diagonal elements g(eµ̂ , eµ̂ ) equal to +1 (−1). The signature of
the metric is defined by s = n+ − n− . Since n+ + n− = n, the dimension of the space, we also have
s = n − 2n− , and det g = (−1)n− . n+ and n− are basis-independent, and so is the signature.
The sign of the signature of an indefinite metric is arbitrary and is set by convention, which can be a source of
confusion. Example 1.8 sets s = +2, a nice choice when spatial indices are often raised/lowered. In more general
spacetimes, s = −2 is often used (but not always. . . see Misner, Thorne and Wheeler’s Gravitation ). Thus, beware!
Definition 1.20. A n-dim space endowed with a metric of signature ±n is called Euclidean. If
n− = 1 (or n− = n − 1), the space is pseudo-Euclidean, or Lorentzian (aka Minkowski when
n = 4). Example 1.8 has a Minkowski metric in four-dimensional space.
Thanks to the metric, we recover the vector gradient of a function defined in calculus. You may have noticed
that throughout our discussion of manifolds and tangent spaces, no mention was made of an inner product, because
none was needed—until now. A metric g pairs the 1-form df with a vector, ∇f ; indeed, from eq. (1.30):
< ∇f, v > = g(∇f, v) = gµν (∇f )µ v ν = (gµν ∂ µ f ) v ν = (∂ν f ) v ν = [df ](v) (1.33)
where v is an arbitrary vector, and the components of ∇f in a coordinate basis are given by: (∇f )µ = gµν ∂ν f .
Only in a Euclidean metric with a standard basis are the components of ∇f the same as those of df .
Example 1.9. In Minkowski spacetime with coordinates (ct, x1 , x2 , x3 ) and metric ηµν = diag(−1, 1, 1, 1):
There is something interesting about the determinant of the metric which we find by writing the transformation
′
law: gµν = ∂µ′ xα gαβ ∂ν ′ xβ , as a matrix equation, and taking the determinant. Defining g = det gαβ , we obtain:
2
∂x
g′ = g (1.34)
∂x′
where |∂x/∂x′ | is the Jacobian of the transformation matrix from x to x′ coordinates. Then g is not invariant!
Definition 1.21. A quantity that has extra powers of |∂x/∂x′ | as factors in its transformation law in
′
addition to the usual ∂µ′ xα and/or ∂α xµ factors is called a tensor density. Thus, g is a scalar density.
This might seem no more than an exotic property until we consider the n-dim volume element as usually written in
an integral. This, as we know from calculus, transforms as : dn x′ = |∂x′ /∂x|dn x (note the position of the prime
in the Jacobian!), so is not invariant. As a result, the volume integral of a scalar function is not
p invariant, yet there
n
should be no memory of the integration variables left after integrating. But if we transform |g| d x, we obtain:
p ∂x p ∂x′ n p
|g ′ | dn x′ = |g| d x = |g| dn x
∂x′ ∂x
Rp
which is seen to be a scalar! Then integrals written as |g|f (x)dn x are invariant. This concept of tensor density
as a notational device has been widely used in General Relativity, although post-1970 literature largely dispenses
with it when p-forms are involved. Indeed, later in section 1.5.2, we shall introduce a deeper definition of the
volume element.
13
Lecture Notes on Mathematical Methods 2022
In general, the wedge product can be used to construct a simple (or decomposable) skew-symmetric
covariant (0, p) tensor out of p 1-forms:
σ1 ∧ . . . ∧ σp = δµ1 1......pµp σµ1 ⊗ · · · ⊗ σ µp (1.35)
Applied to a n-dimensional cooordinate tensor-product cobasis dxρ1 ⊗ · · · ⊗ dxρp , where 1 ≤ ρ1 < . . . <
ρ ... ρ
ρp ≤ n, this becomes dxρ1 ∧ · · · ∧ dxρp = δµ11 ... µpp dxµ1 ⊗ · · · ⊗ dxµp , so that:
ρ ... ρ ρ ... ρ
[dxρ1 ∧ · · · ∧ dxρp ](∂ν1 , . . . , ∂νp ) = δµ11 ... µpp δµ1 ν1 · · · δµpνp = δν11 ... νpp(1 ≤ ρ1 < . . . < ρp ≤ n)
Vp ∗ (1.36)
(0, p) skew-symmetric tensors live in a space denoted by (V ), whose elements are also called p-forms (p is
traditionally used instead of s); very often, V ∗V= Rn .
Thus, from {dxρ } (1 ≤ ρ ≤ n) a basis of p (V ∗ ) can be constructed which contains n!/(p!(n − p)!) indepen-
dent, non-zero elements. In particular, a n-form on a n-dimensional space is a one-component object, a multiple
of the unique basis element, dx1 ∧ dx2 ∧ . . . ∧ dxn , with indices in increasing order. Skew-symmetry forces the
maximum rank of a non-trivial p-formVin n dimensions V to be n (why?).
p q
The
V p+q exterior product of a basis of and a basis of is a basis, dxρ1 ∧ . . . ∧ dxρp ∧ dxρp+1 ∧ . . . ∧ dxρp+q ,
of , again with indices in increasing order, and p + q ≤ n. V V
Then we construct a (p + q)-form out of the antisymmetrised tensor product of σ ∈ p and τ ∈ q :
µ ...µ ν ...νq
[σ ∧ τ ](uρ1 , . . . , uρp+q ) = δρ11...ρp+q
p 1
σ(uµ1 . . . uµp ) τ (uν1 . . . uνq ) µ1 < µ2 . . . < µp , ν1 < . . . < νq
(1.37)
µ ...µ ν ...νq
(σ ∧ τ )ρ1 ...ρp+q = δρ11...ρp+q
p 1
σµ1 ...µp τν1 ...νq µ1 < µ2 . . . < µp , ν1 < ν2 . . . < νq
The exterior product, in contrast to the vector (“cross”) product of vector analysis which it generalises, is
associative: σ ∧ (τ ∧ θ) = (σ ∧ τ ) ∧ θ.
Another very important property of the exterior product of a p-form and a q form is that:
σ ∧ τ = (−1)pq τ ∧ σ (1.38)
ν ...ν µ ...µp µ ...µ ν ...νq
This follows directly from eq. (1.37) by noting that it takes pq transpositions to get δρ11...ρp+q
q 1
into δρ11...ρp+q
p 1
.
It means that the exterior product commutes except when both forms have odd rank.
Eq. (1.37) is easier to use than it might appear. Here are three examples:
Example 1.10. Some people believe that we live in an 11-dimensional world. Let us work out one
component of the 3-form that is the exterior product of a 2-form, σ, and a 1-form, τ :
µνλ
(σ ∧ τ )11,3,6 = δ11 36 σµν τλ µ < ν
3 6 11 3 11 6 6 11 3
= δ11 3 6 σ36 τ11 + δ11 3 6 σ3 11 τ6 + δ11 3 6 σ6 11 τ3
= σ36 τ11 − σ3 11 τ6 + σ6 11 τ3
Example 1.11. In two dimensions, the exterior product of two 1-forms, σ1 and σ2 , is:
σ1 ∧ σ2 = (σ 1 1 dx1 + σ 1 2 dx2 ) ∧ (σ 2 1 dx1 + σ 2 2 dx2 )
= σ 1 1 σ 2 2 dx1 ∧ dx2 + σ 1 2 σ 2 1 dx2 ∧ dx1 = (σ 1 1 σ 2 2 − σ 1 2 σ 2 1 ) dx1 ∧ dx2
= (det S) dx1 ∧ dx2
where S is the 2 × 2 matrix whose two rows are the components of σ1 and σ2 , respectively.
14
Lecture Notes on Mathematical Methods 2022
1 2 1 3 2 3
V 2 In three dimensions, {dx ∧ dx , dx ∧ dx , dx ∧ dx } forms a basis of the space
Example 1.12.
of 2-forms, (V). Therefore, the most general (not necessarily simple!) 2-form can be written as:
1 V
τ = τ12 dx1 ∧ dx2 + τ23 dx2 ∧ dx3 + τ31 dx3 ∧ dx1 = τµν dxµ ∧ dxν ∈ 2 (1.39)
2
The summation on the right of the second equality is now unrestricted.
Three-dimensional simple 2-forms σ1 ∧ σ2 , however, have the coordinate form (EXERCISE):
With eq. (1.35), the output (a number!) resulting from inputting u1 , . . . , up into σ1 ∧ . . . ∧ σp is:
σ1 ∧· · ·∧σ p (u1 , . . . , up ) = δµ1 1......µ
p
p
σµ1 ⊗· · ·⊗σ µp (u1 , . . . , up ) = ǫµ1 ... µp σµ1 (u1 ) · · · σµp (up ) = det [σ i (uj )]
(1.43)
i i i µ
ie. the determinant of the p × p matrix S whose entries are: S j = σ (uj ) = σµ uj , with µ running from 1 to n.
Example 1.13. For a 3-dim V ∗ of which the 2-forms dxi ∧ dxj are basis elements, we have:
ui uj
dxi ∧ dxj (u, v) = dxi (u) dxj (v) − dxj (u) dxi (v) = vi vj
In Rn with Cartesian coordinates, we interpret this (up to a sign—see 1.5.2 below!) as the area of the
parallelogram whose defining sides are the projections of u and v on the xi -xj plane.
ν ...ν
Example 1.14. There is another useful definition of the permutation symbol, δµ11 ...µnn , equivalent to
the one given by eq. (1.26), and given by eq. (1.36):
δµν11...ν ν1 νn
...µn = dx ∧ . . . ∧ dx (∂µ1 , . . . , ∂µn )
n
Another application provides an easy test for the linear independence of p 1-forms: if their wedge product is
different from zero, those p 1-forms are linearly independent. If they were not, one of them at least could be written
as a linear combination of the others and antisymmetry would force the wedge product to vanish. Conversely, if
the p 1-forms are linearly independent, σ1 ∧ · · · ∧ σp cannot vanish.
Example 1.15. In general a p-form F is not simple. If it is, then the 2p-form F ∧ F must vanish by
antisymmetry. But to what extent does the converse hold?
V
Take F ∈ 2 . When n = 2, F = F12 dx1 ∧ dx2 is simple by inspection, with no need to invoke
F ∧ F = 0.
V
For n = 3, F ∧ F still trivially vanishes for F ∈ 2 . That vanishing will lead to F being simple, but
the argument is a little more involved. We can always write:
F = F12 dx1 ∧ dx2 + F13 dx1 ∧ dx3 + F23 dx2 ∧ dx3 ≡ σ + τ ∧ dx3
where σ is a 2-form on the 2-dim subspace, spanned by dx1 ∧ dx2 , of the 3-dim
V space, and τ is a
1-form on the same 2-dim subspace. Then σ is simple, ie., σ = α ∧ β (α, β ∈ 1 ) because n = 2.
Thus, F ∧ F = σ ∧ σ + 2σ ∧ τ ∧ dx3 = 0. Also, σ ∧ σ = 0 because σ is simple. Then
α ∧ β ∧ τ ∧ dx3 = 0. But the only possible linear dependence between the four 1-forms is between
α, β and τ since none of them depends on dx3 . Therefore, λ1 α + λ2 β + λ3 τ = 0.
If λ3 = 0, β is a multiple of α, so σ = 0, leaving F = τ ∧ dx3 , a simple form. If λ3 6= 0, τ = aα + bβ,
and: b
F = α ∧ β + (aα + bβ) ∧ dx3 = (α + β) ∧ (β + adx3 )
a
which is simple. Thus, 2-forms on 3-dim space are always simple! EXERCISE: A 2-form in n = 4 is
simple if, and only if, F ∧ F = 0. Also, when F ∧ F 6= 0 in n = 4, F can be written (EXERCISE) as
the sum of two simple 2-forms. These statements hold whether or not a metric has been introduced.
δν1µ1 · · · δν1µn
.. ..
ǫν1 ... νn ǫµ1 ... µn = (−1)n− δµν11...ν
...µn = (−1)
n n−
. .
δνnµ1 · · · δνnµn
(1.45)
1
ǫν1 ... νp νp+1 ... νn ǫµ1 ... µp νp+1 ... νn = (−1)n− δµν1 ...µ
... νp
(unrestricted sums)
(n − p)! 1 p
In a Euclidean 3-dim space with an orthonormal metric, n− = 0, and the expanded product has six terms.
When contracted over the last or first indices, we obtain (EXERCISE): ǫijk ǫlnk = δi l δj n − δj l δi n . Other
expressions for the product of Levi-Civita tensors in a 4-dim Minkowski space can be found in MTW, pp. 87-88.
. . . ∧ duνn−p , where we introduce the compact notation: |µ1 . . . µp | ≡ µ1 < . . . < µp , and with
components:
1
(⋆σ)ν1 ...νn−p =
⋆ σµ ...µ ǫµ1 ...µp ν1 ...νn−p = σ |µ1 ...µp | ǫµ1 ...µp ν1 ...νn−p (1.47)
p! 1 p
The Hodge dual of a p-form is a pseudo-form, and vice-versa. It can be shown that, given a mostly
positive metric g, ⋆⋆σ = (−1)n− (−1)p(n−p) σ. So Hodge duality is idempotent in Euclidean spaces
⋆ ⋆
(n− = 0) of odd dimension, such as R3 . In 4-dim Minkowski space (n− = 1), it is idempotent only
on 1- and 3-forms.
†
Here, the meaning of “dual” has no relation to its other use in “dual” space or basis.
17
Lecture Notes on Mathematical Methods 2022
One immediate application of eq. (1.47) is that the n-dim volume form is the Hodge dual of the 0-form 1:
⋆ 1
⋆ = ǫ|µ1 ...µn | duµ1 ∧ · · · ∧ duµn
p
= |g| du1 ∧ · · · ∧ dun
A very important consequence of the fact that ⋆⋆σ = ±σ is that a p-form and its Hodge dual contain
⋆ ⋆
exactly the same information! Thus, “dualising” a p-form (or an antisymmetric contravariant tensor) can remove
some (or all!) the redundancy due to anrisymmetry while preserving its information. For instance, in 4-dim
Minkowski space, a 4-form with components σµνλρ is dual to a pseudo-0-form, so one independent number instead
of 44 = 256. Or a 3-form with a ppriori 43 = 64 components can be Hodge-dualised to its dual pseudo-1-form
whose four components are (up to |g|) the only independent components of the 3-form.
and ⋆T3 = T 12 , so that the 1-form dual to T contains only the three independent components of T.
⋆
summations V are unrestricted), in which case the components are not changed—they are just re-allocated to basis
elements of n−p . There are corresponding expressions for Hodge-dualising coordinate bases or the components
of contravariant tensors, as illustrated by the above example.
Example 1.18. If σ and τ are 3-dim 1-forms, the 2-form: σ ∧ τ = (σ2 τ3 − σ3 τ2 ) dx2 ∧ dx3 +
(σ3 τ1 − σ1 τ3 ) dx3 ∧ dx1 + (σ1 τ2 − σ2 τ1 )dx1 ∧ dx2 has as its Hodge dual on a space with metric
g the pseudo-1-form:
p
⋆(σ ∧ τ ) =
⋆ |g| (σ2 τ3 − σ3 τ2 ) dx1 + (σ3 τ1 − σ1 τ3 ) dx2 + (σ1 τ2 − σ2 τ1 ) dx3
If σ corresponds to the vector u and τ to v via the metric, this says that: ⋆(u ∧ v) = u × v, or,
⋆
with eq. (1.47), (u × v)µ = 21 gµρ ǫνλρ (uν v λ − uλ v ν ) = gµρ ǫρνλ uν v λ . So when calculating a vector
product, one is implicitly taking a Hodge dual, the only way that the result can be a pseudo-vector.
It is easy to recover all the relations of vector analysis in Cartesian R3 . For instance:
u · (v × w) = ǫµνρ uµ v ν wρ
= wρ ǫρµν uµ v ν (cyclic permutation of indices on ǫ)
= w · (u × v).
Definition 1.27. A (r, s) tensor field T(p) on a n-dim manifold M n is a function of points p ∈
M n whose components T (p) = T(dxν1 , . . . , dxνr , ∂µ1 , . . . , ∂µs ) are real-valued differentiable
functions of coordinates on the manifold.
Examples: the coordinate vector field ∂µ , the gravitational and electric fields, the metric tensor with
components < ∂µ , ∂ν > in a coordinate basis.
18
Lecture Notes on Mathematical Methods 2022
How do we describe the change of a tensor field at a point? More precisely, how do we differentiate it? We
already know from section 1.2 how to take the directional derivative of a (0, 0) tensor, ie. a function. On a “flat”
(without curvature) manifold, directional derivatives of tensor-field components can be calculated in the same way.
For general (r, s) tensors, however, because of point-dependent bases, defining differentiation requires extra
structure, called a connection, or covariant derivative. Raising this important issue is like opening Pandora’s box
(aka can of worms), because there are a lot of ways to construct such a connection. A few, however, have gained
favour as “natural”. Here we only discuss a particular type of differentiation that offers a neat unification of the
ideas of gradient, divergence and curl in vector calculus, and for which a connection is actually not needed.
p of the 3-dim
By analogy with tensor algebra results, we can recover the contravariant components
curl of a vector, but only in Cartesian coordinates! Only in those coordinates is |g| = 1, with
covariant and contravariant components the same.
As we know all too well, the vector components of the curl of a vector in curvilinear coordinates
can be quite complicated; this is largely due to our insisting on working with objects which are less
natural. Exterior derivatives do not involve raising indices with a metric, and so are more natural.
It is interesting that, in vector calculus with Cartesian coordinates, we could write σ as A · dx, with
A a corresponding vector. Then the right-hand side of eq. (1.51) would correspond to ∇ × A · dS,
where dS is a surface element with Cartesian components dy ∧ dz, dz ∧ dx, and dx ∧ dy. Then we
could write d (A · dx) = ∇ × A · dS.
Example 1.22. Here is an intriguing example: the exterior derivative of a pseudo-2-form τ in R3 with
some metric g. Since this will be a pseudo-3-form, we expect it to be a one-component object. Indeed:
dτ = (∂u τvw du) ∧ dv ∧ dw + (∂v τwu dv) ∧ dw ∧ du + (∂w τuv dw) ∧ du ∧ dv
(1.53)
= (∂u τvw + ∂v τwu + ∂w τuv ) du ∧ dv ∧ dw
Now, in three-dimensions τ can be viewed as the Hodge dual, τ = ⋆σ, of the 1-form σ = σu du +
⋆
σv dv + σw dw. In terms of components, τµν = ǫµνλ σ λ . Inserting and then taking the Hodge dual of
p
the last expression, using ⋆(du ∧ dv ∧ dw) = ǫ123 = (−1)n− / |g| from section 1.5.3, gives:
⋆
1 p
(−1)n− ⋆d ⋆σ = p ∂µ ( |g| σ µ )
⋆ ⋆ (1.54)
|g|
Definition 1.28. Extending to n dimensions, we call the right-hand side the divergence, div B, of
the n-dim vector B with components B µ = σ µ . It holds in any coordinates in a metric-endowed space.
(up to a sign!) as the codifferential operator, δ. We quote without proof the relation between them: When
acting on a p-form in a Euclidean manifold, δ σ = (−1)n(p+1)+1 ⋆d ⋆σ, and δ σ = (−1)n(p+1) ⋆d ⋆σ in a
⋆ ⋆ ⋆ ⋆
pseudo-Euclidean manifold. Actually, these expressions happen to hold also in a Riemannian (curved) or pseudo-
Riemannian manifold!
20
Lecture Notes on Mathematical Methods 2022
Like the exterior derivative, the codifferential operator is nilpotent. Indeed, δ2 = ⋆d ⋆⋆d ⋆ = ±⋆d2 ⋆ = 0.
⋆ ⋆ ⋆ ⋆ ⋆ ⋆
Definition 1.29. We define the divergence of any p-form: div σ := −δσ = (−1)n(p+1)+n− ⋆d ⋆σ. ⋆ ⋆
This ensures consistency between eq. (1.54) and the conversion between ⋆d ⋆ and δ. We extend eq.
⋆ ⋆
From eq. (1.54) follows the definition of the 3-dim Laplacian of a scalar function f in coordinates ui :
1 p 1 p
∇2 f = p ∂i |g|∂ i f = p ∂i |g| g ij ∂j f (1.56)
|g| |g|
1.6.2 Laplace-de Rham operator, harmonic forms, and the Hodge decomposition
It is not hard to show that it reduces to the negative of the Laplacian operator of vector analysis, ie. ∆ = δ d =
−⋆d ⋆d = −∂i ∂ i = −∇2 , when acting on 0-forms on Euclidean R3 with Cartesian coordinates. We shall define
⋆ ⋆
∇2 so that ∇2 = −∆ when acting on any p-form in Euclidean R3 equipped with a standard basis.
Example 1.23. For instance, let it act on a 1-form σ in Euclidean R3 . That is, take ∆σ = ⋆d ⋆dσ −⋆ ⋆
d ⋆d ⋆σ using the conversion formula between δ and ⋆d ⋆. Using eq. (1.52), the first term is the
⋆ ⋆ ⋆ ⋆
curl of a curl, whereas the second is the gradient of a divergence. Thus, we recover the expression
well-known from vector calculus: ∇2 A = ∇(∇ · A) − ∇ × ∇ × A, where A is the vector associated
with the 1-form σ.
When acting on functions (0-forms) in Minkowski space, the Laplace-de Rham operator is related to the
d’Alembertian operator := ∂µ ∂ µ : ∆ = − . This defines the d’Alembertian of any p-form in Minkowski space.
Definition 1.32. A closed form is one whose exterior derivative vanishes. A p-form that can be written
as the exterior derivative of a (p-1)-form is said to be exact.
Clearly, Poincaré’s lemma states that an exact form is closed. But is a closed form exact, ie. if dσ = 0, does
it follow that σ = dτ , with τ uniquely determined? The answer is no, if only because one can always add the
exterior derivative of an arbitrary (p - 2)-form θ to τ and still satisfy dσ = 0. Also, the converse of Poincaré’s
lemma (not proved) states that only in a submanifold in which all closed curves can be shrunk to a point does
dσ = 0 entail the existence in that submanifold of a non-unique (p - 1)-form whose exterior derivative is σ. In
topology, we say that the submanifold must be simply connected (eg. no doughnuts!).
We quote without proof an important result of Hodge: On finite-volume (compact) manifolds without bound-
aries, such as S n , or on a torus, ∆σ = 0 if, and only if, dσ = 0 and d⋆σ = 0 (or δ σ = 0). Harmonic forms
⋆
are both closed and co-closed! This property also holds on open manifolds (eg. Rn ) if σ has compact support (it
vanishes outside a bounded closed region), or if it goes to zero sufficiently fast at infinity.
21
Lecture Notes on Mathematical Methods 2022
Definition 1.33. Assuming a compact manifold without boundaries or, failing that, compact support
(sufficiently fast fall-off at infinity), the unique Hodge decomposition writes a p-form σ as a sum of
exact (closed), co-closed, and harmonic p-forms:
σ = dα + δ β + harmonic p-form (1.57)
non-unique. dα, δ β and the harmonic p-form
where α is a (p-1)-form and β is a (p+1)-form, both V
in the decomposition live in orthogonal subspaces of p .
Example 1.24. Let A be a vector field with compact support on Euclidean R3 . Then its Hodge
decomposition says that its associated 1-form can be written as the exterior derivative of a 0-form (ie.
the gradient of a function), plus the divergence of a 2-form, β, plus some harmonic 1-form. Now, since
⋆β is a pseudo-1-form in R3 , δ β = ⋆d ⋆β is a 1-form. Then, from eq. (1.52) this term corresponds
⋆ ⋆ ⋆
with compact support and Hodge dual 1-form J = ⋆J . Then we have the inhomogeneous equation:
⋆
d ⋆F = 4π J
⋆ (1.59)
If we take the exterior derivative of the equation, the left-hand side vanishes identically, and the right-hand
side becomes: d J = 0. This is better known as the statement that the 4-divergence of J vanishes: ⋆d ⋆ J = 0. ⋆ ⋆
the latter form being more appropriate if we insist on thinking of the source term as a vector. I would argue,
however, that the less conventional form eq. (1.59) is much the more natural. The exterior derivative is metric-
independent, and its index form can be written entirely with covariant indices, the natural ones for p-forms. But
to obtain its equivalent in divergence form, we have to Hodge-dualise the right-hand side, so that the vector J
source depends on the metric (see the paragraph after eq. (1.47)), whereas its 3-form version does not. The price,
of course, is that the 3-form version has 64 explicit components, although still only four independent ones.
It is worth noting that, although dF = 0 and the source equation (1.59) completely determine F, A is deter-
mined only up to an additive term df , where f is an arbitrary differentiable function.
As a 3-form, the homogeneous equation dF = 0 also has a lot of components, and when it comes to solving
the system, we may want to extract only the independent ones. This is the same as d ⋆(⋆F) = 0 whose Hodge ⋆ ⋆
dual is δ⋆F = 0. In other words, the divergence of ⋆ F vanishes, only four equations. Actually, this is a general,
⋆ ⋆
easily shown property (EXERCISE): whenever the exterior derivative of a p-form in some manifold vanishes, so
does the codifferential of its dual, and vice-versa.
Another great advantage of writing Maxwell’s equations as dF = 0 and d ⋆F = 4π J is that, provided the ⋆
source is smoothly varying, they are formally the same in curved spacetime! Only when divergences are written in
index notation are covariant derivatives involving a connection needed. Even in index notation, the first equation
does not involve the connection; it does not even require a metric.
Finally, nothing prevents us from constructing an extended Maxwell-like theory (not describing electromag-
netism) involving F as a 3-form. In the past few decades it has received a good deal of attention in some quarters.
(eg., over some volume in R3 ). The Hodge dual of a scalar function f , of course, is a pseudo-n-form whose single
independent component is f . Then:
Z Z p Z Z
⋆f =
⋆ f (u) |g| du1 ∧ · · · ∧ dun = f (x) dx1 ∧ · · · ∧ dxn = f (x) dn x
V V V V
where u are general coordinates and dn x is the volume pseudo-n-form in Cartesian coordinates. Then we define:
Definition 1.34.
Z Z Z
1 n 1 n
f (x) dx ∧ · · · ∧ dx := f (x) dx · · · dx = f (x) dn x (1.61)
V V V
When a p-dim region R is embedded in a n-dim manifold, it will be described with some coordinates u(x), that
is, n functions ui of the p Cartesian coordinates xj that parametrise Rp . Also, an orientation can be defined for the
region. What is the meaning of the integral of a p-form over such a region? We give two examples in R3 .
The integrals in R2 on the right are over a rectangular region of S in parameter space. The two
coordinate vectors (see section 1.2.3), ∂1 u and ∂2 u, are tangent to S at every point, and are usually
linearly independent, so form a basis for the space tangent to the surface at a point, with no metric
required as yet.
The Hodge dual of β , a pseudo-1-form, has an associated pseudo-vector B with, asp
components, the
i ijk 1
contravariant components of the Hodge dual, B = ǫ βjk (j < k), eg., B = β23 / |g|, etc. Then:
p B1 B2 B3
j k j k i j k j k
βjk (∂1 u ∂2 u − ∂2 u ∂1 u ) = ǫijk B (∂1 u ∂2 u − ∂2 u ∂1 u ) = |g| ∂1 u1 ∂1 u2 ∂1 u3
∂2 u1 ∂2 u2 ∂2 u3
From eq. (1.43), we recognise the last member of the equality as the output obtained from inserting the
three vectors whose components are the rows of the determinant into the three input slots of a simple
3-form—more accurately, a pseudo-3-form which, from definition (1.25) can be identified with the
volume pseudo-form d3 u. Then our integral can be written:
Z Z
3
β = d u(B, ∂1 u, ∂2 u) dx1 dx2
S
This makes it obvious that the integral is independent of the orientation of R3 , since switching it flips
the sign of both B and d3 u. At every point on S, we can choose the unit n̂ normal to the surface so
that n̂ and the vectors ∂1 u and ∂2 u tangent to the surface form a right-handed (positive orientation)
system. We also note that only the normal component of B can contribute to the integral (why?).
Then the scalar function d3 u(B, ∂1 u, ∂2 u) is the normal component of B multiplied by the surface of
the parallelogram defined by the coordinate vectors (see example 1.13) . Defining the surface element
dS ≡ |∂1 u × ∂2 u|, there comes:
Z Z Z
β = Bn dS = B · dS (1.62)
S
where the often used last expression is called the flux of the pseudo-vector B through the surface S.
It does not depend on the parametrisation chosen for S which is integrated out. The same result holds
if β is a pseudo-2-form, with B now a vector.
24
Lecture Notes on Mathematical Methods 2022
A technicality is that both V and ∂V must have compatible orientations. But no metric is required. The boundary
need not be connected, and it can be broken up into non-overlapping parts when it cannot be covered by a single
coordinate patch. Then we simply sum the integrals over each part.
Example 1.27. At the end of example 1.25 we had already worked out an application when ω is a
0-form: the fundamental theorem of calculus. When ω is a 1-form and V a 2-dim surface in Euclidean
R3 parametrised
R with Cartesian
H R C, theRsame example gives
coordinates and bounded by a closed curve
immediately: ∂V ω = C A · du. From eq. (1.51) and example 1.26, S dω = S ∇ × A · dS, and
we recover the well-known Kelvin-Stokes formula.
Finally, when ω is a pseudo-2-form
R in Euclidean
H R3 and S a surface enclosing a volume V , we recover
the divergence theorem: V ∇ · B dV = S B · dS, from examples 1.22 and 1.26.
Note that a metric is required for the translation from the Stokes-Cartan theorem to the divergence and
Kelvin-Stokes theorems in vector calculus.
E := F10 dx1 + F20 dx2 + F30 dx3 B := F12 dx1 ∧ dx2 + F31 dx3 ∧ dx1 + F23 dx2 ∧ dx3 (1.65)
Now, formally, d = ~d+dt ∧∂t . where ~d denotes the 3-dim exterior derivative. Then Maxwell’s dF = 0 becomes:
~d + dt ∧ ∂t E ∧ dt + B = ~dE ∧ dt + ~dB + dt ∧ ∂t B = ~dE + ∂t B ∧ dt + ~dB
= 0
The plus sign in the round brackets is the result of applying the commutation formula eq. (1.38) to the 1-form dt
and the 2-form B. In three dimensions, then, the homogeneous Maxwell equation gives rise to:
~dB = 0 ~dE + ∂t B = 0 (1.66)
Eq. (1.66) is metric-independent, and will thus hold in any spacetime in a coordinate basis.
The Hodge duals of eq. (1.66) can be written as:
div ⋆B = 0
⋆ ⋆ ~
⋆ dE + ∂t ⋆B = 0
⋆
If we identify the contravariant components of the pseudo-1-form ⋆B with the usual components of the magnetic-
⋆
field pseudo-vector, and use eq. (1.52), we see that these are equivalent to the homogeneous Maxwell equations in
their vector-calculus form: ∇ · B = 0 and ∇ × E + ∂t B = 0.
We see that it is much more natural to view the 3-dim magnetic field as a 2-form which is the exterior derivative
of a 1-form, than as a pseudo-vector which is the curl of another vector. and the electric field strength with the
1-form E than with the vector E. It is consistent with force and momentum also being more naturally 1-forms
µ
(consider eipµ x !).
The inhomogeneous Maxwell equation requires much more care, and is treated in Appendix C.
25
Lecture Notes on Mathematical Methods 2022
Appendices
A Tangent Spaces as Vector Spaces
First, let us specify what is meant by addition and s-multiplication on a tangent space TP .
Definition A.1. The addition operation on TP is a map, TP + TP −→ L(C ∞ (M ), R), such that,
∀ f ∈ C ∞ (M ) and any two curves (Γ, Θ) ∈ M intersecting at P ∈ M :
(v(Γ,P) + v(Θ,P) )(f ) := v(Γ,P) (f ) + v(Θ,P) (f )
Again, the addition operation on the left is between mappings, whereas that on the right is on R. As
for s-multiplication, it is a map, R × TP −→ L(C ∞ (M ), R), such that, ∀ a ∈ R:
(a · v(Γ,P) )(f ) := a v(Γ,P) (f )
The question now is: do these operations close? In other words, can we find some curve Θ ∈ M such that:
a · v(Γ,P ) = v(Θ,P) , and perhaps another curve Σ ∈ M such that: v(Γ,P) + v(Θ,P) = v(Σ,P) ?
To construct such a curve for s-multiplication, we first redefine the parameter of the curve Γ as the linear
function, µ : R −→ R, of λ: µ = aλ + λ0 , with λ now the parameter of a curve Θ such that Θ(λ) = Γ(µ)
Therefore, Θ(0) = Γ(λ0 ) = P. As in definition 1.5 we can write: Γ(µ) = Γ ◦ µ(λ). Insert this information into
the expression for the velocity for Θ at P:
v(Θ, P) (f ) = dλ (f ◦ Θ) λ=0
= dλ (f ◦ Γ ◦ µ) λ=0
= dµ (f ◦ Γ) µ(λ=0)=λ0
dλ µ λ=0
= a v(Γ, P) (f )
Therefore, we have found a curve Θ such that the operation a · v(Γ,P ) gives the velocity for that curve at P.
Up to now, in our discussion of tangent spaces, we have not needed any reference to coordinate charts. Unfor-
tunately, when it comes to proving that addition of two velocities in TP gives a velocity in TP , we cannot add the
curve mappings directly since this has no meaning. Instead, as was done in the previous section, assume that both
curves Γ and Θ are in some open subset U ⊂ M parametrised by coordinate functions x. Let Γ and Θ go through
point P at values λ1 and λ2 ot their respective parameter. Then construct a curve Σ parametrised in U by:
(x ◦ Σ)(λ) = (x ◦ Γ)(λ1 + λ) + (x ◦ Θ)(λ2 + λ) − (x ◦ Γ)(λ1 )
Although there might appear to be an obvious cancellation in this expression, it is not allowed because the coor-
dinate functions are not linear and thus do not distribute over the additions in Rn in the arguments on the right.
At λ = 0, however, the cancellation does occur, leaving Σx (0) = Θ(λ2 ) = P, so that our curve Σx runs through
point P at λ = 0.
We also need the derivative of the ν th x coordinate of the curve Σ, evaluated at P:
h
dλ (xν ◦ Σ) 0 = dλ (xν ◦ Γ)(λ1 + λ) + (xν ◦ Θ)(λ2 + λ) − (xν ◦ Γ)(λ1 )
0
ν ν
= dλ1 +λ (x ◦ Γ) λ1
dλ (λ1 + λ) 0
+ dλ2 +λ (x ◦ Θ) λ2
dλ (λ2 + λ) 0
ν ν
= dλ1 +λ (x ◦ Γ) λ1
+ dλ2 +λ (x ◦ Θ) λ2
(A.1)
Now go back to our expression (1.6) for the velocity in coordinates x. The first factor on the right has been
evaluated in eq. (A.1) and, running the chain of equalities in eq. (1.6) backward, there comes:
X h i X h i
vΣ,P) (f ) = ∂ν (f ◦ x−1 ) xν (P) dλ (xν ◦ Γ) λ1 + ∂ν (f ◦ x−1 ) xν (P) dλ (xν ◦ Θ) λ2
ν ν
−1
−1
= dλ (f ◦ x ) ◦ (x ◦ Γ) λ1
+ dλ (f ◦ x ) ◦ (x ◦ Θ) λ2
= v(Γ,P) (f ) + v(Θ,P) (f )
26
Lecture Notes on Mathematical Methods 2022
Thus, adding the velocities for two curves meeting at some point yields the velocity for some other curve in-
tersecting the others at that same point, and the tangent space of a curve at a point can indeed support a vector
space structure! Do note that the result does not depend on whatever coordinate chart we might have used in the
intermediate steps of the proof.†
= ∂xµ y ν xP
∂yν f yP
(B.1)
A vector v ∈ TP must remain invariant under change of chart. That is: v = vxµ ∂xµ x = vyλ ∂yλ y . Inserting the
P P
transformation law for the coordinate bases, we immediately find he transformation law for the components of v:
vyν = ∂xµ y ν xP
vxµ (B.2)
†
For a clear and accessible discussion of differentiability, manifolds and tangent spaces, see Frederic Schullers’s first five lectures at the
2015 International Winter School on Gravity and Light, available on YouTube.
27
Lecture Notes on Mathematical Methods 2022
∗ 1 µν
F = F ǫ dxα ∧ dxβ
4 p µναβ
= − |g| F 10 dx2 ∧ dx3 + F 20 dx3 ∧ dx1 + F 30 dx1 ∧ dx2 + F 12 dx3 + F 31 dx2 + F 23 dx1 ) ∧ dt
Now we must write this in terms of the covariant components of F, and this is where the metric must come in,
since F µν = g µα g νβ Fαβ :
F i0 = (g00 gij − gi0 g0j )Fj0 + g ij g0k Fjk , F ij = (gi0 g jl − g il gj0 )Fl0 + gik gjl Fkl
We know that Fj0 and Fjk are the components of the 3-dim p-forms E and B, respectively. If g0i 6= 0, each
contravariant component of F will involve both E and B, which will lead to very complicated results. When
g0i = 0, however, we are left with F i0 = g 00 gij Fj0 , and F ij = gik g jl Fkl , and lowering the spatial components
of F involves only the spatial sector of the metric (ignoring the g00 factor), the same sector that is used p to raise
00
indices on the Levi-Civita tensor. Also, if we take g = −1 (mostly positive) Minkowski metric, the |g| factor
is the same for the three-dimensional metric determinant as for the 4-dim one. Because of all this, we can now
write:
∗ 1 i0 j k 1 ij k
F = − ǫ F dx ∧ dx + ǫijk F dx ∧ dt
2 ijk 2
where the roman indices run from 1 to 3. Now we can relate the two terms to E and B:
1 1 1
ǫ F i0 dxj ∧ dxk = ǫijk g00 g il Fl0 dxj ∧ dxk = g00 ǫijk E i dxj ∧ dxk = g00 ∗ E = − ∗ E
2 ijk 2 2
Also:
1
ǫ F ij dxk = ∗ B
2 ijk
with no assumption needed for the spatial part of the 4-dim metric. Then our expansion is ∗ F = −∗ B ∧ dt + ∗ E
where it is understood that, on the right-hand side only, the 3-dim Hodge dual is taken. It is not difficult to show
(EXERCISE) that: d∗ F = −(~d∗ B − ∂t ∗ E) ∧ dt + ~d∗ E.
We define the Maxwell source pseudo-3-form as the expansion:
where ρ is the charge scalar density, ρ the three-dim charge-density pseudo-3-form and J the 3-dim current density
1-form. Inserting these expansions in eq. (1.59) yields the two 3-dim Maxwell field equations:
~d∗ E = 4πρ, ~d∗ B = j + ∂t ∗ E (C.1)
Taking the 3-dim Hodge dual of these equations recovers the vector-calculus form of Gauss’s law for electricity
and the Ampère-Maxwell equation.
28
Lecture Notes on Mathematical Methods 2022
By re-labelling b ←→ c in the first table, and a ←→ b in the second, and re-ordering the rows and columns,
we obtain tables which are identical to the cyclic table, and we conclude that they are really those of Z4 .
The last table is genuinely different. It belongs to a group {e, a, b, a⋆b} called the 4-group—aka Felix Klein’s
Vierergruppe V —in which every element is its own inverse (so of order 2), with the fourth element constructed out
of the other two non-identity elements (otherwise V would be cyclic!). An example is D2 , the symmetry group of
a 2-d rectangle centered on the origin:, with the identity, one rotation by π, and two reflections about the axes as
elements.
The foregoing illustrates very nicely two important features of groups:
• Generators of a group
Definition 2.4. A set of generators of a group G is any subset of G from which all other elements
of G can be obtained by repeated compositions of the generators among themselves. G must
contain all the distinct compositions of its generators, including with themselves.
For instance, we can say that if a generates Zn , ap also generates Zn provided p and n have no common
divisor (EXERCISE). Then any such ap can be taken on its own as the generator of Zn . The 4-group is
obtained from two generators. EXERCISE: construct a Cayley table for the group: {e, a, b, b2 , a ⋆ b, b ⋆ a}.
Another example is a rotation by π/6 as the generator of the finite group of rotations by kπ/6 (0 ≤ k ≤ 11)
about the same axis.
EXERCISE: Is it possible to construct a group of order 6 with all its elements of order 2?
• Isomorphisms
We have just been introduced to the important idea that groups which look different may in some sense be
the same because their Cayley tables are identical or can be made to be identical by relabelling. We now
formalise this idea:
Definition 2.5. If there exists a one-to-one mapping between all the elements of one finite group
{G, ◦} and all the elements of another finite group {H, ⋆} such that under this mapping these
groups have identical Cayley tables, then the mapping is an isomorphism, and G and H are
isomorphic: G ∼ = H.
30
Lecture Notes on Mathematical Methods 2022
Another definition is more apt for continuous groups, which do not have a Cayley table as such:
Definition 2.6. If there exists a one-to-one mapping f between all the elements of one group
{G, ◦} and all the elements of another group {H, ⋆} such that under this mapping, f (a), f (b) ∈
H and f (a ◦ b) = f (a) ⋆ f (b) ∀ a, b ∈ G, then f is an isomorphism of G onto H, and G ∼ = H.
– the group of permutations of two objects (S2 ), the group of rotations by π around the z axis, and the
group {1, −1} (under multiplication);
– the group of complex numbers and the group of vectors in a plane, both under addition;
– the groups {R, +} and {R+ , ×} with the exponential as the isomorphism. Later we will see that
because ex ey = ex+y , ex ∈ {R+ , ×} provides a one-dimensional matrix representation of {R, +}.
Definition 2.8. Given a ∈ G, any element b ∈ G which can be obtained as b = x ◦ a ◦ x−1 , where
x ∈ G, is called the conjugate of a by x. This conjugation operation, which consists of two binary
compositions, has the following properties:
{e ◦ ai ◦ e−1 , a1 ◦ ai ◦ a−1 −1 −1
1 , . . . , ai−1 ◦ ai ◦ ai−1 , ai+1 ◦ ai ◦ ai+1 , . . .}
Then e is always in a class by itself, and each element of an Abelian group is the sole element in its class. eg.,
Zn and the four-group.
Classes are disjoint: they have no common element (EXERCISE: show this). Thus, they partition the group.
Elements in the same class share some properties. In particular, they must all be of the same order (EXER-
CISE). In a particularly important type of group, matrix groups, conjugate matrices are similar to one another; they
could represent the same “thing” in different bases.
EXERCISE: obtain the classes for the group: {e, a, b, b2 , a ⋆ b, b ⋆ a}.
†
Actually, conjugacy is only a particular type of equivalence.
31
Lecture Notes on Mathematical Methods 2022
2.2.2 Subgroups
Definition 2.10. A subset H of a group G that behaves as a group in its own right, and under the same
composition law as G is said to be a subgroup of G: H ⊆ G. H is proper if it is non-trivial (ie. not
e) and if H ⊂ G (ie. H 6= G). The subgroups of a group may have more elements than e in common.
We have already seen that any element g of order m < n of G generates a cyclic subgroup Zm ⊂ G.
Example 2.2. The four-group V has the proper Z2 subgroups: {e, a}, {e, b}, and {e, c = a ⋆ b},
which are isomorphic. By inspection, the group of order 6 {e, a, b, b2 , a⋆b, b⋆a} contains the proper
subgroup Z3 = {e, b, b2 }.
Notation alert: Henceforth, we drop the cumbersome star (circle) whenever there is no risk of confusion with
usual multiplication. Also, if H and H ′ are two subsets of {G, ⋆}, we can write H H ′ for {h h′ } h ∈ H, h′ ∈ H ′ .
Let us try out our new notation on the following definition:
Definition 2.11. A subgroup N ⊆ G is invariant (or normal) if N = G N G−1 or, more precisely,
if g h g −1 ∈ N ∀ h ∈ N and ∀ g ∈ G. Alternate notation: N ✁ G, G ✄ N .
EXERCISE: Show that ∀ gi ∈ G the set with distinct elements gi−1 gj−1 gi gj forms an invariant subgroup of G.
Definition 2.11 is sometimes written G N = N G, but it does not mean that an invariant subgroup must be
Abelian (though it can be). It means that if hi ∈ N and g ∈ G, there is some element hj ∈ N such that g hi = hj g.
Example 2.3. Because the four-group V is Abelian, its non-trivial subgroups, {e, a}, {e, b}, {e, a b},
are all invariant. Subgroups of any Abelian group are invariant.
Since classes and normal groups are both defined by conjugation, it is hardly surprising that they are related.
Indeed, let H ⊂ G. Then H is invariant if and only if it contains complete classes, ie. if it is a union of classes of
G. Indeed, if H is invariant, all the conjugates (elements in the same class) of any h ∈ H are also in H; this holds
for all classes, which are disjoint; so only complete classes can be in H. Conversely, let a subgroup H ⊂ G be a
union of complete classes; therefore, g h g −1 ∈ H ∀ g ∈ G, which is precisely the definition of a normal subgroup.
Definition 2.12. A simple group has no invariant subgroup other than itself and the identity.
Definition 2.13. Let H be a subgroup of G, and let g ∈ G. Then g H is a left coset of H for a given
g, and H g is a right coset of H. The set of all left (right) cosets of H is called the left (right) coset
space of H. Every coset g H must contain the same number of elements, equal to the order of H.
If H is invariant, to any ot its left cosets corresponds an identical right coset, and vice-versa, as follows
immediately from Def. 2.11. In particular, the right and left cosets of any Abelian subgroup are identical.
Example 2.4. Let G = R3 under addition, and H be a plane containing the origin. For a given vector
a, a + H ∈ H if a ∈ H; otherwise, a + H is another plane parallel to H, and we would say in this
language that it is a left (or right) coset of H through the origin. And H itself would also be a coset.
The most important property of cosets is that they are either disjoint or else identical. Thus, we can say that
the coset space of a subgroup H ⊂ G provides a partition of G.
Indeed, let g1 h1 = g2 h2 for some (h1 , h2 ) ∈ H and (g1 , g2 ) ∈ G. Therefore, g1 = g2 h2 h−1 1 . Now
consider some other element of the same coset, g1 h3 (h3 ∈ H); then g1 h3 = g2 (h2 h−1 1 h3 ) = g h
2 4 , where
−1
h4 = h2 h1 h3 ∈ H. That is, if two elements of different cosets are the same, then any other element, say g h3 ,
in the first coset, must be equal to some element of the second coset. Since the same argument holds when we
switch g1 and g2 , we conclude that if g1 H and g2 H have one element in common, they have all their elements in
common and are thus identical. The same proof applies to right cosets.
It follows (why?) that e H = H is the only coset of a subgroup H that is a group.
32
Lecture Notes on Mathematical Methods 2022
Definition 2.14. The set of all left cosets of H ⊂ G, each considered as a whole, is called a factor
space for H. Note that the elements of this space are the cosets themselves, each considered as a
whole, not any individual element within a coset.
Factor spaces of a subgroup H are not necessarily groups; but there is one important exception:
Another well-known way of constructing a (this time, external) direct product of, say, two a priori unrelated
matrix groups with elements A ∈ H1 and B ∈ H2 would be:
A 0 A 0 I 0
=
0 B 0 I 0 B
Or we could construct {1, −1}×{1, −1} = {(1, 1), (1, −1), (−1, 1), (−1, −1)}. the external direct product of
Z2 with itself, in this realisation. This, of course, is the four-group (with normal multiplication as group product).
33
Lecture Notes on Mathematical Methods 2022
The horizontal ordering of the initial objects is immaterial. Also as a matter of convention, we agree
that it is the objects in the slots which are rearranged, not the slots. Finally, we do not have to use
numbers as labels, but they offer the greatest range.
In a permutation, an object i may be mapped into itself, ie. it stays in the same slot. But more typically object
i is mapped to j, while j is mapped to k; and so on along a chain that ends back at object a after l steps. When
this occurs, we speak of a l-cycle. More precisely:
Definition 2.18. Let πk ∈ Sn , and let l be the smallest integer for which [πk (j)]l = j, for some
1 ≤ j ≤ n. Then the sequence of objects in [πk (j)]l is called a l-cycle (sometimes a r-cycle. . . ).
This suggests a much more compact notation for πk , one in which we bother to write only the l-cycles (l > 1), and
consider a given permutation as the product of simpler permutations.
As an example, we write:
1 2 3 4 5 6 1 2 3 4 5 6 1 2 3 4 5 6 1 2 3 4 5 6
= ≡ (1 5) (2 4 3)
5 4 2 3 1 6 5 2 3 4 1 6 1 4 2 3 5 6 1 2 3 4 5 6
It is easy to see the advantages of the cycle notation introduced at the end of the line! Note that the cycles are
disjoint. Any permutation can be, and usually is, represented by a sequence of disjoint cycles. Warning: do not
confuse the symbols in a l-cycle with the outcome of a permutation in Sn !
Any πk ∈ Sn can always be written as the product† of transpositions, or two-cycles. Indeed, a l-cycle may
always be decomposed as a product of l − 1 transpositions, but these are not disjoint. An element of Sn and its
inverse have the same cycle structure.
Definition 2.19. A permutation is even (odd) if it is equivalent to an even (odd) number of transposi-
tions, or switches; thus, a l-cycle which contains an even number of symbols is equivalent to an odd
permutation, and vice-versa. An even permutation is said to have parity 1, and an odd permutation
parity −1. We expect that parity will put strong constraints on the group product table of Sn .
Single transpositions always have odd parity. The mapping from Sn to the parities {1, −1} is a nice example of a
homomorphism.
Definition 2.20. A cyclic permutation of length l has a single cycle of length l > 1.
In cycle notation, S2 = {e, (1 2)} and S3 = {e, (1 2), (1 3), (2 3), (1 3 2), (1 2 3)} ≡ {π1 , π2, π3 , π4 , π5 , π6 },
are the smallest non-trivial symmetric groups. For S3 , note the three-cycles (1 2 3) = 12 23 31 and (1 3 2) =
1 2 3
3 1 2 . I have deliberately changed the order of the latter from what it is in BF, but if you write out the corre-
sponding permutation in full notation for BF’s (3 2 1), you will see that it is identical to mine. So long as we cycle
through in the same direction (here, to the right), where we start the cycle does not matter! It can be shown that S3
and Z6 are the only groups of order 6, up to isomorphisms.
†
Since there is little scope for confusion in the context of Sn , we replace “group composition” with “group product”.
34
Lecture Notes on Mathematical Methods 2022
The other unfilled entries in rows and columns for π5 and π6 must be either π2 , π3 , or π4 . For columns π5 and
π6 , applying π2 to π2 π3 gives π3 = π2 π5 , which determines the rest from the table-building rules. Similarly,
π2 π3 π3 = π2 = π5 π3 , and the rest of the π5 and π6 rows is determined. The comes (in two equivalent forms):
e π2 π3 π4 π5 π6 e π5 π6 π2 π3 π4
π2 e π5 π6 π3 π4 π5 π6 e π4 π2 π3
π3 π6 e π5 π4 π2 π6 e π5 π3 π4 π2
≡
π4 π5 π6 e π2 π3 π2 π3 π4 e π5 π6
π5 π4 π2 π3 π6 e π3 π4 π2 π6 e π5
π6 π3 π4 π2 e π5 π4 π2 π3 π5 π6 e
What we have shown is that πai πaj = πai aj ; in other words, by definition 2.6, permutations preserve the group
product of G, and we have Cayley’s Theorem:
35
Lecture Notes on Mathematical Methods 2022
Every group of order n is isomorphic to a subgroup of Sn whose elements (except for e) shuffle all objects in
the set on which it acts.
We have already seen an example of this: the single instance of the cyclic group of order 3 is a subgroup of S3 .
EXERCISE: How many distinct instances of Z4 ⊂ S4 are there? How many of the four-group?
and (b1 , b2 ) 3-cycles, show that because the latter are of order 3, b1 b2 is neither b−1
1 nor b2 , which leaves only one possibility
−1
since b1 and b2 cannot be each other’s inverse. Is that possibility allowed by the parity of these elements?
In the literature, classes of Sn are routinely identified by partitions of n reflecting their cycle structure. Thus,
a given class will be written (iαi . . . j αj ), with (1 ≤ (i, j) ≤ n), where αi is the number of i-cycles in the class.
Start with e, whose cycle structure can be written as a product of n 1-cycles: e = (1) (2) · · · (n). So its class,
which always exists, would be denoted by 1n . A transposition has one 2-cycle and n-2 1-cycles, and Sn must
contain n(n − 1)/2 n−2 ). An arbitrary permutation involves
Pnof them (eg., six for S4 as above); it is denoted by (2 1
αi i-cycles, and i i αi = n. In that sense the cycle structure of a class corresponds to a partition of n.
Once we have noticed this correspondence, it becomes rather easy to find the number and cycle ctructure of
Sn classes. We adopt the usual convention that represents the cycle structure of a class by (λ1 ≥ λ2 ≥ · · · ≥ λn ),
where the λi sum up to n. Thus, the only possible partitions of S3 lead to classes (13 ), (2 1), and (3), ie. a class
with three 1-cycles (the identity), a class with one 2-cycle and one 1-cycle (the transpositions), and a class with
one 3-cycle. As for S4 , the possible partitions of 4 give rise to the five classes (14 ), (2 12 ), (22 ), (3 1), and (4).
It is important not to confuse the cycle notation we first introduced with this standard notation which lists all
the cycles in a class as a whole, including 1-cycles when they occur (whereas the other one ignores them).
‡
Note that this subgroup being Abelian is not sufficient to make it invariant; it must be self-conjugate with respect to all elements in S3 .
36
Lecture Notes on Mathematical Methods 2022
To find the number of elements in a class of Sn , count the distinct ways of partitioning n numbers into its cycle
structure:
n!
(2.1)
α1 ! . . . αn ! 1α1 . . . nαn
where αi ! is the number of non-distinct ways of ordering αi commuting cycles of a given length, and iαi is the
number of equivalent orderings of the symbols inside each i-cycle occurring αi times. From this expression it
should be easy to recover the number of elements in each class of S4 as given above.
Now we can identify (EXERCISE) the invariant subgroups of S4 without writing down its 24×24 Cayley table!
Then we just read off the cycle structure for each: (14 ), (2 12 ), (22 ), (3 1), and (4), respectively. Finding the classes
of such monsters as, say S8 , no longer seems so intimidating. These diagrams are known as Young frames.
2.3.7 Cosets of Sn
Finding the left cosets of the subgroups of S3 is as easy as reading rows in its Cayley table. Take the sub-
group H = {e, π2 }; its left cosets by πk are πk {e, π2 } = {πk , πk π2 } (1 ≤ k ≤ 6). Only three are distinct:
{e, π2 }, {π3 , π6 }, {π4 , π5 }. Following Definition 2.14, this set of cosets is the factor space for H. The same
arguments apply to the subgroups {e, π3 } and {e, π4 }.
Turn now to the remaining non-trivial proper subgroup, A3 = {e, π5 , π6 }, of all even permutations in S3 .
Its left cosets are {πk , πk π5 , πk π6 }. For instance, π2 {e, π5 , π6 } = {π2 , π3 , π4 }, which is identical to the
other cosets π3 {e, π5 , π6 } and π4 {e, π5 , π6 }. Also, e {e, π5 , π6 } = π5 {e, π5 , π6 } = π6 {e, π5 , π6 }, as expected.
So another partition of S3 is provided by {e, π5 , π6 } + π2 {e, π5 , π6 }. Note that these left and right cosets
are identical, another way of saying that {e, π5 , π6 } is invariant, as we had found by simpler means. Then
{e, π5 , π6 }, {π2 , π3 , π4 } is the factor group of S3 . From the Cayley table for S3 , we see that the element
{e, π5 , π6 } is the identity, and that this factor group S3 /A3 is isomorphic to Z2 . It is easy to show that Z2 is a
factor group of Sn ∀ n. Equivalently, An is always a normal subgroup of Sn .
37
Lecture Notes on Mathematical Methods 2022
Definition 2.21. We distinguish between an action from the left, [Tg f ](x) := f (g−1 x), and an action
from the right, [Tg f ](x) := f (x g), ∀ f . Note that, here, the operators Tg act on the functions (not on
x !).
Why did we define the left action of g ∈ G as g−1 x, and not g x? Denote by Tgi gj the transformation
associated with gi gj ∈ G. Then, with gi = i and gj = j in subscripts so as to declutter the notation:
Tij f (x) = f (gi gj )−1 x = f (gj−1 gi−1 x) = Tj f (gi−1 x) = Ti Tj f (x)
which means that the T operators do form a group; but what if instead:
Tij f (x) = f (gi gj x) = Ti f (gj x) = Tj Ti f (x)
Something awkward has happened: if we write the left action as g x, the associated transformations do not form a
group! And, as you should verify, neither do they if we write the right action as x g−1 .
So, as a matter of notational consistency, we should always write x g for the right action and g−1 x for the left
action, which is indeed what BF do (without much explanation) for the left action.
Even addition can be represented by matrix multiplication: Dα Dβ = Dα+β , with α and β two values of a group
parameter, eg. the matrix Dv = v1 01 . Do you recognise the transformation that applies it to the vector xt ?
The matrices GL(n, C) of rank n can be thought of as the set of all invertible linear transformations on a
vector space of complex-valued functions V = {f (x)}. We have: x = xi ei , with {ei } a basis for V and xi the
components of x in the basis; the subscript on basis vectors labels a whole vector, not a component of the vector.
Let us focus on the transformations Tg (x). Then the left action of g ∈ G is expressed as:
Tg (x) = g−1 x = xi g−1 ei = xi ej (DgL−1 )j i = ej (DgL−1 )j i xi (2.2)
It is an instructive exercise to show that the proper way of expressing the right action, x g, of the same group
in terms of its (right) representation DR matrices is:
x g = ei xi g = (DgR )i j xj ei (2.3)
in which DR i
g acts on the the x written as a column vector. Because of this, some see the right action as the more
“natural” one. For a given g, a right DL and DR is in general each other’s inverse.
38
Lecture Notes on Mathematical Methods 2022
We look for a set of functions of φ which, under Rα , transform into linear combinations of themselves.
Try f1 = cos φ, f2 = sin φ. Then:
Rα f1 (φ) = cos(φ + α) = (cos α) cos φ − (sin α) sin φ = (cos −α) f1 (φ) + (sin −α) f2 (φ)
Rα f2 (φ) = sin(φ + α) = (sin α) cos φ + (cos α) sin φ = − (sin −α) f1 (φ) + (cos −α) f2 (φ)
Compare this with Di j (−α) fj (φ), and switch the sign of α to obtain the left D(α) matrix:
(1) cos α sin α
D (Rα ) =
− sin α cos α
Well, that’s the 2-dim left defining (fundamental) representation for SO(2), probably the most often
used. But it is not the only one! If instead f1 = eiφ , f2 = e−iφ , the same procedure would yield:
iα
(2) e 0
D (Rα ) =
0 e−iα
so here is another two-dim
representation. But it is equivalent because the transformation S−1 D(1) S,
√
with the matrix S = 1i 1i / 2, diagonalises D(1) into D(2) , ∀ α, ie. for all elements of SO(2).
And there are more: each linearly independent function eiα and e−iα is also a perfectly acceptable
one-dimensional representation of SO(2)! Both D(1) and D(2) can be viewed as a joining of these
one-dimensional representations, which we shall call D(3) and D(4) . Obviously, there is something
special about e±iα . Before we discover what it is, let us look at another instructive example.
Example 2.7. Let us work out a three-dimensional representation of the left action of S3 , π −1 x, on
R3 . Since Sn merely shuffles the components of x it preserves its length, which is the definition of
−1 i j −1 i j
is their inverse. In fact, S3 ⊂ O(3)! Then, from eq.
orthogonal matrices, ie., those whose transpose
(2.2), πk x = x ej D i (πk ) = x Di (πk ) ej so as to view the permutations as a shuffling of the
components of x (written as row vectors—see the Appendix at the end of the chapter!), and we have:
1 0 0 0 1 0 0 0 1
D(1) (π1 ) = 0 1 0 , D(1) (π2 ) = 1 0 0 , D(1) (π3 ) = 0 1 0 ,
0 0 1 0 0 1 1 0 0
39
Lecture Notes on Mathematical Methods 2022
1 0 0 0 1 0 0 0 1
D(1) (π4 ) = 0 0 1 , D(1) (π5 ) = 0 0 1 , D(1) (π6 ) = 1 0 0
0 1 0 1 0 0 0 1 0
Such a faithful, n-dim left defining (fundamental) representation can be constructed for any Sn .
Now, I claim that there exists another (two-dimensional!) representation of S3 , which is not faithful:
(2) (2) (2) 1 0
D (π1 ) = D (π5 ) = D (π6 ) =
0 1
(2) (2) (2) 0 1
D (π2 ) = D (π3 ) = D (π4 ) =
1 0
Indeed, the products of these matrices are consistent with the group product of S3 in its Cayley table.
Even less faithful, but no less acceptable, is the one-dim representation of Sn obtained by mapping its
permutations to their parity values. For S3 :
D(3) (π1 ) = D(3) (π5 ) = D(3) (π6 ) = 1
D(3) (π2 ) = D(3) (π3 ) = D(3) (π4 ) = − 1
And, of course, we can always map all the πi to 1 and get another (trivial) representation!
On the other hand, we could join D(1) and D(2) into a D(4) = D(1) ⊕ D(2) (direct sum) representation
whose six matrices are 5-dimensional and block-diagonal, each with the submatrices on the diagonal
taken, one from D(2) (the upper one, say), and the other from D(1) , for a given permutation πi .
Definition 2.24. The regular representation of the left action of a finite group G is the set of matrices
DLg , with g ∈ G, derived from a group product, such that:
(
L j j 1 g gi = gj
Dg gi = g gi = gj D i (g) ∀ g ∈ G D i (g) =
0 g gi 6= gj
The regular representation is seen to be closely related to the Cayley table of the group. Its dimension
is equal to NG , the order of the group, and it is faithful. We can also see that D j i (e) = δj i , ie.
DL (e) = I. Also, the other matrices in the representation must have a 1 as their (ji)th element and 0
for all other elements in row j and column i; by inspection, this 1 is never on the diagonal.
Similarly, we define a regular representation, DR
g , for the right action of a group:
(
1 gj g = gi or gi g−1 = gj
DR g gi = gi g
−1
= Di j (g) gj Di j (g) =
0 gj g 6= gi or gi g−1 6= gj
A word of caution: do not confuse the dimension of a representation, ie. of its carrier space (the space of
functions on which group operators act), with the dimension of the coordinate space on which these functions act.
2.4.5 Unitary representations (BF10.6)
A representation Dg is unitary if D†g = D−1 −1 ∗
g ), ∀ g ∈ G. In terms of matrix elements, Dij (g ) = Dji (g). For
example, D(3) and D(4) for SO(2) are unitary. Left and right regular representations are also unitary.
Now, if Dg is not already unitary, we can always find a similarity transformation matrix P S,† the Hermitian
2
square root of the positive semi-definite (ie., with eigenvalues λn > 0) matrix: S = g Dg Dg , such that
D′g = S Dg S−1 is unitary EXERCISE—first, show that D†g′ S2 Dg′ = S2 , then apply S−1 on the left and on
the right . Any representation of a finite group is equivalent to a unitary representation. This is also true for
certain infinite (continuous) groups, such as compact Lie groups.
40
Lecture Notes on Mathematical Methods 2022
Definition 2.25. Let {f (i) } be a subspace H(i) of the carrier space H of functions on which the linear
transformations Tg associated with a group G act. If, ∀ f (i) ∈ H(i) and ∀ g ∈ G, Tg f (i) ∈ H(i) ,
the subspace is invariant under G. Also, it can be shown that if a subspace of the carrier space is
invariant under a unitary representation, its complement must also be invariant.
Definition 2.26. Let H(1) and H(2) be subspaces of a Hilbert space H such that H is the sum of the
two subspaces with zero intersection. Then, if any function in H can be written uniquely as the sum
of a function in H(1) and another in H(2) , H is called the Kronecker (or direct) sum of H(1) and
H(2) , written H = H(1) ⊕ H(2) . The dimension of H is the sum of the dimensions of H(1) and H(2) .
Dg = a1 D(1) (2) (N )
g ⊕ a2 Dg ⊕ · · · ⊕ aN Dg (2.4)
(i)
where ai is the number of times (its multiplicity) the irreducible representation Dg occurs in the
direct sum, and N is the number of different irreducibl;e representations in the direct sum.
It can be shown that every representation of a finite group is either irreducible or fully reducible.
When the n-dimensional function space H has proper invariant subspaces, it means that there are at least
two subspaces in H, each of which has its own set of linearly independent functions that transform among them-
selves. Indeed, let HA be an invariant subspace of dimension d, and let {e1 , . . . , ed , . . .} be a basis of H with
A A
{e1 , . . . , ed } a basis of H . We write vectors of functions in H in block form B , where A ∈ HA has dimen-
sion d, and B belongs to the complement subspace HB , of dimension n - d. When HB is invariant, as it always is
in cases of interest to physics (see section 2.4.5
just then a block-diagonal representation matrix Dg , with
below),
A A′
block submatrices DA B
g and Dg , maps vectors B to B′ where A′ ∈ HA , B ′ ∈ HB . Also, since:
! ! ! !
DA
g 0 DA
g′ 0 DA A
g Dg′ 0 DA
g g′ 0
= =
0 DB
g 0 DB
g′ 0 DB B
g Dg′ 0 DB
g g′
DA B A B
g and Dg do preserve the group product, as they should. Dg has dimension d, and Dg dimension n - d.
Then, if all the matrices Dg in a representation can be brought into diagonal-block form by the same simi-
larity transformation, the representation is reducible to lower-dimensional representations composed of the block
matrices.
Going back to SO(2), the D(2) representation we have obtained is clearly fully reducible to the irreducible
representations D(3) = eiα and D(4) = e−iα , so we can write it as D(2) = D(3) ⊕ D(4) .
Example 2.8. The 5-dimensional representation, D(4) , we have constructed for S3 in Example 2.7 is
(by construction) reducible since it is in block-diagonal form, so D(4) = D(1) ⊕ D(2) . What about that
1 0 0 1
last two-dimensional representation
0 1 and 1 0 ? The first is already in block-diagonal form and
1 0
the second can be diagonalised to 0 −1 , Therefore, we obtain two 1-dim irreducible representations,
41
Lecture Notes on Mathematical Methods 2022
one identical to the identity representation D(5) = 1, and the other the “parity” representation D(3) .
Then we can write: D(4) = D(1) ⊕ D(3) ⊕ D(5) . What about the (left) defining representation of S3 ,
D(1) : is it reducible?
The defining representation of SN has dimension N . This always reducible representation reduces
to a 1-d representation and a (N - 1)-dimensional irreducible representation. To see how this comes
about, let (x1 , . . . , xN ) be a set of coordinates in the carrier space of a defining representation. It is
easy to construct a fully symmetric combination of all those coordinates:
x1 + . . . + xN
X =
N
This function spans the 1-dim subspace of RN invariant under any permutation of the coordinates;
the subspace thus qualifies as the carrier space of the 1-dim irreducible representation of SN that in
section 2.4.8 will be labelled by (N ). Since the defining representation is unitary, the complementary
subspace is itself invariant, and is the carrier space of another irreducible representation. Indeed, let
this (N -1)-dim subspace be spanned by N - 1 functions of the mixed-symmetry form:
These N -1 Jacobi coordinates can be shown to be linearly independent, so that there is no proper
invariant subspace, and the representation is irreducible. The functions are symmetrised with respect
to j - 1 coordinates and then antisymmetrised with respect to the j th one. This allows us to identify
the representation with another irreducible representation of SN that we will label (N -1 1), and the
defining representation can be written as (N ) ⊕ (N -1 1).
The defining representation, D(1) , of S3 is reducible to two irreducible representations, D(5) = 1 and
a set of six 2-dim orthogonal matrices, three with determinant +1 (rotations in a plane by angles 0,
±π/3) and three with determinant −1, thus showing that S3 ⊂ O(2)!. As expected, D(4) is fully
reducible. Can you see why these irreducible representations could not all be one-dimensional?
So this reduction algorithm certainly works, but it would be nice not to have to rely on looking for invariant
subspaces and similarity transformations, which can get quite involved.
(3) (2 1) 13
The sequence of representation labels is the reverse of that for classes! But if they are not cycles, what are they?
To discover the meaning of these Young diagrams we consider how the corresponding permutations act on
functions in the carrier space of the N !-dimensional regular representation of SN . We start by giving ourselves a
set of functions {ψi } (1 ≤ i ≤ N ), each of one variable, where the choice of the same symbol as for particle wave-
functions in quantum mechanics is intentional (some authors use the Dirac notation for them). Then with products
42
Lecture Notes on Mathematical Methods 2022
of these we construct functions of N variables xj . For instance, the product ψ(1...N) := ψ1 (x1 ) · · · ψ1 (xN ) spans
a one-dimensional subspace which contains functions which are obviously completely symmetric and invariant
under any of the N ! possible permutations of the variables. Thus, our subspace qualifies as an invariant subspace
for the regular representation, and it makes sense to associate it with the 1-dim irreducible identity representation
which has the same matrix, 1, for all elements of SN . We shall follow the usual convention by associating it with
the single Young diagram with one row of N boxes. Its label will therefore always be (N ).
With the same set {ψi }, we can also construct the completely antisymmetric function:
ψ1 (x1 ) ··· ψ1 (xN )
.. ..
ψ[1...N] = . ··· .
ψN (x1 ) · · · ψN (xN )
This function changes sign under any transposition in its set of variables, and the 1-d subspace it spans is also
invariant, because the function resulting from multiplying ψ[1...N] by ±1 is obviously in the same subspace. We
associate this subspace with the 1-dim irreducible representation which sends each element of SN to its parity, +1
or −1. Again by convention, this in turn corresponds to the single one-column Young diagram with N rows.
Other irreducible representations, and thus Young diagrams, have a mixed symmetry which can be used to find
their dimension. This is even stronger than eq. (2.14) which is only a constraint on the possible dimensions. Here
is one way to do this.
• Take the Young diagram for each irrep, and fill each of its N boxes with numbers from 1 to N in all possible
permutations to generate N ! Young tableaux. Then assign a function with N subscripts, living in the carrier
space of the regular representation of SN , to each tableau. The order of the subscripts follows the order of
numbers in the first row, then the second row, until the last row. These functions represent products of
functions, each of one coordinate, but we no longer treat them explicitly as such. They form a basis for the
carrier space of the regular representation.
• Symmetrise each function with respect to the numbers in each row of the tableau, and antisymmettrise the
result with respect to the numbers in each column. This yields, for each diagram, a new, mixed-symmetry
function, ψ (i) (1 ≤ i ≤ N ), that is a linear combination of the previous N ! basis functions for the carrier
space of the regular representation.
Example 2.9. For the (2 1) irreducible representation of S3 , the Young tableaux and corresponding
mixed-symmetry functions would be:
1 2 1 3
3 Ψ(1) = ψ123 + ψ213 − ψ321 − ψ231 2 Ψ(2) = ψ132 + ψ312 − ψ231 − ψ321
2 1 2 3
3 Ψ(3) = ψ213 + ψ123 − ψ312 − ψ132 1 Ψ(4) = ψ231 + ψ321 − ψ132 − ψ312
3 1 3 2
2 Ψ(5) = ψ312 + ψ132 − ψ213 − ψ123 1 Ψ(6) = ψ321 + ψ231 − ψ123 − ψ213
The question now is, are these mixed functions independent? Since we expect the regular repre-
sentation to be reducible (fully reducible, in fact), there should exist a lower-dimensional invariant
subspace, the carrier space of our irreducible representation of interest, and we should be able to show
that there are only nα < 6 (for S3 ) independent combinations, where nα will be the number of basis
functions for the invariant subspace, and therefore the dimension of the irreducible representation of
S3 carried by that space.
We note immediately that linear combinations that differ by a transposition of numbers in a column of
their tableaux cannot be independent: they are the negative of one another. So we have at most three
linearly independent combinations. But we also see that Ψ(1) − Ψ(2) − Ψ(3) = 0, leaving only two
independent combinations, which we take to be Ψ(1) and Ψ(2) , and which are the basis functions for
the carrier space of a 2-dim irreducible representation.
43
Lecture Notes on Mathematical Methods 2022
This rather tedious procedure can be made much faster by filling the tableaux in all the possible ways subject
to the following rules: the number 1 fills the uppermost, leftmost box; and the numbers must increase down any
column and to the right along any row. The number of ways this can be done is the dimension of the representation.
For instance, the (2 1) Young diagram of S3 generates the two tableaux with so-called standard numbering:
1 2 1 3
3 2
Ψ(1) Ψ(2)
each corresponding to one basis function in the 2-dimensional invariant subspace carrying the (2 1) irrep of S3 .
There is, however, a much more convenient method for calculating the dimension of the representation associ-
ated with a Young diagram if one does not wish to construct bases for the subspaces:
Definition 2.29. For any box in the Young diagram associated with an irreducible representation,
draw a straight line down to the last box in its column and to the right end of the box’s row. The result
is called a hook and the number of boxes traversed by the hook is the hook length of this box.
Then the dimension of an irreducible representation is the order of SN , N !, divided by the product of the N hook
lengths for the associated diagram.
Definition 2.30. Irreducible representations for which the Young diagrams are the transpose of each
other, ie. for which the length of each row in one is equal to the length of the corresponding column
in the other, are said to be conjugate. Their dimensions are the same.
The Young diagram of a self-conjugate irreducible representation is identical to its transpose.
2.5 Schur’s Lemmas and Symmetry in the Language of Group Theory (BF10.6)
We now present two fundamental results of group theory which provide useful criteria for the irreducibility of
representations as well as insight into symmetries, and which lead to relations that help to classify representations.
2.5.1 What is a symmetry in the language of group theory?
Consider a linear operator L such that, ∀ f ∈ H, [Lx f ](x) = h(x) ∈ H. Under a group G, [Tg Lx Tg−1 ][Tg f ](x) =
[Tg h](x), and [Lx′ f ](x′ ) = h(x′ ), so that L transforms under G as: Lx′ = Tg Lx Tg−1 .
Definition 2.31. When Tg Lx Tg−1 = Lx , ∀ g ∈ G, L is said to be invariant under the action of
group G. Since this condition can also be written as Tg L = L Tg , ∀ g ∈ G, then an operator that
is invariant under a group of transformations must commute with all those transformations. If also
[Tg f ](x) = f (x), f is invariant under G itself as well (eg., f (r) in polar coordinates under rotations).
If L has eigenvalues and eigenfunctions and is invariant under G, then there should exist a set {f i } such that:
Thus, if f i is an eigenfunction of L, so is Tg f i , with the same eigenvalue. Therefore, the distinct Tg f i are all
degenerate with respect to λ. If λ is degenerate, there are N degenerate f i for that λ, which form a basis for a
N -dim subspace of functions, characterised by λ. The Tg f i will then be some linear combination of {f j }. so
that the transformed eigenfunctions Tg f i also form a basis for the same subspace of functions as that spanned by
the eigenfunctions of L: the subspace is invariant under the action of the group, in the sense of Def. 2.25! With
summation up to N over repeated indices implied:
Tg f i = f j (Dg )j i (2.6)
Whenever we find (or observe) a set of degenerate eigenfunctions for some operator, the operator is invariant
under the action of a group, and these functions will be connected with an irreducible representation of the group.
44
Lecture Notes on Mathematical Methods 2022
so that all transformed vectors in H are eigenvectors of M, with the same eigenvalue λ. This can happen only if
M = λ I, and there comes Schur’s First Lemma:
The only complex matrix M that commutes with all the matrices of a given irreducible representation Dg is a
multiple of the identity matrix.
As a corollary, if a matrix can be found which is not a multiple of I and yet commutes with all matrices in a
representation, that representation must be reducible. This provides one handy test for reducibility.
From this Lemma follows an immediate consequence for Abelian groups, where any matrix Dg in a given
representation commutes with the matrices for all other group elements in this representation. Assuming a (n > 1)-
dim irreducible representation, the Lemma requires that Dg = λ I, ∀ g ∈ G. But the n × n identity matrix, which
is diagonal, cannot be irreducible if it represents all group elements, contradicting our assumption. We conclude
that all irreducible representations of an Abelian group are one-dimensional.
(α) (β)
Schur’s Second Lemma: If a non-zero matrix M exists such that Dg M = MDg ∀ g ∈ G, then D(α) and D(β)
must be equivalent irreducible representations. If D(α) and D(β) are inequivalent, M = 0.
This lemma can be proved (pp. BF615–617) by assuming unitary representations. This makes for no loss of
generality for finite or compact Lie groups, since these eg. O(n) have finite-dimensional representations.
2.5.3 An orthogonality relation for the matrix elements of irreducible representations (BF10.6)
Another important consequence of Schur’s Lemmas is the fact that the matrix elements of all the inequivalent irre-
ducible representations of a finite group, or those for infinite groups that have finite-dimensional representations,
form a set of orthogonal functions of the elements of the group. More specifically, if {Dg } is the set of all matrices
Dg in an irreducible representation D, then, for two such representations labelled by α and β::
NG
X i (β) l NG i
Dg(α) k
Dg−1 = δ j δk l δαβ (2.7)
g
j nα
where NG is the order of the group and nα is the dimension of D(α) . The sum is not matrix multiplication! Each
(α) (β)
term is the product of some ik entry of Dg and lj entry of Dg−1 , with ik and lj the same for each term.
In the usual case of unitary representations, this relation simplifies to:
NG
X i l NG i l
Dg(α) k
Dg(β) ∗ j
= δ j δ k δαβ (2.8)
g
nα
45
Lecture Notes on Mathematical Methods 2022
where D(α) and D(β) are m-dim and n-dim inequivalent irreducible matrix representations of G, and X is any
arbitrary operator represented by a m × n matrix X. Then, for some g′ ∈ G,
NG
X
(α) (β) (α) (β)
Dg′ M Dg′ ]−1 = Dg′ g X Dg′ g ]−1
g
The sum on the right-hand side is just a different rearrangement of the sum that defines M, so that:
(α) (β)
M = Dg′ M Dg′ ]−1
(α) (β)
Thus, Dg M = M Dg ∀ g ∈ G, and M meets the condition for Schur’s Lemmas. In particular, Schur’s lemma
requires that M = 0 if α 6= β since D(α) and D(β) are inequivalent. Now let us choose X to be a matrix whose
only non-zero element, 1, is its (kl)th entry. We can write this formally as: (Xkl )m n = δm k δn l . Inserting gives:
NG
X NG
X
i (β) n i (β) l
(Mkl )i j = D(α)
g m
(Xkl )m n Dg−1 = D(α)
g k
Dg−1
j j
g g
When α 6= β Mlk = 0, as we have seen. When α = β, Schur’s First Lemma requires that Mlk = λk l I, leading to:
NG
X NG
X
i (α) n i (α) l
(Mkl )i j = D(α)
g m
(Xkl )m n Dg−1 = D(α)
g k
Dg−1 = λl k δi j
j j
g g
Setting i = j and interchanging the D factors to get a matrix product, there comes:
NG
X XNG
(α) l
(α) j (α) l
Dg−1 Dg k
= Dg−1 g = NG δl k = λl k nα
j k
g g
χg = Tr Dg
Characters of reducible representations are compound; those of irreducible representations are called
simple. Language alert: Mathematicians speak of the “character” of a representation as the set of
traces of the matrices in the representation.
46
Lecture Notes on Mathematical Methods 2022
We establish an interesting fact: In a given representation, all matrices associated with elements of the same
class have the same trace. Recall that the class to which g belongs is made of {g′ g g′−1 } ∀ g′ ∈ G. Then the
trace of Dg′ g g′−1 is equal† to the trace of Dg , or χ. Since matrices for equivalent representations have the same
character, any statement about characters is basis-independent!
Now set k = i and l = j in eq. (2.8):
XNG
i j NG i j NG i
Dg(α) i Dg(β) ∗ j = δ j δi δαβ = δ i δαβ
g
nα nα
where repeated indices are summed over. Since δi i = nα , this can be rewritten as:
NG
X
∗ (β)
χ(α)
g χg = NG δαβ (2.10)
g
This provides our first orthogonality relation between the characters of irreducible representations. It can be viewed
as an inner product on the space of functions of the NG -dim “ character vectors”.
Some of the terms in this sum will be identical since they correspond to group elements in the same class. So
we can collect all terms belonging to the same class, which we label with k, and instead sum over the classes:
Nc
X (α)
nk χk χ∗k (β) = NG δαβ (2.11)
k=1
with nk the number of elements in class k and Nc the p number of classespin the group. This looks for all the world
like an orthogonality relation between two vectors, nk /NG χ(α) and nk /NG χ(β) , each of dimension Nc .
For a given irreducible representation, eq. (2.11) becomes:
Nc
X (α) 2
n k χk = NG (2.12)
k=1
Example 2.10. Take for instance the 3 × 3 representation of S3 found in section 2.4.3. The identity,
with trace 3, is in its own class, the three transpositions are in another class with trace 1, and the two
cyclic permutations) have trace 0. Eq. (2.12) gives: n1 χ21 +n2 χ22 +n3 χ23 = 1 (3)2 + 3 (1)2 + 2 (0)2 =
12. Since this is not equal to 6, the number of elements in S3 , the representation must be reducible.
According to eq. (2.11), the “ character vectors” of the Nr different irreducible representations are orthogonal.
There are Nr such orthogonal vectors, and their number may not exceed the dimensionality of the space, Nc , so
that Nr ≤ Nc . We will need this result a little later.
2.5.5 Multiplicity of irreducible representations and a sum rule for their dimension
Now consider the decomposition of a fully reducible representation into a direct sum of irreducible ones, given in
(α)
eq. (2.4). Taking its trace yields an equation for the compound character χg : χg = aα χg , where the sum runs
over the Nr irreducible representations of the group. The compound character is seen to be a linear combination
of simple characters with positive coefficients equal to the multiplicity of each irreducible representation.
Multiplying this relation by χ⋆ (β) (g) and summing over group elements, we find from eq. (2.10):
XNG NG
X
χg χ⋆g (β) = aα χ(α)
g χg
∗ (β)
= aα NG δαβ = aβ NG
g g
Thus, the multiplicity of each irreducible representation in the decomposition of a reducible representation is:
†
This is because Tr AB = Ai j Bj i = Bj i Ai j = Tr BA.
47
Lecture Notes on Mathematical Methods 2022
NG Nc
1 X 1 X ∗ (α)
aα = χg χg∗ (α) = n k χk χk (2.13)
NG g NG
k
Also, we can exploit the regular representation to obtain other general results for irreducible representations.
As we have seen in section 2.4.4, the matrix elements of the regular representation can only be 1 or 0. Since
only the identity will map a group element to itself, the only matrix with 1 anywhere on the diagonal is the identity
matrix. Therefore, the characters all vanish except by emailing [email protected] for χ(e) = NG .
Now, with nα the dimension of the αth irreducible representation and g = e, eq. (2.13) gives:
1
aα = χe χe∗ (α) = χe∗ (α) = nα
NG
Only χ(e) can contribute to the sum since χg = 0 in the regular representation when g 6= e.
Therefore, the multiplicity of an irreducible representation in the decomposition of the regular representation is
its dimension, and it is never zero. All the irreducible representations of a group must appear in the decomposition
of its regular representation. by emailing [email protected]
Next, taking the P
trace of the Kronecker decomposition (2.4) for the identity element in the regular representa-
tion yields: NG = α aα nα . Combining those results, there comes an important sum rule:
X
NG = n2α (2.14)
α
√ √
This powerful constraint tells us that nα ≤ NG so that any representation of dimension larger than NG
must be reducible. When NG = 2 or 3, all irreducible representations are one-dimensional. When NG = 4, we can
have only four inequivalent 1-d irreducible representations; nα = 2 is ruled out because there would be no identity
1-d representation. When NG = 5, eq. (2.14) does allow the identity representation plus one 2-d irreducible
representation; but we know that this group, Z5 , is Abelian, and so admits only five inequivalent 1-d irreducible
representations. For NG = 6, six 1-d, or two 1-d plus one 2-d irreducible representations, are allowed.
Thus, the characters in a given class k can be considered as components of Nc vectors forming a basis of a
space whose dimension is Nr , the number of irreducible representations, and which, according to eq. (2.15), are
orthogonal. But in a Nr -dimensional space there cannot be more than Nr orthogonal vectors, so Nc ≤ Nr .
In section 2.5.4, however, we had argued that Nr ≤ Nc . These results together lead to the important statement:
The number of inequivalent irreducible representations of a group is equal to the number of classes: Nr = Nc .
Now it can be shown (see Appendix F) that the direct product of an irreducible representation with a 1-d
representation is itself an irreducible representation, which may be the same (when the 1-d representation is the
identity). This goes for their characters also. When the completely antisymmetric (1N ) 1-d representation exists,
as is the case for SN , the characters of an irreducible representation can always be written, class by class, as the
product of the characters of its conjugate representation and the characters in the (1N ) representation. Therefore,
characters for a given class in a pair of conjugate representations are either identical or differ only by their sign.
Characters of a self-conjugate representation in a class that has negative parity must vanish.
48
Lecture Notes on Mathematical Methods 2022
• The first says that complete rows in the table (each for a different representations) are orthogonal, with the
understanding that each term in the sum is weighted by the number of elements in the class (column).
• The second says that complete columns (each belonging to different classes) are othogonal.
Now, if β refers to the identity representation, then, for any irrep α other than the identity, eq. (2.11) becomes:
Nc
X (α)
n k χk = 0 (2.16)
k=1
When invoking the orthogonality constraints to find characters, it is best to apply the linear ones first. Unfortu-
nately, many of these relations will be automatically satisfied and will not yield new information, because of the
strong constraints on the characters imposed by conjugation and self-conjugation of the irreducible representations.
When all possible information has been extracted from eq. (2.16) and (2.15), and there still remain unknowns, one
can try to spot reasonably simple quadratic relations from eq. (2.11) as well as using the normalisation of rows and
columns.
Two last but important remarks: the characters of any 1-dim representations of any group (eg. those of an
Abelian group) must preserve the group product. Also, although the characters of SN are real, characters of other
groups (eg. Zn ) can be complex.
There exist even more sophisticated methods for determining the characters of a group (eg. by generating them
from the characters of a subgroup, or of a factor group), but lack of time and space prevents us from discussing
them here. In fact, character tables for well-known groups can be found in specialised books and on the web.
Let us use these rules to find the characters of S3 as a 3 × 3 table, with classes corresponding to columns and
irreducible representations to rows. The first and last row can be immediately written down from our knowledge
of the parity of each class (−1 for the transpositions and +1 for the cyclic permutations). Note also that the (2 1)
representation is self-conjugate, so we can put 0 for the character in the (2 1) class, because the parity of that class
(last character in the column) is −1. The (2 1) representation is the (N -1 1) representation discussed above, and
its remaining character is determined by its belonging to a class with α = 0; thus, the character must be −1. The
linear constraint (2.16), as well as the other orthogonality rules, are automatically satisfied. Collecting yields:
49
Lecture Notes on Mathematical Methods 2022
(13 ) (2 1) (3)
nk 1 3 2
(3) 1 1 1
(2 1) 2 0 −1
(13 ) 1 −1 1
EXERCISE: work out the character table and irreducible representations of Z4 , the cyclic group of order 4. You
may make the task easier by remembering that products of characters belonging to a 1-d irreducible representation,
which are the actual representation matrices, must mimic the group product of the corresponding elements.
where the D(l) matrix is an irreducible representation of the rotation group SO(3) which acts on the
invariant space spanned by the 2l+1 Ylm for that l. SO(3) will be discussed in chapter 3.
We can simplify things by noting that rotations by an angle α about any axis are all equivalent to (in
the same class as) a rotation by that angle around the z-axis. It will be sufficient to calculate the trace
of the matrix representing rotations around that axis. To find this matrix, notice that [Rα Ylm ](θ, φ) =
eimα Ylm (θ, φ) = Ylm (θ, φ + α) because the dependence of the spherical harmonics on φ is eimφ .
Therefore, D(l) (α) = diag(e−ilα , e−i(l+1)α , . . . , eilα ), and its character is not hard to compute:
l 2l
!
X X 1 − ei(2l+1)α sin (l + 1/2)α
(l) iα m −ilα iα n −ilα
χ (α 6= 0) = (e ) = e (e ) = e =
n=0
1 − eiα sin(α/2)
m=−l
(2.17)
where we have recast the sum as a geometric series by redefining the index as m = n − l.
Now let us turn on a weak interaction whose corresponding potential is no longer rotationally-invariant,
but still retains invariance under rotations by a restricted, finite set of angles, which we collectively
denote by β. This would happen, for instance, if we embed our spherically-symmetric atom in a crys-
tal lattice. Suppose this restricted set of rotations actually is a group, or more precisely, a subgroup of
SO(3). Then the matrix D(l) (β) should be a representation of that subgroup, but that representation
may no longer be irreducible. This will certainly happen for any D(l) whose dimension is too large to
satisfy the sum rule (2.14) that applies to the finite subgroup.
m′
The set of Ylm transform as: Rβ Ylm = Ylm′ D (l) m (β), with summation over repeated indices
implied. If the induced representation D of the restricted-symmetry subgroup is reducible, there
exists a matrix S independent of β which transforms all its matrices into block-diagonal matrices
D′ = S D S−1 , something which was impossible when there was no restriction on the angles.
But we do not have to know S to extract useful information. Indeed, because D and D′ have the
same trace, we can calculate the characters of D(l) (β) for all elements of the restricted-angle subset
in SO(3). Then we find the character table of the restricted-symmetry group, which is finite. If
there is a row in the table that exactly matches the SO(3) characters of D(l) (β), then D(l) (β) is
not only an irreducible representation of SO(3), it is also an irrep of its subgroup defined by the
50
Lecture Notes on Mathematical Methods 2022
angles allowed by the restricted symmetry. The corresponding invariant subspaces are identical, and
the original 2l + 1-fold degeneracy for that value of l is still present after the perturbation has been
turned on. As l increases, however, the dimension 2l + 1 of D(l) (0), which always appears as the first
character corresponding to the identity class of SO(3), will eventually exceed the fixed dimension of
any irreducible representation of the subgroup. Then all the corresponding D(l) (β) will be reducible to
a direct sum of the irreducible representations of the subgroup, given by eq. (2.4), with the multiplicity
of each irrep calculable from eq. (2.13).
For instance, suppose that the perturbation has cubic symmetry. A cube is invariant under1 :
• 6 rotations by ±π/2 around the three axes through its centre that intersect faces through their
centre;
• 3 rotations by π around these same axes;
• 8 rotations by ±2π/3 around the four axes through diagonally opposed corners (vertices).
• 6 rotations by π around the six axes intersecting the centre of two diagonally opposed edges;
With the identity rotation, these add up to 24 elements forming a subgroup of SO(3) isomorphic to
S4 . The correspondence between rotations and permutations is obtained by considering each rotation
as a shuffling of the four pairs of diagonally opposed vertices (or the four principal diagonals through
the centre), each pair labelled 1 to 4. The five classes of S4 are (14 ) (e), (2 12 ) (rotations by π), (22 )
(rotations by π) , (3 1) (rotations by ±2π/3), and (4) (rotations by ±π/2). The S4 character table is:
(14 ) (2 12 ) (22 ) (3 1) (4)
nk 1 6 3 8 6
(4) 1 1 1 1 1
(14 ) 1 −1 1 1 −1
(22 ) 2 0 2 −1 0
(3 1) 3 1 −1 0 −1
(2 12 ) 3 −1 −1 0 1
Here, the irreps of S4 (or of the group of rotational symmetries of the cube) are ordered by increasing
dimension instead of their mixed-symmetry structure. With eq. (2.17), we calculate the characters
of the representations of S4 induced by D(l=1) (β) and D(l=2) (β), with angles β running through the
values corresponding to the five classes of S4 :
(14 ) (2 12 ) (22 ) (3 1) (4) (14 ) (2 12 ) (22 ) (3 1) (4)
(l=1) (l=2)
D 3 1 −1 0 −1 D 5 −1 1 −1 1
The l = 1 irrep of SO(3) restricted to the angles allowed by the cubic-symmetry subgroup has the
same dimension and the same characters as the representation (3 1) of S4 in the above character
table for S4 . The invariant spaces are the same and there is no lifting of the unperturbed 3-fold
degeneracy. The l = 2 irrrep of SO(3), however, has no identical row in the S4 character table,
and must correspond to a reducible representation of S4 . With eq. (2.13), we calculate the following
multiplicity for each irrep of S4 that can appear in the decomposition of D(l=2) (β): a(4) = a(14 ) =
a(2 12 ) = 0, and a(22 ) = a(3 1) = 1. Then we have the S4 decomposition:
D(l=2) (β) = D(22 ) (β) ⊕ D(3 1) (β)
The unperturbed 5-fold degeneracy of the l = 2 states is partially lifted to become two “levels”, one
3-fold and one 2-fold degenerate.
Another example illustrating how symmetry-breaking can remove degeneracy, at least in part, can be
found in Appendix G.
1
See, eg: http://demonstrations.wolfram.com/RotatingCubesAboutAxesOfSymmetry3DRotationIsNonAbelian/.
51
Lecture Notes on Mathematical Methods 2022
Appendices
D The Right and Left Actions of a Group on a Vector, with Sn as Example
First, we recall some important properties of linear transformations. For simplicity we shall take these transforma-
tions to act on vectors x = xi ei ∈ Rn ,, with thestandard basis {ei }. It is customary to distinguish between active
transformations: x −→ x′ = x′i ei = ei Ai j xj , where the transformed coordinates x′i are those of a new vector;
and passive transformations that leave x invariant but transform the basis to {e′i }: e′i = ej P j i .
Paaive transformations, unlike active ones, require the transformed vector x′ to remain the same as the initial
one, because all we have done is change the basis. To preserve x, we must also transform its components with the
inverse transformation:
j
x′ = e′j x′j )P = ei P i j P −1 k xk = ei δi k xk = ei xi = x
Now, although an active transformation is quite different from a passive one, the result should be mathematically
equivalent to the result of transforming the basis with the inverse transformation. In other words, the components,
(x′A )j , of the new vector produced by the active trnsformation should be the same as the those of the initial vector
j
in the transformed basis, that is, (x′A )j = (x′P )j = P −1 k xk . Comparing with (x′A )j = Aj k xk , we see that
that the passive transformation is indeed inverse to the active one.
Notice that the action on the basis vectors involves the matrix for g−1 , as expected for the passive transformation
associated with the active one on the components.
(a) Components
i i
For g = πk ∈ Sn , the expression for xi πk is equivalent to xπk (i) = DgR j
xj , so that DgR j
= δπk (i) j ,
and the ith row of the n-dim DRg matrix is the standard-basis vector eπk (i) . For πk = π5 = (1 3 2) ∈ S3 , for in-
R 1 1 1 1 1 1
stance, we find: Dπ5 1 , Dπ5 2 , DπR5 3 = (0, 0, 1), and for π6 = (1 2 3): DπR6 1 , DπR6 2 , DπR6 3
R
1 n
the standard basis: ei πk = DπR−1 , . . . , DπR−1 . More succinctly, because eπk (i) = ei πk , we get:
k i k i
j
DπR−1 = δj π−1 (i) , or the .ith column is the standard-basis ector eπk (i) .
k i k
R 1 2 3
Using basis vectors to find DR
π5 , for instance, we have: e1 π5 = e3 = (0, 0, 1) = Dπ6 1 , DπR6 1 , DπR6 1 ,
1 2 3
and e2 π5 = e1 = (1, 0, 0) = DπR6 2 , DπR6 2 , DπR6 2 , etc. The resulting matrix is the one that was
obtained from the components for the right action of π6 , and its inverse will be the matrix for the right action
of π5 , the same that we found somewhat more directly by acting on components.
where the last equalities on the right hold when the matrices are orthogonal (eg., rotations). In that case we can
find the left-action matrix for g directly.
0 1 0 0 0 1
DLπ5 = 0 0 1 DLπ6 = 1 0 0
1 0 0 0 1 0
These are the DL matrices found in example 2.7. The left and right representations for each single transposition
are identical.
Components
i i
In the case of g = πk ∈ Sn , we have: xπk (i) = xj DπLk j , so that DπLk j = δj πk (i) . Working with our
trusted friends (π5 , π6 ) ∈ S3 , we obtain:
0 1 0 0 0 1
DLπ5 = 0 0 1 DLπ6 = 1 0 0
1 0 0 0 1 0
53
Lecture Notes on Mathematical Methods 2022
where Nr is the number of irreducible representations. Again, the left-hand side is not matrix multiplication.
Take the equation for each element g of some class k, and sum over all elements of the class; we can also do
this with g′ over the elements of another class k′ . When k 6= k′ , the right-hand side of the double summation must
vanish because classes are distinct; when k = k′ , the double sum collapses into one which adds up to nk .
Pn k (α) i
g Dg j
in the now quadruple sum on the left-hand side is an element of the matrix M constructed by
Pn k
summing all the matrices Dg in the representation that correspond to elements g of class k: M = g Dg .
If g′ is some arbitrary element of G, we have:
X X
Dg′ M Dg′−1 = Dg′ Dg Dg′−1 = Dg′ g g′−1 = M
g g
where the last equality results from the fact that, since g′ g g′−1 is in class k, the left-hand side of the last
equality is just a rearrangement of the sum defining M. Thus, Dg M = M Dg ∀ g ∈ G, and, from Schur’s First
Lemma, M = λ I, with λ a constant that depends on the class and on the n-dim representation. Then Tr M = nλ.
Because all matrices in a class for a given representation must have the same trace, we have: Tr M = nk χk .
Since that trace is also nλ, we find:
nk
M = χI (E.2)
n
With two of its four sums replaced by matrix elements of M, the completeness relation (E.1) now reads:
Nr X
X nα
nα (α) i (α) ∗ i
Mk j
Mk′ j
= nk δk′ k
α
N G
i, j
(α)
Inserting M = (nk /nα )χ(α) I and carrying out the sums over i and j gives another orthogonality relation:
XNr
nk (α) ∗ (α)
χk χk ′ = δk′ k (E.3)
N
α=1 G
Invariance under S means that M−1 K and S commute. In fact, this last statement can be used to obtain the form
of the M−1 K matrix given above.
The eigenvectors of S now satisfy SA = λA. But since S6 = I, we immediately find that the eigenvalues are
the sixth roots of 1, as expected for the cyclic group. Therefore. λ(m) = eimπ/3 , (0 ≤ m ≤ 5). To each value of m
corresponds an eigenvector A(m) with components Aj(m) = λ(m) j−1
= eim(j−1)π/3 .
These eigenvectors are also the normal modes of the system. Inserting into the eigenvalue equation M−1 K A(m) =
2 A
ω(m) (m) with the coupling parameters ω(5) = ω(1) and ω(4) = ω(2) yields the dispersion relation:
6
X
2 2
ω(m) = ωj−1 eim(j−1)π/3 = ω02 − 2 ω12 cos mπ/3 − 2 ω22 cos 2mπ/3 − (−1)m ω32
j=1
We note that A∗(1) = A(5) , and A∗(2) = A(4) . These modes are complex, which is a problem if they are sup-
posed to correspond to real relative amplitudes. But we also note that ω(1) = ω(5) , and ω(2) = ω(4) ; therefore,
the corresponding eigenvectors span two invariant 2-dim subspaces, which allows us to take appropriate linear
combinations of the eigenvectors to turn them into real modes of the same frequency.
The coupling has lifted the original 6-fold degeneracy of the uncoupled system, but there is still some degen-
eracy left because of the two 2-dim subspaces.
This is as far as we can go without knowing the interaction parameters themselves. But we have succeeded in
nailing down the relative amplitudes of motion of the bodies in each normal mode without that explicit knowledge!
55
Lecture Notes on Mathematical Methods 2022
Definition 3.1. Let P be any point in a n-dim manifold M n which is obtained from two other points,
P1 and P2 from invertible mappings P = φi (P1 , P2 ). Let g(P1 ) ⋆ g(P2 ) = g(P ) be the group product
of an infinite group G. If the maps φi and their inverse are differentiable, then G is a Lie group.
The important point to remember here is that since they correspond to points in a manifold, elements
of a Lie group can be parametrised in terms of smooth coordinates on this manifold.
A Lie group is real if its manifold is real and complex if its manifold is complex.
The dimension of a Lie group is the dimension of its manifold.
Definition 3.2. A Lie group is said to be path-connected if any pair of points on its manifold is
connected by a continuous path.
A Lie group is compact when the volume of its manifold is finite.
The subset of all elements in a Lie group whose corresponding points in M n are connected by a continuous path to
the identity is a subgroup. Thus, a Lie group that is not path-connected must contain a path-connected subgroup.
Example 3.1. An infinite line with a coordinate patch −∞ < x < ∞ (x ∈ R) is a 1-dim manifold.
In section 2.1.1 we stated that C was a continuous group under addition. So is R itself, and if we write
a group element as g(x) = ex , we can easily deduce the function corresponding to the group product.
Indeed, g(z) = g(x) ⋆ g(y) = g(x + y), and we are not surprised to find that: z = φ(x, y) = x + y.
Example 3.2. Restrict θ = x ∈ R with 0 ≤ θ < 2π, and define group elements g(θ) = eiθ with product:
The group manifold here is the unit circle, S 1 , whose points are each parametrised by real angle θ,
and φ(θ1 , θ2 ) = θ1 + θ2 . Its elements are complex, but the group is real! It is Abelian, and connected.
Example
3.3. Real invertible 2 × 2 matrices form a group whose elements can be written as g(x) =
x1 x2
x3 x4 . Constraining the matrices to be unimodular (to have determinant 1) lowers the number of
parameters by 1. The group product is:
x1 x2 y1 y2 z1 z2
=
x3 1+xx12 x3 y3 1+yy12 y3 z3 1+zz12 z3
Compute the set of three functions zi = φi (x1 , x2 , x3 , y1 , y2 , y3 ) consistent with this group product:
1 + y2 y3 1 + x2 x3
z1 = x1 y1 + x2 y3 z2 = x1 y2 + x2 z3 = x3 y1 + y3
y1 x1
In this parametrisation, the mappings φi are all differentiable only off the x1 = 0 and y1 = 0 planes.
Whatever the associate manifold is—see later—it cannot be covered with just this coordinate patch.
The inverse mapping corresponding to g−1 (x) can be read off the inverse matrix g−1 .
57
Lecture Notes on Mathematical Methods 2022
Example 3.4. If we demand instead that invertible complex 2 × 2 matrices be not only unimodular,
but unitary as well, the treatment is simpler. Introduce the parametrisation:
z w a0 + i a3 a2 + i a1
=
−w∗ z ∗ −(a2 − i a1 ) a0 − i a3
with the condition |z|2 + |w|2 = a20 + a21 + a22 + a23 = 1 which guarantees that the matrix is unitary
with determinant equal to 1. The group manifold is thus the unit 3-sphere S 3 embedded in R4 with
coordinates (a0 , a1 , a2 , a3 ) ∈ R; this is a real three-dimensional Lie group.
Definition 3.3. Unitary transformations T of a complex matrix M ∈ GL(n, C) are defined by:
M 7→ T M T†
where the subgroup of matrices T leaves the Cartesian n-dim metric M = In invariant: TIn T† =
TT† = In . Thus, T−1 = T† , and we call that subgroup U (n) ⊂ GL(n, C): Both U (n) and its
matrices are unitary. Example 3.2 referred to U (1).
Definition 3.4. Orthogonal transformations T of a real matrix M ∈ GL(n, R) are defined by:
M 7→ T M TT
(TT is the transpose of T), such that T leaves In invariant: T In TT = TTT = In , that is, T−1 = TT .
The group of such matrices is called O(n) and is orthogonal.
Be aware that n in O(n) or U (n) refers to the dimension of the matrices, not that the group which is the number
of coordinates on its manifold! O(n) matrices have determinant ±1, whereas the absolute value of the complex
determinant of U (n) matrices is equal to 1. Thus, (can you see why?) O(n) is not path-connected; neither is U (n).
The group manifolds (and thus these groups themselves) are compact because their matrices define closed,
bounded subsets of the manifolds that parametrise GL(n, C) and GL(n, R). O(n) and U (n) preserve the length
(or norm) of n-vectors in Euclidean Rn , and therefore also angles between those vectors (eg., the angles of any
triangle are determined by the lengths of its sides).
We also have the non-compact groups which preserve the indefinite metric Iqp , defined by the transformations:
T Iqp TT = Iqp O(p, q) (3.1)
T Iqp T† = Iqp U (p, q) (3.2)
A famous example is O(3, 1), aka the full Lorentz group, that leaves the mostly positive Minkowski metric on
R4 (or space-time distance) invariant; equivalently, the norm of a 4-vector x is left invariant by 3-dim rotations,
Lorentz transformations (boosts), and space or time reflections. In principle, from the condition: T I13 TT = I13 ,
one could work out detailed constraints on the elements of the O(3, 1) matrices to find that there are six independent
parameters, but this would be needlessly messy. There are far better ways of parametrising the group to extract all
this information, and much more, as we shall see.
58
Lecture Notes on Mathematical Methods 2022
1. Translations Let f be an analytic function acting on Rn . The left action on f of the operator Ta associated
with Ta x = x + a is:
Except for the identity (a = 0), such transformations leave no x invariant and are called inhomogeneous.
2. Rotations
Parametrise 3-dim rotations in the z = 0 plane of a vector x ∈ R3 by Rα , with Rα φ = φ + α, with
[Rα f ](φ) = f (φ − α), and −π < φ ≤ π. In terms of the left action on the components of x: x′ = Rα x
(ie. x′ obtained by rotating x by +α in the z = 0 plane), the matrix associated with Rα is:
cos α − sin α 0
sin α cos α 0 Then : [Rα f ](x) = f (Rα−1 x) = f (x cos α + y sin α, −x sin α + y cos α, z) .
0 0 1
3. We also have scale transformations x′ = ax, with a ∈ R a non-zero positive constant, and x ∈ Rn in
Cartesian coordinates (think of zooming in or out). The restriction to Cartesian coordinates is important: in
spherical coordinates over R3 , only the radial coordinate would scale.
59
Lecture Notes on Mathematical Methods 2022
x′ = Λx + t
Whereas the homogeneous transformations left the norm of a 4-vector invariant, these inhomogeneous trans-
formations leave invariant only the norm of the difference between two 4-vectors.
If we call Λ the full Lorentz transformation matrix, we can construct the matrix for these transformations by
adding to Λ a fifth row and column whose last element is a 1 that does not do anything, that is:
′
x Λ t x
=
1 0 1 1
These matrices form the 10-parameter inhomogeneous Lorentz group, or Poincaré group, ISO(3, 1). Inci-
dentally, setting Λ = I gives a matrix realisation of the 4-dim translation group.
These examples illustrate the isomorphism between physical transformations and matrix Lie groups. We can
then identify, say, a rotation with a SO(3) matrix, and even call SO(3) the rotation group.
1. Translations
We can first look just at smooth functions f (x), x ∈ R. Then the result of a translation Ta x = x + a, a ∈ R,
on f , with a ≪ x, can be Taylor-expanded about x:
[Ta f ](x) = f (Ta−1 x) = f (x − a) = (1 − a dx + . . .) f (x) = exp(−a dx ) f (x)
In R3 this generalises to:
" ∞
#
X 1
[Ta f ](x) = f (Ta−1 x) = f (x − a) = (−ai ∂i )n f (x) = exp(−ai ∂i ) f (x) (3.3)
n!
n=0
The operators −∂i are called the infinitesimal generators of translations. Quantum mechanics uses instead
the Hermitian momentum operator p = −i~∂ and writes the translation operator as: Ta = e−ia·p/~ .
We note that the Cartesian infinitesimal generators −∂i (or pi ) commute amongst themselves.
2. Rotations
For rotations Rα φ = φ + α in the (z = 0) plane by a small angle α:
[Rα f ](φ) = f (Rα−1 φ) = f (φ − α) = (1 − α dφ + . . .) f (φ) = exp(−α dφ ) f (φ)
As we have seen in the last section, in R3 with Cartesian coordinates, this gives for the left action of a rotation
Rα x = (x cos α − y sin α, x sin α + y cos α, z): f (Rα−1 x) = f (x cos α + y sin α, −x sin α + y cos α, z).
If we Taylor-expand the right-hand side we obtain:
[Rα f ](x) = 1 + α (y ∂x − x ∂y ) + . . . f (x) = exp(α Mz ) f (x) (3.4)
where Mz = y ∂x − x ∂y . Similarly for rotations about the x and y axes, the general rotation operator is:
Rα,β,γ = exp(αMx ) exp(βMy ) exp(γMz ), where:
Mx = z ∂y − y ∂z , My = x ∂z − z ∂x , Mz = y ∂x − x ∂y (3.5)
60
Lecture Notes on Mathematical Methods 2022
or: Mi = −ǫijk xj ∂ k = 21 ǫijk J jk , where Jjk := x[k ∂j] , with xj and ∂k defining a 2-dim plane of rotation.
The pseudovector operator M is the Hodge dual of the more natural simple 2-form operator J. In quantum
mechanics, it is redefined as L = i~M and interpreted as the (Hermitian) angular-momentum operator.
These infinitesimal generators do not commute. Indeed: [Mi , Mj ] = ǫij k Mk , or [Li , Lj ] = i~ ǫij k Lk .
(Note: we could have written — some do! — defined M as the negative of the above. The cost, however,
would be an extra minus sign in the commutation relations.)
We can now find the infinitesimal generators of an arbitrary group of transformations with m parameters ai
near the identity, such that ai = 0 ∀ i for the identity group element. These transformations map a point in a
manifold M n (not the group manifold!) to another one nearby that can be described by the same coordinate chart.
Let the transformations act (left action!) on a space (aka carrier space) of differentiable functions f on M n :
Focus on Ta f , and take f as a function of the parameters ai . As before, Taylor-expand the right-hand side to first
order around the identity parametrised by a = 0:
h i
[Ta f ](x) = 1 + ai ∂ai (Ta−1 x)j ∂j + . . . f (x)
a=0
where i runs over the number of parameters, ie. the dimension of the group, and j from 1 to the dimension of the
space on which the functions f act.
For example, rotations in the z = 0 plane in Cartesian R3 involve one parameter (angle) a1 = α, and only x and
y derivatives can occur since z does not depend on α. Then the second term in the square bracket of eq. (3.4) is
recovered.
Ma = In + ai Xi Xi = ∂ai Ma (3.8)
a=0
where we understand that differentiating a matrix means differentiating each of its elements. The matrices Xi are
the infinitesimal generators of the group. Again, some prefer the definition Xi = −i ∂ai Ma a=0 .
61
Lecture Notes on Mathematical Methods 2022
cos θ − sin θ
Example 3.5. Let Mθ ∈ SO(2): sin θ cos θ , for 0 ≤ θ < 2π, that effects rotations in a plane.
1 −θ
Taylor-expand to first order: Mθ ≈ θ 1 = I2 + θ X
Then the infinitesimal generator of SO(2) is:
0 −1
X = ∂θ Mθ θ=0
=
1 0
a matrix fully consistent with the constraints on SO(n) generators as we shall discover in section
3.3.4. We shall write the space it spans as:
0 −θ
so(2) =
θ 0
Another example (EXERCISE) that is quite easy to work out is SL(2, R); it will have three infinitesimal
generators. Similarly, using the parametrisation of example 3.4, we see that an element of SU (2) may be written
as a0 I2 + ai Xi , where Xi = i σi are the generators of SU (2), with σi the Pauli matrices.
When the group matrices are not known we must resort to other methods to be discussed a little later.
An infinitesimal generator is an operator that effects an infinitesimal transformation away from the identity.
We want to reconstruct a finite transformation out of a succession of infinitesimal transformations that use only the
generators, ie., the first-order contribution in the expansion of a transformation or of a matrix:
n
Ma = lim (Ma/n )n = lim 1 + (ai /n)Xi
n→∞ n→∞
since the right-hand side of the first relation is equal to its derivative when n → ∞. Therefore, Ma = exp(ai Xi ).
This exponential map, then, is the tool that reconstructs finite transformations from infinitesimal ones. But it must
be handled with some care as we shall discover.
Note that the inverse of a group element eA is e−A , and that a generator matrix A = ai Xi need not be invertible.
g g′ ≈ e + ǫ (aX + bY ) + . . .
Manifestly, aX + bY is a generator for the product g g′ , and the generators indeed form a linear vector space.
62
Lecture Notes on Mathematical Methods 2022
Now expand the product h = g g′ (g′ g)−1 ∈ G to first non-vanishing order, this time writing g ≈ e + ǫ1 X +
ǫ21 X 2 /2,
and g′ ≈ e + ǫ2 Y + ǫ22 Y 2 /2, with (ǫ1 , ǫ2 ) arbitrarily small:
1 1 1 1
g g′ (g′ g)−1 ≈ e + ǫ1 X + ǫ21 X 2 e + ǫ2 Y + ǫ22 Y 2 e − ǫ1 X + ǫ21 X 2 e − ǫ2 Y + ǫ22 Y 2 + . . .
2 2 2 2
≈ e + ǫ1 ǫ2 (XY − Y X) + . . .
All other contributions of order ǫ2i and ǫ1 ǫ2 cancel out. We define [X, Y ] := XY − Y X, the commutator of the
generators X and Y . As the generator for g g′ (g′ g)−1 , [X, Y ] must be an element of the same vector space as X
and Y . When h = e, g g′ = g′ g, and the commutator of the generators vanishes. Thus, mathematicians often refer
to g g′ (g′ g)−1 as the “commutator” for the group product, but we shall reserve the term for [X, Y ].
It is straightforward to show that the Jacobi identity holds, just by expanding it:
X, [Y, Z] + Y, [Z, X] + Z, [X, Y ] = 0 (3.10)
Definition 3.6. An algebra g is a vector space equipped with, on top of the generic addition operation,
a bilinear product g × g −→ g. When the product is the Lie bracket [·, ·], which:
• is linear: [aX + bY, Z] = a [X, Z] + b [Y, Z] ∀ a, b ∈ R or C;
• is antisymmetric: [X, Y ] = − [Y, X];
• satisfies the Jacobi identity: X, [Y, Z] + Y, [Z, X] + Z, [X, Y ] = 0.
we say that g is a Lie algebra. In physics, the Lie bracket is the commutator XY − Y X. Many,
because they
alwaysdeal with the algebra,
not the group, use G to denote g, which can be confusing.
Because X, [Y, Z] − [X, Y ], Z] 6= 0, Lie algebras are non-associative.
It is crucial to keep in mind that the action of a Lie-algebra element X on another one, Y , is not XY , but their
commutator! The closure property of a Lie group in effect translates into the existence of its algebra.
The algebra ±i g is said to be essentially real. Example: the linear and orbital angular-momentum operators
of quantum mechanics related to real infinitesimal generators.
Sometimes, however, it proves very convenient to construct a complex extension of a real or essentially real
algebra, by allowing basis redefinitions that involve complex coefficients. For instance, we might wish to construct
J± = Jx ± iJy . This provides more flexibility in constructing useful bases.
The dimension n of a Lie algebra is the number of parameters of its associated group.
The coefficients Cij k are called the structure constants of the Lie algebra, whose structure they are said to
specify. In fact, with some rarely relevant caveats, they pretty much tell us everything about the group itself.
The structure constants inherit the antisymmetry of the commutators: Cji k = −Cij k . When the structure
constants all vanish, ie., when [X, Y ] = 0 ∀ (X, Y ) ∈ g, we say that the algebra is Abelian.
The Jacobi identity on elements of an algebra induces (EXERCISE) a relation between the structure constants:
Defining a matrix (Di )j k = −Cij k , we find (EXERCISE) that D satisfies the commutation relation (3.11). If
we can take the group’s representations to be unitary, as for compact groups such as SU (n) and SO(n), the
corresponding representations of the algebra are anti-Hermitian and we immediately find (EXERCISE), since they
must satisfy the commutation relations, that the structure constants are real.
The structure constants for the essentially real algebra ±i g are just (exercise) ±i Cij k . Very often, in the case
of essentially real algebras, people will call the Cij k themselves the structure constants instead of ±i Cij k .
Two Lie algebras are said to be isomorphic when they have the same dimension and structure constants, up to
a redefinition (eg. rescaling) of their generators.
Expanding, we arrive (exercise) at three conditions on the block matrices: S† = −S, V† = −V, T† = U.
Both the S and V diagonal blocks are antihermitian. The off-diagonal blocks are each other’s adjoint. Over R, this
means that A has two antisymmetric diagonal block matrices, one q × q and one p × p; the off-diagonal blocks
are the transpose of one another. The number of parameters of the indefinite orthogonal group O(p, q) is then
p(p − 1)/2 + q(q − 1)/2 + pq = n(n − 1)/2, the same as for the compact orthogonal group O(n).
There only remains to notice that the non-zero elements of the infinitesimal generator matrices can only be ±1
(over R) and also ± i (over C) because of the linearisation.
Another important constraint can be imposed on a group matrix M: det M = 1, which defines SL(n, R or C).
Since the determinant of a product of matrices is equal to the product of the determinants of the matrices, and
because—when a matrix A is diagonalisable— there exists a similarity transformation SAS−1 which takes A to
A′ = diag(λ1 , . . . , λi , . . .), we conclude that det A is equal to the product of the eigenvalues of A.
Also, if M = eA , it transforms as:
1 1 ′
S eA S−1 = S IS−1 + S A S−1 + S A S−1 S A S−1 + . . . = I + A′ + (A′ )2 + . . . = eA
2! 2!
′
where eA is a diagonal matrix with eλi as entries. In other words, the eigenvalues of eA are just eλi . Then:
Y X ′
det eA = eλi = exp λi = eTr A
i i
But Tr A′ = Tr(SAS−1 ) = Tr A. We obtain via this elegant (but limited to diagonalisable matrices!) deriva-
tion an important basis-independent relation, valid for any square matrix:
This extends to det (eA eB · · · ) = eTr (A+B+...) , and since all SL(n, R) matrices can be written as a prodiuct eA eB
(to be shown later), we immediately deduce that all matrices in the algebra sl(n, R) must have vanishing trace,
including those in su(n) and so(n). Thus, it can be said that sl(n, R) is the set of all traceless matrices of rank n.
Since antisymmetric real matrices are traceless, o(n) and so(n) are identical. This is very much related to the
absence of a continuous path from the O(n) identity (which is unimodular) to orthogonal matrices with determinant
−1: O(n) is not path-connected. Spatial inversions cannot be linearised; one cannot invert axes by a “small”
amount! So the infinitesimal generators of O(3) are those of its path-connected SO(3) subgroup of rotations.
We quote an important but difficult to prove expression which says that the familiar rule ea eb = ea+b does not
hold for matrices unless they commute! This is the so-called Baker-Campbell-Hausdorff (BCH) formula:
1 1
eA eB = eC C = A + B + [A, B] + A, [A, B] + [A, B], B + . . . (3.14)
2 12
Example 3.6. To find the matrix realisation of the generators of SO(3), which live in a three-
parameter algebra, consider counterclockwise rotations by a small angle θ around an axis whose di-
rection is specified by the vector n̂. An active transformation rotates a vector x by adding a small
vector that is perpendicular to both the axis and to x, with only vectors along the axis unchanged. By
geometry, we find that, to first-order, the transformed vector is x′ = x + θn̂ × x. Expanding gives:
where θ = θn̂. The matrix is an element of the so(3) algebra. How does this compare to the operator
algebra as laid out in eq. (3.5)? By identifying α = θz , etc., we can write the first order in the
expansion of the general rotation operator as:
0 −θz
θy ∂x
x y z θz 0 −θx ∂y
−θy θx 0 ∂z
The matrix is indeed the so(3)-algebra matrix. A rotation by a finite angle θ around axis n̂ can be
k
written as: R(θ) = eθ Mk , with generators:
0 0 0 0 0 1 0 −1 0
Mx = 0 0 −1 , My = 0 0 0 , Mz = 1 0 0
0 1 0 −1 0 0 0 0 0
The operator and matrix algebras have the same commutator structure, [Mi , Mj ] = ǫij k Mk , estab-
lishing their isomorphism.
Often, SO(3) generators are written as Jij = ǫijk M k , which is arguably more natural. Since
(Mi )jk = −ǫijk , the matrix elements are: (Jij )lm = −ǫijk ǫklm = −(δi l δj m − δi m δj l ). The labels
(ij), i < j for J refer to the plane of rotation. To obtain their commutators, compute (EXERCISE):
i
Jmn , Jpq j = (Jmn )i k (Jpq )k j − (Jpq )i l (Jmn )l j and rearrange the eight resulting terms, yielding† :
This result is important because it applies to rotations in dimensions N > 3, for which a plane of
rotation does not uniquely define an axis, as it does for N = 3. But a rotation in a 2-dim plane in
N -dim space is always about a well-defined point where all axes perpendicular to the plane meet.
Two other important and often useful results: scalar operators, ie., those that are invariant under 3-dim
rotations, must commute with the SO(3) generators (eg., the Hamiltonian for a spherically-symmetric
potential). As for a vector operator V, ie., one that transforms as a vector under rotations, it is shown
in Appendix H that it satisfies [Mi , Vj ] = ǫijk V k , or [Li , Vj ] = i ǫijk V k .
Example 3.7. The 6-dimensional so(4) Lie algebra is the set of all antisymmetric 4 × 4 real matrices,
which can be parametrised in the following way:
0 −a3 a2 −b1
a3 0 −a1 −b2
so(4) = ai Mi + bi Ni =
−a2 a1 0 −b3
b1 b2 b3 0
It is now appropriate to use the 4(4 − 1)/2 = 6 Jij generators, introduced in example 3.6, that
generate rotations in the (ij)-plane. With eq. (3.15) it is easy to compute the nine non-trivial so(4)
commutators, by taking Ji4 = Ni and Jij = ǫijk M k (1 ≤ i, j < k ≤ 3). Alternatively, we could use
the isomorphism with differential operators. With R4 coordinates x, y, z, u, there are six of these:
M1 = z ∂y − y ∂z , M2 = x ∂z − z ∂x , M3 = y ∂x − x ∂y
N1 = x ∂u − u ∂x , N2 = y ∂u − u ∂y , N3 = z ∂u − u ∂z
Whether with eq. (3.15) or the operator realisation, we obtain:
[Mi , Mj ] = ǫijk Mk , [Mi , Nj ] = ǫijk Nk , [Ni , Nj ] = ǫijk Mk (3.16)
1 1
The generators can be decoupled by transforming to the basis: Yi = (Mi + Ni ), Zi = (Mi − Ni ),
2 2
from which we immediately obtain the decoupled relations:
[Yi , Yj ] = ǫijk Yk , [Yi , Zj ] = 0, [Zi , Zj ] = ǫijk Zk (3.17)
By inspection, the Yi and Zi are generators of two separate su(2) (or so(3)) algebras, and so(4) =
su(2) ⊕ su(2). In terms of dimensions, 6 = 3 ⊕ 3. At group level, we say that SO(4) is locally
iso-
morphic to the direct product SU (2)×SU (2); it is globally isomorphic to SU (2)×SU (2) /Z2 since
a pair consisting of an element of SU (2) and its negative corresponds to the same SO(4) rotation;
SO(4) is globally isomorphic to SO(3) × SO(3).
i i
Then so(4) = ai Yi + bi Zi , and since [Yi , Zj ] = 0, an element of SO(4) takes the form: ea Yi eb Zi .
Example 3.8. The so(3, 1) algebra of the group SO(3, 1) derived from the metric-preserving con-
straint is:
0 ζx ζy ζz
ζx 0 −θz θy
so(3, 1) =
ζy θz 0
−θx = θ µ Mµ + ζ ν Kν (3.18)
ζz −θy θx 0
where the infinitesimal generators can be read off:
0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 1 0 0 −1 0
Mx = My = Mz =
0 0 0 −1 0 0 0 0 0 1 0 0
0 0 1 0 0 −1 0 0 0 0 0 0
(3.19)
0 1 0 0 0 0 1 0 0 0 0 1
1 0 0 0 0 0 0 0 0 0 0 0
Kx = Ky = Kz =
0 0 0 0 1 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 1 0 0 0
66
Lecture Notes on Mathematical Methods 2022
One shows (EXERCISE) that the commutators of the infinitesimal generators are:
Although the number of generators is identical to so(4), there is an important difference between these
relations and the ones derived in example 3.7: the minus sign in the relation for the K, which can also
be obtained by letting N → i K. Then the complex basis in which the commutators decouple is:
L± ± ± k ± ± ∓
i = (Mi ± iKi )/2, yielding (EXERCISE) : [Li , Lj ] = ǫij Lk and [Li , Lj ] = 0.
As in example 3.6, by defining Jij = ǫij k Mk and J0i = Ki , 1 ≤ i ≤ 3, one rewrites the commutator
relations (3.20) as a relation valid for any so(p, q) algebra in N = p + q dimensions.
Jµν , Jαβ = ηµα Jνβ + ηνβ Jµα − ηµβ Jνα − ηνα Jµβ 0 ≤ (µ, ν) ≤ N − 1 (3.21)
where Jνµ = −Jµν (subscripts label generators J, not their components!), and ηµν is the Cartesian
Minkowski metric: diag (∓1, ±1, . . . , ±1), depending on the metric sign convention.
One very important realisation of this algebra interprets θi as the three angles rotating around Cartesian
axis 1 ≤ i ≤ 3, and ζi = β̂i tanh−1 β the rapidity parameters for pure Lorentz boosts along the x,
y and z axes, written in terms of the relative velocity β betweeen two inertial frames. Then so(3, 1)
is called the Lorentz algebra for Minkowski spacetime. The relation (3.20) can also be derived
(EXERCISE) in the differential-operator realisation: Jµν = xν ∂µ − xµ ∂ν .
Symmetries under rotations and Lorentz transformations, as well as under translations, are prime example of
global symmettries, in the sense that the transformations have the same form at all points. Local symmetries
involve transformations that can vary arbitrarily from one point to another, so are said to be point-dependent.
3.3.5 Hard-nosed questions about the exponential map — the fine print
Three theorems by Lie, which we have implicitly used, show that for any Lie group an algebra can be found,
characterised by the structure constants. At best only the path-connected part of a group can be recovered from its
algebra. We have relied on the exponential map to do this, but it is not always possible,
at least with just one map.
Here is a counter-example (provided by Cartan). Take: Z = x2x+x 1 x2 −x3
3 −x1
∈ sl(2, R), whose trace vanishes.
Exponentiating gives the basis-independent result (EXERCISE):
sinh r
I2 cosh r + Z r2 > 0
X 1 r
eZ = Zn = I2 + Z r2 = 0
n!
n
I2 cos r + Z sin r
r2 < 0
r
where r 2 = x21 + x22 − x23 = −det Z, which makes the results basis-independent. The structure is reminiscent
of the light-cone structure obtained by endowing the parameter space R3 with an indefinite metric invariant under
SO(2, 1). Inside the light-cone, for any value of x3 , the values of the other two parameters are confined inside
a circle of radius smaller than x3 . The corresponding generators map to compact group elements. Outside the
light-cone, however, r can grow without restriction and maps to non-compact elements of SL(2, R).
So far, so good. But a glance at the above expressions shows that Tr Z
e ≥ −2 always. Yet SL(2, R) has a
large subset of elements with trace smaller than −2: matrices of the type −λ0 −1/λ0 (λ > 1), for instance. These
cannot be reached with the above exponential map.
Cartan argued that all the group elements could nevertheless be reached by writing:
x1 x2 0 −x3
Z = Za + Zb = x2 −x1
+
x3 0
67
Lecture Notes on Mathematical Methods 2022
and taking the product of the exponentials of Za and Zb , which is not eZ since [Za , Zb ] 6= 0. Then (EXERCISE):
Za Zb z+y x cos x3 − sin x3
e e = x z−y sin x3 cos x3
where z ≡ cosh r ′ ≥ 1, x ≡ xr2′ sinh r ′ , and y ≡ xr1′ sinh r ′ , with r ′2 = x21 + x22 . Each matrix is unimodular, and
the trace of the product is now 2z cos x3 = 2 cosh r ′ cos x3 , which is unrestricted.
In example 3.3 we noted that we needed more tools to tell us what the manifold of SL(2, R) was. Now we
know! The parameters of the non-compact matrix satisfy z 2 − (x2 + y 2 ) = 1 which is the positive-z hyperboloid.
Topologically, it is equivalent to R2 . The parameter values −π ≤ x3 ≤ π map the Zb subalgebra to SO(2) ⊂
SL(2, R), whose manifold is S 1 . We conclude that SL(2, R) is non-compact, and that its manifold is R2 × S 1 .
Every point is path-connected to the origin (x1 = x2 = 0) of R2 and x3 = 0 on S1 , so SL(2, R) is path-connected.
3.4.3 The regular (adjoint) representation and the classification of Lie algebras
We have already noted how eq. (3.12) for the structure constants could be written as the commutator of matrices
which we now recognise as providing a new representation of the algebra:
Definition 3.9. The regular (adjoint) representation of a Lie algebra associates with each element
j
Z of the algebra a matrix RZ (or adZ ) such that RZ (Xi ) = [Z, Xi ] = Xj RZ i , where the Xi are
j
the basis generators of the algebra. Some authors use the definition [Z, Xi ] = RZ i Xj .
Clearly, the regular representation of a basis generator is just the structure constants: [Xi , Xj ] =
k
RXi j Xk = Cij k Xk . Its dimension is that of the algebra, the number of generators (or parameters).
68
Lecture Notes on Mathematical Methods 2022
We confirm that R is a representation (EXERCISE, with the Jacobi identity): [RXi , RXj ]Xk = R[Xi , Xj ] Xk .
Example 3.9. Take the defining, two-dimensional representation of the essentially real version of the
su(2) algebra with basis set Si = σi /2, where σi are the three Hermitian Pauli matrices, and whose
commutators are: [Si , Sj ] = i ǫij k Sk . Then (adSi )k j = i ǫij k , and we have† ;
0 0 0 0 0 i 0 −i 0
adS1 = 0 0 −i adS2 = 0 0 0 adS3 = i 0 0
0 i 0 −i 0 0 0 0 0
Definition 3.10. A subalgebra of an algebra g is just a subspace that closes under commutation. A
subalgebra gsub is invariant if [gsub , g] ⊆ gsub , ie. if, ∀ X ∈ gsub and ∀ Y ∈ g, [X, Y ] ∈ gsub . An
invariant subalgebra is sometimes called an ideal, but we shall not be using this term.
The centre z of an algebra is the largest subalgebra that commutes with all elements of the algebra.
The centre of a commutative (Abelian) algebra is itself. z is always an Abelian invariant subalgebra.
Like the structure constants, the regular representation summarises the structure of the Lie algebra. This
algebra is a vector space spanned by a basis of generators. But we can decide to transform to another basis via a
similarity transformation. The question is: can we transform the regular representation to a basis where it takes a
form that might help classify the algebra?
Definition 3.11. If a sequence of transformations exists that puts the regular representation of a non-
Abelian Lie algebra into block-diagonal form, with the blocks irreducible non-zero subrepresentations,
the representation is said to be fully reducible. In this case, the regular representation can be written
as a direct sum of irreducible representations. Of course, these irreducible representations cannot all
be one-dimensional. In this basis, the block submatrices commute with one another.
Definition 3.12. If an algebra has no non-trivial invariant subalgebra, its regular representation is
irreducible (it leaves no proper subspace of its carrier space invariant), and the algebra is called simple.
Definition 3.13. A Lie algebra that contains no Abelian, invariant subalgebra is said to be semisimple,
ie. it has zero centre (no non-zero element commutes with all other elements). A semisimple algebra is
either simple or the sum of simple algebras (that may occur more than once in the sum). A semisimple
algebra always has at least two complementary invariant subalgebras, and there is a basis in which all
the generators of one commute with all the generators of the other(s), but not amongst themselves.
From these two definitions it follows that all simple algebras are semisimple since they are already in (single)
block form. Non-simple semisimple algebras must contain a proper, non-Abelian, invariant subalgebra.
Abelian Lie algebras (eg. u(1), so(2)) are not semisimple, and therefore not simple. Apart from so(4) (see
example below), the non-Abelian so(n) algebras are all simple, and so are the su(n) and sl(n, R) algebras.
†
The cummutation relations for the adjoint representation are: adSi , adSj = i ǫij k adSk < ahref = ”p154022a ss1.pdf ” >. With
k
our convention, adXi j = Cij k , for the adjoint representation, the structure constants for adjoint and defining represeentations are
k
always identical. With the other convention, adXi j = Cij k , they would differ by a minus sign.
69
Lecture Notes on Mathematical Methods 2022
Example 3.10. From eq. (3.16), no basis generator of so(4) commutes with all others: the algebra has
no non-zero centre! It is therefore† semisimple. Its structure constants determine the 6-dim regular
representation of a generic element of so(4) in block-diagonal form:
0 −a3 a2 0 0 0
a3 0 −a1 0 0 0
−a2 a1 0 0 0 0
R =
0 0 0 0 −b3 b2
0 0 0 b3 0 −b1
0 0 0 −b2 b1 0
The blocks cannot be further reduced, so(3) being simple; so(4) is semisimple, but not simple.
(Y, Z) = Tr Y Z
Definition 3.14. The Cartan-Killing form (CK-form) is a symmetric, bilinear form whose compo-
nents are the inner product of all pairs of elements of a Lie algebra in their adjoint representation:
(Y, Z) := Tr RY RZ = (RY )kl (RZ )lk (3.22)
The CK-form for basis generators Xi is easily calculated: (Xi , Xj ) = Cil k Cjk l . If the algebra has n
parameters, the CK-form has n(n + 1)/2 components.
An important property of the CK-form is its invariance under the action of any element g in the Lie group associated
with a Lie algebra. Let X and Y be elements of a Lie Algebra. Then:
g X g−1 , g Y g−1 = Tr Rg RX Rg−1 Rg RY Rg−1 = Tr (RX RY ) = (X, Y )
where we have used the property Tr A B = Tr B A. Linearising after writing: g = eǫZ , we obtain (EXERCISE):
[Z, X], Y + X, [Z, Y ] = 0 (3.23)
Definition 3.15. A CK-form is degenerate (or singular) if there exists at least one element Z in the
algebra for which (Z, Y ) = 0 ∀ Y ∈ g, ie., if the matrix (Xi , Xj ) has a row and column entirely pop-
ulated with zeros, which forces its determinant to vanish. Otherwise, the CK-form is non-degenerate.
Equivalently, a CK-form is non-degenerate if there exists a basis in which it is diagonal with all entries
non-zero. Then we say that it induces a Cartan metric g on a Lie algebra, with components gµν =
(Xµ , Xν ), where {Xµ } is that basis. If the algebra is compact, we can transform to an orthonormal
Cartan metric g = kIn ; if the algebra is non-compact, we can transform to an indefinite metric kIqp ,
with p + q = n, the dimension of the algebra. In these two cases, it is habitual to call In and Iqp
themselves the metric, which is then manifestly orthonormal.
Like all metrics, an orthonormal Cartan metric can raise and lower indices. In particular, introduce
fµνλ := Cµν ρ gρλ . Inserting gρλ = (Xρ , Xλ ), one can show (EXERCISE) with eq. (3.23) that fµνλ is antisymmet-
ric.
Now, if an algebra has a non-zero centre z (ie. an Abelian invariant subalgebra that commutes with all the
elements of the algebra), its CK-form is degenerate because the adjoint representation of any element of z vanishes
trivially. Cartan’s criterion asserts that the converse is also true, which leads to a useful alternate definition:
†
For a given i, j Yi and Zj in the decoupled basis of eq. (3.17) form an Abelian subalgebra, but it is not invariant.
70
Lecture Notes on Mathematical Methods 2022
Definition 3.16. A Lie algebra is semisimple if, and only if, its CK-form is non-degenerate,
Example 3.11. xi ∂j is a basis of the operator realisation of gl(3, R). Then xi ∂i commutes with every
other element of the algebra, and gl(3, R) has a non-zero centre. Therefore, it is not semisimple.
Example 3.12. In example 3.9, we have already obtained the adjoint representation for the generators
of su(2) — and the one for so(3) because the structure constants for the two algebras are now identical.
With S in the adjoint representation, eq. (3.22) then gives:
The CK-form is then 2I. This confirms that the CK-form for su(2) induces an invertible definite
(Euclidean) orthonormal metric, g = I. Therefore, the group is compact as well as semisimple, and
we can write the structure constants as the skew-symmetric fijk = iǫijk .
The Cartan metric as defined above in terms of the regular representation for each generator can be tedious to
calculate when the matrices are huge. So long as it is non-degenerate, however, we can extract useful information
about it with much less work by instead calculating (R, R), with R = aµ Xµ , the aµ being the parameters:
and (R, R) contains information about the Cartan metric gµν —more specifically, whether the algebra is compact.
Example 3.13. Go back to the defining representation used for Z ∈ sl(2, R) in section 3.3.5:
x1 x2 + x3 1 0 0 1 0 1
Z = = x1 + x2 + x3
x2 − x3 −x1 0 −1 1 0 −1 0
The corresponding independent non-zero structure constants are: C123 = 2, C312 = −2, and C231 =
−2. From these we build the regular-representation matrix:
0 2x3 −2x2
RZ = −2x3 0 2x1
−2x2 2x1 0
Now, we only need to calculate the diagonal elements of R2 and sum them to get: (R, R) = 8(x21 +
x22 − x23 ). We deduce that the algebra is non-compact. That X1 and X2 are non-compact, while X3 is
compact, was determined earlier in section 3.3.5.
Interestingly enough, using the defining representation directly, we would find (EXERCISE) 2(x21 +
x22 − x23 ). This is because for semisimple algebras the defining and regular representations are both
faithful, and thus contain the same information, opening up the possibility of calculating aµ aµ in eq.
(3.24) directly from the defining representation instead of the more unwieldy regular representation.
for some Y ∈ g, which makes Y an eigengenerator of Hi . In fact, we would like to know the maximal subset of
elements of an algebra that commute between themselves, thus forming an Abelian (non-invariant!) subalgebra.
71
Lecture Notes on Mathematical Methods 2022
Definition 3.17. A maximal Abelian subalgebra of a semisimple Lie algebra is called a Cartan sub-
algebra h. Its dimension r < n defines the rank of the algebra. It is unique up to isomorphism. The
elements of a Cartan subalgebra are called its Cartan generators. Being Abelian, its irreducible rep-
resentations are one-dimensional, and there exists a basis in which all Cartan generators are diagonal.
Example 3.14. An ordered basis of the complex extension of su(2) (Example 3.9) in its defining
representation is {S− , S0 , S+ }, where S± = √12 (S1 ± iS2 ) and S0 = S3 , with [Si , Sj ] = iǫij k Sk ,
or: [S0 , S± ] = ±S± , and [S+ , S− ] = S0 . Then the adjoint representation for S0 and S± is:
−1 0 0 0 0 0 0 1 0
adS0 = 0 0 0 adS+ = 1 0 0 adS− = 0 0 −1
0 0 1 0 −1 0 0 0 0
Because adS0 is diagonal, S0 is a Cartan generator; comparing with eq. (3.25), adS0 has a complete set
{S− , S0 , S+ } of eigengenerators for the corresponding eigenvalues {−1, 0, 1}, which form a basis
of the algebra. But neither S+ nor S− is diagonalisable and they are not Cartan generators, Thus, the
algebra contains only one Cartan generator and is of rank 1.
Another important thing we learn from this is that the structure constants in the complex extension of
an algebra can be quite different from those of the algebra itself, even in its essentially real version.
Indeed, the adjoint representation of S3 found in example 3.9 is not diagonal, and has only zeros on
its diagonal, in contrast with with adS0 , although adS3 does diagonalise to adS0 . Of course, this does
not affect the CK-form which, being a trace, is basis-independent.
It can be shown that the rank of a su(n) algebra is n − 1; also, so(2n) and so(2n + 1) have rank n.
Definition 3.18. Let |µi be an eigenvector common to all Cartan basis generators Hi , living in the
carrier space of some representation D of the generators. Then Hi |µiD = µi |µiD . The set {µi }
corresponding to each eigenvector can themselves be viewed as the components of a r-dimensional
vector called a weight µ of the representation. The number of these weights is the dimension of D.
To find the n weights (often called a multiplet) of a representation D with matrices of rank n, simply identify
a set of r Cartan generators Hi in D , and diagonalise them if they are not in diagonal form. The ith (1 ≤ i ≤ r)
component of the j th weight (1 ≤ j ≤ n) is the (jj)th entry of the n × n matrix representing Hi . These weights
correspond to a point on a r-dimensional weight diagram, or lattice.
Definition 3.19. In a semisimple algebbra there exists a basis in which the non-Cartan generators Eα
of a semisimple algebra satisfy: [Hi , Eα ] = αi Eα , 1 ≤ i ≤ r The Eα ∈ g are then eigengenerators
(often confusingly called root vectors by mathematicians) of the element Hi of the Cartan subalgebra.
Then the set {αi } of eigenvalues can be viewed as the components of a r-dimensional vector called
the root α. We can also write [H, Eα ] = αEα . In any representation (defining, adjoint), this basis is
called the Cartan-Weyl basis.
Do keep in mind the crucial distinction between the eigengenerators, whose associated eigenvalues are the root
components, and the eigenvectors that live in the carrier space, whose eigenvalues are the components of the
weights. Also, the roots do not depend on the representation D, whereas the weights do. Indeed, one often speaks
of the weights of D as being the representation itself.
We can write an algebra g as the sum of its Cartan subalgebra, with roots zero, and the non-Cartan generators
with non-zero roots. The set of all non-zero roots define the root system of the algebra in a r-dim space.
72
Lecture Notes on Mathematical Methods 2022
As we are soon to discover, all the information about a semisimple algebra is encoded in its root system. A
Euclidean metric is induced on their space by the metric of the Cartan subalgebra, so that we can represent it as
having r Cartesian axes, each associated with a Cartan generator Hi . The root vectors can then be represented in
a root diagram. The ith component of each root is the projection of the root along the Hi axis. Being of smaller
dimension, this root space is almost always much easier to work with than the algebra itself.
If α + β is not a root, [Eα , Eβ ] = 0. Otherwise, [Eα , Eβ ] is an eigengenerator with root α + β, so we can write:
Using the definition, [Hi , Eα ] = αi Eα , one should now be able to see that in the Cartan-Weyl basis, the
adjoint representation of Hi is a diagonal matrix, with as entries r zeros and the αi component of the n−r roots.
Also, all diagonal entries of the adjoint representation of any other generator Eα must be zero. From this, the
following statements about the CK-form of a semisimple algebra can be derived:
Also, according to a theorem from linear algebra about nondegenerate symmetric bilinear forms, there exists a
basis of h in which hij = (Hi , Hj ) = kD δij , where hij are the metric components of the Cartan subalgebra.
These results, although derived in the adjoint representation, apply to the defining representation as well.
To go further, work with Hermitian Cartan generators: Hi† = Hi of the essentially real algebra. Then, if
† † †
[Hi , Eα ] = αi Eα , we immediately find that [Hi , Eα ] = −αi Eα , so that Eα = E−α . Thus, non-Cartan genera-
tors and non-zero roots always come in pairs, {Eα , E−α }. In fact, −α is the only possible multiple of α which
is a root; it always exists, otherwise (Eα , Z) = 0 ∀ Z ∈ g, and the CK-form would be degenerate. Now we
know how to compute the non-Cartan generators in the Cartan-Weyl basis from the pairs Xk and Xl of non-Cartan
generators of the algebra: E±α = A(Xk ± iXl ), with A a normalisation contant.
When β = −α, eq. (3.26) maps [Eα , Eβ ] to a generator with zero root, ie. one that lives in the Cartan
subalgebra. Therefore, [Eα , E−α ] = λi Hi for 1 ≤ i ≤ r. Taking the inner product with Hj , one quickly shows
(EXERCISE), using the cyclicity of the trace, that λi = αi Eα , E−α , so that:
[Eα , E−α ] = Eα , E−α hij αj Hi = Eα , E−α k αi Hi
where we have noted that hij = k δij for a semisimple algebra. Now [Hi, Eα ] = αi Eα determines Eα only up to
a normalisation constant, which can be chosen so as to make Eα , E−α cancel k, leaving the more simple:
[Eα , E−α ] = αi Hi := α · H summation implied (3.27)
Now is a good time to discover what those non-Cartan generators do for a living. We have:
Hi E±α |µi = [Hi , E±α ]|µi + E±α Hi |µi = (µi ± αi ) E±α |µi (3.28)
We see that E±α |µi is an eigengenerator of Hi with eigenvalue µi ± αi . The E±α act as raising/lowering
operators on the carrier space of the Cartan generators, changing weights µ by ±α. Often, the quickest way
to obtain the roots is to work out all the possible differences between neighbouring weights of a low-dimension
representation.
73
Lecture Notes on Mathematical Methods 2022
When the representation is irreducible, C has to be a multiple of the identity by Schur’s lemma. All elements of
an invariant subspace of the carrier space of the representation will be eigenvectors of C with the same eigenvalue.
When the algebra is semisimple, work by C. Chevalley amd Harish-Chandra (1951) guarantees the existence of a
set of Casimir operators as polynomials in the generators, whose eigenvalues may be used to label the irreducible
representations of the algebra. More precisely, each invariant subspace of the carrier space has a set of basis
vectors, each labelled by an eigenvalue of each Casimir operator. The number of Casimir operators is the rank of
the algebra.
In other words, if f (x) is in an invariant subspace of the carrier space of the algebra, then for each Casimir
operator Ci , Ci f (x) = g(x) is also in that same invariant subspace.
Because a metric can always be defined for a semisimple algebra, I claim that C2 := g µν Xµ Xν is a Casimir
operator, called the quadratic Casimir invariant, and where the Xµ are basis generators of the algebra. Indeed:
µν
g Xµ Xν , Xρ = gµν Xµ [Xν , Xρ ] + [Xµ , Xρ ] Xν = gµν Cµρλ (Xν Xλ + Xλ Xν )
= gµν gαλ fµρα (Xν Xλ + Xλ Xν )
= 0
since g µν gαλ fµρα is antisymmetric in, and the term in round brackets is symmetric in, ν and λ.
Example 3.15. From example 3.12, the metric for so(3) is, up to a constant, gµν = δµν . Then:
where J is the angular momentum operator of quantum mechanics. Since so(3) is of rank 1, this is the
only Casimir invariant. Then the eigenvalues of J 2 each label an irreducible representation of so(3).
Note that because of its construction, C2 is not in the algebra. In a Cartan-Weyl basis, it takes the form:
X X
C2 = gij Hi Hj + (E−α Eα + Eα E−α ) = gij Hi Hj + 2 E±α E∓α ∓ α · H (3.29)
+roots +roots
Then J+ raises, and J− lowers, the weights m by 1, but they cannot transform | mi to an eigenvector of J 2
with a different eigenvalue λ. All the weights in a given invariant subspace are eigenvectors of J 2 with the same λ.
Next, we write relation (3.29) between the C2 Casimir operator and the generators for so(3):
Since an irreducible representation must be finite-dimensional, we expect that for a given λ there exists a
highest weight, mmax ≡ j, and also a lowest weight, mmin ≡ j ′ . Then J+ | ji = 0 and J− | j ′ i = 0. There comes:
J 2 | ji = j 2 | ji + j | ji = j(j + 1) | ji = λ | ji
J 2 | j ′ i = (j ′ )2 | j ′ i − j ′ | j ′ i = j ′ (j ′ − 1) | j ′ i = λ | j ′ i
Comparing yields λ = j(j + 1) = j ′ (j ′ − 1), and thus j ′ = −j. It follows that the weights m go from −j to j
in N integer steps, ie, j = −j + N , so j = N/2.
We conclude that the eigenvalues of the Casimir operator J 2 are j(j + 1), where j is a positive integer or
a half-integer, and that for a given value of j, the weights m can take 2j + 1 values, from −j to j. Therefore,
odd-dimensional irreducible representations correspond to integer j and even-dimensional ones to half-integer j.
With the help of eq. (3.30), we can now exhibit the full action of J− on a weight | jmi of J 2 and J0 . Let
J− | jmi = c− | j, m − 1i. Then, if | jmi is normalised:
j j δm′ ,m±1 p
Dm ′ m (J0 ) = m δm′ m Dm ′ m (J± ) = √ (j ∓ m)(j ± m + 1) |m| ≤ j (3.32)
2
This form for the coefficients is often quoted, but the equivalent form in eq. (3.31) is often
√ easier to use since only
√
the second factor in the root changes. The representation matrices for Jx = (J+ +J− )/ 2, Jy = (J+ −J− )/(i 2)
and Jz = J0 are easily recovered if needed. Keeping in mind that the rows and columns are labelled by the values
of m from −j to j, we have for the defining representation of so(3), labelled by j = 1:
0 1 0 −1 0 0 0 0 0
D 1 (J+ ) = 0 0 1 D 1 (J0 ) = 0 0 0 D 1 (J− ) = 1 0 0
0 0 0 0 0 1 0 1 0
Any other irreducible representation for integer values of j can be calculated in the same way with eq. (3.32).
Another approach relies on the actual form of the generators. In the defining, irreducible 3-dim representation
of the Cartesian basis, the three generators, which we choose to be Hermitian, are:
0 0 0 0 0 i 0 −i 0
J1 = 0 0 −i J2 = 0 0 0 J3 = i 0 0 (3.33)
0 i 0 −i 0 0 0 0 0
75
Lecture Notes on Mathematical Methods 2022
Diagonalise, say, J3 with the transformation Ji −→ A−1 Ji A, where A is a unitary matrix so as to preserve
Hermiticity, and construct the non-Cartan generators in the Cartan-Weyl basis as before:
0 −i 0 −1 0 0 0 0 0 −i 0 i
1
J+ = 0 0 i J0 = 0 0 0 J− = i 0 0 A = √ 1 √0 1
0 0 0 0 0 1 0 −i 0 2 0 2 0
Although these generators are different from the D 1 matrices obtained from eq. (3.32), they are perfectly acceptable
as an irreduciblerepresentation
since they satisfy
both [J0 , J± ] = ±J± and [J+ , J− ] = J0 . Indeed, any pair of
0a0 0 0 0
the form J+ = 0 0 b and J− = a∗ 0∗ 0 , wih (a, b) ∈ C, satisfies these commutation relations! All these
00 0 0 b 0
irreducible representations in the Cartan-Weyl basis are equivalent.
by j), is the name given to the irreducible representations of SU (2), and the matrix elements are called Wigner
functions. They can be rather complicated, except when n̂ = ẑ and sz = s0 is diagonal, in which case (Dθj )m m =
′
′
eimθ δm m (|m| ≤ j). They are tabulated in many places for small j and are easily calculated by computer.
76
Lecture Notes on Mathematical Methods 2022
3.6.5 su(2) substructure of a semisimple algebra and constraints on its root system
Because they live in a r-dim space, only r of the n − r roots of a semisimple algebra can be linearly independent.
Definition 3.21. A positive root is one whose first non-zero component is positive; otherwise, it is
negative. The r positive roots which cannot be obtained from a linear combination of other positive
roots are called simple, fundamental, or independent. The other positive roots can be obtained as
linear combinations of the simple roots, with positive coefficients.
√
Each pair e±α := 2E±α /|α| of normalised non-Cartan generators of a semisimple algebra, together with
the combination: hα = 2α · H/|α|2 , forms a su(2) subalgebra. There is a su(2) subalgebra for each pair
of non-zero roots (Chevalley 1955). {hα , e±α } is called the Chevalley basis of the su(2) subalgebra . Indeed,
[eα , e−α ] = hα , but also:
√ √
2 2α 2 2 |α|2
hα, , e±α = · [H, E±α ] = ± E±α = ± 2 e±α
|α|3 |α|3
With hα = 2s0 and e±α = s± , we recover the su(2) structure constants in the Cartan-Weyl basis. Thus, a
semisimple algebra of dimension n and rank r contains (n − r)/2 generally non-distinct su(2) subalgebras, each
associated with a different root and having as Cartan generator a different element of the Cartan subalgebra, plus
two non-Cartan generators corresponding to that root.
Roots are tightly constrained by the su(2) substructure described above. Consider some other root β. Then:
h i √ √
2 2α 2 2α· β α· β
hα , e±β = 2
· H, E±β = ± 2
E±β = ± 2m e±β m :=
|α| |β| |α| |β| |α|2
Since hα /2 is a Cartan generator of su(2), we may have found another su(2) subalgebra if we can make sense of
m. Now let β + kα (k ∈ Z) be a non-zero root. Then, in the same way as above:
hh i α · (β + kα)
α
, eβ+kα = eβ+kα = (m + k) eβ+kα
2 |α|2
Then let p and q be two non-negative integers, with p the largest number for which β + pα is still a root, and q
be the largest number for which β − qα is still a root. So we have a string, or chain eβ−qα , . . . , eβ , . . . , eβ+pα
of su(2) generators that act in the root space, raising or lowering in integer steps. All elements in the set {β +
kα; k = −q, . . . m, . . . p} are roots. We can associate each of these generators with one in a (2j +1)-dim su(2)
representation labelled by j. It takes p unit steps to go from m to the highest root, and q steps to go to the lowest
root in the chain, so that −j + q = m and j − p = m, leading to q − p = 2m, and p + q = 2j. As expected for a
su(2) algebra, m (and j) is an integer or half-integer. We arrive at the master formula† :
α·β
−p≤ 2 = − (p − q) < q (3.34)
|α|2
If we had started instead with eα and added/subtracted integer multiples of β to α, we would have found that
2β · α/|β|2 = − (p′ − q ′ ). Multiplying the two master formulae yields the important expression:
(α · β)2 1
2 2
= cos2 θαβ = (p − q)(p′ − q ′ ) ≤ 1 (3.35)
|α| |β| 4
p
The relative length of the roots is seen to be constrained to |α|/|β| = (p′ − q ′ )/(p − q). Also, if α and β are
simple roots, ±(α − β) cannot be a root; otherwise, one of the two must be positive, and a simple root could be
constructed out of two different positive roots: eg., β = (β − α) + α. Thus, β − kα is not a root for any k 6= 0,
†
A derivation that does not rely on the su(2) substructure can be found in Appendix I; but it involves rather heavier calculations.
77
Lecture Notes on Mathematical Methods 2022
including k = q, and q = 0 for simple roots. Therefore, from the master formula (3.34), the angle between two
simple roots satisfies cos θαβ ≤ 0, so that π/2 ≤ θαβ ≤ π.
Since (p − q)(p′ − q ′ ) must be an integer, There are only five possible values allowed for cos2 θαβ in eq.
(3.35), and this, for any two roots of any semisimple algebra: 0 ⇒ θαβ = ±90◦ ; 1/4 ⇒ θαβ = (60◦ , 120◦ );
1/2 ⇒ θαβ = (45◦ , 135◦ ); 3/4 ⇒ θαβ = (30◦ , 150◦ ); and 1 ⇒ θαβ = 180◦ (0◦ is forbidden because no two
roots can be a positive multiple of each other).
Thanks to all these constraints, a systematic and exhaustive procedure exists to construct the root space for
all four families of classical semisimple groups, and for the five so-called exceptional groups. The root diagrams
exhibit a high degree of symmetry. All positive roots can be generated by linear combinations of the simple roots.
So-called Weyl reflections about hyperplanes perpendicular to the simple roots through the origin generate the
rest.
With the subscript denoting the rank of the algebra, the four families of semisimple groups are:
• An−1 (n > 1), corresponding to SU (n), SL(n, R), SU (p, q), with p + q = n (not the p and q above!)
SU (2), SL(2, R), both A1 , and SO(3) (B1 ) , all have the same one-dim root space with the two roots ±1. Only
five two-dimensional root spaces (four classical and one exceptional) can satisfy all our constraints; but B2 and
C2 are rotated from each other by 45◦ , so are taken to be the same. And there are only four three-dimensional
root spaces. Beyond three dimensions, root spaces can no longer be represented on root diagrams. Instead, one
uses Dynkin diagrams, which are planar and represent only the simple roots and the angle between them. They are
equivalent to a root diagram.
Finally, a few words about weight diagrams. One of the Cartan generators, say H1 , will always be the Cartan
generator of a su(2) (and so(3) - see section 3.6.2) subalgebra. Then weight points are arranged on lines parallel
to the H1 axis, with each line corresponding to an irreducible representation (multiplet) of su(2) labelled with j,
an integer multiple of 1/2, and containing 2j + 1 weights. These weights can be generated by starting from the
highest weight of the representation, defined as the weight µ for which µ + α is not a weight when α is any
positive root. and applying the lowering non-Cartan generator of su(2) to the weights in each su(2) multiplet, ie.,
by repeated addition of the r-dim root, (−1, 0, . . . , 0), to that highest weight. This root, as well as (1, 0, . . . , 0)
(which moves up from the lowest to the highest weight), is always a root of the semisimple algebra. Needless to
say, as one moves parallel to the H1 axis, all other components in the weights remain the same. Subtracting a
simple root from the highest weight yields the highest weight of a neighbouring su(2) multiplet.
The number of weights for these different su(2) multiplets must add up to the dimension of the multiplet of
the semisimple algebra. The su(2) multiplets must fit snugly inside this multiplet. For instance, take the 10-
dim representation (decuplet) of su(3) of rank 2; thus the weights are 2-component vectors. The weights lie on
an inverted-triangle lattice with one horizontal su(2) quadruplet, triplet, doublet and singlet, in the direction of
decreasing H2 eigenvalues.
Definition 3.22. Let fj1 m1 and fj2 m2 be two basis functions in the carrier space of irreducible repre-
sentations Dgj1 and Dgj2 , respectively, of g ∈ SU (2) or SO(3), such that:
′ ′
Sg fj1 m1 = fj1 m′1 (Dgj1 )m1m , Sg fj2 m2 = fj2 m′2 (Dgj2 )m2m
1 2
78
Lecture Notes on Mathematical Methods 2022
In Dirac notation, the product of the basis functions would read: |j1 m1 , j2 m2 i = |j1 m1 i | j2 m2 i.
Such a product is needed when a system responds to transformations in more than one way, either because of
the coupling of two separate systems (eg. particles) or because two distinct dynamical variables of one system get
coupled. A common transformation on the whole system is to be written as a direct product of transformations on
each of its parts in its own subspace.
Linearise eq. (3.36) using the generic expansion D = I + ai Xi , where X stands for a generator of SU (2)
or SO(3) in that representation. We find that the generators of the composite representation are the sums of the
generators of the distinct terms in the tensor product, so that:
X(1⊗2) (fj1 m1 fj2 m2 ) = (X (1) fj1 m1 ) fj2 m2 + fj1 m1 (X (2) fj2 m2 ) (3.37)
that is: X(1⊗2) = X (1) ⊗ I + I ⊗ X (2) or, more sloppily, X = X(1) + X(2) . When the generators have diagonal
representations, as happens with J0 (SO(3)) or s0 (SU (2)), we find, eg.:
Definition 3.23. Let {Tjm } be a set of operators on the carrier space that transform as:
′
Rθ Tjm Rθ−1 = Tjm′ (Dθj )m m
(3.39)
Then we say that they are the components of a rank-j irreducible (or spherical) tensor.
If we linearise this equation, we obtain (EXERCISE) a more useful alternative definition of irreducible tensors in
terms of generators J (j) of an irreducible representations of the algebra, preferably in the Cartan-Weyl basis:
′
[J (j) , Tjm ] = Tjm′ (J (j) )m m (no summation on j) (3.40)
79
Lecture Notes on Mathematical Methods 2022
As a direct consequence of these commutation relations, the matrix element of Tjm , hj2 m2 |Tjm |j1 m1 i, vanishes
unless m2 = m1 + m and |j1 − j| ≤ j2 ≤ j1 + j. These are a version of the famous vector addition rules.
The importance of the Wigner-Eckart theorem resides in its separating symmetry-related (“geometrical”) as-
pects of matrix elements from other (“dynamical”) aspects stored in the possibly unknown reduced matrix element.
T ′ij = Ri k Rj l T kl (3.43)
The 6-dim symmetric part of T ij rotates into a symmetric object, and the 3-dim antisymmetric part into an anti-
symmetric one. Thus, we have easily found invariant subspaces. Moreover, the trace of T ij , T ii , is invariant under
rotations, forming a 1-dim invariant subspace that should be separated out from the symmetric part.
Note that the trace is obtained by contracting T ij with the metric of the carrier space, with components gij ,
which here is just the identity matrix invariant under rotations. Similarly, the antisymmetric part can be obtained
with the Levi-Civita symbol that is also invariant under rotation. Thus, we can write:
ij 1 ij ji
1 ijk lm 1 2 ij k 1 1
T = T + T + ǫ ǫklm T = T + T − g T k + (T ij − T ji ) + gij T kk (3.44)
ij ji
2 2 2 3 2 3
The numerical coefficient of the trace term has been chosen so as to make the symmetric term traceless.
But we can also think of eq. (3.43) as a 3 ⊗ 3 exterior direct product of a rotation with itself, so a 9 × 9 matrix,
with each row labelled by a pair {ij} and each column labelled by a pair {kl}, acting on a 9×1 matrix with entries
T kl labelled by the pairs {kl}. The direct-product matrix is a representation of SO(3). Indeed, under a rotation
80
Lecture Notes on Mathematical Methods 2022
R1 followed by R2 , T ij −→ (R2 R1 )i m (R2 R1 )j n T mn , where now the 9 × 9 matrix is formed from the matrix
product R2 R1 . Being reducible, the representation can be transformed via an angle-independent similarity matrix
to a block-diagonal matrix with a symmetric traceless 6 × 6 block (which acts only on the symmetric traceless part
of T) , an amtisymmetric 3 × 3 block acting only on the antisymmetric part of T , and a 1 acting only on the trace.
We obtain the following decomposition into irreducible representations: 9 = 5 ⊕ 3 ⊕ 1.
As expected, the total dimensions on the left and right match. The result is also consistent with what we would
find by decomnposing a j1 ⊗ j2 = 1 ⊗ 1 SO(3) product representation with the method of section 3.7.1 to obtain
a direct sum of three irreducible representations labelled by j = 2, j = 1, and j = 0.
Appendices
H Commutators of Angular Momentum with Vector Operators
Take a unit vector û and a vector operator V with components Vu with respect to û. In example 3.6 we found that
under a rotation R(θ) by a small angle θ about an axis n̂, û′ = û + θ n̂ × û. Then:
Vu′ = V · û′ = V · û + θ V · n̂ × û = Vu + θ V · n̂ × û
Also,
Vu′ = R(θ) Vu R† (θ) = e−i θn̂·L Vu ei θn̂·L ≈ (1 − i θn̂ · L) Vu (1 + i θn̂ · L) ≈ Va − i θ n̂ · L, Vu (H.1)
Consistency then demands that: n̂ · L, Vu = i V · n̂ × û = i ǫijk V k ni uj . With n̂ along the the x-axis and
û along the y-axis, there comes:
Li , Vj = i ǫijk V k (H.2)
81
Lecture Notes on Mathematical Methods 2022
The Hi being linearly independent, this can only be satisfied if: α Cβγ + β Cγα + γ Cαβ = α(Cβγ − Cαβ ) +
β(Cγα − Cαβ ) = 0, which yields additional symmetries on the structure constants of a semisimple algebra:
Eα , [Eβ+kα , E−α ] + Eβ+kα , [E−α , Eα ] + E−α , [Eα , Eβ+kα ] = 0
=⇒ [Eα , Eβ+(k−1)α ] Cβ+kα,−α − [Eβ+kα , α · H] + [E−α , Eβ+(k+1)α ] Cα,β+kα = 0
=⇒ Cα,β+(k−1)α Cβ+kα,−α + α · (β + kα) + C−α,β+(k+1)α Cα,β+kα = 0
Applying relations (I.1) and then (I.2) to the first and last term on the left yields the recursion relation:
2 2
Cα,β+(k−1)α = Cα,β+kα + α · (β + kα)
2
We already know that, by definition of p, Cα,β+pα = 0. Then, from our recursion relation, Cα,β+(p−1)α =
2 2 2 2 2
α · β + p|α| , Cα,β+(p−2)α = Cα,β+(p−1)α + α · β + (p − 1)|α| = 2α · β + (p − 2)|α| , etc. Generically:
2 p+k
Cα,β+(k−1)α = (p − k + 1) α · β + |α|2
2
The recursion stops when k = −q, ie. when C−α,β−qα = −Cα,−(β−qα) = −Cβ−(q+1)α,α = 0:
2 p−q
0 = Cα,β−(q+1)α = (p + q + 1) α · β + |α|2
2
or:
α·β
2 = − (p − q) (I.3)
|α|2
82
Lecture Notes on Mathematical Methods 2022
Also, solving a nth -order ODE requires specification of boundary conditions (B.C.) on the solution Thus, the
domain of L also depends on the BC imposed on the functions on which it is allowed to act. Then invertibility of
L is also very much dependent on those BC. We address each each of these questions in turn.
(1) One-Point (Initial) conditions, aka Initial-Value Problem (IVP): In the formal theory, one matrix in Afa +
Bfb = C, say B, is set to zero, so that only one point, the initial “time” a, is involved, and A is diagonal.
Therefore, f and its n-1 derivatives take known values (or can be set arbitrarily) at t = a. Then a theorem
shows that the solution to the one-dim IVP exists and is unique.
(2) Two-point boundary conditions, or Boundary-Value Problem (BVP): this time the n known or specified
values of f and its derivatives can be at both end-points a and b. This is a much more complicated situation,
with neither existence of a solution nor its uniqueness guaranteed.
In the most prevalent case, f (a) and f (b) are known (Dirichlet problem), or its first derivatives at a and b are
known (Neumann problem). Periodic B.C., where f (a) = f (b) and f˙ a = f˙ b , can also occur.
84
Lecture Notes on Mathematical Methods 2022
Notice that fh = i(a)f (a)/i(t) solves the homogeneous equation: dt f + β(t)f = 0, and that the inhomogeneous
term in eq. (4.4) satisfies homogeneous B.C., as expected. Thus, fh (a) = f (a). When f (a) = 0, f (t) = 0.
With f (a) specified, the solution is unique. Indeed, let g(t), with g(a) = f (a), also satisfy L[g] = F . Then
h = g − f solves the homogeneous ODE with homogeneous B.C. h(a) = 0, which forces h(t) = 0 for t > a.
where t0 is any point in the interval [a, b]. In the important case that β = α̇, eq. (4.6) leads to αW being constant.
Also, if the Wronskian of two homogeneous solutions vanishes anywhere, it vanishes everywhere, because the
exponential cannot vanish in a finite interval.
Given one solution, f1 , of eq. (4.5), an immediate useful application
of the Wronskian generates a second
linearly independent solution. Noticing that W (t)/f12 = dt f2 /f1 and integrating, we find with eq. (4.6) that:
Z t
W (t0 ) − R t′ (β/α)dt′′ ′
f2 (t) = f1 (t) e dt (4.7)
a f12 (t′ )
Discarding any term proportional to f1 leaves a solution that is linearly independent from f1 .
And now comes a surprising fact, courtesy also of the Wronskian: given two independent solutions of the
homogeneous equation, a solution of the inhomogeneous equation: L[f (t)] = F (t), can be generated which
satisfies homogeneous B.C. Appendix J presents a simplified version, leading to eq. (J.1), of this variation of
parameters method discovered by Euler and Lagrange. Shortly, however, we shall explore another method which
yields the same resultsR while providing much deeper insight.
Note also that if β(t)/α(t)dt exists within the interval of interest, it is always possible to eliminate the
first-order derivative term in any linear second-order ODE, with a redefinition of the form f (t) = g(t)eµ(t) (the
substitution f (t) = µ(t)g(t) also works), to arrive (EXERCISE) at the normal Sturm-Liouville form:
β Z t
1 1 β2 F (t) −µ(t) F (t) β ′
g̈(t) + γ − dt − g(t) = e = exp dt (4.8)
α(t) 2α 4 α α(t) α(t) 2α
85
Lecture Notes on Mathematical Methods 2022
as determined by the requirement that the transformed ODE have no first-order derivativeof g. In the frequent case
α = 1 and β and γ constants, this assumes the much simpler form: g̈(t) + γ − (β/2)2 g(t) = F (t)eβt/2 .
Let f1 (t) and f2 (t) be two independent solutions of L[f ] = 0. Then fh = c1 f1 + c2 f2 , with c1 and c2 deter-
mined from the B.C. on fh , is the general solution of the homogeneous equation (principle of linear superposition).
where θ(t − t′ ) is the step-function which vanishes for t < t′ and equals 1 when t > t′ , and:
f˙2 (a) f (a) − f2 (a) f˙(a) f˙1 (a) f (a) − f1 (a) f˙(a)
c1 = , c2 = −
W (a) W (a)
4.2 Solving One-dimensional Second-order Equations with Green Functions (BF 7.3)
4.2.1 Solutions in terms of Green Functions
We shall now investigate the conditions that allow the existence and uniqueness of finh formally written as
finh (t) = [L−1 F ](t), where L−1 is an integral operator whose action on F (t) is:
Z
−1
L F (t) = G(t, t′ ) F (t′ ) dt′ (4.10)
Assuming that F is square-integrable over some interval, we want L−1 to return a square-integrable result L−1 F (t),
ie., finh (t); this is the case if the two-point function G(t, t′ ) is itself square-integrable over the interval (see the
end of section BF7.1 for more details). R
Now, acting on the above equation with L gives: [Lf ](t) = F (t) = [Lt G](t, t′ ) F (t′ ) dt′ . This is satisfied
provided G(t, t′ ) obeys:
Lt G (t, t′ ) = δ(t − t′ ) (4.11)
We shall refer to this as the defining equation for a Green function G(t, t′ ) of L. It should be expected that any
indefinite solution of eq. (4.11) must be supplemented with B.C. related to those on f (t).
For the Green function to exist, L must be invertible, which we have seen requires that there be no non-trivial
homogeneous solution with homogeneous B.C.
The link between the existence of the Green function and the criterion for invertibility of L can be made more
tangible if L has a complete set of orthonormal eigenfunctions φj of L, with associated eigenvalues λj , on the
interval. A version of the spectral theorem of operator theory asserts that such a set exists if L is in self-adjoint
86
Lecture Notes on Mathematical Methods 2022
form, ie., if β = α̇ in eq. (4.5). Even if it isn’t, it can always be put in such a form by multiplying it by a function
w and imposing dt (wα) = wβ, which determines w up to a constant.
Then finh can be expanded over the subset R that satisfies homogeneous B.C. (with unknown coefficients aj ),
and so can F , with known coefficients bj = φ∗j (t′ )F (t′ ) dt′ . Both sets of coefficients are, as usual, projections of
f and F on the eigenfunctions. The eigenvalue equation then yields a relation between them, and, assuming that
integral and summation signs can be interchanged, there comes (EXERCISE) the inhomogeneous solution:
X φj (t) Z Z hX i
∗ ′ ′ ′
finh (t) = [φj (t ) F (t ) dt = φj (t) φ∗j (t′ )/λj F (t′ ) dt′
λj
j j
R
We can write this solution as f (t) = G(t, t′ )F (t′ ) dt′ , so long as the Green function:
X φj (t) φ∗ (t′ )
′ j
G(t, t ) = (4.12)
λj
j
exists, ie., only if there is no non-trivial φj satisfying homogeneous B.C. such that L[φj ] = 0. Note, however, that
even if G(t, t′ ) defined as obeying equation (4.11) does not exist, the solution f (t) might still exist, provided
that the φj associated with λj = 0 satisfy bj = 0. But such a solution would be far from unique, because any
multiple of φj could be added to it (see Appendix L for more details).
(a) G(t, t′ ) is a continuous function of t everywhere, including at t = t′ , otherwise its second derivative at t = t′
would be the derivative of a δ-function, and the differential equation would not be satisfied. Note, however,
that the Green function for a first-order operator is discontinuous, eg., L = −idt has as Green function the
step-function i θ(t − t′ ).
(b) Ġ must have a discontinuity at t = t′ . To see this, integrate eq. (4.11) from t = t′ − ǫ to t = t′ + ǫ. Since the
coefficients in L are continuous, they hardly vary when the interval is arbitrarily small (ǫ → 0). In that limit,
the integrals of G and Ġ both vanish because G is continuous, and only the integral of G̈ contributes:
t=t′ +ǫ 1
lim Ġ(t, t′ ) =
ǫ→0 t=t′ −ǫ α(t′ )
Because of the discontinuity in its derivative at t = t′ , G should be different on either side while satisfying
[LG](t, t′ ) = 0, so that it can be written in terms of f1 and f2 :
a1 (t′ ) f1 (t) + a2 (t′ ) f2 (t) t′ < t
G(t, t′ ) =
b (t′ ) f (t) + b (t′ ) f (t) t′ > t
1 1 2 2
The continuity of G and the discontinuity in Ġ at t = t′ then yield the matrix equation at t′ :
f1 (t′ ) f2 (t′ ) a 1 − b1 0
=
f˙1 (t′ ) f˙2 (t′ ) a 2 − b2 1/α
For the system to have a solution, the determinant of the matrix, ie. the Wronskian, W ≡ f1 f˙2 − f˙1 f2 , cannot
vanish anywhere, or else it would vanish everywhere, and f1 and f2 would not be independent as postulated. Then:
a1 − b1 1 −f2 (t′ )
=
a2 − b2 α(t′ ) W (t′ ) f1 (t′ )
87
Lecture Notes on Mathematical Methods 2022
Eliminating a1 and a2 with this equation, the Green function for L without B.C. must take the general form:
′ ′
b1 (t′ ) f1 (t) + b2 (t′ ) f2 (t) − f1 (t) f2 (t ) − f2 (t) f1 (t )
t′ < t
G(t, t′ ) = [α W ](t′ ) (4.13)
b (t′ ) f (t) + b (t′ ) f (t) ′ >t
1 1 2 2 t
The term with the Wronskian vanishes at t = t′ , ensuring the continuity of G as required. The adjustable parame-
ters b1 and b2 can now be chosen so that G satisfies suitable boundary conditions.
The step-function does not make Givp (t, t′ ) discontinuous because the rest of the expression vanishes at t′ = t.
Eq. (4.9) can now be written as:
Z ∞
f˙2 (a) f (a) − f2 (a) f˙(a) f1 (a) f˙(a) − f˙1 (a) f (a)
f (t) = f1 (t) + f2 (t) + Givp (t, t′ ) F (t′ ) dt′
W (a) W (a) a
(4.15)
with Givp given by eq. (4.14). A physicist is pleased that the B.C. guarantee causality: Givp (t′ > t) = 0.
(2) Two-point boundary-value problem
Although superficially similar, the two-point boundary-value problem (BVP) requires a little more care. Many
treatments enforce homogeneous B.C. on Gbvp through B.C. on f1 and f2 (or their derivatives) at the end-
points, a strong restriction. Here, we follow a slightly different approach that, as in the IVP, initially only
assumes linear independence of f1 and f2 , without B.C. on f1 and f2 .
We focus on the Dirichlet problem, where fh (x) is specified at x = a and x = b, with a < b. But we do
impose homogeneous B.C. on the Dirichlet Green function GD : GD (a, x′ ) = 0 (a < x′ ) immediately leads to:
b2 (x′ ) = − b1 (x′ )f1 (a)/f2 (a), whereas GD (b, x′ ) = 0 (x′ < b) gives:
f2 (a) f2 (b) f1 x′ ) − f1 (b) f2 (x′ ) f1 (a) f1 (b) f2 (x′ ) − f2 (b) f1 (x′ )
b1 (x′ ) = =⇒ b2 (x′ ) =
[α W ](x′ ) f1 (a) f2 (b) − f1 (b) f2 (a) [α W ](x′ ) f1 (a) f2 (b) − f1 (b) f2 (a)
The resulting Dirichlet Green function factorises (EXERCISE) in x and x′ :
′ 1 f2 (b) f1 (x> ) − f1 (b) f2 (x> ) f2 (a) f1 (x< ) − f1 (a) f2 (x< )
GD (x, x ) = (4.16)
[α W ](x′ ) f1 (a) f2 (b) − f1 (b) f2 (a)
where x> := max(x, x′ ) and x< := min(x, x′ ). Linear independence of f1 and f2 guarantees the non-
vanishing of W , but unlike an IVP, GD exists only if f1 (a)f2 (b) − f1 (b)f2 (a) 6= 0.
The most simple case occurs when f1 (a) = f2 (b) = 0; then f1 (b) and f2 (a) drop out, leaving: GD (x, x′ ) =
f1 (x< )f2 (x> )/α(x′ )W (x′ ).
We can now write down the general solution to the Dirichlet problem:
Z b
f2 (b) f (a) − f2 (a) f (b) f1 (a) f (b) − f1 (b) f (a)
f (x) = f1 (x) + f2 (x) + GD (x, x′ ) F (x′ ) dx′
f1 (a) f2 (b) − f1 (b) f2 (a) f1 (a) f2 (b) − f1 (b) f2 (a) a
(4.17)
If the homogeneous B.C. allow the existence of a non-zero homogeneous solution, ie., an eigenfunction of L
with eigenvalue zero satisfying these same conditions, eq. (4.12) forbids a Green function. 88
Lecture Notes on Mathematical Methods 2022
Note the dependence of the IVP Green function on the difference t − t′ . Indeed, it can be shown
(EXERCISE) that for the second-order linear differential equation: [Lf ](t) = F (t) with constant
coefficients, Green functions for a one-dim IVP must satisfy G(t, t′ ) = G(t − t′ ), just by using the
general form of the homogeneous solutions: f± (t) = eλ± t . This is a manifestation of the invariance
of the differential operator with constant coefficients under translations of the variable t (eg. time).
By contrast, for the same L, f1 and f2 , (with ω0 = k), but with a Dirichlet problem at a = 0 and b,
we immediately obtain from eq. (4.16):
1
GD (x, x′ ) = sin k(x> − b) sin kx< (4.18)
k sin kb
and, provided kb 6= nπ, the unique inhomogeneous part of the solution to (d2x + k2 )f (x) = F (x) is:
Z x Z b
sin k(x − b) ′ ′ ′ sin kx
finh (x) = sin(kx ) F (x ) dx + sin k(x′ − b) F (x′ ) dx′
k sin kb a k sin kb x
If kb = nπ (n ∈ Z), ie., if b is an integer multiple of the half-period, the condition for the existence
of a Dirichlet Green function, f1 (a)f2 (b) − f1 (b) f2 (a) = − sin kb 6= 0, is violated.
Note that the same result would have followed from the initial choice f1 (0) = f2 (b) = 0, where
f1 = sin kx and f2 = sin k(x − b), with now W = −k sin kb.If k = nπ/b for some integer n 6= 0,
W = 0, so that f1 and f2 are linearly dependent. φ0 (x) = sin nπ(b − x)/b satisfies homogeneous
B.C. (at x = 0 and b) and solves the homogeneous equation. Thus, L = d2x + (nπ/b)2 is not
invertible and the standard Green-function approach fails. As discussed in Appendix L, a modified
Green function could still be constructed if φ0 (x)F (x) integrates to zero over the interval, but the
complete solution would not be unique unless an extra normalisation condition is imposed on the
homogeneous solutions.
Example 4.2. Eq. (4.13) has no explicit dependence on the coefficient of the first–order derivative
in L. This reflects the option we know we have to eliminate it from a second-order equation. For
instance, invoking eq. (4.8) with constant coefficients transforms the homogeneous equation for a
damped harmonic oscillator, (d2t + 2γdt + ω02 )f (t) = 0, into d2t g(t) + (ω02 − γ 2 )g(t) = 0, with
f (t) = g(t)e−γt .Inserting
p a solution
of the form eλtp
, we find the independent homogeneous
p solutions:
−γt ′
f1 (t) = e sin ω0 − γ t , f2 (t) = e−γt cos
2 2 ω02 − γ 2 t . Now W = − ω02 − γ 2 e−2γt , and
a straightforward substitution into eq. (4.14) for an IVP gives:
p 2
′ ′ −γ(t−t )′ sin ω0 − γ 2 (t − t′ )
G(t, t ) = θ(t − t ) e p (4.19)
ω02 − γ 2
89
Lecture Notes on Mathematical Methods 2022
Example 4.3. While we are talking about the damped harmonic oscillator, let us use it to illustrate
another way to solve differential equations that combines Fourier and Green techniques. The idea is
to write the equation:
f¨(t) + 2γ f˙(t) + ω02 f (t) = F (t)
in the frequency domain, assuming that the driving force dies at t → ±∞ or, alternatively, is turned
on at, say, t = 0, and then off at some later time. In this case the Fourier transform of F (t) exists and,
writing the Fourier representation of a function and of its drivative:
Z Z
1 iωt
iω
dt f (t) = √ dt f (ω) e dω = √ f (ω) eiωt dω
2π 2π
it is easy to see that our differential equation becomes:
Z h i
1
√ f (ω) − ω 2 + 2i γ ω + ω02 − F (ω) eiωt dω = 0
2π
Then, because the Fourier transform of the zero function vanishes everywhere, the differential equa-
tion is turned into the algebraic equation:
F (ω)
f (ω) =
−ω 2 + 2i γ ω + ω02
To go back to the time domain, we just write a solution to the inhomogeneous equation:
Z ∞ Z " Z #
1 1 eiω(t−t′ ) dω
f (t) = √ f (ω) eiωt dω = F (t′ ) dt′
2π −∞ 2π −ω 2 + 2i γ ω + ω02
Z ∞
= G(t, t′ ) F (t′ ) dt′
−∞
where: Z ′
∞
′ ′ 1 eiω(t−t )
G(t, t ) = G(t − t ) = − dω
2π −∞ (ω − ω+ )(ω − ω− )
p
with ω± = ± ω02 − γ2 + iγ.
To calculate G for t > t′ , we use contour integration in the complex ω plane, with the contour C
chosen to be counterclockwise around the upper infinite half-plane. Both poles ω = ω± lie in the
upper half-plane. Breaking up the contour into the real axis plus the semi-circle at infinity, we have:
Z ∞ ′ I ′ Z ′
1 eiω(t−t ) 1 eiω(t−t ) 1 eiω(t−t )
− dω = − dω + dω
2π −∞ (ω − ω+ )(ω − ω− ) 2π C (ω − ω+ )(ω − ω− ) 2π |ω|→∞ (ω − ω+ )(ω − ω− )
With t − t′ > 0, the numerator in the second integral on the right goes to zero as |ω| → ∞, and the
integral vanishes. The contour integral is evaluated with the Residue theorem:
Z ∞ ′ ′ ′
!
′ 1 eiω(t−t ) −1 eiω+ (t−t ) eiω− (t−t )
G(t − t ) = − dω = 2πi −
2π −∞ (ω − ω+ )(ω − ω− ) 2π ω+ − ω− ω+ − ω−
p 2
′ sin ω − γ 2 (t − t′ )
= e−γ(t−t ) p0
ω02 − γ 2
When t−t′ < 0, we must use a contour enclosing the lower infinite half-plane. But the integrand in the
contour integral is analytic in this region, and the integral vanishes by the Cauchy-Goursat theorem.
Thus, G(t, t′ ) = 0 for t < t′ , and we have recovered the result obtained in eq. (4.19) for an IVP. Here,
however, no knowledge of the homogeneous solutions was needed to find the Green function! As for
a BVP, if we can find a particular solution to eq. (4.11), we can enforce, eg., GD = 0 at the end-points
for a Dirichlet problem, by adding a suitable term G̃ that satisfies L[G̃) = 0.
90
Lecture Notes on Mathematical Methods 2022
4.3.1 Green’s second 1-dim identity and general solution of a BVP in terms of Green functions
Assume that a second-order linear operator Lx = αd2x + βdx + γ has been put in self-adjoint form, that is, β = α′ ,
with α′ = dx α. Then a few manipulations (EXERCISE) lead to Lagrange’s identity:
′
v Lx [u] − u Lx [v] = (α (v u′ − u v ′ )
where u, v ∈ D of L. Integrate over an interval [a, b] to obtain Green’s second identity in one dimension:
Z b
x=b
v Lx [u] − u Lx [v] dx = α (v u′ − u v ′ ) (4.20)
a x=a
Thanks to this identity, the homogeneous part, fh , of the solution to a BVP for L[f ] = F can be expressed in terms
of the same Green function that appears in the inhomogeneous solution, and for any B.C., homogeneous or not.
Indeed, suppose that u = G(x, x′ ) and that v = f (x) is the general solution to the inhomogeneous equation.
Then one easily shows from Green’s identity that for x′ ∈ [a, b]:
Z b
′
x=b
f (x ) = G(x, x′ ) F (x) dx − α G(x, x′ ) ∂x f − f ∂x G(x, x′ ) (4.21)
a x=a
where G(x, x′ ) is a Green function for Lx . We are already familiar with the first (inhomogeneous) term, but the
second one warrants careful examination. Obviously, it must be related to the homogeneous solution. But wait—is
f (x′ ) actually the general solution? Not yet! It is still just an identity. The second term is evaluated at the end-
points of the interval, so it depends on the boundary conditions for f . We cannot freely specify f and f ′ at both
a and b as this would be in general inconsistent. If f is specified at the end-points, then we must first find the
solution for f in order to know what its derivatives are at the end-points.
For a Dirichlet problem, however, we know that GD = 0 at the end-points. After interchanging x and x′ , and
using the symmetry, proved below, of GD in its arguments, there comes the general solution:
Z b h ix′ =b
f (x) = GD (x, x′ ) F (x′ ) dx′ + α f ∂x′ GD ′ GD (x, a) = GD (x, b) = 0 (4.22)
a x =a
Compare this form of the general solution, which explicitly depends only on F (x) and GD , plus f (a) and f (b), to
the solution (4.17) in terms of the linearly independent homogeneous solutions. It is a very instructive EXERCISE
to show their equivalence. We also see that if f happens to obey homogeneous B.C., f (a) = f (b) = 0, there is no
homogeneous part, which we saw in section 4.1.2 was essential to the existence of GD .
One important property of Dirichlet Green functions may be derived by letting v = GD (x′′ , x′ ) and u =
GD (x′′ , x) in Green’s second 1-dim identity (4.20), which holds
for differential operators of the form Lx′′ =
dx′′ (αdx′′ ) + γ. Because GD = 0 at the end-points and Lx′′ G (x , y) = δ(x′′ − y), we immediately find that
′′
91
Lecture Notes on Mathematical Methods 2022
By transforming the Cartesian components of L to spherical coordinates (r, θ, φ), we obtain (the calculation
is rather tedious, but Maple/Mathematica will readily do it for us):
The derivatives with respect to r have cancelled out! We also find that:
2 1 1 2
L = − ∂θ (sin θ ∂θ ) + ∂ (4.27)
sin θ sin2 θ φ
So L2 depends only on the angular coordinates. Eq. (4.26) makes it obvious that the commutator [∇2 , L2 ] = 0.
Now one readily shows (see section 3.2.4) that the following important relations hold:
Now we require that H (and f ) be single-valued, that is, H(φ + 2π) = H(φ). Thus:
which constrains m to be any integer. Therefore, l := mmax must also be an integer. This is what rules out
the possibility of half-integer eigenvalues allowed for a self-adjoint operator J that just satisfies the canonical
commutation relations: [Ji , Jj ] = iǫijk J k .
The θ dependence of the eigenfunctions must be derived from the eigenvalue equation for L2 . Call f (θ, φ) =
m
Yl (θ, φ) = F (θ)H(φ); these must satisfy:
2 m 1 1
L Yl (θ, φ) = − ∂θ (sin θ ∂θ ) + 2
∂ Y m (θ, φ) = l(l + 1) Ylm (θ, φ)
sin θ sin2 θ φ l
as well as Lz Ylm (θ, φ) = mYlm (θ, φ), ie., Lz H(φ) = mH(φ). Then Ylm (θ, φ) = F (θ)ei mφ , and:
1 m2
− dθ (sin θ dθ ) − F (θ) = l(l + 1) F (θ)
sin θ sin2 θ
93
Lecture Notes on Mathematical Methods 2022
Instead of solving this equation by brute force, we use a clever technique involving the ladder operators L± :
L± = ± eiφ ∂θ ± i cot θ ∂φ
whose solution is F (θ) = (sin θ)l . Therefore, Yll = (sin θ)l eilφ . Applying L− the requisite number of times
generates the other Ylm (0 < m < l): Ylm ∝ Ll−m l
− Yl . When normalised, these are the spherical harmonics:
s
(−1)m 2l + 1 (l − m)!
Ylm (θ, φ) = (1 − x2 )m/2 dl+m
x (x2 − 1)l eimφ x = cos θ (4.31)
2l l! 4π (l + m)!
L2 Ψ 1
∇2 Ψ + γ(x)Ψ = − 2
+ ∂r Ψ + ∂r (r ∂r Ψ) + γ(x) Ψ
r r
L2 F (θ, φ) F (θ, φ)
= − R(r) 2
+ dr R(r) + dr (r dr R(r)) + γ(x) R(r)F (θ, φ)
r r
Multiplying the second line by r 2 /(R(r)F (θ, φ)), we see that the equation is separable provided γ(x) = γ(r):
R(r)
L2 F (θ, φ) = λ F (θ, φ) dr R(r) + dr (r dr R(r)) + r γ(r) R(r) = λ
r
The first equation is the eigenvalue equation for L2 , whose eigenvalues are λ = l(l + 1) (l ≥ 0 ∈ Z), with the
spherical harmonics Ylm (θ, φ) as eigenfunctions. The radial equation can thus be written:
1 2
l(l + 1)
dr r dr Rl (r) + γ(r) − Rl (r) = 0
r2 r2
When γ(r) = 0, this is the radial part of the Laplace equation which becomes, after the change of variable
r = ex : d2x R+dx R−l(l+1)R = 0. Inserting a solution of the form epx turns the equation into p2 +p−l(l+1) = 0,
that is, p = l or p = −(l + 1), which leads to R = Aelx + Be−(l+1)x = Ar l + Br −(l+1) . Therefore, the general
solution to the Laplace equation in spherical coordinates is:
X∞ X l
Blm
Ψ(r, θ, φ) = Alm r + l+1 Ylm (θ, φ)
l
(4.32)
r
l=0 m=−l
The coefficients Alm and Blm are determined from boundary or matching conditions. In regions either containing
the origin, or extending all the way to infinity, Blm = 0 or Alm = 0, respectively. Clearly, if this solution is to
be regular, and if it holds everywhere, it must vanish. In other words, if the Laplace equation is valid everywhere,
it has no non-vanishing regular solution. For a non-trivial solution, there must be a region of space where there
exists an inhomogeneous term acting as a source.
Note, however, that the general solution holds at any point where there is no source. The effect of sources is
encoded in the coefficients Alm and Blm .
94
Lecture Notes on Mathematical Methods 2022
When γ(r) = k2 > 0, we get the radial part of the Helmholtz equation in spherical coordinates:
2 l(l + 1)
d2r Rl (r) + dr Rl (r) + k2 − Rl (r) = 0
r r2
Defining dimensionless x = kr readily transforms it into a form of the Bessel equation whose solutions are the
spherical Bessel functions of the first and second (Neumann) kind, usually written as (see also Jackson’s Classical
Electrodynamics, section 9.6):
l xl x ≪ (1, l)
1 d sin x
jl (x) = (−x)l ∼ (4.33)
x dx x 1 sin(x − lπ/2) x ≫ l
x
1
l
− l+1 x ≪ (1, l)
l 1 d cos x x
nl (x) = − (−x) ∼ (4.34)
x dx x
1
− cos(x − lπ/2) x ≫ 1
x
The general solution of the Helmholtz equation is a linear combination of the jl and nl .
Here are a few spherical Bessel and Neumann functions as plotted on Maple, with ρ = x:
The nl diverge at the origin and thus are excluded from any solution regular at the origin.
(1,2)
Spherical Bessel functions hl (x) = jl (x) ± i nl (x), aka Hankel functions of the first and second kind, can
(1,2)
come in handy. One can express the general solution of the Helmholtz equation in terms of the hl .
95
Lecture Notes on Mathematical Methods 2022
where ∂V is the closed boundary of the volume V of integration, and the unit vector n̂ normal to ∂V , by con-
vention, always points outward from the volume. This is Green’s first identity in three dimensions; when α is a
constant, and introducing the normal derivative ∂n = n̂ · ∇, it reduces to the more familiar form:
Z h i I I
f ∇2 g + ∇f · ∇g d3 x = f ∇g · dS = f ∂n g dS (4.36)
V ∂V ∂V
Interchanging f and g in the first identity (4.35) and subtracting, adding and subtracting γf g in the volume
integral yields the second Green identity in three dimensions—compare with one-dim eq. (4.20):
Z I I
3
f L[g] − g L[f ] d x = α f ∇g − g ∇f · dS = α f ∂n g − g ∂n f dS (4.37)
V ∂V ∂V
Example 4.4. Uniqueness and existence of solutions for the Poisson equation with B.C.
The Poisson (inhomogeneous Laplace) equation is of the form ∇2 Ψ(x) = F (x). We also specify
B.C. for either Ψ or ∂n Ψ on ∂V . With f = g = Ψ3 and α constant, eq. (4.36) becomes:
Z I
2 2 3
[Ψ3 ∇ Ψ3 + (∇Ψ3 ) ] d x = Ψ3 ∂n Ψ3 dS
V ∂V
Suppose there exist two solutions, Ψ1 and Ψ2 , of ∇2 Ψ(x) = F (x) that satisfy the same conditions
on the surface. Define Ψ3 := Ψ2 − Ψ1 . Then ∇2 Ψ3 = 0 inside the R volume. The surface integral
2 3
is zero because either Ψ3 = 0 or ∂n Ψ3 = 0 on the surface; and (∇Ψ3 ) d x = 0 everywhere.
Also, Ψ3 being twice differentiable at all points in the volume, ∇Ψ3 is continuous and therefore zero
everywhere inside, so that Ψ3 is a constant. It follows immediately that if Ψ3 = 0 on ∂V , Ψ1 = Ψ2
everywhere. On the other hand, when ∂Ψ3 /∂n = 0 on ∂V , Ψ3 can be a non-zero constant inside.
We conclude that Ψ1 = Ψ2 inside the volume (up to a possible additive constant), and that the solution,
if it exists, is uniquely determined. The importance of this result cannot be overstated: any function
that satisfies the inhomogeneous Laplace equation and the B.C. is the solution, no matter how it was
found! Moreover, we see that we cannot arbitrarily specify both Ψ and ∂Ψ/∂n on the boundary since
one suffices to determine the solution.
The B.C. determine the solution, but only if it exists. Further conditions must be met for this to happen.
Indeed, integrate ∇2 Ψ(x) = F (x) over (connected!) V ; the divergence theorem yields the condition:
Z Z
F (x) d3 x = ∂n Ψ(x) dS (4.39)
V ∂V
Another condition for the existence of a solution is that the enclosing boundary be “reasonably”
smooth (eg. no spikes . . . ) if we wish to specify ∂n Ψ on ∂V .
Finally, if ∇2 φn = λn φn , and taking f = φ∗n and g = φn in eq. (4.36), one shows (EXERCISE) that
the eigenvalues of the Laplacian are always negative.
96
Lecture Notes on Mathematical Methods 2022
4.7.2 Green function for the 3-d Elliptic Helmholtz operator without boundary conditions
We proceed to find a Green function for the operator ∇2 + λ, with λ a constant. The Fourier transform of
(∇2 + λ) Ψ(x) = F (x) is (−k2 + λ)ψ(k) = F (k). We must distinguish between two possibilities:
1. λ = −κ2 ≤ 0, κ ≥ 0
Then, similarly to what happens in one dimension (example 4.3), an “inhomogeneous” solution is:
Z ZZ ′
1 F (k) ik·x 3 1 3 ′ −ik·x′ F (x )
Ψ(x) = − e d k = − d x e eik·x d3 k
(2π)3/2 k2 + κ2 (2π)3 k2 + κ2
R
Compare with the Green-function form of the inhomogeneous solution, V F (x′ )G(x, x′ ) d3 x′ (EXER-
CISE):
Z ik·(x−x′ ) Z ∞ ′
1 e i k eik|x−x |
G(x, x′ ) = − d 3
k = dk
(2π)3 k2 + κ2 (2π)2 |x − x′ | −∞ k2 + κ2
This integral is easily evaluated as part of a contour integral around a semi-circle at infinity in the upper
complex k half-plane. As in the one-dimension example, the contribution at infinity vanishes, and the
′
residue due to the pole at k = iκ is e−κ|x−x | /2. The Residue theorem then yields the (sometimes called
fundamental, or singular) solution:
′
′ 1 e−κ|x−x |
G(x, x ) = − (4.43)
4π |x − x′ |
98
Lecture Notes on Mathematical Methods 2022
2. λ = κ2 ≥ 0
In order to invert the algebraic equation for ψ(k), we write λ = (q ± iǫ)2 (ǫ ≥ 0). Then we arrive at:
Z ′ ′
(±) ′ 1 eik·(x−x ) 3 1 e±iq|x−x |
Gq (x, x ) = − lim d k = − (4.44)
(2π)3 ǫ→0 k2 − (q ± iǫ)2 4π |x − x′ |
For details of the calculation, see pp. BF415–416.
Do check that these Green functions satisfy (∇2 + λ)G(x, x′ ) = δ(x − x′ ). But note that they are not the
general solution of this equation, since any function that satisfies the homogeneous equation can be added to them!
If the volume integral extends over all space, the surface integral in the Dirichlet solution for the case λ < 0
certainly vanishes at infinity for fairly weak conditions on Ψ(x), because of the exponential factor in Green’s
function. When λ ≥ 0, the surface integral also vanishes provided Ψ(x) → 0 faster than 1/|x − x′ |, (since
dS ∼ |x − x′ |2 ), and we are left with just the inhomogeneous integral:
Z ′
(±) 1 F (x′ ) e±iq|x−x | 3 ′
Ψq (x) = − d x (4.45)
4π V |x − x′ |
If, however, Ψ(x) does not vanish fast enough at infinity, it is more convenient to write it in terms of the
solution of the homogeneous equation (∇2 + q 2 )Ψ(x) = 0, plus the volume integral:
Z ′
(±) iq·x 1 F (x′ ) e±iq|x−x | 3 ′
Ψq (x) = A e − d x (4.46)
4π V |x − x′ |
Note that these expressions for Green’s functions assume no boundary surfaces (except at infinity)!
Example 4.5. Solution of the Dirichlet problem on a sphere for the Laplacian
Consider a sphere of radius a centered on the origin. We want: GD (an̂, x′ ) = GD (x, an̂′ ) = 0
Symmetry of GD dictates that x′′ and x′ be collinear, which means that, at |x| = r = a, we can write:
!
′ 1 1 g
GD (an̂, x ) = − + ′′ a
4π a n̂ − ra′ n̂′ r r′′ n̂ − n̂′
where rn̂ = x, etc. By inspection, we see that if GD (an̂, x′ ) is to vanish for n̂ in an arbitrary direction,
we must have: 1/a = −g/r ′′ and r ′ /a = a/r ′′ . Then:
g = − a/r ′ , r ′ r ′′ = a2 (4.47)
99
Lecture Notes on Mathematical Methods 2022
Thus, x′′ does lie outside the sphere if x′ is inside, and vice-versa. Replacing an̂ by rn̂ = x yields:
" #
1 1 1
GD (x, x′ ) = − −
4π |x − x′ | (r ′ /a)x − (a/r ′ )x′
(4.48)
1 1 1
= − p − q
4π 2
r 2 + r ′ − 2rr ′ cos γ 2 ′ 2 2 2 ′
r r /a + a − 2rr cos γ
The second form makes is most easy to see that not only GD (x, an̂′ ) = 0, but also GD (an̂, x′ ) = 0,
as desired. In spherical coordinates centered on the sphere, the angle γ between x amd x′ is, from
spherical trigonometry: cos γ = cos θ cos θ ′ + sin θ sin θ ′ cos(φ − φ′ ). The Dirichlet Green function
we have found is valid for any ball since it does not care about which particular B.C. is specified for
Ψ(x) on its spherical boundary.
When Ψ(r ′ = a) = 0, the surface integral in eq. (4.41) vanishes; the volume integral remains the
same since it is independent of the B.C. for Ψ. If Ψ(r ′ = a) 6= 0, we must evaluate ∂n′ GD on the
sphere. In spherical coordinates, this is:
depending on whether dS′ , the normal to the surface which always points out of the volume, is in the
direction of x′ or in the opposite direction. Then the general solution of the inhomogeneous Laplace
equation with B.C. specified on the surface r = a for Ψ is:
Z
1 1 1
Ψ(x) = F (x′ ) q − p d 3 x′
4π 2 ′ 2 2 2 ′ r 2 + r ′ 2
− 2rr ′ cos γ
r r /a + a − 2rr cos γ
I
1 a2 − r 2
± Ψ(r ′ = a) dS ′ (4.49)
4π a (r 2 + a2 − 2ar cos γ)3/2
where the (+) sign refers to the solution for r < a and the (−) sign applies to r > a. In the latter
case, there is an implicit assumption that the integrand, Ψ∂n′ GD , of the surface integral vanishes at
infinity faster than 1/r ′2 . When F (x) = 0 everywhere inside the volume where the solution is valid,
we are left with the Laplace equation ∇2 Ψ = 0, with solution:
I
′ ′ 1 a(a2 − r 2 )
Ψ(x) = ± Ψ(a, θ , φ ) dS ′ (4.50)
4πa2 (r 2 + a2 − 2ar cos γ)3/2
Clearly also, if Ψ(a, θ ′ , φ′ ) 6= 0 and r > a, F (x) 6= 0 somewhere in the region r < a, and vice-versa.
One can also show that for a ball of radius a and surface Ωn−1 in Rn :
1
ln |x − x′ | − ln r ′ x − a x′ (n = 2)
2π a r′
′
GD (x, x ) =
1 1 1
− − (n > 2)
(n − 2) Ωn−1 |x − x′ |n−2 |(r ′ /a)x − (a/r ′ )x′ |n−2
(4.51)
which leads to a unified expression, valid for n ≥ 2, for the normal derivative of GD on the sphere:
1 a2 − r 2
∂n′ GD = ± an−2 (4.52)
r ′ =a Ωn−1 |x − x′ |n
r ′ =a
100
Lecture Notes on Mathematical Methods 2022
where the completeness relation for spherical harmonics has been invoked:
∞ X
X l
∗
Ylm (θ ′ , φ′ )Ylm (θ, φ) = δ(x − x′ )δ(φ − φ′ ) (x = cos θ) (4.54)
l=0 m=−l
Inserting into eq. (4.53), we immediately find with eq. (4.26) that gl (r, r ′ ) must satisfy the radial equation:
r 2 ∇2r gl (r, r ′ ) = dr r 2 dr gl (r, r ′ ) − l(l + 1) gl (r, r ′ ) = δ(r − r ′ )
We now find ourselves in the familiar territory of 1-dim Green-function problems and self-adjoint operators. For
instance, we can connect with eq. (4.13) for a 1-dim Dirichlet Green function. We take two concentric spheres of
radius a and b, with b > a.
We have α(r ′ ) = r ′2 and, with f1 = r l and f2 = r −(l+1) , W (r ′ ) = −(2l + 1)/r ′2 . Also, let r< ≡ min(r, r ′ )
and r> ≡ max(r, r ′ ). It takes only a straightforward computation using eq. (4.16) to arrive at (EXERCISE):
∞ X
X l ! !
∗ (θ ′ , φ′ ) Y (θ, φ) l
Ylm lm a2l+1 r> 1
GD (x, x′ ) = l
r< − l+1 − (4.55)
(2l + 1) 1 − (a/b)2l+1 r< b2l+1 l+1
r>
l=0 m=−l
Inspection of the radial factors shows that this expression vanishes at r = a and r = b (and when r ′ = a or
r ′ = b), as it should. We did not have to require this since it is built in the derivation of the 1-dim Dirichlet Green
function. Two important cases:
∞ X
X l !
′ Y ∗ (θ ′ , φ′ )Y
lm lm (θ, φ) l
l
r> 1
GD (x, x ) = r< − (a = 0) (4.56)
(2l + 1) b2l+1 l+1
r>
l=0 m=−l
∞ X
X l !
Y ∗ (θ ′ , φ′ ) Y lm (θ, φ) 1 a2l+1
GD (x, x′ ) = lm
l+1 l+1
l
− r< (b → ∞) (4.57)
(2l + 1) r> r<
l=0 m=−l
The first expression gives the Green function inside a sphere of radius b; the second one, outside a sphere of radius
a and all the way to infinity. When there are no boundary surfaces, we obtain over all space:
∞ X
X l
1 l
r<
G(x, x′ ) = − Y ∗ (θ ′ , φ′ ) Ylm (θ, φ)
l+1 lm
(4.58)
2l + 1 r>
l=0 m=−l
101
Lecture Notes on Mathematical Methods 2022
This also yields a useful expansion of the ubiquitous distance factor 1/|x − x′ |.
When 0 ≤ r ≤ b (interior case) we can rewrite (EXERCISE) the surface integral in eq. (4.49) as:
∞ X
X l Z
′ ′ ∗ r l
Ψ(b, θ , φ ) Ylm (θ ′ , φ′ ) dΩ′ Ylm (θ, φ)
b
l=0 m=−l
where Ψ(b, θ ′ , φ′ ) is specified on the surface r = b. The normal derivative of the Green function on the surface,
∂G/∂n′ = ∂G/∂r ′ r′ =b , has been evaluated for r< = r and r> = r ′ since r < r ′ = b. Also, the surface element
on a sphere of radius b is dS ′ = b2 dΩ′ . This expression is still rather complicated, but it simplifies considerably
if Ψ(b, θ ′ , φ′ ) exhibits a symmetry (eg. azimuthal). Also, if one can write Ψ(b, θ ′ , φ′ ) as a linear combination of
spherical harmonics, the angular integration becomes trivial due to the orthonormality of the harmonics, and only
a few terms in the sums might contribute.
102
Lecture Notes on Mathematical Methods 2022
The time-independent Schrödinger equation (TISE) for a potential V (x) takes the following suggestive form:
2m
(∇2 + λ) ψ(x) = V (x) ψ(x) (4.59)
~2
where λ = 2mE/~2 . Although the right-hand side is not inhomogeneous, our previous results still hold.
For bound states (E < 0) of an attractive potential, λ = −κ2 < 0, and we have the integral equation:
Z −κ|x−x′ |
m e
ψ(x) = − V (x′ ) ψ(x′ ) d3 x′
2π~2 |x − x′ |
A somewhat simpler integral expression may be derived from the convolution [V ∗ ψ](k):
Z
1
[V ∗ ψ](k) := V (k − k′ )ψ(k′ ) d3 k′
(2π)3/2
According to the convolution theorem, the Fourier transform of [V ∗ ψ](k) is just V (x)ψ(x) in eq. (4.59). Then
the Fourier representation of this equation can be written as:
Z Z Z
2 2 ik·x 3 2m 1
− (k + κ )ψ(k) e d k = 2 ′ ′ 3 ′
V (k − k )ψ(k ) d k eik·x d3 k
~ (2π)3/2
and there comes: Z
2m V (k − k′ )ψ(k′ ) 3 ′
ψ(k) = − d k
(2π)3/2 ~2 k2 + κ2
For unbound states (E > 0), κ2 > 0, and we can immediately write the Lippmann-Schwinger equation:
Z ±iq|x−x′ |
(±) A iq·x m e
ψq (x) = e − 2
V (x′ ) ψq(±) (x′ ) d3 x′ (4.60)
(2π) 3/2 2π~ |x − x′ |
p
with q = 2mE/~2 .
The asymptotic√form of the Lippmann-Schwinger equation is of particular interest. When r >> r ′ , we can
expand |x − x′ | = r 2 − 2x · x′ + r ′2 ≈ r − n̂ · x′ , with n̂ = x/r. Inserting into the integral equation yields:
Z
(±) A iq·x m e±iqr ′ (±)
ψq (x) = 3/2
e − 2
e∓iqn̂·x V (x′ ) ψq (x′ ) d3 x′
r→∞ (2π) 2π~ r
A iq·x e±iqr
= e + f± (q)
(2π)3/2 r
This expression represents the spatial dependence of a superposition of a plane wave and a scattered spherical
wave propagating inward or outward from the origin. The function f± (q) is called the scattering amplitude; it
also obeys an integral equation in q (momentum) space, eq. BF7.75, and its square modulus is directly related to
experimental data. See BF p. 414–420 for more details and an application to the Yukawa potential.
Just as for the Laplacian, there exist Green functions for ∇2 + k2 ; we have found them earlier in eq. (4.44):
1 e±ikR
G(±) (R) = − R ≡ |x − x′ | (4.63)
4π R
Now we are ready to derive the full Green functions for the d’Alembertian operator, which satisfy:
′ ′
x G(x, t; x , t ) = δ(x − x′ ) δ(t − t′ ) (4.64)
Thus, in non-elliptic problems, Green functions can contain δ-functions and so may not be actual functions!
Using eq. (4.64), we also recognise that:
Z Z ∞ Z Z ∞
3 ′ (±) ′ ′ ′ ′ ′ 3 ′
x d x G (x, t; x , t ) F (x , t ) dt = d x F (x′ , t′ ) x G(±) (x, t; x′ , t′ )dt′ = F (x, t)
all
space
−∞ −∞
has the generic form Ψ(x, t) = F (x, t), which shows that the general solution of a wave equation with sources
can be written either as the retarded or advanced solutions:
Z Z ∞
Ψ{ ret } (x, t) = Ψ{ in } (x, t) + G(±) (x, t; x′ , t′ ) F (x′ , t′ ) d3 x′ dt′
adv out −∞
Z
1 F (x′ , t ∓ R/c) 3 ′
= Ψ{ in } (x, t) − d x (4.67)
out 4π x − x′ { ret }
adv
where in the integral the position x′ must be evaluated at the retarded time t − R/c, or at the advanced time
t + R/c. This ensures the proper causal behaviour of the solutions, in the sense that, eg., the solution at time t only
depends on the behaviour of the source point x′ at time t − R/c. Ψin and Ψout are possible plane-wave solutions
of the homogeneous wave equation for Ψ. Often they can be dropped.
104
Lecture Notes on Mathematical Methods 2022
where ke and km are constants that depend on the system of units, and ke /km = c2 , with c the speed of light. The
1
source terms ρ and J satisfy a continuity equation: ∂t ρ = 4πk e
∇ · ∂t E = − ∇ · J.
The two homogeneous equations are equivalent to:
E = − ∂t A − ∇Φ B = ∇×A (4.69)
∇2 Φ + ∂t (∇ · A) = −4π ke ρ
(4.70)
1
A − ∇ ∇ · A + 2 ∂t Φ = − 4π km J
c
.
∇u + ∇ × w
A = |{z}
| {z }
AL AT
AL = ∇u, whose curl vanishes identically, is the longitudinal part (or projection) of A; AT = ∇ × w, whose
divergence vanishes identically, is the transverse projection of A. This allows us to decompose Maxwell’s equa-
tions for the fields and the potential into longitudinal and tranverse parts, which are perpendicular to each other.
Project the second equation (4.70). The transverse projection immediately gives:
AT = − 4π km JT (4.71)
105
Lecture Notes on Mathematical Methods 2022
where we have used the fact a gradient is a longitudinal object. The two transverse components AT satisfy a
proper wave equation and correspond to physically observable quantities, in the sense that being transverse, they
are unaffected by A → A + ∇f , which can change only the longitudinal component AL . Therefore, the time
evolution of the two transverse AT is not arbitrary and they have a well-posed Cauchy problem.
Now, remembering that = ∇2 − (∂t2 )/c2 , take the divergence of the longitudinal projection of (4.70):
1 ∂t ∇2 Φ
∇· AL − ∇ ∇ · AL + 2 ∂t Φ + 4π km JL = (∇ · AL ) − ∇2 (∇ · AL ) − + 4π km ∇ · JL
c c2
1 h i
= − 2 ∂t ∂t (∇ · AL ) + ∇2 Φ + 4π ke ρ
c
where the continuity equation has been invoked in the second line. But the terms in the square bracket on that line
are just the first of equations (4.70). Therefore, the second Maxwell equation for the 3-vector potential contains
no information about ∇ · A that is not in the first equation. But that is really an equation for Φ, with ∇ · Ȧ (more
precisely, ∇ · ȦL ) as a source together with ρ. Therefore, Maxwell’s theory cannot uniquely determine the time
evolution of the divergence of the 3-vector potential. Nor can it uniquely determine the time evolution of Φ, since
Φ is gauge-variant. Systems whose time-evolution involves arbitrary functions are often called singular.
E = 4πke ∇ρ + 4πkm ∂t J
(4.72)
B = − 4π km ∇ × J
These look like wave equations for six quantities. But only those of their solutions which also satisfy the
first-order field equations (4.68), including at initial time t0 , are acceptable.
The two first-order divergence equations contain no time derivatives and are thus constraints on E and B at
t = t0 . The constraint equation on E can be rewritten ∇2 u = ρ, a Poisson-type equation which can be solved for
u at initial time so long as ρ falls off faster than 1/r 2 at infinity). In the case of B, the scalar field u satisfies a
Laplace equation everywhere and is therefore zero. So B has no longitudinal component, only transverse ones. In
both cases, the longitudinal component is either zero or can be solved for at t0 , so cannot be freely specified.
Now look at the two first-order equations (4.68) which contain time derivatives. Suppose we specify E and
∂t E at t = t0 , so as to solve the 2nd -order equations, eq. (4.72). Then the two transverse components of B are
determined by ∇×B = 4πkm J+∂t E/c2 ; ∂t B is determined, also at t = t0 , by the curl equation for E. Therefore,
once we have specified the two transverse components of E and their time derivatives, the first-order equations
take over and determine the others at t = t0 . Alternatively, we could have specified the two transverse components
of B and their time derivatives at t = t0 to constrain all the other field components and time derivatives.
You can also use (EXERCISE) the transverse/longitudinal projections of the first-order equations (4.68) to
show that in source-free space, only the transverse components of E and B obey a classical wave equation.
Thus, the results of the first-order Cauchy-data analysis are fully consistent with the second-order analysis
on A: only two transverse components correspond to independent, physical dynamical degrees of freedom, This
Cauchy analysis does not rely on some particular solution, but is valid for any electromagnetic field and potential.
Since Maxwell’s theory contains no information about ∇ · A, this must be supplied by a so-called gauge
condition. One that is frequently used is the Lorenz condition: ∇ · A = −∂t Φ/c2 ,. Inserting it into Maxwell’s
equation (4.70) for A could lead you to believe that Φ and the three components of A propagate to infinity, whereas
I hope to have convinced you that only the transverse components of A do. In Appendic M, moreover, we show
that AL can be made to disappear without affecting Maxwell’s equations for the fields and the potentials.
106
Lecture Notes on Mathematical Methods 2022
Appendices
J Solving an Inhomogeneous Equation in Terms of Homogeneous Solutions
Let f1 and f2 be independent solutions to the homogeneous differential equation (4.5). We use them to derive a
particular solution finh (t) to the inhomogeneousequation. The key step is to insert finh (t) = f1 (t)g(t) to obtain
a first-order equation for ġ: g̈ + dt (ln f12 ) + β/α ġ = F/αf1 . Then the general first-order solution (4.4), together
with Abel’s formula (4.6) and W (x)/f12 = dt (f2 /f1 ), yields:
Z t
f2 f1 (t′ ) F (t′ ) ′
ġ(t) = dt B + ′
dt
f1 a [α W ](t )
Z t Z t
f2 f1 (t′ ) F (t′ ) ′ f2 f1 (t′ ) F (t′ ) ′
= dt B + ′
dt − dt ′
dt
f1 a [α W ](t ) f1 a [α W ](t )
Z t
f2 f1 (t′ ) F (t′ ) ′ f2 (t) F (t)
= dt B + ′)
dt −
f1 a [α W ](t [α W ](t)
Because we have not implemented homogeneous boundary conditions (B.C.) on this inhomogeneous solution, we
must include the terms Af1 + Bf2 , even though they look like belonging to the homogeneous solution.
We know that finh (t) must satisfy homogeneous (B.C.). We consider the two most important cases.
With one-point B.C. (IVP), finh (t) and its derivative must vanish at t = a. The integral term and its derivative
are automatically zero at t = a. The other contribution also vanishes because the IVP has no non-zero homoge-
neous solution for homogeneous B.C.. In other words, the integral term satisfies the B.C. without any help from
the adjustable constants A and B. Therefore, the inhomogeneous solution to an IVP Is:
Z t Z ∞
f1 (t′ ) f2 (t) − f2 (t′ ) f1 (t) ′ ′ ′ f1 (t′ ) f2 (t) − f2 (t′ ) f1 (t)
finh (t) = ′)
F (t ) dt = θ(t−t ) ′)
F (t′ ) dt′
a [α W ](t a [α W ](t
(J.2)
and it is the general solution when the boundary conditions on the general solution are homogeneous. When they
are not, we must add to finh (t) the homogeneous solution with appropriate non-zero constants A and B. Of course,
α should not vanish for t > a, and neither can the Wronskian, but the latter is guaranteed by our assumption that
f1 and f2 are linearly independent. We conclude that a unique solution to the IVP always exists, provided that
α 6= 0 and that the source term, F (t), is piecewise continuous for t > a.
We should check that our solution (J.2) satisfies the inhomogeneous equation. A surprise awaits us: because
L[f1 ] = L[f2 ] = 0, the integrand does not contiribute to L[finh ] for a ≤ t′ < t; the sole contribution must come
from the point t′ = t. This suggests that the Direc delta-function must somehow be involved, and this is indeed
what happens if we we use the second expression with the step-function, whose derivative is the delta-function.
The other case we wish to address is the Dirichlet problem, ie. the boundary-value problem (BVP) with f
specified at the two end-points. While the integral term in eq. .(J.1) satisfies a homogeneous B.C. at t = a, it does
not at the other end of the interval, at t = b. Enforcing finh (b) = 0 requires adjusting the constants A and B.
finh (a) = 0 immediately leads to B = −Af1 (a)/f2 (a). Then finh (b) = 0 determines A, and we arrive at:
Z t
f1 (t′ ) f2 (t) − f2 (t′ ) f1 (t)
finh (t) = F (t′ ) dt′
a α(t′ ) W (t′ )
Z
f2 (a)f1 (t) − f1 (a)f2 (t) b f1 (t′ ) f2 (b) − f2 (t′ ) f1 (b)
+ F (t′ ) dt′
f2 (a)f1 (b) − f1 (a)f2 (b) a [α W ](t′ )
107
Lecture Notes on Mathematical Methods 2022
This expression looks more symmetric if we combine the two integrals from a to t. Some tedious algebra yields:
Z
f2 (b)f1 (t) − f1 (b)f2 (t) t f1 (t′ ) f2 (a) − f2 (t′ ) f1 (a)
finh (t) = F (t′ ) dt′
f1 (a)f2 (b) − f2 (a)f1 (b) a [α W ](t′ )
Z
f2 (a)f1 (t) − f1 (a)f2 (t) t f1 (t′ ) f2 (b) − f2 (t′ ) f1 (b)
+ F (t′ ) dt′
f1 (a)f2 (b) − f2 (a)f1 (b) b [α W ](t′ )
which can be written in the compact form:
Z b " #
f2 (b)f1 (t> ) − f1 (b)f2 (t> ) f2 (a) f1 (t< ) − f1 (a) f2 (t< )
finh (t) =
′ ) W (t′ ) f (a)f (b) − f (a)f (b)
F (t′ ) dt′ (J.3)
a
α(t 1 2 2 1
where t> := max(t, t′ ) and t< := min(t, t′ ). The existence of the inhomogeneous solution depends on the
denominator of the integrand not vanishing, as well as piecewise continuity of F (t).
As with the IVP, checking the validity of this solution by calculating L[finh ] reveals the same behaviour: only
the point t′ = t contributes. The presence of the delta-function is easier to see if we write the expression in terms
of step-functions which split the integral over two intervals.
The Green-function formalism introduced in the main body of these notes will shed more light on these results.
Thus, a solution exists only if the driving term is itself orthogonal to φ0 over the interval. To discover what form
that solution takes, consider the modified Green function:
X φj (x) φ∗ (x′ )
′ j
G(x, x ) := λj 6= 0 (L.2)
λj
j6=0
While superficially identical to eq. (4.12), this expression specifically omits the now non-zero φ0 (x) φ∗0 (x′ ) term
which simply did not exist in eq. (4.12).
Not surprisingly, and although Pthey satisfy the same homogeneous B.C., G(x, x′ ) and G(x, x′ ) solve different
defining equations: [LG](x, x ) = all j φj (x) φ∗j (x′ ) − φ0 (x) φ∗0 (x′ ), that is:
′
LG (x, x′ ) = δ(x − x′ ) − φ0 (x) φ∗0 (x′ ) (L.3)
because {φj } with φ0 included is a complete set. Then one quickly shows that if condition (L.1) holds, the form:
Z b
f (x) = C φ0 (x) + G(x, x′ ) F (x′ ) dx′ (L.4)
a
is the solution to [Lf [(x) = F (x). But since C is an arbitrary constant, we have lost unicity—in fact the number
of solutions is infinite. Of course, if f must satisfy non-homogeneous B.C., we must also add the homogeneous
solution that satisfies them.
When solving eq. (L.3) to find a modified Green function, we can proceed as in sections 4.2.2 and 4.3. Adding
the particular solution of [LG](x, x′ ) = −φ0 (x) φ∗0 (x′ ) to its homogeneous solution does not change the conditions
on G(x, x′ ) at x = x′ , but that particular solution must be added to (4.13), which will alter the b1 and b2 coefficients
calculated by imposing the relevant homogeneous B.C. on G. Unfortunately, without an explicit φ0 , it becomes
impossible to write a general result for the modified Green function.
Example L.1. The one-dim Laplace equation, d2x f (x) = 0, has for general solution fh (x) = Ax+B,
with A and B constants. The Dirichlet B.C., fh (a) = fh (b) = 0, lead to fh (x) = 0 everywhere,
and the Green function always exists. The homogeneous Neumann B.C., dx fh a = dx fh b = 0,
however, do not determine B, and the homogeneous equation is solved by the non-trivial φ0 (x) = B.
√ once, the same Neumann B.C. on the
Integrating eq. (L.3) modified Green function GN are consistent
only if B = 1/ L, with L := b − a. The equation dx GN = −1/L for x 6= x′ is then solved by:
2
GN (x, x′ ) = −x2 /2L + b1 (x′ )x + b2 (x′ ). Implementing the homogeneous Neumann B.C. leads to:
GN (x, x′ ) = −x2 /2L + ax/L + θ(x − x′ )(x − x′ ) + b2 (x′ ), with b2 (x′ ) arbitrary.
Provided the source in d2x f = F integrates to zero over [a, b], as required by eq. (L.1), a general
solution to this Neumann problem (with homogeneous B.C.!) is given by eq. (L.4), but it is unique
only up to an arbitrary constant.
109
Lecture Notes on Mathematical Methods 2022
1
∇ · (A + ∇f ) = − ∂t (Φ − ∂t f ) (M.1)
c2
This shows that gauge-transformed potentials still satisfy the Lorenz condition! As noted before, it is important to
keep in mind that since the transformation shifts A by a gradient, which is a longitudinal object, it does not affect
the transverse components of A.
Now, for the first time, we shall have to look at actual solutions of the wave equations for A and Φ. To make
things as simple as possible, take plane-wave solutions A = A0 ei(kx−ωt) , where the x-axis has been aligned along
the direction of propagation, and Φ = Φ0 ei(kx−ωt) . Then:
Inserting into the Lorenz condition with ω/k = c yields, as expected, a relation between the longitudinal compo-
nent Ax and Φ: A0x = Φ0 /c.
Now fold in f = f0 ei(kx−ωt) into eq. (M.1) for the gauge-transformed potentials, to get:
ω
ik (A0x + ik f0 )ei(kx−ωt) = i (Φ0 + i ω f0 ) ei(kx−ωt)
c2
Since f0 is arbitrary, we can choose it to cancel A0x , which at the same time gets rid of Φ0 , leaving us with only
the transverse components of A!
Tte conclusion is the same as that of the analysis of the field equations: only the two transverse components of
A propagate, in the sense that they carry energy to infinity.
110