Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Lecture I: Vectors, Tensors, and Forms in Flat Spacetime

Download as pdf or txt
Download as pdf or txt
You are on page 1of 6

Lecture I: Vectors, tensors, and forms in flat spacetime

Christopher M. Hirata
Caltech M/C 350-17, Pasadena CA 91125, USA∗
(Dated: September 28, 2011)

I. OVERVIEW

The mathematical description of curved spacetime is an involved subject and we will spend much of the first term
on developing and applying the associated machinery. Key to all of this work are the concepts of vectors, forms,
and tensors. I’m sure all of you have worked extensively with vectors and vector fields (the electric field in E&M,
the velocity field in hydrodynamics, etc.) in previous coursework, and have studied the operations of vector algebra
(norms, dot and cross products) and of vector calculus (grad, div, curl). All of these have extensions to curved
spacetime, but are customarily dressed in the language of equations with tensors and funny-looking up and down
indices (Aαβ γ ). Before we use this machinery to work in curved spacetime, we will repeat the description of flat
spaces (Euclidean space Rn and Minkowski space M4 ) and special relativity to gain familiarity.
The recommended reading for this lecture is:
• MTW Ch. 1 – This is to give you a taste of where we’re headed in the first term. Don’t worry if you can’t
follow every equation in Ch. 1 yet; it will all be explained in due course! You should read this chapter as you
might the first chapter of a mystery novel ...
• MTW §2.1–2.5 – This is where the mathematical material begins. It covers the same material as in lecture, and
you should make sure you follow it. Since you’ve seen the underlying physics before, albeit in different language,
this is the place for you to get a firm grasp of the notation and how to manipulate tensor expressions. So please
if something in this section doesn’t make sense to you – speak up! Probably someone else in class has a similar
question.

II. VECTORS

[Reading: §2.3]
In freshman physics, you learned about the concept of representing a vector as an arrow (an object with a magnitude
and direction – as distinct from a scalar, with only a magnitude). In special relativity, you probably learned about
the notion of 4-vectors. Several types of vectors were considered, such as:
• The displacement of an object: if something moves from point A to B, then we can write its displacement vector,

sAB = B − A. (1)

• The velocity of an object: if an object is moving along a curve P(λ), where λ is a parameter, then we define its
velocity vector,
dP
v= . (2)

To define a velocity in Newtonian mechanics, P ∈ R3 and λ is the static, globally defined time. In relativity, we
will take λ to be the proper time, and P ∈ M4 .
In this lecture, since we are working in flat spacetime, both Eqs. (1) and (2) are equally valid. When we go to
curved spacetime, only Eq. (2) will make sense. Drawn on the 2-dimensional surface of the Earth, the velocity of a
boat and the velocity of the ocean current are valid 2-dimensional vectors at every point, and they can be added,
subtracted, etc. like any other vector. But the concept of “displacement” as a vector will no longer be meaningful

∗ Electronic address: chirata@tapir.caltech.edu


2

(addition doesn’t commute: think about marching 20 km north and then 20 km east, versus 20 km east and then
20 km north). So for the moment you could think of a vector being tangent to the 4-dimensional surface defining
spacetime, or of a vector as being defined by an expression such as Eq. (2) – e.g. a vector defined at a point P on
Earth’s surface is a possible velocity of an object moving on the surface of Earth that is instantaneously at P. We will
make this more mathematically precise in the context of curved spacetime later (for now we will use the traditional
flat-spacetime notion of vector, and redefine it when we consider curved spacetime).

A. Bases

For the moment, let us proceed: we work in an n-dimensional space, and the vectors v live in some n-dimensional
vector space V. Sometimes we will describe it with an arrow, but in order to do calculations you know it is usually
most convenient to work with coordinates. One constructs coordinates by defining a set of basis vectors. In Euclidean
geometry, these could be the basis vectors of R3 ,

{e1 , e2 , e3 }, (3)

or in special relativity they could be those of M4 :

{e0 , e1 , e2 , e3 }. (4)

(By convention, the vector pointed in the “time” direction is usually denoted as e0 . Aside from this label, we haven’t
done anything yet that singles out “time” as different from “space”.) You are probably accustomed to working with
orthonormal bases, i.e. bases where each vector has unit length and is orthogonal to the others, but we won’t do this
yet. We haven’t even defined “orthogonal” or “unit length” yet! All that is required to be a basis is that every vector
be representable in a unique way by an expression:
3
X
v = v 0 e0 + v 1 e1 + v 2 e2 + v 3 e3 = v α eα , (5)
α=0

where v 0 , v 1 , v 2 , and v 3 ∈ R are components.


Some nomenclature: if there is a way of writing every vector in the form of Eq. (5), then the set of vectors
{e0 , e1 , e2 , e3 } is said to be complete. If no vector can be written in this form with more than one different sets of
components, then {e0 , e1 , e2 , e3 } is said to be linearly independent. Bases are both complete and linearly independent.

B. Very important notation

You will notice that in Eq. (5), we wrote the indices on v α in the “up” rather than “down” position. This is very
important. The location of the index will denote how a quantity transforms under changes of basis (this could be a
rotations in R3 , but it could also be much more general – and in the more general cases these are not equivalent). In
texts, the “up” indices are called contravariant and the “down” indices are covariant. I’ll call them up and down.
The Greek indices are used in 4-dimensional spacetimes and can take any value. At various times we will consider
cases where we want to allow an index to take only spatial values (1, 2, or 3) and deliberately exclude the time – at
those times we will use Latin indices.
A final, very important bit of notation: the index α in Eq. (5) is summed over all of its legal values (0...3).
Expressions of this form are so common in GR that we will often use the Einstein summation convention (EΣC) in
which the sum is implied. That is, we will usually write Eq. (5) in the form

v = v α eα . (6)

The EΣC applies to one up and one down index. Summed indices are dummy indices: one may consistently change
their labels without consequence:

v = v α eα = v β e β . (7)
3

C. Linear algebra

In order for vectors to be useful, we need operations on them, and then we need to come up with laws of physics
using these operations that agree with experiment. We can introduce now the standard operations of linear algebra.
These are addition and scalar multiplication: in the case that we want the vector

w = au + bv (8)

(with a, b ∈ R and u, v ∈ V) the corresponding components are

wα = auα + bv α . (9)

D. Changes of basis

[Not quite covered at this point in the book, but I think it’s a useful reminder.]
Sometimes one wants to convert the components of a vector from one (“unprimed”) basis to another (“primed”)
basis. As you learned in linear algebra, the way to do this is to write the unprimed basis vectors in terms of the
primed basis vectors. If the primed vectors form a basis this is always possible – one writes
X
eα′ = Lβ α′ eβ = Lβ α′ eβ , (10)
β

where L is an n × n linear transformation matrix. There must also be an inverse transformation,



eβ = [L−1 ]α β eα′ , (11)

where L−1 is the matrix inverse of L (a change of basis is always invertible). [Homework: Prove that the transformation
matrix back to unprimed coordinates is actually L−1 .]
If v = v α eα , then it follows that
′ ′
v = v β eβ = v β [L−1 ]α β eα′ = v α eα′ , (12)

where we have set


′ ′
v α = [L−1 ]α β v β . (13)

You can see here that the vector components transform according to the inverse of the transformation matrix of the
basis vectors. Generally the down and up indices are used to describe whether the transformation is of the type of
Eq. (10) or (13).

III. METRIC TENSOR AND DOT PRODUCT

[Reading: §2.4]
In ordinary electrodynamics, we learned that a new operation was useful: the dot product ·. This is an operation
that takes two vectors and returns a scalar. The familiar dot product had the following properties:
• Symmetric: u · v = v · u.
• Distributive: (au + bv) · w = a(u · w) + b(v · w).
One can define the square-norm |u|2 ≡ u · u, and define two vectors to be orthogonal if their dot product is zero.
You may remember an additional property from freshman physics that |u|2 ≥ 0 with equality holding only if u = 0.
Such is not the case in relativity, as we will discuss in more detail below.
If we want to define a dot product satisfying the distributive property, it is sufficient to describe its behavior on
the basis vectors. Let us define gαβ = eα · eβ ; then for general vectors u and v, we find

u · v = eα · eβ uα v β = gαβ uα v β . (14)
4

The symmetry property tells us that we should have gαβ = gβα . Thus in any basis the dot product can be represented
by a symmetric matrix. This matrix changes if we do a transformation of basis – it is:
gα′ β ′ = eα′ · eβ ′ = Lα α′ Lβ β ′ eα · eβ = Lα α′ Lβ β ′ gαβ . (15)
The nature of this transformation is the reason why we write gαβ with down indices.
Another common notation for the dot product is to write it as a tensor: a “linear” (i.e. satisfying the distributive
property over addition and scalar multiplication) function that takes in 2 vectors and outputs a scalar. That is, we
could write
g(u, v) ≡ u · v (16)
and consider gαβ to be the components of the object g.
It is a matter of taste whether we write the dot product as a function or with a · symbol: they are different notations
for the same thing.

A. Example: R3

As a first example, let’s consider the metric tensor in Euclidean space R3 with the standard orthonormal basis
consisting of the unit vectors e1 , e2 , and e3 . The dot products of these form the identity matrix
 
1 0 0
gij = δij ≡  0 1 0  . (17)
0 0 1

(The δij is the Kronecker delta symbol and is defined as 1 if i = j and 0 otherwise.)
In this case, the square norm of any vector is
|u|2 = gij ui uj = (u1 )2 + (u2 )2 + (u3 )2 . (18)
So in this case the square norm is positive for all vectors except the zero vector.

B. More general case

I mentioned earlier that it is not always true that square norms have to be positive for nonzero vectors. One can
see this by applying a change of basis: on the homework, you will show that (i) a change of basis can always bring
the metric tensor into the form of a diagonal matrix with diagonal entries of +1, 0, or −1; and (ii) the number of
+1, 0, and −1 diagonal entries cannot be changed. The number of each of these entries is called the signature of the
metric. Metric tensors with 0s in their signatures are pathological – there are nonzero vectors that are orthogonal to
all vectors, a situation that voids most of the theorems we will use – and so we will disallow them in further study.
Thus the important aspect of the signature is how many +1s and how many −1s it contains. Euclidean R3 has
signature + + +.

C. Example: M4

Now we move to the spacetime of special relativity: M4 . The physically useful dot product in this space has
signature − + ++: this means that we can choose a basis such that
−1 0 0 0
 
 0 1 0 0
gαβ = ηαβ ≡  . (19)
0 0 1 0
0 0 0 1
The matrix defines the symbol ηαβ :

 −1 if α = β = 0
ηαβ = 1 if α = β 6= 0 . (20)
 0 if α 6= β
5

The change in sign of one number seems like it would be trivial, but it has far-reaching consequences. To see this,
let’s consider the square norm of a vector u:

|u|2 = gαβ uα uβ = −(u0 )2 + (u1 )2 + (u2 )2 + (u3 )2 . (21)

This is:
p
• Positive if the spatial components of u dominate, i.e. (u1 )2 + (u2 )2 + (u3 )2 > |u0 |. Such a vector is called
spacelike.
p
• Negative if (u1 )2 + (u2 )2 + (u3 )2 < |u0 |. Such a vector is called timelike.
p
• Zero if (u1 )2 + (u2 )2 + (u3 )2 = |u0 |. Such a vector is called null or lightlike.
One can see in special relativity that the (4-dimensional) velocity vector of an object moving slower than the speed of
light is timelike, and that of an object moving faster than the speed of light is spacelike. The infamous cosmic “speed
limit” is encoded into the geometry of spacetime in the form of the definition of the dot product!
[Warning: Some references, including most of the particle physics literature, define the metric with signature
+ − −−. All the physics works out the same, except that there are differing minus signs in many equations. It’s like
SI versus CGS units in E&M ... you’ll have to get used to it. The − + ++ signature is however in almost universal
use in relativistic astrophysics.]

IV. 1-FORMS

[Reading: §2.5]
So far we’ve defined vectors and the dot product. We will need a few more types of objects in order to formulate
the laws of special (and then general) relativity. The most elementary new construct is the 1-form. Some of our
favorite objects, such as the electromagnetic vector potential or wave vector, are naturally thought of as 1-forms. You
are probably familiar with them as vectors; we will clarify the relationship in a few minutes.
A 1-form is a linear operator that takes a vector v and returns a scalar φ. We will sometimes denote it by a tilded
letter, e.g. k̃, and write the operation as:

φ = hk̃, vi. (22)

Linearity means that the 1-form is distributive:

hk̃, au + bvi = ahk̃, ui + bhk̃, vi. (23)

If the vectors live in a vector space V, then the 1-forms are in a vector space called the dual space V ⋆ . The operation
h, i is called contraction.

A. Description

How are we to describe 1-forms? Again, either by geometric pictures or by components. The geometric picture
presented by MTW is of a set of equally spaced parallel wave crests. Given a vector v, the contraction hk̃, vi is the
number of times the vector pierces a crest. This number is signed (piercing a crest the “wrong way” incurs a − sign)
and interpolated (there are fractional crests). Physically, if k̃ is a wavenumber, then hk̃, si represents the number of
wave cycles one traverses in a displacement s.
The component description of a 1-form is straightforward: we know that a 1-form is completely described by its
action on the basis vectors eα . We may therefore write the components:

k̃α = hk̃, eα i. (24)

You should be able to prove that in a change of basis, k̃α′ = Lβ α′ kβ , so it is appropriate to use the lower index. Since
v = v α eα , it is easily seen that

hk̃, vi = hk̃, v α eα i = hk̃, eα iv α = k̃α v α . (25)

Just as we had basis vectors eα , so we can construct basis 1-forms ω α with k̃ = kα ω α .


6

B. Relation to vectors

In a universe where we did not define the dot product (or metric tensor), vectors and 1-forms would be completely
different beasts. But in a universe with a metric tensor there is a natural correspondence between vectors and 1-forms
that is extremely useful in GR. Let’s take any vector k (think of a wave vector) and associate to it the 1-form k̃
defined by

k · v = hk̃, vi. (26)

The components of k̃ are found from Eq. (24):

k̃α = hk̃, eα i = k · eα = k β eβ · eα = gαβ k β . (27)

Recall that we disallowed degenerate metrics, i.e. we required gαβ to form an invertible matrix (it need not be positive
definite). Then the mapping from k to k̃ is one-to-one and onto (establishing this was in fact the reason why we
required a nondegenerate metric). The inverse mapping is

k α = g αβ k̃β , (28)

where we have defined g αβ to be the matrix inverse of gαβ : g αβ gβγ = δγα . On the homework, you will find the
transformation properties of g αβ and see that it makes sense for it to have upper indices. You will also prove that:

g αβ k̃α l̃α = k · l. (29)

The mapping between vectors and 1-forms is so fundamental that we will usually drop the tilde and use k to
describe both. When doing actual computations, the location of the index – up or down – will tell us whether a vector
or 1-form is used in evaluating an expression.
Since in GR the concept of a dot product exists, why are we even bothering with 1-forms? The reason is two-fold.
First, some calculations that don’t actually use the dot product are most easily carried out with 1-forms. Second,
while many of the more complicated manipulations we will do in curved spacetime are formally possible with just
vectors and tensors, the 1-form will dramatically simplify our lives.

You might also like