Walker - Intro To Abstract Algebra
Walker - Intro To Abstract Algebra
Walker - Intro To Abstract Algebra
TO ABSTRACT
ALGEBRA
Elbert A. Walker
New Mexico State University
Las Cruces, New Mexico, USA
c 1998 by Elbert A. Walker, New Mexico State University.
All rights reserved. No part of this document may be reproduced, stored in a re-
trieval system, or transcribed, in any form or by any means— electronic, mechanical,
photocopying, recording, or otherwise— without the prior written permission of the
author.
Revision of: Introduction to Abstract Algebra by Elbert A. Walker; c 1987 by Random
House, Inc.
This document was produced by Scienti…c WorkPlace TM .
Contents
1 Sets 1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Sets and Subsets . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Mappings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.4 Families . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.5 Equivalence Relations . . . . . . . . . . . . . . . . . . . . . . . . 15
1.6 The Integers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2 Groups 25
2.1 De…nitions and Examples . . . . . . . . . . . . . . . . . . . . . . 25
2.2 Subgroups and Cosets . . . . . . . . . . . . . . . . . . . . . . . . 33
2.3 Homomorphisms . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
2.4 Permutation Groups . . . . . . . . . . . . . . . . . . . . . . . . . 52
2.5 The Groups Sn . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
2.6 Direct Products and Semi-direct Products . . . . . . . . . . . . . 62
2.7 Finite Abelian Groups . . . . . . . . . . . . . . . . . . . . . . . . 69
3 Vector Spaces 79
3.1 De…nitions and Examples . . . . . . . . . . . . . . . . . . . . . . 79
3.2 Homomorphisms of Vector Spaces . . . . . . . . . . . . . . . . . . 88
3.3 Linear Independence and Bases . . . . . . . . . . . . . . . . . . . 94
3.4 Dual Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
iii
iv CONTENTS
6 Fields 237
6.1 Sub…elds and Extension Fields . . . . . . . . . . . . . . . . . . . 237
6.2 Splitting Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . 243
6.3 Galois Theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
6.4 Solvability by Radicals . . . . . . . . . . . . . . . . . . . . . . . . 261
Bibliography 311
Preface
In teaching a beginning course in abstract algebra, one must suppress the urge
to cover a lot of material and to be as general as possible. The di¢ culties in
teaching such a course are pedagogical, not mathematical. The subject matter
is abstract, yet it must be kept meaningful for students meeting abstractness for
perhaps the …rst time. It is better for a student to know what a theorem says
than to be able to quote it, produce a proof of it detail by detail, and not have
the faintest notion of what it’s all about. However, careful attention must be
paid to rigor, and sloppy thinking and incoherent writing cannot be tolerated.
But rigor should be ‡avored with understanding. Understanding the content of
a theorem is as important as being able to prove it. I have tried to keep these
things in mind while writing this book.
The speci…c subject matter chosen here is standard, and the arrangement of
topics is not particularly bizarre. In an algebra course, I believe one should get
on with algebra as soon as possible. This is why I have kept Chapter 1 to a bare
minimum. I didn’t want to include it at all, but the material there is absolutely
essential for Chapter 2, and the students’knowledge of it is apt to be a bit hazy.
Other bits of “set theory” will be expounded upon as the need arises. Zorn’s
Lemma and some of its applications are discussed in the Appendix.
Groups are chosen as the …rst algebraic system to study for several reasons.
The notion is needed in subsequent topics. The axioms for a group are simpler
than those for the other systems to be studied. Such basic notions as homomor-
phisms and quotient systems appear in their simplest and purest forms. The
student can readily recognize many old friends as abstract groups in concrete
disguise.
The …rst topic slated for a thorough study is that of vector spaces. It is
this material that is most universally useful, and it is important to present
it as soon as is practical. In fact, the arrangement of Chapters 3 through 5
is a compromise between mathematical e¢ ciency and getting the essentials of
linear algebra done. Chapter 4 is where it is because its results are beautifully
applicable in Chapter 5 and contain theorems one would do anyway. Besides,
there are some nice applications of Chapter 3 in Chapter 4. Chapter 6 is basic
and should not be slighted in favor of 7 or 8. A feature of Chapter 7 is an
algebraic proof of the fundamental theorem of algebra.
There are many exercises in this book. Mathematics is not a spectator sport,
and is best learned by doing. The exercises are provided so that students can
test their knowledge of the material in the body of the text, practice concocting
proofs on their own, and pick up a few additional facts.
vi CONTENTS
There is no need here to extol the virtues of the abstract approach, and
the importance of algebra to the various areas of mathematics. They are well
known to the professional, and will become fully appreciated by the student
only through experience.
Elbert A. Walker
Las Cruces, New Mexico
March, 1986
Chapter 1
Sets
1.1 Introduction
The purpose of this book is to present some of the basic properties of groups,
rings, …elds, and vector spaces. However, some preliminaries are necessary.
There are some facts about sets, mappings, equivalence relations, and the like
that will be used throughout the text, and which are indispensable. This chapter
presents those facts. It is not meant to be an introduction to “set theory.” The
amount of material is kept to a minimum — just what is needed to get started
on algebra.
A few fundamental facts concerning the integers will be assumed. These are
spelled out in Section 1.6.
1
2 CHAPTER 1. SETS
one way to de…ne a set is to list its elements and put braces around the list.
If S is a big set, this procedure can be cumbersome. For example, it would be
tiresome to describe the set of positive integers less that 1,000,000 in this way.
The sentence we have just written would be more e¢ cient. However, there is
a convention that is universally used that is most convenient. Suppose S is a
set, and that it is the set of all elements s which satisfy some property P . Then
S = fs : s satis…es property P g means that S is the set of all elements s which
enjoy property P . For example, S = fn : n is an integer, 0 < n < 1; 000; 000g
is the set of all positive integers less than 1; 000; 000. Here, property P is the
property of being a positive integer less than 1; 000; 000. If we already knew that
Z was the set of integers, we could write S = fn : n 2 Z, 0 < n < 1; 000; 000g.
This is sometimes written S = fn 2 Z : 0 < n < 1; 000; 000g. The set of even
integers could be written in any of the following ways:
a. fn 2 Z : n is eveng,
b. fn : n is an even integerg,
c. f2n : n 2 Zg, or
The letter Z will be used throughout this book to denote the set of integers.
Suppose S and T are sets. If every element of S is an element of T , then
S is a subset of T . This is denoted S T , or T S. Note that S T does
not rule out the possibility that S = T . The symbol S $ T means S T and
S 6= T , but we will not have much occasion to use this latter notation. If S $ T ,
then S is a proper subset of T . Note that S = T implies both S T and
T S. Also, if S T and T S, then S = T . Thus, S = T if and only if both
S T and T S.
It is convenient to allow the possibility that a set have no elements. Such a
set is called the empty set. There is only one such, since, if S and T are both
empty, then S T and T S, so S = T . The empty set is denoted ;. It has
the property that ; S for any set S.
There are various ways to make new sets from old. For example, given sets
S and T , there are various ways to associate with them a third set. The more
common ones are listed below.
For example, if R denotes the set of real numbers, S = fx 2 R : 0 < x < 2g,
and T = fx 2 R : 1 x 3g, then
S [ T = fx 2 R : 0 < x 3g,
S \ T = fx 2 R : 1 x < 2g,
and
S n T = fx 2 R : 0 < x < 1g:
Let’s note some properties of [, \, and n. First, keep in mind what they
mean. S [ T is that set consisting of all the elements of S along with all the
elements of T ; S \ T is that set consisting of the elements that are in both S
and T ;. and SnT is that set consisting of the elements that are in S but not in
T.
Let A, B, and C be sets. The following are immediate from the de…nitions.
a. A [ A = A; A \ A = A; AnA = ;.
b. A [ B = B [ A; A \ B = B \ A.
c. (A [ B) [ C = A [ (B [ C); (A \ B) \ C = A \ (B \ C).
d. A [ ; = A; A \ ; = ;.
e. A \ (B [ C) = (A \ B) [ (A \ C).
f. A [ (B \ C) = (A [ B) \ (A [ C).
and
(B [ C)0 = B 0 \ C 0 :
Let’s back up and consider (e). It asserts that
A \ (B [ C) = (A \ B) [ (A \ C),
with nothing said about the sets A, B, and C. That is, this equality holds for
all sets A, B, and C. That is what we proved. In our proof, nothing was used
about A, B, and C except that they were sets. No matter what sets we take for
A, B, and C, the equality (e) holds. For example, let B, C, and D be any sets,
and let A = B [ D, so that A \ (B [ C) = (B [ D) \ (B [ C). From (e) we get
the equality
(B [ D) \ (B [ C) = B [ ((B [ D) \ C),
PROBLEMS
1. List all the subsets of the set fag.
2. List all the subsets of the set fa; bg.
3. List all the subsets of the set fa; b; cg.
4. List all the subsets of the set ;.
5. List all the subsets of the set f;g.
6. List all the subsets of the set fa; b; fa; bgg
7. List the elements in fn 2 Z : mn = 100 for some m 2 Zg.
8. List the elements in fn 2 Z : n2 n < 211g.
9. Prove directly that A [ (B \ C) = (A [ B) \ (A [ C).
10. Prove that A [ ((A [ B) \ C) = A [ (B \ C).
11. Prove that (AnB) \ (BnA) = ;.
12. Prove that AnB = ; if and only if A B.
1.3. MAPPINGS 5
1.3 Mappings
The most important concept in mathematics is that of a mapping, or function.
We will use the two words interchangeably, with preference going to the former.
Suppose that A and B are sets, and that with each element a 2 A is associated
a unique element f (a) 2 B. Then we say that f is a mapping, or function, from
A into B. We wish to be a little more precise, but that is the idea of a mapping.
Let A and B be sets. We can form the set of all ordered pairs (a; b), where
a 2 A and b 2 B. It is just the set of all pairs whose …rst member is an element
of A and whose second member is an element of B. This is a fundamental
construction.
A B = f(a; b) : a 2 A, b 2 Bg
A B = f(1; 1); (1; 4); (2; 1); (2; 4); (3; 1); (3; 4)g.
This de…nition is just a careful way of saying that with each element of A is
associated exactly one element of B. We say that ”f is a mapping from A into
B,” that is, we refer to the mapping as just f , although A and B are integral
parts of the mapping.
Suppose f is a mapping from A into B. The set A is the domain of f ,
and B is the co-domain, or the range of f . The set f is the graph of the
function. Thus two mappings are equal if they have the same domain, the
same co-domain, and the same graph.
If (a; b) 2 f , then we think of f as having associated the element b 2 B with
the element a 2 A, or as having taken the element a onto the element b. In fact,
if (a; b) 2 f , we say that “f takes a onto b.” Instead of writing (a; b) 2 f , we
will write f (a) = b. If f (a) = b, then the element b is the image of the element
a under the mapping f . The set ff (a) : a 2 Ag is the image of the mapping
f , and is denoted Im f . If S is any subset of A, then f (S) = ff (a) : a 2 Sg.
Thus f (A) = Im f . If Im f = B, then f is onto, or surjective. If f (x) = f (y)
implies x = y, then f is one-to-one, or injective. If f is both one-to-one and
onto, then it is bijective, or a one-to-one correspondence between A and
B, or an equivalence between A and B.
To de…ne a function we must specify two sets A and B, and a subset f of
A B which satis…es conditions (a) and (b) of 1.3.2. Thus, having A and B,
we must tell what f (a) is for each a 2 A and make sure that this association of
f (a) with a satis…es (a) and (b) of 1.3.2. This means that with each a 2 A we
must have exactly one f (a).
A common situation is for f to be de…ned by a formula. To illustrate this,
suppose that A and B both are the set of real numbers R, and that we are
only considering functions from R to R. “The function f (x) = x2 ” means that
f is the function from R into R consisting of the set f(x; x2 ) : x 2 Rg. That
is, f associates with each real number x the real number x2 . This is clearly a
function. In this vein, g(x) = x 1, h(x) = 69, and k(x) = x=2 determine
functions g, h, and k from R into R, but we will freely use phrases like “the
function g(x) = x 1.”
1.3. MAPPINGS 7
f : A ! B : a ! f (a):
g : R!R:x!x 1,
h : R ! R : x ! 69,
and
k:R!R:x! x=2,
respectively. Thus, f : A ! B : a ! f (a) means that f is the function from A
to B that takes a onto the element f (a). Let’s consider some examples.
Example 1.3.3
g. Let a and b be real numbers with a < b, and let I be the set of functions
from the interval [a; b] to R which are integrable on [a; b]. Then
Z Z b
:I!R:f ! f (x)dx
a
h. Let D be the set of functions from [a; b] to R which are di¤ erentiable at all
points x in [a; b], and let Map([a; b], R) be the set of all functions from
[a; b] to R. Then
C ! C : z = a + bi ! z = a bi
a. (h g) f = h (g f ).
e. f 1A = 1B f = f.
1.3. MAPPINGS 9
Proof. Consider (a). The two functions clearly have the same domain and
range, so it remains to show that for each a 2 A, ((h g) f )(a) = (h (g f ))(a).
We have
((h g) f )(a) = (h g)(f (a)) = h(g(f (a))),
and
(h (g f ))(a) = h((g f )(a)) = h(g(f (a))).
Thus (a) is proved. To prove (b), we need that if c 2 C, then there is some
a 2 A such that (g f )(a) = c. Suppose c 2 C. Then since g is onto, there is
some b 2 B such that g(b) = c. Since F is onto, there is some a 2 A such that
f (a) = b. Thus g(f (a)) = (g f )(a) = c, and g f is onto.
To prove (c), we must show that if (g f )(x) = (g f )(y) then x = y,
where we know that both g and f are one-to-one. If (g f )(x) = (g f )(y), then
g(f (x)) = g(f (y)), and since g is one-to-one, f (x) = f (y). Since f is one-to-one,
x = y.
Part (d) follows from (b) and (c). Part (e) is easy.
Additional properties of composition of functions are in Problems 8-14.
Suppose that f : A ! B is an equivalence. That is, f is one-to-one and
onto. Since f is onto, for b 2 B there is an a 2 A such that f (a) = b. Since
f is one-to-one, there is at most one such a. Hence, associating a with b gives
a mapping g : B ! A. Notice that (g f )(a) = g(f (a)) = a, so g f = 1A .
Also, (f g)(b) = b, so f g = 1B . The function g is denoted f 1 , and called
the inverse of f . Hence with every equivalence f : A ! B, we have associated
another function f 1 : B ! A, and they are related by f f 1 = 1B and
f 1 f = 1A .
Conversely, suppose that f : A ! B and g : B ! A, with g f = 1A and
f g = 1B . Then for b 2 B;
PROBLEMS
1. Let A = f1:2g and B = f3; 4; 5g. Write down all the mappings from A to
B. (Just write down the graphs.)
2. How many functions are there from a set with m elements into a set with
n elements? Remember that sets can be empty.
6. Let A be the open interval (0; 1) = fx 2 R : 0 < x < 1g, and let R+
be the set of all positive real numbers. Prove that the function given by
f : A ! R+ : x ! (1 x)=x is an equivalence, and write down a formula
for its inverse.
15. Let A be a …nite set, and let Perm (A) be the set of all permutations of
A. For f in Perm (A) and n a positive integer, let f n be f composed with
itself n times.
(a) Prove that for f in Perm (A), there exist distinct positive integers m
and n such that f m = f n .
(b) Prove that there exists a positive integer n such that f n = 1A .
(c) Prove that there exists a positive integer n such that f n = 1A for
all f in Perm (A).
A B C = f(a; b; c) : a 2 A, b 2 B, c 2 Cg:
f : Map(A; B) ! Map(A; C) : g ! f g
f : Map(C; A) ! Map(B; A) : g ! g f
1.4 Families
This section is devoted to some notation that will be useful throughout. Let A
and B be sets. A family of elements of B indexed by A is a mapping f : A ! B.
There is no new concept here. A family is a mapping. A notation for such a
family is fba ga2A , where f (a) = ba . Note that bx can be by with x 6= y. Indeed,
if f is not one-to-one, this is the case. We write such sentences as “Let fba ga2A
be a family of elements of B indexed by A.” . The set A is called the indexing
set. The ba ’s are called the members of the family. If the mapping is onto,
then B is indexed by A. A familiar example of a family is a sequence of real
numbers. It is just a family of real numbers indexed by the set Z+ of positive
integers, that is, just a map Z+ ! R. We do not think of sequences as mappings,
but rather as real numbers with positive integers attached. Even though the
concept of family is the same as that of mapping, families are thought of a little
bit di¤erently.
What does the sentence “Let fSi gi2I be a family of sets.” mean? It means
that for each i 2 I there is associated a set Si . That is, fSi gi2I is a mapping
f from I into a set of sets and f (i) = Si . Remember that Si may be Sj with
i 6= j. A family of sets is not a set of sets.
With every family fba ga2A there is associated a set, namely the image of
the mapping. This set is fba : a 2 Ag. With every set B there is associated a
family, namely the family fbb gb2B which is the identity mapping 1B . For most
purposes, there is no di¤erence between the family fbb gb2B and the set B. From
either we can recapture the other. Thus, in a sense, the concept of family is
more general than that of set. (Family was de…ned in terms of set, however.)
The following situation arises fairly often. With each element a 2 A, there
is associated a subset Sa of a set B. That is, we have a map from A into the
set P (B) of subsets of B, or a family of subsets of B indexed by A. We de…ne
the intersection of such a family fSa ga2A to be
\
Sa = fx : x 2 Sa for all a 2 Ag,
a2A
and \
A = fx : x 2 A for all A 2 Sg,
A2S
14 CHAPTER 1. SETS
PROBLEMS
1. Let fBi gi2I be a family of sets. Prove that
!
\ [ [ \
A Bi = A Bi :
i2I i2I
Prove that !
\ X X \
A Bi = A Bi :
i2I i2I
Show that the members of the family fZn gn2Z are not mutually disjoint,
and that the sets in the set fZn : n 2 Zg are mutually disjoint. Determine
how many elements there are in the set fZn : n 2 Zg
5. For each x 2 Q, the set of rational numbers, let Bx = fy 2 Q : x y 2 Zg.
Show that the members of the family fBx gx2Q are not mutually disjoint,
and that the sets in the set fBx : x 2 Qg are mutually disjoint. Prove
that
fx 2 Q : 0 x < 1g ! fBx : x 2 Qg : x ! Bx
is an equivalence.
1.5. EQUIVALENCE RELATIONS 15
Example 1.5.2
Theorem 1.5.7 Let A be a set. Let E( ) be the set of equivalence classes of the
equivalence relation on A. Then ! E( ) is a one-to-one correspondence
between the set of equivalence relations on A and the set of partitions of A.
1.5. EQUIVALENCE RELATIONS 17
PROBLEMS
1. Determine which of the properties of “re‡exive,”“symmetric,”and “tran-
sitive” the following relations on Z satisfy.
(a) m n if m n.
(b) m n if m n is odd or is 0.
(c) m n if m divides n.
(d) m n if jm nj 10.
4. Let two complex numbers be equivalent if their real parts are equal. Prove
that this is an equivalence relation, and that the equivalence classes are
in one-to-one correspondence with the real numbers.
11. Let R be the set of non-zero real numbers, and let R+ be the set of
positive real numbers. De…ne on R by a b if a=b 2 R+ . Prove that
is an equivalence relation and determine the equivalence classes of .
12. Do Problem 11 with the real numbers R replaced by the rational numbers
Q.
13. Determine all partitions and all equivalence relations on the empty set.
16. Let S be a non-empty set, and let Perm(S) be the set of all one-to-one
mappings of S onto S. Let G be a non-empty subset of Perm(S) such that
if f , g are in G then g f and f 1 are in G. De…ne on S by a b if there
is an element g 2 G such that g(a) = b. Prove that is an equivalence
relation on S. If G = Perm(S), what are the equivalence classes of ? If
G = f1S g, what are the equivalence classes of ?
F : A= ! Im f : Cl(a) ! f (a)
is an equivalence.
24. Let S be a partition of a set A, and let T = fAnB : B 2 Sg. Prove that
(a) \C2T C = ;;
(b) if C, D 2 T and C 6= D, then C [ D = A, and
(c) if C 2 T then C 6= A.
25. Prove that the re‡exive, symmetric, and transitive properties of an equiva-
lence relation are independent; that is, prove that no two of the properties
imply the third.
Now nk is the greatest common divisor of n1 and n2 . From the last equation
above, we see that nk divides nk 1 , from the next to last, that nk divides nk 2 ,
and so on, yielding the fact that nk divides both n1 and n2 . Starting with the
…rst equation, if an integer n divides both n1 and n2 , the second equation shows
that n divides n3 . Continuing, we see that n divides nk . Thus nk is the greatest
common divisor of n1 and n2 . Starting with the equation ni = ni+1 qi + ni+2 ,
we see that nk is the greatest common divisor of ni and ni+1 for any i < k.
22 CHAPTER 1. SETS
p1 = q1 , p2 = q2 , . . . , pj = qj
with the pi ’s distinct primes, and that any such representation of n is unique up
to the order of the pi ’s. In particular, the representation is unique if we insist
that p1 < p2 < : . . < pk :
In Chapter 4 we will prove such “unique factorization theorems”for systems
more general than that of the integers Z:
Let p1 , p2 , . . . , pn be primes. Then p1 p2 pn + 1 is divisible by a
prime, but is not divisible by any of the pi . Thus there are an in…nite number
of primes.
PROBLEMS
1. Prove that for all positive integers n;
13 + 23 + + n3 = (1 + 2 + + n)2 .
n i
2. Prove that for all positive integers n, i=1 1=2 < 1.
3. Prove that an < (7=4)n for all terms an of the Fibonacci sequence.
5. Use the Euclidean Algorithm to …nd the greatest common divisors of the
following pairs of integers. Express the greatest common divisors as linear
combinations of the two integers.
9. Let a, b, and c be non-zero integers. Prove that (a; b; c) = ((a; b); c).
How can you use the Euclidean Algorithm to write (a; b; c) as a linear
combination of a, b, and c?
10. Let m and n be non-zero integers. The least common multiple of m
and n is the smallest positive integer that both m and n divide. It is
denoted [m; n]. Let m = a(m; n), and let n = b(m; n). Prove that
11. Let Z+ be the set of all positive integers, and let Zn = fm 2 Z+ : m ng,
and let R be the set of all real numbers. Let P be the set of all positive
integers n such that there exists a unique function fn : R Zn ! R such
that
Prove that P = Z+ .
12. Let Z+ be the set of all positive integers, and let R be the set of all real
numbers. Prove that there is a unique function f : R Z+ ! R such that
Groups
25
26 CHAPTER 2. GROUPS
with binary operations on them which enjoy these three properties arise often
and in many di¤erent places in mathematics (and elsewhere). They are called
groups. They have been, and are still being, studied extensively.
As is the custom, the image of the pair (g; h) under the mapping is denoted
g h instead of (g; h). Condition (a) is expressed by saying that is associative,
condition (b) by saying that has an identity element, and condition (c) by
saying that each element has an inverse.
There is only one identity element e 2 G. Indeed, if e0 were another such
element, then e = e e0 = e0 e = e0 . This unique element is called the identity
element of G. It will typically be denoted by e, or eG if we need to call attention
to G. Note also that each g 2 G has only one inverse h. If h0 were another, then
h = e h = (h0 g) h = h0 (g h) = h0 e = h0 .
This unique element associated with g is denoted g 1 and is called the inverse
of g.
It is important to realize that g h need not be h g. However, if g h = h g, for
all g, h 2 G, then the group is called commutative, or Abelian. A group that
is not commutative is called non-commutative, or non-Abelian. Suppose
that S is a subset of G. Consider the mapping restricted to S S. If maps
S S into S, then it is a binary operation on S, and S together with this binary
operation might be a group. If it is, we say that S is a subgroup of G. Note
that for any group G, G itself is a subgroup of G, and feg is a subgroup of G.
A subgroup of G that is not feg or G is called a proper subgroup of G.
In order to give an example of a group, we must specify two things, a set G
and a binary operation on that set. Then we must make sure that satis…es
(a), (b), and (c) of 2.1.1.
Example 2.1.2 (Groups) a. We have seen two examples already — the set
Z of integers with ordinary addition as the operation, and the set Perm(S)
of all permutations of a set S with composition of mappings as the opera-
tion. If S = f1; 2; 3, . . . ; ng, then Perm(S) is denoted Sn and is called
the symmetric group of degree n.
1 1 i i
1 1 1 i i
1 1 1 i i
i i i 1 1
i i i 1 1
The examples above should convince you that groups occur in many math-
ematical contexts. Since they do occur so frequently, it is worthwhile to study
them “in the abstract." The knowledge gained in such a study can then be
applied in many places, namely wherever a group occurs.
In spite of the fact that a group is a set G together with a binary operation
on G satisfying certain conditions, most often just G itself is referred to as the
group. Thus when we say “G is a group,” we mean that G is a set on which
there is a binary operation satisfying the requirements (a), (b), and (c) of 2.1.1.
There are many binary operations on a set which make that set into a group.
For example, if Gis a group with a binary operation , and f is a one-to-one
mapping of G onto G, then x y = f 1 (f (x) f (y)) de…nes a new binary
operation on G, and G together with is a group. (See Problem 17 below.)
For a, b in a group G, it is the usual custom to write ab instead of a b, and
ab is read “a times b.” We will follow this custom.
Suppose G is a group and a, b, c, d 2 G. Then we can form the prod-
ucts a(b(cd)), ((ab)c)d), (a(bc))d, and a((bc)d). Now using the associative law
((a) in 2.1.1), we can conclude that all these possibilities are equal. For exam-
ple, a(b(cd)) = a((bc)d) = (a(bc))d = ((ab)c)d by repeated application of the
associative law. Thus we may as well write abcd for these products.
The “generalized”associative law holds. That is, suppose a1 , a2 , : : : , an 2 G.
We may form the products
a1 (a2 (a3 ( an )) ),
( ((a1 a2 ) )an 1 )an ,
and
(a1 a2 )(a3 (a4 ( an )) ),
for example, and if n is very large then there are many, many such possibilities.
We can get from any one of these products to another by repeated applications
of 2.1.1 (a). A formal proof may be e¤ected by induction on n, and can be
found in [N. Jacobson, page 39], for example. Any product such as those above
can be written simply a1 a2 an since there is no possible ambiguity.
30 CHAPTER 2. GROUPS
Suppose a 2 G. Then we have the products aa, aaa, aaaa, and so on.
It is reasonable to denote these by a2 , a3 , a4 , and so on. Also a to negative
powers may be de…ned, and certain useful laws of exponents hold. The following
de…nition is an example of a simple de…nition by induction. Remember that a 1
is the inverse of the element a.
Proof. To prove (a), we …rst prove that aam = am+1 for all m 0. To do
this, we induct on m. For m = 0, it is true. If m 0 and aam = am+1 , then
then
1 m+1 1 m 1 m 1
(a ) = (a ) a =a a = (am ) 1
a 1
= (aam ) 1
= (am+1 ) 1
=a (m+1)
:
aam = am+1
2.1. DEFINITIONS AND EXAMPLES 31
for all m < 0. To show this we need to use the fact that
1 k
a a = ak 1
for all k > 0. Since k 1 0, this follows from what we have already proved:
ak 1 = (a 1 a)ak 1 = a 1
(aak 1 ) = a 1 ak . Now if m < 0, then aam =
(a 1 ) 1 (a 1 ) m = (a 1 ) m 1
= am+1 .
Finally, we show that
am an = am+n :
First we show that for all n, it is true for all m 0. If m = 0, it is true. If
am an = am+n for all n, then
and so am an = am+n for all m 0 and for all n. Now suppose that m < 0.
Then am an = (a 1 ) m (a 1 ) n = (a 1 ) m n = am+n , and (a) is proved.
To prove (b), we show …rst that
(am )n = amn
for all n 0. For n = 0, it is true. If n 0 and (am )n = amn , then (am )n+1 =
(am )n am = amn am = amn+m = am(n+1) , whence (am )n = amn for all m 2 Z
and for all n 0. If n < 0, then (am )n = ((am ) n ) 1 = (a mn ) 1 = amn .
To prove (c), note that (ab) 1 = b 1 a 1 . Then for any n 2 Z, we have
(ab)n = ((ab) 1 ) n = (b 1 a 1 ) n .
PROBLEMS
1. Prove that Z is not a group under subtraction.
7. Let Z(n) = Z(n)nf0g. Prove that Z(n) under n is a group if and only
if n is a prime.
14. For elements a, b in a group G, prove that the equation ax = b has exactly
one solution for x in G. Prove a similar fact for the equation xa = b.
15. Prove that the cancellation laws hold for groups. That is, if G is a
group, and a, b, c 2 G, prove that ab = ac implies that b = c, and that
ab = cb implies that a = c.
20. Let G be the non-zero real numbers, and for x and y in G, de…ne x y =
jxjy. Prove that this is an associative binary operation on G, there is a
left identity, elements have right inverses, but that G is not a group under
.
21. Let G and H be groups. Prove that G H with multiplication given
by (a1 ; b1 )(a2 ; b2 ) = (a1 a2 ; b1 b2 ) is a group. (G H is called the direct
product of G and H. Direct products are studied in 2.6.)
22. Give an example of a group with 4 elements. Give an example of a group
with 5 elements. Give an example of a group with 51000 elements.
2.2. SUBGROUPS AND COSETS 33
Theorem 2.2.2 Let fSi gi2I be a family of subgroups of the group G, with I 6=
;. Then \i2I Si is a subgroup of G.
hfg1 , g2 , , gn gi
De…nition 2.2.7 Let G be a group and g 2 G. If hgi is …nite, then the order
of g is the number of elements in hgi. If hgi is in…nite, then the order of g
is in…nite. The order of g is denoted o(g), and if o(g) is in…nite, we write
o(g) = 1 . The number of elements in a …nite group G is called the order of
G, and denoted o(G). If G is in…nite, we write o(G) = 1 .
Theorem 2.2.8 Let G be a group and g 2 G. Then o(g) is …nite if and only
if there is a positive integer n such that g n = e. In the case that o(g) is …nite,
o(g) is the smallest such n.
Example 2.2.9 Let D be the set of all 2 2 matrices with entries from the
complex numbers C, and of the form
a b
b a
where x denotes the complex conjugate of x. It is easily checked that the product
of two such matrices has the same form. The determinant aa+bb is not 0 unless
36 CHAPTER 2. GROUPS
both a and b are 0. Thus a non-zero matrix of this form has an inverse, and
that inverse turns out to be of the same form. It follows that the set D of all
such non-zero matrices is a group under matrix multiplication. Note also that
if a matrix
a b
M=
c d
is in D , then so is
a b
M=
c d
The matrices
0 1 0 i
I= and J =
1 0 i 0
are in D . What is the subgroup hI; Ji of D generated by I and J? By 2.2.4,
it is the set of all possible products of powers of I and J. Denoting
1 0 i 0
and
0 1 0 i
The elements I, J, and K are all of order four, while 1 is of order one
and 1 is of order two. In the dihedral group D4 (2:1:2(l));, which is also of
order eight, there are only two elements of order four, so that these two groups,
Q8 and D4 , both non-Abelian of order eight, are not alike.
Suppose that hgi is …nite. If S is a subgroup of hgi, then S = hg m i, where m
is the smallest positive integer such that g m 2 S. (See 2.2.6.) If o(g) = n, then
n = mk + r, with 0 r < m, and g n mk = g r 2 S. Thus r = 0, and hence m
divides n. We have then the fact that if hgi is …nite, then the number of elements
in any subgroup of hgi divides the number of elements in hgi. However, this is
a very special case of the more general fact that the number of elements in a
subgroup of a …nite group divides the number of elements in that …nite group.
We begin now the development of this fact.
Of course, if G is Abelian, then the left coset gS is the same as the right coset
Sg. Distinct elements of G may determine the same left coset of a subgroup S.
For example, if g 2 S and g 6= e, then gS = S = eS.
2.2. SUBGROUPS AND COSETS 37
For the group Z of integers under +, the subgroup of all multiples of 3 has
just the three cosets
f:::; 9; 6; 3; 0; 3; 6; 9; :::g,
f:::; 8; 5; 2; 1; 4; 7; 10; :::g,
f:::; 7; 4; 1; 2; 5; 8; 11; :::g.
These cosets are disjoint from each other, and their union is all of Z. This is a
fundamental property of left (or right) cosets of any subgroup of any group.
Theorem 2.2.11 Let S be a subgroup of G. Then the set of left (or right)
cosets of S is a partition of G.
Corollary 2.2.14 If o(G) is prime, then G is cyclic, and the only subgroups of
G are feg and G.
1 1
For any subset S of a group G, let S = fs : s 2 Sg.
Theorem 2.2.15 Let L be the set of all left cosets of a subgroup S of a group G,
and R the set of all right cosets of S. Then L ! R : C ! C 1 is a one-to-one
correspondence.
38 CHAPTER 2. GROUPS
G : K = (G : H)(H : K):
Theorem 2.2.17 Let H be a subgroup of G, and let fgi gi2I be a set of represen-
tatives of the left cosets of H in G. Let K be a subgroup of H, and let fhj gj2J be
a set of representatives of the left cosets of K in H. Then fgi hj : (i; j) 2 I Jg
is a set of representatives of the left cosets of K in G.
a 0
diag(a) =
0 a
with positive a gives distinct left cosets of SL2 (R). Simply note that every
element in diag(a)SL2 (R) has determinant a2 .
PROBLEMS
1. Write down all subgroups of the group f1; 1; i; ig of fourth roots of 1.
6. In the dihedral group D4 , write down the left cosets and the right cosets
of the subgroup fE; D1 g.
7. In the quaternion group Q8 , write down the left cosets and the right cosets
of the subgroup f1; 1g.
11. Find a set of representatives of the left cosets of SL2 (R) in GL2 (R). Find
such a set for the right cosets.
(a) g n = e, and
(b) g m = e implies that n divides m.
30. Give an example of an in…nite group such that every subgroup 6= feg is of
…nite index.
31. Give an example of a group G which has subgroups 6= feg of in…nite index
and subgroups 6= G of …nite index.
33. Let S be a …nite set, and let G be the group of all permutations of S. Let
s 2 S, and let H = ff 2 G : f (s) = sg. What is G : H? If S is in…nite, is
G : H in…nite?
34. Let G be the group of all permutations of a …nite set S, and let T S.
Let H = f' 2 G : '(t) = t for all t 2 T g, and let K = f' 2 G : '(t) 2 T
for all t 2 T g. What is G : H? What is G : K? What is K : H?
2.3 Homomorphisms
A central idea in algebra is that of homomorphism. Here is a simple example.
Consider the group G = f1; 1g, where the operation is ordinary multiplication,
and the group S2 , of all permutations of the set A = f1; 2g. The group S2 has
two elements, 1A and the mapping f that interchanges 1 and 2. Intuitively,
these two groups are just alike, even though they have no elements in common.
The element 1A in S2 behaves just like the element 1 does in G, and the element
f behaves in S2 just like the element -1 does in G. Here there is actually a one-
to-one correspondence between the two groups, but homomorphisms allow more
general circumstances. Consider the group Z of integers under addition, and the
mapping f : Z ! G given by
a. f (eG ) = eH .
1 1
b. f (x ) = (f (x)) , for all x 2 G.
c. If A is a subgroup of G, then f (A) is a subgroup of H. In particular,
f (G) = Im f is a subgroup of H.
1
d. If B is a subgroup of H, then f (B) is a subgroup of G.
e. f (x1 x2 :::xn ) = f (x1 )f (x2 ):::f (xn ) for xi 2 G.
f. f (xn ) = f (x)n for all n 2 Z and all x 2 G.
g. If g : H ! K is a homomorphism, then g f : G ! K is a homomorphism.
1
h. If f : G ! H is an isomorphism, then f : H ! G is an isomorphism.
Proof.
1
a. f (eG ) = f (eG eG ) = f (eG )f (eG ), and multiplying through by f (eG )
yields eH = f (eG ).
1 1 1 1
b. eH = f (eG ) = f (xx ) = f (x)f (x ), so f (x ) = (f (x)) .
c. For f (x), f (y) 2 Im f , f (x)f (y) = f (xy) 2 Im f , so Im f is closed under
multiplication. Since f (x) 1 = f (x 1 ), Im f is a subgroup of H.
d. Let x, y 2 f 1 (B). Then f (xy) = f (x)f (y) 2 A, and f (x 1 ) = f (x) 1
2
A since f (x) and f (y) are in A. Thus f 1 (A) is a subgroup of G.
e. This follows readily by induction on n.
f. This holds for n > 0 by (e). For n < 0, f (xn ) = ((f (xn ) 1
)) 1
=
f ((xn ) 1 ) 1 = (f (x n )) 1 = ((f (x)) n ) 1 = f (x)n .
g. Let x, y 2 G. Then (g f )(xy) = g(f (xy)) = g(f (x)f (y)) = g(f (x))g(f (y)) =
(g f )(x)(g f )(y).
h. Let a; b 2 H. We need only that f 1 (ab) = f 1 (a)f 1 (b). For suitable
x, y 2 G, a = f (x) and b = f (y). Thus f 1 (a) = f 1 (f (x)f (y)) =
f 1 (f (xy)) = xy = f 1 (a)f 1 (b).
2.3. HOMOMORPHISMS 43
Several conditions which are equivalent to being normal are given in Problem
21.
1
De…nition 2.3.5 Let f : G ! H be a homomorphism. Then f (eH ) is de-
noted Ker f and called the kernel of f .
Notice that for any group G, feg and G itself are normal subgroups of G.
If G is Abelian, then any subgroup of G is normal. Keep in mind that one
way to prove that a subgroup is normal is to get it to be the kernel of some
homomorphism. For example, the kernel of the homomorphism in 2.3.3 (f) is the
special linear group SL2 (R). In particular then, SL2 (R) is a normal subgroup
of GL2 (R).
A useful fact is that any subgroup of index two is normal. If G : H = 2,
then the only left or right coset of H besides H itself is the complement of H in
G. In the group Q8 of quaternions, the subgroups hii, hji, and hki are distinct
subgroups of index 2, having order four in a group with 8 elements. Thus these
subgroups are all normal. It happens that all of the subgroups of Q8 are normal
(Problem 13).
In the dihedral group D4 , the subgroup N = fE; R1 ; R2; R3 g is of index
two, hence normal, and so is the subgroup M = fE; V; H; R2 g. The subgroup
fE; R2 g of N is normal in N , and indeed normal in D4 , while the subgroup
fE; Hg is normal in M , but it is not normal in D4 since D1 fE; Hg = fD1 ; R3 g,
while fE; HgD1 = fD1 ; R1 g.
Let N be a normal subgroup of a group G, and let G=N denote the set of
all left cosets of N in G. Since xN = N x, G=N is also the set of all right
cosets of N . We will omit “left”or “right”when talking about cosets of normal
subgroups. We wish to make G=N into a group. In the case where N = Ker f
for some homomorphism f , we know that we can do it by de…ning
xN yN = xyN .
and
1 1 1 1
(xN ) x N = xx N = eN = x xN = x N (xN ) ,
so that G=N is a group. Now notice that G ! G=N : x ! xN is an epimorphism
with kernel N .
G ! G=N : x ! xN
is an isomorphism.
f is an isomorphism.
Thus given any homomorphism f : G ! H, it can be decomposed into a
composition
G ! G= Ker f ! Im f ! H
2.3. HOMOMORPHISMS 47
of three homomorphisms, the …rst the natural homomorphism, the second the
isomorphism in 2.3.8, and the third an inclusion.
Let N be a normal subgroup of G. We need to know what the subgroups of
G=N are.
(Z=Zmn)=(Zm=Zmn) Z=Zm.
De…nition 2.3.12 Let n 0. The group Z=Zn is called the group of integers
modulo n, and will also be denoted Z(n).
We have already called another group the integers modulo n, denoted the
same, namely the group 2.1.2 (d). The two groups are isomorphic (Problem 2 ).
Let G = hgi. That is, G is a cyclic group. The mapping f : Z ! G : n ! g n
is a homomorphism, in fact an epimorphism. Just note that f (m+n) = g m+n =
g m g n = f (m)f (n). If o(g) = m, then g m = e and Zm = Ker f . If o(g) = 1 ,
then Ker f = 0 and f is an isomorphism. In any case, Z= Ker f G and Ker
f = Zm for some m 0. We thus have the following.
Corollary 2.3.14 Two cyclic groups are isomorphic if and only if they have
the same order.
One way to express 2.3.13 is to say that the cyclic groups, up to isomorphism,
are just the groups Z=Zn, n 0. Notice that for n > 0, Z=Zn has exactly n
elements.
Let’s see what 2.3.11 says in a familiar case. Consider two subgroups Zm
and Zn of Z with m and n positive. We know that Zm \ Zn and Zm + Zn
are subgroups of Z. The subgroup Zm \ Zn is the set of common multiples
of m and n, and hence is generated by the least common multiple [m; n] of
m and n. (See 1.6, Problem 10.) Thus Zm \ Zn = Z[m; n]. The subgroup
Zm + Zn = fam + bn : a; b 2 Zg is generated by the smallest positive integer
in it. But the smallest positive integer of the form am + bn is the greatest
common divisor (m; n) of m and n. Thus Zm + Zn = Z(m; n). Applying 2.3.11,
we get
Zn=Z[m; n] Z(m; n)=Zm.
Now suppose that a, b 2 Z+ , and a divides b. That is, suppose that 0 6=
Zb Za. Then o(Za=Zb) = b=a. Applying this to Zn=Z[m; n] and Z(m; n)=Zm
gets [m; n]=n = m=(m; n), or mn = [m; n](m; n).
Let Aut(G) be the set of all automorphisms of a group G. Then Aut(G) is
itself a group under composition of mappings. In fact, Aut(G) is a subset of
Perm(G), the set of all permutations of G, (2.1.2 (a)), and so 2.3.2 (g) and 2.3.2
(h) imply that Aut(G) is a subgroup of Perm(G).
2.3. HOMOMORPHISMS 49
Theorem 2.3.15 Let G be a group, and let '(g) be the inner automorphism of
G induced by the element g. Then
G ! Aut(G) : g ! '(g)
PROBLEMS
1. Let G be the group R R under the operation (a; b)(c; d) = (a + bc; bd).
Prove that f : G ! G : (a; b) ! (0; b) is an endomorphism with f 2 = f .
Is G ! G : (a; b) ! (a; 1) an endomorphism?
2. Prove that Z(n) as de…ned in 2.1.2 (d) and in 2.3.12 are isomorphic.
50 CHAPTER 2. GROUPS
10. Let G be the additive group of all complex numbers of the form m + ni,
where m and n are integers. Let H be the multiplicative group of all
rational numbers of the form 2m 3n , again where m and n are integers.
Prove that G H.
12. Prove that there are exactly two groups of order four, up to isomorphism.
Which one is Q8 =f1; 1g?
13. Prove that every subgroup of the quaternion group Q8 is normal. Find all
the normal subgroups of D4 .
14. Prove that the set of elements of GL2 (R) of the form
a 0
c d
15. Prove that SL2 (R) is a normal subgroup of GL2 (R), and that GL2 (R)=SL2 (R)
is isomorphic to the multiplicative group R of non-zero real numbers.
(a) N is normal in G.
1
(b) gN g = N for all g 2 G.
1
(c) gN g N for all g 2 G.
1
(d) gN g N for all g 2 G.
(e) (xy)N = (xN )(yN ).
(f) N g gN for all g 2 G.
(g) gN N g for all g 2 G.
22. Prove that if G has exactly one subgroup of order 50, then that subgroup
is normal. Generalize.
26. Let G be an Abelian group. Let T be the set of elements of G which have
…nite order. Prove that T is a subgroup of G, and that in G=T , only the
identity element has …nite order.
27. Prove that the intersection of any family of normal subgroups of a group
G is a normal subgroup of G.
52 CHAPTER 2. GROUPS
subgroup of Perm(S) for some S. Our …rst object is to show that every group
is a permutation group — more precisely, that every group is isomorphic to a
permutation group. This fact is known as Cayley’s Theorem. It has the virtue
of representing an abstract group as something concrete. Every group can be
represented as a group of one-to-one mappings of some set onto itself. In fact,
groups arose in this way. Permutations groups were studied before the notion
of an abstract group was formulated.
How is Cayley’s Theorem proved? If G is a group, we must get a monomor-
phism from G into Perm(S) for some S. Where do we get a suitable set S? We
use G. Then with each g 2 G we must associate a permutation of G. How can
we do that? Multiplying each element of G on the left by g is a permutation of
G.
Proof. For each g 2 G, let '(g) be the map from G into G de…ned by
'(g)(x) = gx. That is, '(g) is multiplication on the left by g. If '(g)(x) =
'(g)(y), then gx = gy, so that x = y. Hence '(g) is one-to-one. If y 2 G,
then '(g)(g 1 y) = g(g 1 y) = y, so that '(g) is onto. Thus '(g) 2 Perm(G).
Therefore we have a mapping ' : G ! Perm(G). We will show that ' is a
monomorphism. If '(g) = '(h), then '(g)(x) = '(h)(x) for all x 2 G. In
particular, '(g)(e) = ge = g = '(h)(e) = he = h. Thus ' is one-to-one. For g,
h 2 G;
since x 1 hx 2 N H. Thus h 2 Ker '. Therefore Ker ' contains all the
normal subgroups of G contained in H, and since Ker ' is a normal subgroup
of G contained in H, it is the largest one. In particular, every subgroup H of a
group G has a subgroup N which is normal in G and contains any other such
subgroup. In fact, N is just the subgroup generated by all the subgroups of H
which are normal in G. We have proved the following generalization of Cayley’s
Theorem.
Theorem 2.4.2 Let H be a subgroup of G, and let L(H) be the set of left cosets
of H. Then ' de…ned by '(g)(xH) = gxH is a homomorphism from G into
Perm(L(H)), and Ker ' is the largest normal subgroup of G contained in H.
Note that 2.4.1 follows from 2.4.2 by taking H = feg. In 2.4.2 (and 2.4.1 of
course), each element g 2 G induced a permutation of a set, and associating g
with that permutation was a homomorphism.
The subgroup S = fE; Hg of D4 (2:1:2(l)) is not normal, so that the largest
normal subgroup of D4 in S is fEg. The homomorphism D4 ! Perm(L(S))
given by 2.4.2 then has kernel fEg, so is a monomorphism. Thus we get a
representation of D4 as a group of permutations of the set L(S) with four
elements.
In Chapter 8 we will see how 2.4.2 is useful in proving that certain groups
have non-trivial normal subgroups. Several of the following exercises also make
use of 2.4.2.
PROBLEMS
1. Let a, b 2 R with b 6= 0. De…ne a;b by
a;b : R ! R : x ! bx + a:
Prove that the set of all such a;b is a group of permutations of R. Prove
that this group is isomorphic to the group G = R R with multiplication
given by (a; b)(c; d) = (a + bc; bd).
2. Let G be the group in Problem 1. Let H be the subgroup of elements of G
of the form (0; b). Use 2.4.2 to show that H has no non-trivial subgroups
which are normal in G.
3. Let G be the group in Problem 1. Let N be the subgroup of elements of
the form (a; 1). Prove that N is normal in G.
4. Let G be the group in Problem 1. Prove that G is isomorphic to the
multiplicative group of matrices of the form
b a
0 1
1
5. Let G be a group. Prove that ' : G ! Perm(G) given by '(g)(x) = xg
is a monomorphism. What happens if ' is de…ned by '(g)(x) = xg?
is an equivalence.
12. Let H be a subgroup of G, and let ' be the mapping ' : G ! Perm(L(N (H)))
given by '(g)(xN (H)) = (gx)N (H), as in 2.4.2. Let f be the mapping in
Problem 8, f the isomorphism Perm(L(N (H))) ! Perm(Cl(H)) induced
by f , and the mapping in Problem 9. Prove that f ' = . Here is a
picture.
G
'
. &
f
Perm (L (N (H))) ! Perm(Cl(H))
13. Prove that if H is a proper subgroup of the …nite group G such that
o(G) does not divide (G : H)!, then G has a nontrivial normal subgroup
contained in H.
14. Prove that if p is the smallest prime dividing o(G), then any subgroup of
G of index p is normal.
56 CHAPTER 2. GROUPS
1 2 n
(1) (2) (n)
That is, under each integer k, 1 k n, put its image. Thus the elements of
S3 are the six permutations
1 2 3 1 2 3 1 2 3
, , ,
1 2 3 1 3 2 2 1 3
1 2 3 1 2 3 1 2 3
, ,
2 3 1 3 1 2 3 2 1
For example,
1 2 3
3 1 2
is the mapping such that (1) = 3, (2) = 1, and (3) = 2.
Multiplication in Sn is composition of mappings. We have de…ned composi-
tion of two mappings and by ( )(x) = ( (x)). That is, apply
and then apply . For the rest of this section we will adopt the convention that
means apply and then apply . We will write simply as . Thus
1 2 3 1 2 3 1 2 3
=
2 1 3 3 1 2 1 3 2
1 2 3 4 5
(1; 4; 3; 5; 2) =
4 1 5 3 2
is the product (1; 5)(2; 3)(4; 7; 6). The way to write a permutation as a product
of disjoint cycles is as follows. Let 2 Sn , and k 2 f1; 2; : : : ; ng. There is a
smallest m 1 such that m (k) = k. This yields the cycle
2 m 1
(k; (k); (k); :::; (k)).
Two such cycles are either equal or disjoint. (Keep in mind that (1; 2; 3) =
(2; 3; 1) = (3; 1; 2).) The cycles so obtained are the cycles of . It should be
clear that is the product of its cycles. Thus we have the following theorem.
Disjoint cycles commute. That is, if and are disjoint cycles, then
= . Thus if is a product 1 2 ::: k of disjoint cycles, then =
(1) (2) ::: (k) , where is any permutation of f1; 2; : : : ; kg. If also =
0
1 2 ::: j , with the i s disjoint cycles, then j = k, and (after rearrangement)
1 = 1 , 2 = 2 , : : : , k = k . This becomes clear by noting that each i is a
cycle of . Thus up to the order of the factors, each element of Sn is uniquely
a product of disjoint cycles, these disjoint cycles being simply the cycles of .
Note that
(1; 2; 3; : : : ; n) = (1; 2) (1; 3) (1; 4) (1; n)
or more generally,
That is, any cycle is a product of 2-cycles. The 2-cycles are called transposi-
tions. Therefore, by 2.5.1 we have the following fact.
allows us to move (a; br ) over to the left next to (a; bi ), and since i = r, the
product of these two is the identity. For example,
For example,
1 2 3 4 5 6 7
=
5 3 2 7 1 4 6
is even since = (1; 5)(2; 3)(4; 7; 6) = (1; 5)(2; 3)(4; 7)(4; 6), while = (1; 2) is
odd.
The subset of even permutations of Sn is a subgroup. (See 2.5.6 below.)
Proof. The mapping from Sn into the multiplicative group f1; 1g that
takes 2 Sn onto 1 or 1 depending on whether is even or odd is an epimor-
phism with kernel An .
The groups Sn and An have many interesting properties. One such property
that we will need in a later chapter in our study of …elds is that for n 5, An
is simple. A group is simple if it has no nontrivial normal subgroups.
is also in N .
Let 2 N with 6= e, and write = 1 2 ::: k , with the 0i s disjoint cycles.
Let 1 = (i1 ; i2 ; : : : ; ir ). First suppose that r > 3, and let = (i1 ; i2 ; ir ). Then
2 An , and
1 1 1
= (i1 ; i2 ; ir )(i1 ; i2 ; : : : ; ir ) 2 k (ir ; i2 ; i1 )
1 1
= (i1 ; i3 ; i4 ; i5 ; : : : ; ir 1 ; i2 ; ir ) 2 k k 2 (ir ; : : : ; i2 ; i1 )
= (i1 ; i2 ; ir 1)
is in N .
60 CHAPTER 2. GROUPS
is in N , whence
(a; i3 ; i2 )2 = (a; i2 ; i3 )
is in N .
We have then that if (i1 ; i2 ; i3 ) 2 N , then for any a 6= i1 , i2 , i3 , the element
(a; i2 ; i3 ) 2 N . Suppose now that (a; b; c) is any 3-cycle, and that (i1 ; i2 ; i3 ) 2 N .
If a is not one of i1 , i2 , or i3 , then (a; i2 ; i3 ) 2 N , and
(a; i2 ; i3 ) = (i2 ; i3 ; a)
and that
(i1 ; i2 )(i3 ; i4 ) = (i1 ; i3 ; i4 )(i1 ; i3 ; i2 ),
and remembering that every element of An is a product of an even number
of 2-cycles, we have that every element of An is a product of 3-cycles. Thus
An = N , and our theorem is proved.
PROBLEMS
1. Write the following permutations as a product of disjoint cycles and as a
product of transpositions.
1 2 3 4 5
(a)
5 4 3 2 1
1 2 3 4 5 6 7
(b)
7 5 6 3 2 4 1
1 2 3 4 5 6 7
(c)
1 2 5 4 3 7 6
(d) (1; 2; 5; 3)(1; 2; 4; 5)(6; 3; 4)
(e) (1; 2; 3)(1; 2; 3; 4)(1; 2; 3; 4; 5)
(f) (1; 2; 3; 4; 5; 6; 7; 8)100 .
2. Prove that the order of a product of disjoint cycles is the least common
multiple of the lengths of those cycles.
62 CHAPTER 2. GROUPS
3. Prove that in Sn ,
1
(i1 ; i2 ; :::; ir ) = ( (i1 ); (i2 ); :::; (ir )).
4. Prove that the elements and of Sn are conjugates if and only they
have the same cycle structure, that is, if and only if and have the
same number of disjoint cycles of the same length.
5. View Sn 1 as a subgroup of Sn , that is as the subgroup of Sn …xing n.
Write down a set of representatives of the left cosets of Sn 1 . Write down
a set of representatives of the right cosets of Sn 1 .
6. What are the conjugates of Sn 1 in Sn ? (See Problem 7.)
7. What is Sn 1 \ An ?
8. A4 has non-trivial normal subgroups. Find them all.
9. Prove that if f is any endomorphism of Sn , then f (An ) An .
10. Prove that the commutator subgroup of Sn is An .
11. Show that every automorphism of S3 is inner.
12. Find the automorphism group of S3 .
13. Find all subgroups of Sn of index two.
14. Prove that Sn is generated by the n 1 transpositions (1; 2), (1; 3), ::: ,
(1; n).
15. Prove that Sn is generated by the n 1 transpositions (1; 2), (2; 3), ::: ,
(n 1; n).
16. Prove that S5 is generated by f(1; 2); (1; 2; 3; 4; 5)g.
17. Prove that Sn is generated by two elements.
This makes G H into a group (2.1, Problem 21). The identity element is
(eG ; eH ), and (g; h) 1 = (g 1 ; h 1 ). (From now on, we will write simply e for
2.6. DIRECT PRODUCTS AND SEMI-DIRECT PRODUCTS 63
the identity of any group under consideration. Thus (eG ; eH ) may be written
(e; e) or even more simply just e.) The associative law holds since it holds in
both G and H. This group is denoted G H, and called the external direct
product of G and H. In G H, let
The mappings
G H ! H : (g; h) ! h and
G H ! G : (g; h) ! g
are epimorphisms with kernels G and H , respectively. Therefore G and H
are normal subgroups of G H. For (g; h) 2 G H, (g; h) = (g; e)(e; h), so
that every element in G H can be written as the product of an element from
G with one from H . Further, this can be done in exactly one way since
(g; h) = (g 0 ; e)(e; h0 ) implies that (e; e) = (g 1 g 0 ; h 1 h0 ), so that g = g 0 , and
h = h0 . Thus G H has normal subgroups G and H such that every element
of G H can be written uniquely in the form gh with g 2 G and h 2 H , and
so G H = G H = fgh : g 2 G , h 2 H g. We say that G H is the direct
product of the normal subgroups G and H .
It is worth observing here that G \ H = (e; e), and that gh = hg for
g 2 G and h 2 H . In fact, if M and N are normal subgroups of a group
which intersect in just the identity element, then the elements of M commute
with those of N . (See Problem 19 in section 2.3.)
This discussion carries over to any …nite number of groups. That is, let G1 ,
G2 , : : : , Gn be groups, and let G1 G2 Gn be the group obtained from
the set
by de…ning
G1 G2 Gn = G1 G2 Gn ,
although strictly speaking, the ’s on the left have a di¤erent meaning than the
’s on the right. The appropriate interpretation of can always be determined
from the context. In general, when we write G = G1 G2 Gn , we mean it
in the sense of 2.6.1. The group G1 G2 Gn is also written G = ni=1 Gi .
If G = G1 G2 Gn , then Gi \ Gj = feg for distinct i and j. Since
Gi and Gj are normal, we have that for gi 2 Gi and gj 2 Gj , gi gj = gj gi .
Therefore, if gi , hi 2 Gi , then
G = (G1 G2 ) G3 = G1 (G2 G3 ).
(H K)=H K.
Theorem 2.6.5 Let N and H be groups, and let ' be a homomorphism from
H into the automorphism group Aut(N ) of N . Then
1 1
Proof. The identity is (e; e), and the inverse of (n; h) is (n '(h )(n),
1
h ). The equations
In this case, that is, when H = Aut(N ), the semi-direct product is call the
holomorph of N , and denoted Hol(N ). Elements of a group N produce per-
mutations of the set N by multiplication on the left, say. Automorphisms of N
are special kinds of permutations of N . The subgroup of the group of all per-
mutations of N generated by the left multiplications and the automorphisms of
N is isomorphic to the holomorph of N . In this connection, see Problem 23.
2.6. DIRECT PRODUCTS AND SEMI-DIRECT PRODUCTS 67
Let N be any group, let H be any subgroup of Aut(N ), and let ' be the
homomorphism H ! Aut(N ) : h ! h. The semi-direct product of N and H
over ' is then the Cartesian product G = N H with multiplication given by
PROBLEMS
1. Let Q be the multiplicative group of non-zero rational numbers, and let
Q+ be the subgroup of positive rational numbers. Prove directly from the
de…nition that Q = f1; 1g Q+ . Prove also using 2.6.2.
8. Prove that G is the direct product of the two normal subgroups H and K
if and only if every g 2 G can be written g = hk with h 2 H and k 2 K,
and H \ K = feg.
15. Prove that the group of quaternions Q8 is not a non-trivial split extension.
21. Construct all possible groups of order 30 which are split extensions of a
cyclic group of order 10 by a group of order 3.
23. Let N be any group, and let P be the group of all permutations of the
set N . Identify N with the set of left multiplications in P . Let G be the
subgroup of P generated by N and the automorphisms of N . Prove that
G is a split extension of N by Aut(N ). Prove that G is isomorphic to
Hol(N ).
26. Let fGi gi2I be a family of groups. Let G = i2I Gi be the set of all
mappings f of I into [i2I Gi such that f (i) 2 Gi for all i 2 I. For f ,
g 2 G, let (f g)(i) = f (i)g(i). Prove that G with this multiplication is a
group. let ei be the identity of Gi . Prove that i2I Gi = ff 2 G : f (i) = ei
for all but …nitely many ig is a subgroup of G. (G is the direct product
of the family fGi gi2I , and the subgroup described is the direct sum,
or the coproduct of the family.)
To realize our objective, we will show that every …nite Abelian group is a
direct product of cyclic groups, and in essentially only one way. Once this is
accomplished, a completely satisfying Inv(G) will be at hand.
When dealing with Abelian groups, it is customary to use additive notation.
That is, the group operation will be called “addition,” and we will “add” two
elements rather than “multiply” them. We write g + h instead of g h or
gh, and ng replaces g n . The identity element is denoted 0 instead of 1 or
e. Direct products are called direct sums, and direct factors are called direct
summands. If A1 , A2 , : : :, An are Abelian groups, then their direct sum is
written A1 A2 An , or ni=1 Ai . These are merely notational changes.
For example, 2.6.1 and 2.6.2 become, in additive notation, the following.
For example, Z(2) Z(2) and Z(4) are 2-groups, and Z(27) is a 3-group.
However, Z(6) is not a p-group for any prime p.
G ! G=S ! hg + Si ! hgi ! G,
where the …rst map is the natural homomorphism, the second is the projection
of G=S onto its summand hg + Si, the third takes m(g + S) to mg, and the last
is the inclusion map, is an idempotent endomorphism of G with image hgi. By
2.7.2, hgi is a summand of G.
Theorem 2.7.6 Every …nite Abelian p-group is a direct sum of cyclic groups.
Thus G can be written in several ways as the direct sum of cyclic groups. Even
for a given cyclic subgroup, say H = f(0; 0), (1; 1)g, there were two distinct
subgroups K and L such that G = H K = H L. This simple example shows
that there will be, in general, many ways to write a …nite Abelian group as a
direct sum of cyclic groups. Is there anything worthwhile that is unique about
such a direct sum decomposition? The answer is a resounding yes. Let G be a
…nite Abelian p-group, and let G = G1 G2 Gm = H1 H2 Hn
with each Gi and Hi cyclic and non-zero. We will show that m = n, and after
a possible renumbering, Gi Hi . Another way to say this is that for each
positive integer k, the number of Gi of order pk is the same as the number of
Hi of order pk . That is, this number depends only on G, not on a particular
direct sum decomposition of it. These numbers will be our magic invariants.
G = G1 G2 Gm = H1 H2 Hn
with each Gi and Hi non-zero and cyclic, then m = n, and after suitable renum-
bering, Gi Hi for all i.
kg1 = k(h1 + + hn ) = x2 + + xn
2.7. FINITE ABELIAN GROUPS 73
G = G1 H2 Hn :
So we have
G=G1 G2 Gm H2 Hn :
Now let be an isomorphism from H2 Hn to G2 Gm . We get
G2 Gm = (H2 ) (Hn ). By the induction hypothesis, after
suitable renumbering, Gi (Hi ) for all i 2. But Hi (Hi ) for all i 2,
and so Gi Hi for all i. The theorem follows.
We now have the desired invariants for …nite Abelian groups. If G is a …nite
p-group, then G = G1 Gm with Gi cyclic. For convenience, let’s arrange
the Gi so that
o(G1 ) o(G2 ) o(Gm ).
Now Gi , being cyclic, is determined by o(Gi . Thus, the family fo(G1 ), o(G2 ),
: : :, o(Gm )g determines G. Furthermore, by 2.7.8, it does not depend on the
particular decomposition G1 Gm . That is, fo(G1 ); : : : ; o(Gm )g is an
invariant of G. Denote it I(G). What all this means is that if H is a …nite
Abelian p-group, H = H1 Hn with Hi cyclic and
Now the subgroups in square brackets are cyclic and the order of the …rst is
divisible by the order of the second, the order of the second is divisible by the
order of the third, and so on. These orders are, in fact, 53 33 24 , 52 33 , 52 32 , and
5. Therefore, we have associated the family 53 33 24 ; 52 33 ; 52 32 ; 5 with G. We
can reconstruct G (up to isomorphism) from this family, namely
G = G1 Gm = H1 Hn ,
with the Gi and Hi non-zero cyclic, and with o(Gi ) divisible by o(Gi+1 ) and
o(Hi ) divisible by o(Hi+1 ) for all appropriate i, then m = n and o(Gi ) = o(Hi ),
whence Gi Hi . To see that this is the case, consider a p-component K of
G. Then K = m i=1 (Gi )p . That is, the p-component of G is the sum of the
p-components of the Gi ’s. But (Gi )P is cyclic since Gi is cyclic. Similarly K =
m
i=1 (Hi )p . But o((Gi )p ) o((Gi+1 )p ), and similarly for the (Hi )p . Therefore,
from the discussion of the invariant I(K) for p-groups K, (Gi )p (Hi )p for all
i. Since the p-component of Gi is isomorphic to the p-component of Hi for all
p, it follows that Gi Hi .
The result of all this discussion is this. With each …nite Abelian group
G there is associated, in exactly one way, a family fn1 ; n2 ; ; nk g of positive
integers such that ni is divisible by ni+1 for i < k. The association is made by
writing G = ki=1 Gi with Gi cyclic of order ni , and we have shown that the ni ’s
are unique. If we know fn1 ; n2 ; ; nk g, we can retrieve G, up to isomorphism:
k
G i=1 Z(ni ). So for a …nite Abelian group G we have a complete invariant
I(G), the family fn1 ; n2 ; ; nk g of positive integers. In case G is a p-group, it
is the same invariant as discussed for p-groups. In case G is cyclic, I(G) is the
family whose only member is o(G).
2.7. FINITE ABELIAN GROUPS 75
De…nition 2.7.9 The family I(G) is called the family of invariant factors
of the …nite Abelian group G. The invariant factors of the p-components GP of
G are the elementary divisors of G.
Theorem 2.7.10 Let G and H be …nite Abelian groups. The following are
equivalent.
a. G H.
Now that we have the invariant I(G) for …nite Abelian groups G, we can
calculate how many non-isomorphic Abelian groups there are of a given order.
Let n be any positive integer, and write n = n1 n2 nk , where the ni are
powers of distinct primes pi . If a …nite Abelian group G has order n then its pi -
component must have order ni . Now two …nite Abelian groups are isomorphic
if and only if their corresponding p-components are isomorphic. Therefore, if
N (n) denotes the number of non-isomorphic Abelian groups of order n, then
N (n) = ki=1 N (ni ). So we must calculate N (pn ), where p is an arbitrary prime.
If G is an Abelian p-group of order pn , then
I(G) = fi1 , i2 , , ik g;
PROBLEMS
All groups in these problems are Abelian.
1. Prove that G is the direct sum of the two subgroups H and K if and only
if every g 2 G can be written g = h + k with h 2 H and k 2 K, and
H \ K = f0g.
2. Prove that G is the direct sum of the subgroups G1 , G2 , : : : , Gn if and
only if every g 2 G can be written g = g1 + g2 + + gn with gi 2 Gi for
all i, and Gi \ (G1 + G2 + + Gi 1 + Gi+1 + +Gn ) = f0g for all i.
3. Let G = ni=1 Gi , and g = gi with gi 2 Gi . Prove that o(g) = lcmfo(g1 ),
o(g2 ) , : : : , o(gn )g if o(g) is …nite. Prove that if G is a p-group, then o(g)
is the maximum of the o(gi ).
4. Prove that every …nite group G is a direct sum of p-groups in the following
way. Let p divide o(G), p a prime. Let Gp be the p-component of G, and
let Gp = fg 2 G : ((o(g), p) = 1g. Prove that G = Gp Gp . Complete by
inducting on the number of primes that divide the order of G.
5. Prove that every …nite group is a direct sum of p-groups as follows.
n : G ! G : g ! ng
is an automorphism.
(b) Let o(G) = i1 i2 ir , with the ij powers of distinct primes pj . Let
q = o(G)=i1 . Then q is an automorphism of the p1 -component of G,
and hence has an inverse f . Now q is an endomorphism of G, and f q
is an idempotent endomorphism of G with image the p1 -component
of G. Apply 2.6.2 and induct.
6. Let G be a group, not necessarily …nite, all of whose elements are of …nite
order. Suppose that only …nitely many primes are relevant for G. That
is, there are …nitely many primes p1 , p2 , : : : , pr such that if g 2 G, then
o(g) is a product of powers of these primes. Prove that G is the direct
sum of its p-components.
7. Prove that the plane is the direct sum of any two distinct straight lines
through the origin.
8. Let : G ! H be an isomorphism. Prove that if A is a summand of G,
then (A) is a summand of H. Prove that it is not enough to assume that
is a monomorphism, nor enough to assume that is an epimorphism.
9. Prove that if Gi Hi for i = 1; 2; : : : ; n, then
n n
i=1 Gi i=1 Hi .
2.7. FINITE ABELIAN GROUPS 77
10. Prove that the subgroups of a cyclic p-group form a chain. That is, if A
and B are two subgroups of Z(pn ), then A B or A B.
12. Prove that for each m n, Z(pn ) has exactly one quotient group isomor-
phic to Z(pm ).
13. Prove that for each m n, Z(pn ) has exactly one subgroup isomorphic to
Z(pm ).
14. Prove that a …nite p-group G is a direct sum of cyclic groups as follows.
Let g 2 G be an element of maximum order. If G 6= f0g, then o(G=Zg) <
o(G), so that G=Zg = ni=1 (Ai =Zg), with Ai =Zg cyclic. Each Ai =Zg =
Z(ai + Zg) with o(a) = o(ai + Zg). Then Ai = Zai Zg, and
15. Prove that a …nite p-group G is a direct sum of cyclic groups as follows.
Let g 2 G be an element of maximum order. Let H be a subgroup of G
such that H \ Zg = f0g and such that if K is any subgroup of G properly
containing H, then K \ Zg 6= f0g. Then G = Zg H. Induct on o(G).
16. Find the number of non-isomorphic groups of order 71, 712 , 492 , 10011,
and p51 p52 p43 p34 p25 p6 , where the pi ’s are distinct primes.
18. Let G = H K, with G …nite. How does one get the invariant factors of
G directly from those of H and K? Given the invariant factors of G and
H, calculate those of K.
20. Let G and H be …nite. How does one tell from the invariant factors of G
and H whether or not
21. Let G be a …nite group and let pn divide o(G). Prove that G has a
subgroup of order pn . (This is true even if G is not necessarily Abelian,
but is considerably more di¢ cult to prove in that case.)
22. Prove that every …nite group of order greater than 2 has a non-trivial
automorphism. Prove this also for non-Abelian groups.
78 CHAPTER 2. GROUPS
24. Let Hom(G, H) be the group of all homomorphisms from G into H. Prove
that
Generalize.
25. Suppose that there is an epimorphism G ! Z. Prove that G = A B
with A Z. Generalize.
Chapter 3
Vector Spaces
Now real numbers themselves can be added and multiplied. We have four op-
erations at hand — addition of elements of P , addition and multiplication of
real numbers in R, and multiplication of elements of P by elements of R to get
elements of P . These four operations satisfy some basic rules. First, P is an
Abelian group under addition of its elements. The real numbers R, together
with ordinary addition and multiplication of its elements, constitute what is
called a …eld, an algebraic system which we will formally de…ne shortly. The
multiplication of elements of P by elements of R satis…es several basic rules
involving the other operations. For example, r((a; b) + (c; d)) = r(a; b) + r(c; d).
It is a setup such as this that arises over and over again in mathematics (and
elsewhere). It has been formalized and called a vector space. We proceed now
to its formal de…nition, and we …rst need to de…ne a …eld.
a. F is a group under +,
b. F = F n f0g is a group under , where 0 is the additive identity, and
79
80 CHAPTER 3. VECTOR SPACES
a. 0 a = a 0 = 0.
b. (a + b)c = ac + bc.
c. a( b) = (ab) = ( a)b.
d. ab = ( a)( b).
e. (e)a(b c) = ab ac.
f. ( 1)a = a.
g. 1 6= 0.
Proof.
a. a 0 = a (0 + 0) = a 0 + a 0, so a 0 a 0 = 0 = a 0 + a 0 a 0 = a 0.
Also, 0 a = a 0 since multiplication is commutative.
g. 1 2 F and 0 2
=F .
g. Let p be a prime. Then the “integers modulo p" form a …eld. Speci…cally,
in the Abelian group Z=Zp, multiply by the rule (m + Zp)(n + Zp) =
mn + Zp. This makes Z=Zp into a …eld. Two things need checking —
that multiplication is well de…ned, and that every non-zero element has an
inverse. No tricks are needed to do either, but the fact that p is a prime
is necessary to get the latter.
Now that we have the de…nition of a …eld, we can formally de…ne a vector
space. A vector space consists of several things. Recall the situation with the
plane. We had the Abelian group P of pairs of real numbers, the …eld R of real
numbers, and a way to multiply elements of P by elements of R to get elements
of P . This last multiplication satis…es some basic rules. Here is the de…nition.
a. a (v + w) = a v + a w for a 2 F , and v, w 2 V .
b. (a + b) v = a v + b v for a, b 2 F , and v 2 V .
d. 1 v = v for all v 2 V .
The elements of V are called vectors, and the elements of F are called
scalars. The mapping F V ! V is called scalar multiplication. Vectors
may be multiplied by elements of F (scalar multiplication). Scalars may be
added and multiplied since F is a …eld. We denote both addition in V and
addition in F by+. The context will make it clear which operation is meant.
For example, in (a), both + signs denote addition in V . In (b), however, the +
on the left is in F , and the one on the right is in V . We will habitually drop
the denoting scalar multiplication. Thus a v will be written simply av. There
are two zeroes around, the additive identity of F and the identity of V . Both
will be denoted by 0. Again the context will make it clear which zero is meant.
A vector space may be viewed as a special kind of Abelian group, namely one
where a multiplication of its elements by those of a …eld is given which satis…es
(a) –(d) above. A vector space is sometimes denoted simply by one letter, such
as V , but V actually is just the underlying set of vectors. It is understood that
V is an Abelian group and that there is a …eld F of scalars around. Also we say
that V is a vector space over F .
The following theorem gives some elementary consequences of the vector
space axioms.
3.1. DEFINITIONS AND EXAMPLES 83
a. a 0 = 0.
b. 0 v = 0.
c. ( a) v = (a v) = a ( v).
d. ( a) ( v) = a v.
e. ( 1) v = v.
f. a(v w) = av aw.
g. (a b)v = av bv.
Proof.
b. 0 v = (0 + 0)v = 0 v + 0 v, whence 0 = 0 v.
e. ( 1)v = (1 v) = v.
Theorem 3.1.7 Let V be a vector space over F and let S be a non-empty subset
of V . Then S is a subspace of V if and only if S is closed under vector addition
and scalar multiplication, that is, if and only if s + t 2 S whenever s, t 2 S,
and as 2 S whenever a 2 F and s 2 S.
84 CHAPTER 3. VECTOR SPACES
Proof. Just note that the intersection of any family of subspaces is closed
under addition and scalar multiplication, so is a subspace by 3.1.7.
Thus by 3.1.8, if S is any subset of a vector space V , then there is a unique
smallest subspace of V containing S, namely the intersection of all the subspaces
of V which contain S. As in the case for groups (2.2.3), we de…ne this subspace
to be the subspace generated by S. A vector space is …nitely generated if
it is generated by a …nite set. If S = fvg, where v is an element of V , then the
subspace generated by S is F v = fav : a 2 F g. If V1 , V2 , : : : , Vn are subspaces
of V , then the smallest subspace containing them all is
V1 + V2 + + Vn = fv1 + v2 + + vn : vi 2 Vi g.
a. Let V be the Abelian group of all pairs of real numbers with addition given
by
(a; b) + (c; d) = (a + c, b + d).
Let R be the …eld of real numbers, and let R V ! V be given by
De…ne F V ! V by
c(a; b) = (ca; cb):
This makes V into a vector space over F . Every subspace of V not V is
of the form F v for some v in V . (See PROBLEM 10.)
c. Let F be any …eld. For any integer n 1, let V be the set of all n-tuples
(a1 , a2 , : : : , an ) with ai 2 F . Add by the rule
De…ne F V ! V by
We will show later that any (…nite dimensional) vector space is essentially
one of these. Notice that when n = 1, V is just the additive group of
F , and scalar multiplication is just multiplication in F . Thus any …eld
may be regarded as a vector space over itself.
d. Let F be any …eld, and let F [x] be the set of all polynomials in x with
coe¢ cients in F . Add polynomials just as you would add polynomials with
real coe¢ cients. Scalar multiplication is given by
This makes V into a vector space over the …eld R of real numbers. (It is
proved in calculus that f +g and rf , as de…ned above, are again continuous
functions.) Instead of all continuous functions from R into R, we could
just as well take all integrable functions, or all di¤ erentiable functions, or
even all functions from R into R. The set of di¤ erentiable functions is a
subspace of the vector space of continuous functions, of course.
86 CHAPTER 3. VECTOR SPACES
PROBLEMS
1. Prove that there is no …eld with six elements.
7. Let V be the set of all functions f from R into R such that f (x) = f (x+2 ).
Prove that V is a vector space with the usual de…nitions of addition and
scalar multiplication.
8. Let V be the set of all functions f from the closed interval [0; 1] into itself
such that f (x) = f (1 x). Prove that V is a vector space under the usual
de…nitions of addition and scalar multiplication.
9. Let V be the set of all functions f from R into R which have second
derivatives f 00 , and such that f 00 = f . Prove that V is a vector space
under the usual de…nitions of addition and scalar multiplication. (It is
actually a subspace of the vector space in Problem 7.)
3.1. DEFINITIONS AND EXAMPLES 87
13. Prove that the subspace of R3 generated by f(1; 2; 3); (4; 5; 0); (0; 6; 7)g
is R3 itself.
14. Find the intersection of the two subspaces R(1; 2; 3)+R(4; 5; 6) and R(1; 2; 0)+
R(4; 5; 0) of R3 .
15. Let F = Z=Z2. Write down all the subspaces of F 2 . Write down all the
subspaces of F 3 .
16. Let F = Z=Z3. Write down all the subspaces of F 2 .
17. Let S and T be subspaces of a vector space V . Prove that
S + T = fs + t : s 2 S, t 2 T g
a. f (0) = 0.
b. f ( v) = f (v).
c. f is one-to-one if and only if Ker f = f0g.
d. Ker f is a subspace of V .
e. Im f is a subspace of W .
Proof. Parts (a), (b), and (c) hold because f is a homomorphism from
the group V into the group W . For the same reason, Ker f is a subgroup
of V and Im f is a subgroup of W . Let a 2 F and let v 2 Ker f . Then
f (av) = af (v) = 0. Thus av 2 Ker f , and so Ker f is a subspace of V . Let
a 2 F and let w 2 Im f . Then there is some v 2 V such that f (v) = w. But
f (av) = af (v) = aw 2 Im f . Thus (d) and (e) hold.
Example 3.2.3 In these examples, it is straightforward to verify that what are
claimed to be linear transformations are indeed linear transformations.
a. Let V and W be vector spaces over any …eld F . The maps V ! W : v ! 0
and 1V : V ! V : v ! v are linear transformations.
b. Let V be a vector space over F , and let a 2 F . The map f : V ! V
de…ned by f (v) = av is a linear transformation. Thus each scalar induces
a linear transformation V ! V . If a 6= 0, then this linear transformation
is non-singular.
c. Let F be any …eld, let F n be the vector space of n-tuples of elements of F ,
and let (a1 , a2 , : : : , an ) 2 F n . The map f : F n ! F n de…ned by
f (b1 , b2 , : : : , bn ) = (a1 b1 , a2 b2 , : : : , an bn )
is a linear transformation. It is non-singular if and only if each ai is
non-zero.
is a linear transformation.
h. Rotation of the plane through any number of degrees about the origin
is an automorphism of the vector space R2 . The formula for rotation
counterclockwise through is given by (x; y) ! (x cos + y sin ,
y cos x sin ). This example is a special case of (f ).
Kernels of linear transformations are subspaces. It will turn out that every
subspace of a vector space is the kernel of a linear transformation from that
vector space to another one. In fact, given a subspace S of a vector space V
over a …eld F , we will construct a vector space V =S over F , the quotient space
of V with respect to S. There will be a natural epimorphism V ! V =S, with
kernel S. This construction is similar to that of quotient groups, in fact, uses
that of quotient groups, and is fundamental.
a subgroup of the Abelian group V , so we have the quotient group V =S. Its
elements are cosets v + S, and one adds by the rule
(v + S) + (w + S) = (v + w) + S:
av = a(w + s) = aw + as;
V =(Ker f ) Im f:
S=(S \ T ) (S + T )=T .
PROBLEMS
1. Prove that the composition of two linear transformations (when de…ned)
is a linear transformation.
3. Let V be the vector space of all polynomials with real coe¢ cients. For
p in V , let d(p) be the derivative of p. Let fx (p) = xp, that is, just
multiplication by x. Prove that d and fx are linear transformations, and
that d fx fx d = 1V . What does this problem have to do with real
numbers?
10. Let V = F 3 , and let S = F v for some non-zero v. Prove that there is a
subspace T of V strictly between S and V . Prove that there is no subspace
strictly between T and V .
11. For the vector spaces in Problem 10, prove that V =S F 2 and V =T F.
12. In R3 , let
S = R(1; 2; 3) + R(4; 5; 2),
and
T = R(2; 1; 4) + R(1; 1; 1)3 .
Find (a; b; c) in R3 such that S \ T = R(a; b; c).
13. In Problem 12, …nd a set of representatives for the cosets of S in R3 . Find
a set of representatives for the cosets of S \ T in S.
17. Prove that the set of automorphisms of a vector space, with multiplication
given by composition of maps, is a group.
94 CHAPTER 3. VECTOR SPACES
20. Let V be the vector space of all mappings of a set S into a …eld F . (See
Problem 12, section 3.1.) Let be a mapping of S into S. Prove that the
mapping
V !V :v!v
is a linear transformation. Prove that this linear transformation is non-
singular if and only if is a permutation of S.
In fact, (a; b) = a(1; 0) + b(0; 1), and if (a; b) = c(1; 0) + d(0; 1), then (a; b) =
(c; 0) + (0; d) = (c; d), so that a = c and b = d. One of our principal aims in this
section is to show that this phenomenon holds in any …nitely generated vector
space 6= f0g. Speci…cally, we want to show that if V is a …nitely generated
vector space and V 6= f0g, then there are vectors v1 , v2 , : : : , vn of V such that
every element of V can be written uniquely in the form
a1 v1 + a2 v2 + + an vn :
is a basis of F n . This basis is called the natural basis of Fn . Two other bases
of F 3 are
f(1; 0; 0); (0; 1; 0); (0; 0; 1)g, and
f(1; 1; 1); (0; 1; 1); (0; 0; 1)g.
At this point we have the following things on our mind. We want to show
that every …nitely generated vector space 6= f0g has a basis and that any two
bases of such a vector space are the same size. The number of elements in a
basis of V will be called the dimension of V , and it will turn out that two
vector spaces over a …eld F are isomorphic if and only if they have the same
dimension.
Theorem 3.3.2 Every …nitely generated vector space 6= f0g has a basis.
fwn , v1 , v2 , : : : , vn g
is dependent. Remove from this family the …rst member that is a linear combi-
nation of the preceding ones. There is such a member, and it is not wn . This
new family
fwn , v1 , v2 , : : : , vi 1 , vi+1 , : : : , vm g
still spans V . Do the same thing with this generating family and wn 1 . Continue
the process. At each step, a w gets added and a v gets removed. When a w is
added, a dependent family results. But fwk , wk+1 , : : : , wn g is independent for
every k n. Hence there are at least as many v’s as w’s, and so m n.
Letting fv1 , v2 , : : : , vm g and fw1 , w2 , : : : , wn g be bases, we conclude that
m = n, and hence get the following corollary.
Corollary 3.3.6 Any two bases of a …nitely generated vector space have the
same number of elements.
F w1 + F w2 + + F wm :
Theorem 3.3.8 Two …nitely generated vector spaces V and W over a …eld F
are isomorphic if and only if dim(V ) = dim(W ).
Im = F w1 + F w2 + + F wn ,
Consider all the …nitely generated vector spaces over the …eld F . By 3.3.8,
a complete isomorphism invariant of such a vector space V is dim(V ). That
is, V is determined up to isomorphism by the number dim(V ). This does not
conclude the study of vector spaces though. The real things of interest about
vector spaces are linear transformations, and we have hardly begun to study
them.
Let V1 , V2 , : : : , Vn be vector spaces over a …eld F . Let V1 V2 Vn
be the vector space whose elements are the elements of V1 V2 Vn and
whose vector addition and scalar multiplication are de…ned by
and
a(v1 , v2 , : : : , vn ) = (av1 , av2 , : : : , avn ):
This construction is the same as that for Abelian groups except that here
we also have scalar multiplication involved. That V1 V2 Vn is indeed
a vector space follows easily. This vector space is the external direct sum of
V1 , V2 , : : : , Vn . For example, F n is the external direct sum of n copies of the
vector space F .
If Vi consists of those elements of V1 V2 Vn of the form
where vi is in the ith place, then Vi is a subspace, and as in the case for
Abelian groups, every element in V1 Vn can be written uniquely in the
form ni=1 vi , with vi 2 Vi .
Theorem 3.3.12 Every …nite dimensional vector space 6= f0g is the direct sum
of one dimensional subspaces.
PROBLEMS
1. Prove that if fv1 , v2 , : : : , vn g is independent, then v1 6= v2 , and v1 6= 0.
2. Show by direct computation that in Q3 , the set
S = f(3; 0; 3), ( 1; 1; 2), (4; 2; 2), (2; 1; 1)g
is linearly dependent. Does S generate Q3 ? Does S contain a basis of Q3 ?
3. Do the following.
(a) Show that f(1; 0; 1), (1; 2; 1), (0; 3; 2)g is a basis of Q3 .
(b) Is f(1; 1; 2; 4), (2; 1; 5; 2), (1; 1; 1; 0), (2; 1; 1; 0)g linearly inde-
pendent in Q4 ?
4. Let F = Z=Z2. Write down all bases of F 2 . Write down all bases of F 3 .
5. Let F = Z=Z3. Write down all bases of F 2 .
6. In R4 , let V = R(1; 1; 1; 1), and let W = V + R(1; 1; 1; 0). Let V ? =
f(a1 ; a2 ; a3 ; a4 ) 2 R4 : ai bi = 0 for all (b1 ; b2 ; b3 ; b4 ) 2 V g, and de…ne
W ? similarly. Find a basis for V ? . Find a basis for W ? .
7. Suppose that v1 6= 0. Prove that fv1 , v2 , : : : , vn g is dependent if and only
if some vi+1 is a linear combination of v1 , v2 , : : : , vi .
8. Prove that fv1 , v2 , : : : , vn g is a basis if and only if it is linearly independent
and every family fw1 , w2 , : : : , wm g with m > n is dependent.
9. Prove that fv1 , v2 , : : : , vn g is a basis of V if and only if it spans V and
no family fw1 , w2 , : : : , wm g with m < n spans V .
10. Prove that fv1 , v2 , : : : , vn g is a basis if and only if it is linearly independent
and every family properly containing it is dependent.
11. Prove that fv1 , v2 , : : : , vn g is a basis of V if and only if it spans V and
no proper subfamily spans V .
12. Prove that (V W )=W V.
13. Let fv1 , v2 , : : : , vm g and fw1 , w2 , : : : , wn g be bases of V . Prove that
m = n in the following way. Write v1 = ni=1 ai wi . We may suppose that
a1 6= 0. Then V = F v1 F w2 F wn , so
V =F v1 F v2 F vm F w2 F wn :
Inducting on m yields m = n.
3.3. LINEAR INDEPENDENCE AND BASES 101
17. Let F be any …eld. As a vector space over F , what are the bases of F ?
What is dim(F )? Prove that f(a1 , a2 ), (b1 , b2 )g is a basis of F 2 if and
only if a1 b2 a2 b1 6= 0.
18. Suppose that F is a …eld with 9 elements. (There is such a …eld.) Suppose
that P is a sub…eld with 3 elements. (There is such a sub…eld.) Consider
F as a vector space over P . What is dim(F )?
19. Consider the complex numbers as a vector space over the real numbers.
What is its dimension?
20. Consider the real numbers as a vector space over the rational numbers.
Prove that it is not …nite dimensional.
22. Let V be the vector space over the …eld of real numbers consisting of all
functions from the real numbers into the real numbers. Prove that V is
not …nite dimensional.
23. Let V1 , V2 : : : , Vn be subspaces of V . Prove that the map from the external
direct sum V1 Vn into V given by (v1 , v2 , : : : , vn ) = v1 + v2 + : : :
+vn is a linear transformation. Prove that V is the direct sum of the
subspaces V1 , V2 , : : : , Vn if and only if is an isomorphism.
n
24. Prove that if V = V1 ::: Vn , then dim(V ) = i=1 dim(Vi ).
25. Let V be …nite dimensional, and let W and X be subspaces of V . Let fv1 ,
v2 , : : : , vr g be a basis of W \ X, fv1 , : : : , vr , w1 , : : : ; ws g a basis of W ,
and fv1 , : : : , vr , x1 , : : : , xt g a basis of X. Prove that fv1 , : : : ; vr , w1 ,
: : : , ws , x1 , : : : , xt g is a basis of W + X. (This proves 3.3.16.)
28. Let V be the vector space of all sequences fa1 ; a2 , : : : :g of real numbers.
Let S be the subspace consisting of those sequences with only …nitely
many non-zero terms. Prove that S has a basis. Try to prove that V has
a basis.
and
and
a(f + g) = af + ag;
and
a(bf ) = (ab)f
hold. The …rst holds because for v 2 V;
De…nition 3.4.1 The vector space Hom(V; F ) is called the dual space of V ,
and denoted V .
Other names for V are conjugate space, and adjoint space. The …rst
thing we will do is decide what the dimension of V is. It is about as easy
to compute dim(Hom(V; W )). We will assume throughout that all our
vector spaces are …nite dimensional.
fij (vi ) = wj ,
and
fij (vk ) = 0 if i 6= k:
Thus f23 (v1 ) = 0, f23 (v2 ) = w3 , f23 (v3 ) = 0, : : : , f23 (vm ) = 0.
These functions are more conveniently de…ned using the Kronecker delta.
Let ij = 0 2 F if i 6= j, and ij = 1 2 F if i = j. Then for i = 1, 2, : : : , m and
j = 1, 2, : : : , n,
fij (vk ) = ik wj .
104 CHAPTER 3. VECTOR SPACES
We claim that the fij constitute a basis for Hom(V; W ). At least it has the size
we want.
Suppose i;j aij fij = 0. Then
i;j aij fij (vk ) =0= i;j (aij fij (vk )) = i;j aij ik wj = i;j akj wj .
Hence k;j akj fkj = f , so that the fij span Hom(V; W ). The proof is complete.
vi (vj ) = ij
vi (vj ) = ij
Proof. We do not know yet that is even into V . To see this, we need
that for a 2 F and x, y 2 V ;
and
(v)(ax) = a( (v)(x)).
But this follows from the equations
and
(v)(ax) = (ax)(v) = a(x(v)) = a( (v)(x)).
Thus maps V into V . Now must preserve vector addition and scalar
multiplication. That it does is left to the reader. (Problem 2) To get to be an
isomorphism, it su¢ ces to get it to be one-to-one since V is …nite dimensional
and dim(V ) = dim(V ). Suppose that (v) = 0. Then (v)(v ) = 0 = v (v)
for all v 2 V . But if v 6= 0, then v is an element of a basis of V , and the
appropriate element of the basis dual to it takes v to 1 2 F . Thus v = 0, and
we are done.
Let S be a subset of V . Consider all those v 2 V such that Ker(v ) S.
This subset of V is called the annihilator of S in V . Denote it by K(S).
Thus
K(S) = fv 2 V : v (s) = 0 for all s 2 Sg.
Similarly, if T is a subset of V , let
c. K(S) is a subspace of V .
f. K (K(S)) S; K(K (T )) T.
106 CHAPTER 3. VECTOR SPACES
fv1 , v2 , : : : , vn g
n
be dual to it. Now i=1 ai vi is in K(S) if and only if ( ai vi )(vj ) = 0 for j = 1,
2, : : : , m. But
( ai vi )(vj ) = ai (vi (vj )) = aj :
Hence
(w ) = w .
There are some important relations between and . One is that and
have the same rank, where the rank r( ) of a linear transformation is de…ned
to be dim(Im ). We will relate this to the rank of matrices at the end of this
section.
Hom(V; W ) ! Hom(W ; V ) : !
is an isomorphism.
Theorem 3.4.16 The row rank of a matrix equals its column rank.
Proof. This theorem is a re‡ection of the fact that the rank of a linear
transformation is the rank of its dual (3.4.13). Let fv1 , v2 , : : : , vm g be the
natural basis of F m , and fw1 , w2 , : : : , wn g the natural basis of F n . Let fv1 ,
: : : , vm g and fw1 , : : : , wn g be bases dual to these bases. Let : F m ! F n be
given by
(vi ) = (ai1 , ai2 , : : : , ain ).
Then r( ) is the row rank of (aij ). Let be the isomorphism de…ned by
(vi ) = vi . Then (wi ) = j aji vj since
Thus
(wi ) = ( aji vj ) = aji vj ) = (a1i , a2i , : : : , ami ).
Hence r( ) is the column rank of the matrix (aij ). Since is an isomorphism,
r( ) = r( ), and we know that r( ) = r( ). The theorem is proved.
The row rank of a matrix is called simply the rank of that matrix. Let (aij )
be a matrix and (aij )0 its transpose. That is, the ijth entry in (aij )0 is aji .
Then the rows of (aij ) are the columns of (aij )0 and the columns of (aij ) are
the rows of (aij )0 . Sometimes 3.4.16 is expressed by saying that the rank of a
matrix equals the rank of its transpose.
PROBLEMS
1. Give a complete proof that Hom(V; W ) is a vector space.
Hom(V; W ) ! Hom(W ; V ) : !
is a linear transformation.
( ) = .
110 CHAPTER 3. VECTOR SPACES
13. Let vi be the vector in F n with 1 in the ith place and 0 elsewhere. Let
fv1 , : : : vn g be the basis dual to fv1 , : : : , vn g. Identify F n with F n via
the isomorphism given by (vi ) = vi . Prove that
(a1 , a2 , : : : , an )((b1 , b2 , : : : , bn )) = ai bi .
De…nition 4.1.1 A ring is a set R with two binary operations, called addition
and multiplication, and denoted + and respectively, such that
b. is associative, and
The two equalities in (c) are called the left and right distributive laws, re-
spectively. As usual for additively written Abelian groups, the additive identity
of R is denoted by 0, and a denotes the element of R such that a + ( a) = 0.
It may happen that is commutative, that is, that a b = b a for all a, b in
R. In that case, R is called a commutative ring. If R 6= f0g and there is an
element x in R such that x a = a x = a for all a in R, then that element is
unique, is always denoted by 1, and R is called a ring with identity. As in the
case of …elds, which are special kinds of rings, we will tend to write a b as ab.
Before proceeding with examples, some immediate consequences of the ax-
ioms will be proved that are needed in making elementary calculations in rings.
111
112 CHAPTER 4. RINGS AND MODULES
a. a 0 = 0 a = 0.
b. a( b) = (ab) = ( a)b.
c. ab = ( a)( b).
f. 1 6= 0.
Proof.
a. a 0 = a(0 + 0) = a 0 + a 0. Thus a 0 a 0 = 0 = (a 0 + a 0) a 0=
a 0 + (a 0 a 0) = a 0. Similarly, 0 a = 0.
Theorem 4.1.5 A …nite ring with identity and no zero divisors is a division
ring.
4.1. DEFINITIONS AND EXAMPLES 113
Example 4.1.7 These examples should be examined carefully, and the reader
should verify that each example is what it is claimed to be.
a. Any …eld is a ring. Thus the rational numbers Q with the ordinary opera-
tions of addition and multiplication is a ring. Similarly, the real numbers
and the complex numbers are rings, being …elds.
1. The set Z2 of even integers is a subring of the ring Z. This ring does not
have an identity, but is commutative and has no zero divisors.
e. Let Z2 be the set of all 2 2 matrices over Z. That is, Z2 is the set of all 2
2 matrices whose entries are integers. De…ne addition and multiplication
by the equations
and
a11 a12 b11 b12 a11 b11 + a12 b21 a11 b12 + a12 b22
=
a21 a22 b21 b22 a21 b11 + a22 b21 a21 b12 + a22 b22
114 CHAPTER 4. RINGS AND MODULES
0 1 1 0 1 0 0 1
0= 6= ,
0 0 0 0 0 0 0 0
f. Let R be any ring, and n any positive integer. Let Rn be the set of all
n n matrices over R. An element
2 3
a11 a12 ::: a1n
6 a21 a22 ::: a2n 7
6 7
6 .. .. .. 7
4 . . . 5
an1 an2 ::: ann
of Rn has all the aij in R, and is denoted simply by (aij ). The element
aij denotes the (i; j) entry of (aij ). Addition and multiplication then are
de…ned by the equations
and
n
(aij ) (bij ) = ( k=1 aik bkj ):
g. Let R be the …eld of real numbers. Let C be the set of 2 2 matrices with
entries from R of the form
a b
.
b a
h. Let R be any commutative ring. Let R[x] be the set of all polynomials
in x with coe¢ cients in R. That is, R[x] is the set of all expressions of
the form an xn + + a1 x + a0 , with ai in R. Under the usual rules for
adding and multiplying polynomials, R[x] is a commutative ring, and has
an identity if and only if R does. (See 4.5.)
i. Let R be any ring. Let S be the set of all mappings f of the set of non-
negative integers into R such that f (n) = 0 for all but …nitely many n.
De…ne (f + g)(n) = f (n) + g(n), and (f g)(n) = i+j=n f (i)g(j). Then
S is a ring with this de…nition of + and . What is the di¤ erence between
(i) and (j)? Is commutativity necessary in (h)?
4.1. DEFINITIONS AND EXAMPLES 115
j. Let G be any Abelian group and let R be the set of all endomorphisms of
G. For f , g 2 R, de…ne f + g and f g by
and
(f g)(x) = f (g(x)).
Then R is a ring with identity, called the endomorphism ring of G. If
G is the additive group of Z, then R is an integral domain (4.2, Problem
23). If G is the additive group Z=pZ of the integers modulo the prime p,
then R is a …eld (4.2, Problem 24).
k. Let V be any vector space, and let R be the set of all linear transformations
from V into V . De…ne addition and multiplication as in (k). Then R is
a ring with identity, called the ring of linear transformations of V .
We will look at this ring in some detail in Chapter 8.
l. Let p be a prime and let
PROBLEMS
1. Do the following.
(a) Prove that a ring with identity has just one identity.
(b) Find an example of a ring with identity that has a subring with a
di¤erent identity.
(c) Find an example of a ring with no identity that has a subring with
identity.
4. Let R be a ring with identity. Let U be the set of units of R. That is,
u 2 U if and only if there is a v 2 R such that uv = vu = 1. Prove that U
is a group under multiplication. (U is called the group of units of R.)
What is the group of units of Z?
5. Do the following.
9. Let S be a subset of the ring R. Prove that there is a subring T such that
S T and such that every subring of R containing S contains T .
15. Let G be an Abelian group. Prove that the group ring Z(G) is an integral
domain if and only if G is torsion free, that is, if and only if G has no
nonzero elements of …nite order.
118 CHAPTER 4. RINGS AND MODULES
d. Let R be a ring and S a non-empty set. Let RS be the ring of all mappings
from S into R, as in Example 4.1.7 (q). Let s 2 S. De…ne ' : RS ! R
by '(f ) = f (s). Then ' is an epimorphism.
e. Let R(G) be the ring in Example 4.1.7 (o). De…ne the mapping ' :
R(G) ! R by '(f ) = g2G f (g). Then ' is an epimorphism.
r 0
R ! R2 : r 7!
0 r
r 0
R ! R2 : r 7!
0 0
Kernels of homomorphisms are ideals. It will turn out that every ideal of a
ring R is the kernel of a homomorphism from R to another ring. In fact, given
an ideal I of a ring R, we will now construct a ring R=I, the quotient ring
of R with respect to I. There will be a natural homomorphism from R onto
R=I whose kernel is I. This construction is fundamental.
(r + I) + (s + I) = (r + s) + I:
(r + I)(s + I) = rs + I:
120 CHAPTER 4. RINGS AND MODULES
But r1 j, is1 , and ij are all in I since i and j are in I and I is an ideal. Thus
r1 j + is1 + ij is in I, so that rs + I = r1 s1 + I.
It is easy to verify that with the multiplication
(r + I)(s + I) = rs + I;
so that the left distributive law holds. The other axioms of a ring are just as
easy to check. If R has an identity, then 1 + I is the identity for R=I. If R is
commutative, then R=I is commutative. The mapping : R ! R=I de…ned by
(r) = r + I is an epimorphism.
De…nition 4.2.6 The ring R=I constructed in 4.2.5 is called the quotient ring
of R with respect to the ideal I, or more simply “R modulo I.” The ho-
momorphism : R ! R=I : r ! r + I is called the natural homomorphism
from R to R=I.
R=(Ker f ) Im f
(r + I)(j + I) = rj + I 2 J=I:
(r + I)(j + I) = rj + I 2 J=I
ax + Zp = 1 py + Zp = 1 + Zp,
so that Z=Zp is a …eld. We have then that n is prime if and only if Z=Zn is
a …eld.
Thus a maximal ideal is one that is not the whole ring, and such that there
are no ideals strictly between it and the whole ring. The maximal ideals of Z
are Zp with p prime, and these were also the ideals I of Z such that Z=I was a
…eld.
fra + m : r 2 R; m 2 M g
PROBLEMS
1. Find all rings with two elements. (First, decide what this problem means.)
2. Prove that the rings Z2 and Z3 are not isomorphic. Prove that if m
and n are distinct positive integers, then the rings Zm and Zn are not
isomorphic.
3. Prove that the …elds R and C are not isomorphic.
p p p
4. Let Q 2 = fa + b 2 : a; b 2 Qg, and de…ne Q 5 similarly. Prove
that these are rings, and prove that they are not isomorphic.
5. Let R be the set of 2 2 matrices of the form
a 2b
b a
with a and p
b elements of Q. Prove that R is a ring and is isomorphic to
the ring Q 2 of Problem 4.
4.2. HOMOMORPHISMS AND QUOTIENT RINGS 123
6. Find all homomorphisms from the ring Z=Z12 to the ring Z=Z20.
10. Let R be a ring with identity and suppose that x is a nilpotent element
of R. Prove that 1 x is a unit.
11. Prove that the set of nilpotent elements in a commutative ring R forms
an ideal N , and that R=N has no non-zero nilpotent elements.
15. Prove that the set Aut(R) of all automorphisms of a ring R is a group
under composition of automorphisms.
17. Let R be a ring with identity, and let Inn(R) be the set of all inner au-
tomorphisms of R. Prove that Inn(R) is a subgroup of Aut(R), and is
normal in Aut(R).
18. Let R be a ring with identity, and let U (R) be the group of units of R.
Prove that the map
U (R) ! Inn(R) : u ! fu
is an ideal of R.
29. Let I and J be ideals of R. Prove that
IJ = fi1 j1 + i2 j2 + ::: + in jn : i1 ; i2 ; :::; in 2 I, j1 ; j2 ; :::; jn 2 J, n 2 Z+ g
is an ideal of R. (IJ is the product of I and J.)
30. Let I and J be ideals of R. What is the relation between IJ and I \ J?
31. Let P be a set of primes. Let
ZP = fa=b : a; b 2 Z; b 6= 0; (b; p) = 1 for all p 2 P g.
Find all the maximal ideals M of the ring Zp , and in each case determine
the …eld Zp =M .
4.2. HOMOMORPHISMS AND QUOTIENT RINGS 125
32. Find an example of an integral domain that has exactly one maximal ideal.
Now …nd such an example that is not a …eld.
33. Find the maximal ideals of Z=Zn.
34. For which n does Z=Zn have exactly one maximal ideal?
35. Let S be any set, and let P (S) be the ring of subsets of S. (See 4.1.7 (r).)
Let T S. Prove that
P (S) ! P (T ) : a ! a \ T
(a) S 2 U and ; 2
= U.
(b) If u, v 2 U , then u \ v 2 U .
(c) If u 2 U , v 2 P (S), and u v, then v 2 U .
(d) For each a 2 P (S), either a or its complement Sn a is in U .
38. Let S be a set, and let P (S) be the ring of subsets of S. A subset U of
P (S) which satis…es (a) – (d) of Problem 37 is called an ultra-…lter on
S. Prove that a subset M of P (S) is a maximal ideal of P (S) if and only
if its complement is an ultra-…lter on S.
39. Let S be a set, and let F be the …eld with two elements. Prove that the
ring F S of all mappings from S into F is isomorphic to the ring P (S) of
all subsets of S.
40. Let R be a ring with identity. De…ne a new ring S as follows. The elements
of S are those of R. If + and are addition and multiplication in R, de…ne
and on S by
a b = a + b + 1;
and
a b = a b + a + b:
Prove that S is a ring, and that it is isomorphic to the ring R.
41. Let G = fe; a; bg be the cyclic group of order 3. Calculate the following
products in the group ring Q(G).
(b) (e a)3 + (e a2 )3 .
(c) (1=3(e a) + 1=3(e a2 ))(x(e a) + y(e a2 )), where x and y are
any elements of Q.
We have two things in mind. We want to show that every integral domain D
has a …eld of quotients, and that any two …elds of quotients of D are essentially
the same. We must decide what this last part should mean.
This de…nition just says that F and F 0 must be isomorphic via an isomor-
phism that respects the embeddings ' : D ! F and '0 : D ! F 0 . If we
think of D as being contained in F and F 0 , then it says that there must be an
isomorphism : F ! F 0 …xing D elementwise. Here is a picture.
D
' '0
?
F - F0
4.3. FIELD OF QUOTIENTS OF AN INTEGRAL DOMAIN 127
Theorem 4.3.3 Any integral domain D has a …eld of quotients, and any two
…elds of quotients of D are equivalent.
Proof. This proof is a little long and formal, but conceptually it is easy.
A …eld F will be constructed from D in the same way that Q is constructed
from Z. A rational number is a quotient a=b = ab 1 of integers, but two such
quotients a=b and c=d may be the same rational number with a 6= c and b 6= d.
For example, 1=2 = 5=10. In fact, a=b = c=d if and only if ad = bc. We will
follow this lead in building our …eld. Recall that D = Dnf0g. In the set
D D , let
(a; b) (c; d) if ad = bc.
This is an equivalence relation on D D . Only transitivity threatens to be
di¢ cult. If (a; b) (c; d) and (c; d) (e; f ), then ad = bc and cf = de. We need
af = be. But ad = bc implies adf = bcf , and cf = de implies bcf = bde. Since
D is an integral domain, af = be. Thus is transitive.
The equivalence relation ~ on D D partitions D D into equivalence
classes. Let F be the set of these equivalence classes. We will make F into
a …eld. Denote the equivalence class containing (a; b) by a=b. Then we have
a=b = c=d if and only if ad = bc. We need to de…ne addition and multiplication
on F . Let
(a=b)(c=d) = ac=bd,
and
a=b + c=d = (ad + bc)=bd.
First, we must make sure that we have really de…ned binary operations on
F . It is conceivable that a=b = a0 =b0 , c=d = c0 =d0 , and ac=bd 6= a0 c0 =b0 d0 , and
similarly for addition. So suppose a=b = a0 =b0 . We will show that (ad+bc)=bd =
(a0 d + b0 c)=b0 d. This holds if and only if (ad + bc)b0 d = (a0 d + b0 c)bd. Since
ab0 = ba0 , the left side is (a0 bd + bb0 c)d and so is the right. It follows that
addition is well de…ned. The proof that multiplication is well de…ned is similar.
Note that both operations are commutative. Under this addition and mul-
tiplication, F is a …eld. First, 0=1 is the additive identity since 0=1 + a=b =
(0 b + a 1)=b = a=b. The negative of a=b is ( a)=b. The associative laws are
not di¢ cult. The multiplicative identity is 1=1. Let’s check the distributive law.
while
Thus the distributive law holds. If a=b 6= 0=1, then a 6= 0, and b=a 2 F . Further,
(a=b)(b=a) = ab=ba = 1=1. Therefore
1
(a=b) = b=a:
128 CHAPTER 4. RINGS AND MODULES
In short, one computes with these a=b just as with rational numbers.
We now have a …eld F . The map
' : D ! F : a ! a=1
and
'(ab) = ab=1 = (a=1)(b=1) = '(a)'(b):
If '(a) = 0=1, then 0=1 = a=1, and so 1 a = 0 1 = a = 0. We have
already observed that 1=b = (b=1) 1 . Thus, for a=b 2 F , a=b = (a=1)=(b=1) =
'(a)('(b)) 1 .
It remains to show that any two …elds of quotients of D are equivalent. Let
' : D ! F and '0 : D ! F 0 be such. De…ne
: F ! F 0 : '(a)('(b)) 1
! '0 (a)('0 (b)) 1
.
It is not entirely obvious that is well de…ned. To make sure that it is,
suppose that '(a)('(b)) 1 = '(a0 )('(b0 )) 1 . Then '(a)'(b0 ) = '(a0 )'(b) =
'(ab0 ) = '(a0 b), and so ab0 = a0 b. Therefore '0 (ab0 ) = '0 (a0 b), and hence
'0 (a)('0 (b)) 1 = '0 (a0 )('0 (b0 )) 1 . Obviously, ' = '0 .
We need to be an isomorphism. First, if a, b 2 D with a 6= 0 6= b, then
'(ab) = '(a)'(b), so ('(ab)) 1 = ('(a)'(b)) 1 = ('(a)) 1 ('(b)) 1 . Now
1 1 1
(('(a)('(b)) )('(c)('(d)) )) = ('(ac)('(bd)) ) = '0 (ac)('0 (bd)) 1
0 0 1 0 0 1
= (' (a)(' (b)) )(' (c)(' (d)) )
1 1
= ('(a)('(b)) ) ('(c)('(d)) ).
This de…nes an equivalence relation, and the set of equivalence classes is a ring
under the operations de…ned as in the integral domain case. This ring is denoted
RS , R is “contained”in RS , and every element of S has an inverse in RS . Even
if R is an integral domain, S does not have to be taken to be all the non-zero
elements of R. For example, in Z let S be all those integers relatively prime
to a …xed prime p. The resulting ring of quotients consists of all those rational
numbers whose denominators are relatively prime to p.
There are generalizations to the non-commutative case. If R is a ring with no
zero divisors, there is a condition due to Oré which insures that the construction
above can be carried out, yielding an embedding of R in a division ring.
PROBLEMS
1. Find the …eld of quotients of the integral domain
p
fa + b 3 : a, b 2 Zg.
classes of ~ . Let a=b denote the equivalence class containing (a; b). Prove
that
(a=b)(c=d) = ac=bd,
and
a=b + c=d = (ad + bc)=bd
make RS into a commutative ring with identity. Let s 2 S. Prove that
'S : R ! RS : a ! (as=s)
Ra = fra : r 2 Rg
y with coe¢ cients in F is not a PID. The ideal generated by the set fx; yg is
not principal. The ring Z[x] is not a PID, since the ideal generated by f2; xg is
not principal. We have not studied polynomial rings yet, and these facts may
not be entirely obvious.
Our purpose in this section is to discuss primes, greatest common divisors,
factorization of elements into primes, and the like in PID’s. Application of these
results will be mainly to F [x] in 4.5.
We will assume throughout that R is an integral domain. A number of
de…nitions are needed.
The element a in R is an associate of the element b in R, if a = ub for some
unit u in R. An element a not a unit and not 0 is irreducible if whenever
a = bc, either b or c is a unit. The element a divides the element b, or b is
a multiple of a, denoted ajb, if b = ac for some c. The non-zero element p
is a prime if p is not a unit, and whenever pjab, then pja or pjb. A greatest
common divisor (abbreviated gcd) of the set of elements fa1 ; a2 , : : : ; an g is
an element d such that djai for all i, and such that if ejai for all i, then ejd.
Two elements a and b are relatively prime if any gcd of fa; bg is a unit. A
least common multiple (abbreviated lcm) of the set of elements fa1 ; a2 , : : :
; an g with not all ai = 0 is an element a such that a is a multiple of each ai ,
and such that if e is a multiple of each ai , then e is a multiple of a.
Being “an associate of”is an equivalence relation. This follows from the fact
that the set U of units of R is a multiplicative group. The set of equivalence
classes of this equivalence relation partitions R. The equivalence classes on Z
are f ng, n = 0, 1, 2, : : : . There are only two units, 1, in Z, so each
equivalence class has at most two elements in it. For a …eld F , f0g and F are
the only equivalence classes. For any integral domain R and element a in R,
the equivalence class containing a is aU = fau : u 2 U g, where U is the group
of units of R. If F is the …eld of quotients of R, a coset xU of U in F is either
contained in R or is disjoint from R, according to whether x 2 R or x 2 = R.
Thus the equivalence classes are the cosets of U which are in R.
Suppose that the element a is irreducible and that u is a unit. If au = bc,
then a = b(cu 1 ). Therefore, either b or cu 1 is a unit, and hence either b or
c is a unit. Therefore if an element a is irreducible and u is a unit, then au is
irreducible.
Notice that if ajb and u is a unit, the aujb. In fact, ac = b implies that
(au)(u 1 c) = b. Suppose that a is a prime and that u is a unit. If aujbd, then
ajbc, so ajb or ajc. Thus aujb or aujc. Therefore au is prime if a is prime.
Equivalently, if a is prime, then so is any associate of a.
Suppose d and d0 are both gcd’s of fa1 ; a2 , : : : ; an g. Then djd0 and d0 jd.
Hence de = d0 and d0 e0 = d for some e and e0 . Thus dee0 = d, so ee0 = 1, whence
e and e0 are units. Clearly, if d is a gcd of fa1 ; a2 , : : : ; an g, then so is du for any
unit u. Therefore, if a set fa1 ; a2 , : : : ; an g has a gcd d, then the set of all gcd’s
of that set is the equivalence class of associates of d. A similar story is true for
lcm’s. One class of associates is the set U of units of the ring. This class has a
natural representative, namely 1. If the set of gcd’s of a set fa1 ; a2 , : : : ; an g is
U , then we way that the gcd of the set is 1, and write gcdfa1 ; a2 , : : : ; an g = 1.
132 CHAPTER 4. RINGS AND MODULES
De…nition 4.4.5 Let R be any ring. Then R satis…es the ascending chain
condition, or R is Noetherian, if every ascending sequence I1 I2 I3
of ideals of R becomes constant.
That is, R satis…es the ascending chain condition (abbreviated acc) if R has
no in…nite chain I1 I2 I3 of ideals with the inclusions all proper ones.
Theorem 4.4.6 A PID is Noetherian.
Corollary 4.4.9 Every non-zero element a in a PID can be written in the form
where v is a unit and the qi are nonassociate primes, then j = k, and after
suitable renumbering, pi and qi are associates for all i.
PROBLEMS
1. Prove that being “an associate of” is an equivalence relation.
2. Let p be a prime in Z, and let R = fa=pn : a, n 2 Zg. Prove that R is
a PID. What are the primes, and what are the units in R? Prove that if
p = 3, then a gcd of f5; 7g is 9.
3. Prove that any subring of Q which contains Z is a PID.
4. Let S be a nonempty multiplicatively closed subset of Z with 0 2
= S. What
are the primes, and what are the units of ZS ?
5. Let R be a PID, and let S be a non-empty multiplicatively closed subset
of R with 0 2
= S. Prove that RS is a PID. What are its units?
6. Do the following.
with u and v units, the pi ’s primes, and ni and mi 0. Write down a gcd
and an lcm of fa; bg.
10. Let a and b be in R. Prove that if R is a PID, and if d is a gcd of fa; bg,
then ab=d is an lcm of fa; bg.
11. Prove that the ideals of a PID R are in one-to-one correspondence with
the associate classes of elements of R.
4.4. PRINCIPAL IDEAL DOMAINS 135
12. Let I and J be ideals of the PID R. If I = Ra and J = Rb, prove that
IJ = Rab. (See 4.2, Problem 29, for the de…nition of the product IJ.)
13. Let I and J be ideals of the PID R. Prove that if I = Ra and J = Rb,
then I + J = Rd, where d is any gcd of a and b. Prove that I \ J = Rc,
where c is any lcm of a and b.
14. Let I and J be ideals in the PID R. Let I = Ra, and let J = Rb. Prove
that IJ = I \ J if and only if ab is an lcm of a and b.
15. Call an ideal I in a commutative ring R with identity a prime if I 6= R
and if ab 2 I implies that a 2 I or b 2 I. Prove that I 6= R is a prime ideal
in R if and only if R=I is an integral domain. Prove that every maximal
ideal of R is a prime ideal.
16. Let R be a PID, and let r 6= 0 nor be a unit. Prove that R=Rr is either a
…eld or has zero divisors.
17. Find those r in a PID R such that R=Rr has no non-zero nilpotent ele-
ments.
18. Let R be a PID. Prove that an ideal I 6= 0 is a prime ideal if and only if
I = Rp for some prime p.
19. Let R be a PID, and let I be an ideal of R such that 0 6= I 6= R. Prove
that I = P1 P2 Pn , where each Pi is a prime ideal. Prove that if
I = Q1 Q2 Qm with each Qi prime, then m = n, and after suitable
renumbering, Pi = Qi for all i. That is, prove that every ideal 6= R or 0
is uniquely a product of prime ideals.
20. Call an ideal I in a commutative ring R with an identity primary if I 6= R
and if ab 2 I and a 2 = I, then bn 2 I for some n. Prove that prime ideals
are primary. Prove that if R is a PID, then I is primary if and only if
I = P n for some prime ideal P . Prove that if R is a PID and if P m = Qn
for prime ideals P and Q of R, then P = Q and m = n.
21. Let I be an ideal in the PID R, with I 6= R. Prove that
I = Q1 \ Q2 \ \ Qm
Qi = Pini
for all i.
136 CHAPTER 4. RINGS AND MODULES
22. Let R be a PID, and let a and b be in R with (a; b) = 1. Prove that
1 + 2x + 0x2 , and
1 + 2x
represent the same element in Z[x]. More generally, the symbols
a0 + a1 x + + an xn
and
b0 + b1 x + + bm x m
represent the same element in R[x] if and only if ai = bi whenever either ai or
bi 6= 0. This de…nes an equivalence relation, and we could de…ne R[x] to be the
set of its equivalence classes, and addition and multiplication accordingly. This
would give us what we want, but we prefer another way that is equally precise.
Let R be any ring, and let R[x] be the set of all mappings f from the set
fx0 , x1 , x2 , g into the set R such that f (xn ) = 0 for all but …nitely many
4.5. POLYNOMIAL RINGS 137
n. The largest n such that f (xn ) 6= 0, if such exists, is called the degree of f .
If no such n exists, then f has degree 1 . In R[x], add and multiply by the
rules
(f + g)(xn ) = f (xn ) + g(xn ),
and X
(f g)(xn ) = f (xi )g(xj ).
i+j=n
This makes R[x] into a ring. There are many things to check, but they are all
easy. Note …rst that f + g and f g are actually in R[x]. The most di¢ cult thing
to check is the associative law for multiplication, which we will do now.
!
X X X
n i j
((f g)h)(x ) = (f g)(x )h(x ) = f (x )g(x ) h(xj )
k m
Similarly, X
(f (gh))(xn ) = f (xk )g(xm )h(xj ),
k+m+j=n
and so the associative law for multiplication holds. The remaining details in
showing that R[x] is a ring are left to the reader.
De…nition 4.5.1 The ring R[x] is the ring of polynomials in x with coef-
…cients in R.
whenever f (xm ) = 0 for all m > n. That is, every element of R[x] is a sum of
monomials. When are two sums
a0 x0 + a1 x1 + + a n xn
and
b0 x 0 + b1 x 1 + + bm x m
equal? By the de…nition of addition in R[x], the …rst takes xi to ai for i n
and xk to 0 for k > n, and similarly for the second. Thus these two sums of
monomials are equal if and only if ai = bi , with the convention that ai = 0
for i > n and bi = 0 for i > m. That is, they are equal if and only if they
are equal as polynomials as we ordinarily think of polynomials. How do such
138 CHAPTER 4. RINGS AND MODULES
sums of monomials add and multiply? Noting that (axm )(bxn ) = (ab)xm+n ,
and that axm + bxm = (a + b)xm by their very de…nition, we see that they add
and multiply just as we should add and multiply polynomials.
Some notational changes are in order. If R has an identity, then the mono-
mial 1xn is written simply as xn . In any case, x1 is written x, and the monomial
ax0 is written simply as a. Hence we can write any element in R[x] in the form
a 0 + a 1 x + a 2 x2 + + an xn
with the ai 2 R, and we have already observed how unique such a representation
is. The letter x is called the indeterminate, and the ai the coe¢ cients. We
could just as well have used y, or z, or anything for the indeterminate.
Having R[x], we can form R[x][y], or more generally, R[x1 ][x2 ] [xn ]. How-
ever, R[x][y] R[y][x] in the following special way. An element in R[x][y] is a
polynomial in y with coe¢ cients in R[x], and is the sum of monomials. Let
(a0 + a1 x + + an xn )y m
be a monomial in R[x][y]. Since
(a0 + a1 x + + an xn )y m = a0 y m + a1 xy m + + an xn y m ,
every element in R[x][y] is a sum of monomials of the form aij xi y j with aij 2 R.
Associate with aij xi y j the monomial aij y j xi in R[y][x]. Now the mapping
R[x][y] ! R[y][x] that sends aij xi y j to aij y j xi is an isomorphism. The rings
R[x][y] and R[y][x] are usually identi…ed and denoted R[x; y]. Since any element
in R[x][y] can be written in the form aij xi y j with aij 2 R, the elements in
R[x][y] are “polynomials in x and y” with coe¢ cients in R. Similar remarks
apply to R[x1 ][x2 ] [xn ] . We leave the details to the reader and will not further
concern ourselves with “polynomials in several variables” at the moment.
We are not going to study general polynomial rings in any depth. Our
primary interest is in F [x] where F is a …eld, and at this point it is convenient
to restrict attention to rings D[x] where D is an integral domain. So from now
on, D will be an integral domain.
The map D ! D[x] : a ! ax0 is clearly a ring monomorphism. In fact, we
are already writing ax0 simply as a. Therefore, we consider D as a subring of
D[x]. The elements of D in D[x] are called constants. The reason is this. For
any element a0 + a1 x + an xn 2 D[x], there is associated a mapping from D
into D de…ned by
d ! a0 + a1 d + + an dn .
The constant polynomial a0 corresponds to the constant function D ! D : d !
a0 .
An element of D[x] will typically be denoted p(x). If p(x) = a0 + a1 x +
+ an xn and d 2 D, then p(d) denotes the element a0 + a1 d + + an dn of
D. That is, p(d) is “p(x) evaluated at d.” Thus with each p(x) 2 D[x] we have
associated a map from D into D. Letting Map(D; D) denote the set of all maps
from D into D, we have then a map
: D[x] ! Map(D; D).
4.5. POLYNOMIAL RINGS 139
Theorem 4.5.2 If p(x) and q(x) are in D[x] and d 2 D, then p(x) + q(x)
evaluated at d is p(d) + q(d), and p(x)q(x) evaluated at d is p(d)q(d).
: D[x] ! Map(D; D)
a0 , a1 , a2 , , an 2 D such that
f (d) = a0 + a1 d + + an dn
for all d 2 D. Now one could de…ne D[x] as the set of all these polynomial
functions on D, and then verify that it is a subring of the ring Map(D; D). In
fact, in elementary analysis, we are used to thinking of polynomials (with real
coe¢ cients, say) as functions. In any case, if D is an in…nite integral domain,
the ring D[x] and the ring of polynomial functions on D may be identi…ed— they
are isomorphic.
If
a0 x0 + a1 x1 + + an xn 6= 0,
140 CHAPTER 4. RINGS AND MODULES
1+n=n+ 1= 1+ 1= 1
for all non-negative integers n. We write deg(p(x)) for the degree of the poly-
nomial p(x).
Theorem 4.5.3 If p(x) and q(x) are in D[x] and if D is an integral domain,
then
deg(p(x)q(x)) = deg(p(x)) + deg(q(x)).
Corollary 4.5.5 If D is an integral domain, then the units of D[x] are the
units of D.
Corollary 4.5.6 If F is a …eld, then the units of F [x] are the non-zero elements
of F .
g(x) = a0 + a1 x + + am xm 2 D[x]
f (x) = b0 + b1 x + + bn xn 2 D[x];
we want to write f (x) = g(x)q(x) + r(x) with q(x) and r(x) in D[x] and
deg(r(x)) < deg(g(x)). If deg(f (x)) < deg(g(x)), it is easy. Just let q(x) = 0
and r(x) = f (x). If not, n m, and
f (x) am1 bn xn m
g(x)
with deg(h1 (x)) < deg(f (x)). If deg(h1 (x)) deg(g(x)), write h1 (x) = q2 (x)g(x)+
h1 (x) with deg(h2 (x)) < deg(h1 (x)). Eventually we get
with
deg(r1 (x)) < deg(g(x));
then
g(x)(q(x) q1 (x)) = r1 (x) r(x):
Unless
q(x) q1 (x) = 0;
deg(g(x)(q(x) q1 (x))) > deg(r1 (x) r(x)):
Therefore q(x) = q1 (X) and r(x) = r1 (x). Thus we have
Proof. Write f (x) = (x a)q(x) + r(x) with deg(r(x)) < deg(x a). We
can do this because the leading coe¢ cient of x a is 1, which is a unit in D.
Since deg(x a) = 1, r(x) = d 2 D. Hence f (a) = (a a)q(a) + d = d.
Corollary 4.5.9 Let D be an integral domain, let a 2 D, and let f (x) 2 D[x].
Then x a divides f (x) in D[x] if and only if f (a) = 0.
142 CHAPTER 4. RINGS AND MODULES
Theorem 4.5.12 Let F be a …eld. Then every non-zero element p(x) 2 F [x]
can be written in the form
with b 2 F and each q(x) monic prime, then a = b, m = n, and after suitable
renumbering, pi (x) = qi (x) for all i.
Theorem 4.5.13 Let F be a …eld. Then every non-zero element p(x) 2 F [x]
can be written in the form
Then pk is the greatest common divisor of p1 and p2 , and from the equations
above, pk may be written as a linear combination of p1 and p2 . The procedure
is entirely analogous to that for the integers, and again is called the Euclidean
Algorithm (1.6, page 19).
PROBLEMS
1. Prove that R[x] has an identity if and only if R does.
is a homomorphism.
9. Let F be a …eld, and let K be the quotient …eld of F [x]. For a 2 F , let
Ka ! F : f (x)=g(x) ! f (a)=g(a)
10. Prove that if F is a …eld with n elements, and if the non-zero element
p(x) 2 F [x] has degree m, then F [x]=(p(x)) has nm elements.
22. Let F be a …eld and let S be the subring of the quotient …eld of F [x]
generated by 1=x and F [x]. Let G be the in…nite cyclic group. Prove that
F (G) S.
23. De…ne R[x1 ; x2 ; x3 ; : : : ] and show that it is a ring.
24. Show that R R[x] is possible. (R 6= 0:)
25. Let R be any ring. De…ne the ring Rfxg of “power series” a0 + a1 x +
a2 x2 + and show that it is a ring.
26. Prove that if F and K are …elds and F [x] K[x] as rings, then F K
as rings.
27. Prove that if F is a …eld and the rings F [x] and R[x] are isomorphic, then
so are the rings F and R.
28. Prove that if the rings Z[x] and R[x] are isomorphic, then so are Z and
R.
29. Prove that R[x]=(x r) and R are isomorphic as rings.
30. Let F be any …eld. Prove that the rings F [x]=(x2 +1) and F [x]=(x2 +4x+5)
are isomorphic.
31. Prove that the rings R[x; y]=(x) and R[y] are isomorphic.
32. Find a non-zero prime ideal in Z[x] that is not maximal.
33. Let D be an integral domain. Find a non-zero prime ideal in D[x; y] that
is not maximal.
34. Use the Euclidean algorithm to …nd the greatest common divisor of x6 1
and x14 1 in Q[x]. Express that greatest common divisor as a linear
combination of the two polynomials.
35. Use the Euclidean algorithm to …nd the greatest common divisor of x4 + 1
and x6 + x4 + x3 + 1 in Q[x]. Express that greatest common divisor as a
linear combination of the two polynomials.
4.6 Modules
The concept of module generalizes both that of an Abelian group and that of a
vector space. The purpose of this section is to introduce the reader to this more
general notion and to lay the groundwork for the theorems in section 4.7.
A vector space is an Abelian group V , a …eld F , and a map F V ! V
satisfying certain rules (3.1.5). Let G be any Abelian group. We know how to
“multiply”elements of Z times elements of G, getting elements of G. In fact, if
n is a positive integer and g 2 G, ng = g + g + +g, the sum of g with itself
n times. Letting ( n)g = (ng) and 0g = 0 completes the de…nition. Thus
4.6. MODULES 147
we have a map Z G ! G. This map satis…es the same rules that the map
F V ! V does. That is, it satis…es
d. 1g = g for all g 2 G.
Thus a module would be a vector space if the ring involved were a …eld.
In the de…nition above, the scalars, that is, the elements of R, are written on
the left, and we say that M is a left module over R, or that M is a left
R-module. Right modules are de…ned in the obvious way. Unless speci…cally
stated otherwise, module will mean left module. Thus when we say “Let M be
a module over R.”, we mean that M is an Abelian group, R is a ring with an
identity, and there is a map R M ! M on hand satisfying (a)–(d) above. From
the remarks made before 4.6.1, we see that every Abelian group is a module over
Z, and vector spaces are modules over …elds. Another example is this. Let R
be any ring with identity. Now R is an Abelian group under +, and there is a
map R R ! R satisfying (a)–(d), namely the ring multiplication of R. Thus
any ring is a module over itself.
For the rest of this section, ring means ring with identity, and by a subring
of R we mean one that contains the identity of R:
Let M be a module over R, let r and s be in R, and let m and n be in M .
Then the following hold, and can be proved just as they were for vector spaces
(3.1.6).
(a) r0 = 0.
(b) 0m = 0.
a. If f : M ! N is a homomorphism, then
M=(Ker f ) Im f .
(S + T )=T S=(S \ T ).
There are a couple of new concepts that we need. Let R be any ring. Recall
that R is module over itself. What are the submodules of R? They are subgroups
I of R closed under multiplication on the left by elements of R, that is, subgroups
I such that ri 2 I for all r 2 R and i 2 I.
0 1
m= ,
0 0
then Rm is the left ideal consisting of those matrices with …rst column all zeroes.
However, it is not a right ideal since m is in Rm but m m0 is not, where m0 is
the transpose of m.
Now let M be a module over R and let m 2 M . Consider the set 0 : m = fr 2
R : rm = 0g. Since r, s 2 0 : m implies (r s)m = rm sm = 0 0 = 0, 0 : m
is a subgroup of R. If r0 2 R and r 2 0 : m, then (r0 r)m = r0 (rm) = r0 0 = 0.
so that 0 : m is a submodule of R, that is, a left ideal of R.
De…nition 4.6.3 Let M be a module over R, and let m 2 M . The left ideal
0 : m = fr 2 R : rm = 0g is the order ideal of m.
150 CHAPTER 4. RINGS AND MODULES
Theorem 4.6.4 If M is a cyclic R-module, then M R=I for some left ideal
I of R. In fact, if M = Rm, then
M R=(0 : m).
M1 M2 Mk
(m1 , m2 , , mk ) + (n1 , n2 , , nk ) =
(m1 + n1 , m2 + n2 , , mk + nk );
and
r(m1 , m2 , , mk ) = (rm1 , rm2 , , rmk ).
This construction is the same as the one we did for vector spaces. It is
straightforward to show that M = M1 M2 Mk is a module over R.
This module is called the external direct sum of the modules M1 , M2 , ,
Mk . If we let Mi be those elements of M of the form (0, 0, , mi , 0, 0 ,
, 0), where mi is in the ith place, then Mi is a submodule, and as in the
case of vector spaces, every element of M can be written uniquely in the form
m1 + m2 + +mk with mi 2 Mi . When we have such a situation, namely
when M has submodules N1 , N2 , , Nk such that every element in M can
be written uniquely in the form n1 + n2 + + nk with ni 2 Ni , we say that
the module M is the internal direct sum of these submodules. We also write
M = N1 N2 Nk when M is the internal direct sum of the submodules N1 ,
4.6. MODULES 151
and
M = Ker ' + Im '.
Let m 2 (Ker ') \ (Im '). Then m = '(m0 ) for some m0 2 M , and '(m) =
'2 (m0 ) = '(m0 ) = m = 0 since m 2 Ker ' and '2 = '. Let m 2 M .
Then m = (m '(m)) + '(m). Clearly '(m) 2 Im '. Also '(m '(m)) =
'(m) '2 (m) = '(m) '(m) = 0. Thus m '(m) is in Ker '. We have
proved the following theorem.
Good theorems about modules are ones that say that certain modules are
direct sums of simpler ones. We have two examples already. Every …nite Abelian
group, that is, every …nite module over the ring Z, is a direct sum of cyclic
modules (2.2.7). Every …nitely generated module over a …eld is a direct sum of
cyclic modules (3.3.12). In the next section (4.7.13) we will prove that a …nitely
generated module over a PID is a direct sum of cyclic modules, of which both
the examples above are special cases.
We pause here to look at a simple example illustrating some of our results
so far. Let R be the ring of all 2 2 matrices over the …eld Q. Let M be the set
of all 2 3 matrices over Q. Now M is an Abelian group under matrix addition,
and matrix multiplication makes M into a left module over R. Let Mi be the
matrices in M with zeroes o¤ the ith column. That is,
whose entries in the ith column are zero. Therefore, Mi is a direct summand of
M by 4.6.5. However, it should be clear directly from the de…nition of internal
direct sum that
M = M1 M2 M3 .
Further, Mi is isomorphic to the module C of 2 1 matrices over Q, and M is
isomorphic to the (external) direct sum of three copies of C.
Let
1 0 0 0 1 0 0 0 1
m1 = , m2 = , m3 = .
0 0 0 0 0 0 0 0 0
Then Mi = Rmi , so that the Mi are cyclic. The left ideal 0 : mi consists of
those matrices in R with …rst column all zeroes. By 4.6.4, R=0 : mi Mi ,
and it follows that the Mi are isomorphic to one another. Using the fact that
Q is a …eld, it can be shown that Mi has no submodules other than 0 and Mi
itself. Thus, Mi = Rm for any non-zero element m of Mi . Therefore R=0 : m
is isomorphic to Mj for any m in Mi , i, j = 1, 2, 3. The 0 : m are not all the
same. For example, if
0 0 0
m= ,
0 1 0
then 0 : m consists of those matrices of R whose second column have only zero
entries. Still, R=0 : m R=0 : mi for any i.
Consider the R-module R. It has a special element in it, namely 1. Since
R = R 1, R is cyclic. Furthermore, if r 1 = s 1, then r = s. Thus every
element r 2 R can be written in exactly one way as an element of R times 1,
namely r 1. Suppose that M is a cyclic R-module with a generator m such
that rm = sm only if r = s. Then
R ! M : r ! rm
Therefore, f = g.
Theorem 4.6.8 Let M be any …nitely generated R-module. Then there exists
an epimorphism F ! M with F a free R-module.
IM = fi1 m1 + i2 m2 + + ik mk : ij 2 I, mj 2 M , k > 0g
PROBLEMS
1. Let S be a subring of the ring R. Prove that an R-module is an S-module.
In particular, R is an S-module.
2. Let ' : S ! R be a ring homomorphism with '(1) = 1. Let M be an
R-module. Prove that scalar multiplication de…ned by s m = '(s)m also
makes M into an S-module.
3. In Problem 2, let R = S, and let ' be an automorphism of the ring
R. Thus M is a module over R in two ways. Are these two modules
isomorphic?
4.6. MODULES 155
5. Prove that there is only one way to make an Abelian group into a module
over Z.
6. Prove that there is at most one way to make an Abelian group into a
module over Q.
8. Prove that the additive group of Z cannot be made into a module over Q.
is a submodule.
14. Let M be an R-module, and let EndZ (M ) be the ring of all group endo-
morphisms of M . For r 2 R, de…ne 'r by
'r : M ! N : m ! rm.
R ! EndZ (M ) : r ! 'r
is a ring homomorphism.
15. Let R be a ring, and let r 2 R. Let 'r : R ! R : s ! sr. Prove that 'r
is an endomorphism of the R-module R. Prove that every endomorphism
of the module R is of this form. Prove that f : R ! EndR (R) : r ! 'r
is a ring anti-isomorphism. That is, f is one-to-one and onto, preserves
addition, and reverses multiplication.
156 CHAPTER 4. RINGS AND MODULES
16. Let HomR (M; N ) be the set of all homomorphisms from the R-module M
into the R-module N . Prove that
I : r = fx 2 R : xr 2 Ig.
Theorem 4.7.2 Let R be a PID, and let M be an R-module. The set Mt of tor-
sion elements of M is a submodule of M . (It is called the torsion submodule
of M .)
Lemma 4.7.4 Let R be a PID, and let F be its quotient …eld. Then any …nitely
generated R-submodule of F is cyclic (and hence free).
Therefore
Theorem 4.7.6 Two …nitely generated torsion-free modules over a PID are
isomorphic if and only if they have the same rank.
Theorem 4.7.7 Let M be a free module of rank n over a PID R. Then any
submodule of M is free and of rank n.
Now let’s turn to the case of …nitely generated torsion R-modules. Sup-
pose that M is such a module, and suppose that fm1 , m2 , , mk g generates
M . Since M is torsion, there exist non-zero ri ’s such that ri mi = 0. Then
(r1 r2 rk )mi = 0 for all i. If m 2 M , then m = s1 m1 + s2 m2 + + sk mk for
some si 2 R, and it follows that (r1 r2 rk )m = 0. That is, (r1 r2 rk )M = 0.
But r1 r2 rk 6= 0. We have that
0 : M = fr 2 R : rM = 0g =
6 f0g.
xi = (r=pni i )mi .
Since
pni i 1
xi = (r=pi )mi 6= 0;
and
pni i xi = rmi = 0;
then 0 : xi = Rpni i . By 4.7.10,
Theorem 4.7.12 Every …nitely generated torsion module over a PID is a direct
sum C1 C2 Ct of cyclic modules such that 0 : C1 0 : C2 0 : Ct .
and
n n
1;q+1
0 : m12 = Rpq+1 2;q+2
pq+2 pnt 1t ,
with n1i < ni for i > q. Thus by Problem 9,
n
0 : (m12 + xq+1 + q+1
+ xt ) = Rpq+1 pnt t .
with Mi =Rm cyclic and 0 : (M1 =Rm) 0 : (M2 =Rm) 0 : (Mv =Rm).
Let Mi =Rm = R(wi + Rm), and let 0 : (wi + Rm) = Rsi . Then si wi = ui m for
some ui 2 R. Also si divides r. We have si ti = r, si ti wi = ti ui m = 0, so that
ti ui = rri , and si ti ui = si rri = rui . Hence si ri = ui . We have si wi = ui m =
si ri m, so si (wi ri m) = 0. Let wi ri m = yi . Then Mi = Ryi + Rm and Ryi \
Rm = f0g. therefore, Mi = Ryi Rm, and hence M = Rm Ry1 Ryv .
Since Ryi Mi =Rm, then 0 : Rm 0 : Ry1 0 : Ry2 Ryv , and the
theorem is proved.
C1 C2 Cn = D1 D2 Dm
Proof. Let M be a …nitely generated module over the PID R, and let Mt
be the torsion submodule of R. then M=Mt is a torsion-free …nitely generated
module. Hence, by 4.6.7, M = F Mt with F free of …nite rank. Now Mt is
…nitely generated since it is a homomorphic image of M . Since F = C1 C2
Ck with Ci R, then 0 : Ci = 0, and 4.7.12 yields
M = C1 C2 Cn
M = D1 D2 Dm
4.7. MODULES OVER PRINCIPAL IDEAL DOMAINS 163
rj M = rj C1 rj C2 rj Cj 1 = rj D1 rj D2 :. rj Dn :
0 : rj C1 0 : rj C2 0 : rj Cj 1
Theorem 4.7.18 Every …nitely generated torsion module over a PID is a direct
sum of primary modules.
Proof. It su¢ ces to prove the theorem for cyclic modules. So let M = Rm,
0 : m = Rr, and
r = pn1 1 pn2 2 pnk k ;
where the pi are non-associates. Let qi = r=pni i . Then
0 : qi m = Rpni i ,
Since the qi0 s are relatively prime, there exist ri 2 R such that ri qi = 1. Hence
m = ri qi m, so that
M = Mp1 M p2 M pk .
M = M p1 M p2 M pk .
Proof. Only the last assertion needs verifying, and it should be obvious at
this point.
Corollary 4.7.20 Two …nitely generated torsion modules M and N over a PID
R are isomorphic if and only if Mp Np for all primes p 2 R.
The invariant factors of the various Mp are called the elementary divisors
of M .
PROBLEMS
1. Prove that a …nitely generated torsion module over Z is …nite.
2. Prove that a …nitely generated torsion module over a PID is not necessarily
…nite.
3. Prove that quotient modules of torsion modules are torsion modules.
4. Prove that direct sums of torsion modules are torsion modules.
5. Prove that direct sums of torsion-free modules are torsion-free modules.
6. Prove that if M is a free module (of …nite rank) over a PID and if S is
a submodule of M , then S is contained in a summand of M having the
same rank as S.
7. Let R be a ring such that every left ideal is …nitely generated as an R-
module. Prove that submodules of …nitely generated R-modules are …-
nitely generated.
8. Let R be a PID, and let p be a prime in R. Prove that if R=Rpi R=Rpj ,
then i = j.
9. If M is a module over a PID R, and x and y are in M with 0 : x =
Rpm 1 m2
1 p2 pm n1 n2
t , 0 : y = Rp1 p2
t
pnt t , and 0 mi < ni for all i, then
0 : (x + y) = 0 : (y).
10. Prove that if R is a commutative ring with identity, I and J are ideals of
R, and the R-modules R=I and R=J are isomorphic, then I = J.
11. Let S be a submodule or a quotient module of the direct sum C1 C2
Cn of cyclic modules Ci over a PID, where 0 : C1 0 : C2 0 : Cn .
Prove that if
S = S1 S2 Sm
with each Si non-zero cyclic and 0 : S1 0 : S2 0 : Sm , then
m n and 0 : Ci 0 : Si for all i m.
4.7. MODULES OVER PRINCIPAL IDEAL DOMAINS 167
13. How does one tell from the invariant factors of two …nitely generated
torsion modules whether or not one is isomorphic to a summand of the
other? to a submodule of the other?
14. Let C be a cyclic module over the PID R. Let M and N be …nitely
generated R-modules. Prove that if C M C N , then M N . (This
is true for arbitrary R-modules M and N .)
16. Prove that submodules, quotient modules, and direct sums of p-modules
are p-modules.
18. Let p be a prime in the PID R, and let C be a cyclic p-module over R.
Prove that C R=Rpn for some n.
19. Let C be a cyclic p-primary module with 0 : C = Rpn . Prove that the
submodules of C are exactly the submodules in the chain
C pC pn C = 0.
20. Let M be a module over the PID R, and let p be a prime in R. Let
M [p] = fm 2 M : pm = 0g. Prove that M [p] is a submodule of M , and
hence a vector space over the …eld R=pR.
21. Let M be a …nitely generated module over a PID R. Prove that for any
non-zero prime p in R,
22. Let p be a prime in the PID R, and let C be a cyclic p-module. Prove
that dim(C[p]) = 1. Prove that
23. Let M be a …nitely generated p-primary module over the PID R. Prove
that
dim((pk M )[p]=(pk+1 M )[p])
is the number of summands which are isomorphic to R=pk+1 R in any
decomposition of M into a direct sum of cyclic modules.
24. Let F be the …eld of integers modulo 2. Use the Euclidean algorithm to
…nd the greatest common divisor of x2 +1 and x5 +1 in F [x]. Express that
greatest common divisor as a linear combination of the two polynomials.
Chapter 5
Linear Transformations
Summing up, Hom(V; V ) is at the same time a ring and a vector space over
F . The ring addition is the same as the vector space addition. Furthermore,
scalar multiplication and ring multiplication are connected by the relations
a( ) = (a ) = (a ):
169
170 CHAPTER 5. LINEAR TRANSFORMATIONS
De…nition 5.1.1 Let V be a …nite dimensional vector space over the …eld F .
Let fv1 , v2 , : : : , vn g be a basis of V , let be a linear transformation from V
into V , and let (vj ) = ni=1 aij vi . The matrix of relative to the basis
fv1 , v2 , : : : , vn g is 0 1
a11 a12 a1n
B a21 a22 a2n C
B C
B .. .. .. C
@ . . . A
an1 an2 ann
Let Fn denote the set of all such matrices. That is, Fn is the set of all n n
arrays of elements of F . Elements of Fn are typically denoted (aij ). Thus (aij )
is the matrix in Fn whose entry in the ith row and jth column is the element
aij of F . Given any (aij ) 2 Fn , we may de…ne the linear transformation
by (vj ) = aij vi . This, together with our previous remarks, shows that
the map A(V ) ! Fn gotten by associating with 2 A(V ) its matrix relative
5.1. LINEAR TRANSFORMATIONS AND MATRICES 171
The units of a ring are those elements with inverses. For example, is a
unit in A(V ) if there is an element in A(V ) with = = 1, the identity
map on V . Elements in A(V ) and elements in Fn that are units are called
non-singular. This is special linear algebra terminology.
There are some facts about non-singular linear transformations and matrices
that we need. First, if (aij ) is the matrix of relative to some basis, it should
be clear from 5.1.3 that is non-singular if and only if (aij ) is non-singular.
Indeed, if = = 1, then in the notation of 5.1.3,
( )= ( )=1= ( ) ( )= ( ) ( ):
i. is non-singular.
ii. has a right inverse.
iii. has a left inverse.
iv. is onto.
v. is one-to-one.
i. (aij ) is non-singular.
ii. (aij ) has a right inverse.
5.1. LINEAR TRANSFORMATIONS AND MATRICES 173
Therefore (aij )(cij ) = (cij )(bij ) since each side is the matrix of with
respect to fv1 , v2 , : : : , vn g. Now is non-singular, so therefore (cij ) is non-
singular. Hence the relation (aij )(cij ) = (cij )(bij ) can just as well be expressed
as
(aij ) = (cij )(bij )(cij ) 1
This is expressed by saying that (aij ) is similar to (bij ). Similarity of matrices
is an equivalence relation on Fn . In the same vein, two linear transformations
and in A(V ) are called similar if there is a non-singular 2 A(V ) with
1
= . Similarity is an equivalence relation on A(V ).
a. and are similar if and only if they have the same matrix relative to
appropriate bases.
b. (aij ) and (bij ) are similar if and only if they are matrices of the same
linear transformation relative to appropriate bases.
Proof. What has been shown already is one half of (b), namely that if (aij )
and (bij ) are the matrices of relative to two bases, then (aij ) and (bij ) are
similar. To get the other half of (b), suppose that
1
(aij ) = (cij )(bij )(cij ) .
Let fv1 , v2 , : : : , vn g be any basis of V . Let (aij ), (bij ), and (cij ) be the matrices
of , , and respectively, relative to this basis. Since is non-singular,
1
is a basis of V . We have = by 5.1.3. Since
1
( (vj )) = ( (vj )) = (vj ) = ( bij vi ) = bij (vi );
(bij ) is the matrix of relative to the basis f (v1 ), (v2 ), : : : , (vn )g. This
completes the proof of (b).
1
The proof of (a) is about the same. Suppose that = . Let fv1 , v2 ,
: : : , vn g be any basis of V . Then f (v1 ), (v2 ), : : : , (vn )g is also a basis of
V . Let (vj ) = bij vi . Then
whence the matrix of relative to f (v1 ), (v2 ), : : : , (vn )g is the same as the
matrix of relative to fv1 , v2 , : : : , vn g
Now suppose that the matrix (aij ) of relative to a basis fv1 , v2 , : : : , vn g is
also the matrix of relative to a basis fw1 , w2 , : : : , wn g. Let be the element
of A(V ) de…ned by (wj ) = vj . Then
1
(vj ) = (wj ) = ( aij wi ) = aij ( wi ) = aij vi :
1
Therefore = . The proof is complete.
What does all this really say? Given a basis of V , we have an isomorphism
A(V ) ! Fn :
PROBLEMS
1. Prove that (a + bi) = a bi is a linear transformation of the complex
numbers regarded as a vector space over the …eld of real numbers. Find
the matrix of relative to the basis f1; ig.
176 CHAPTER 5. LINEAR TRANSFORMATIONS
14. Suppose that is in A(V ) and that is idempotent, that is, that 2 = .
Prove that V = V1 V2 with (v) = v for all v 2 V1 and (V2 ) = 0.
15. Suppose that (aij ) is idempotent. Prove that (aij ) is similar to a matrix
of the form 0 1
1 0 0 0 0 0
B 0 1 0 0 0 0 C
B C
B .. .. .. C
B . . . C
B C
B 0 0 0 1 0 0 C
B C
B 0 0 0 0 0 0 C
B C
B . .. .. C
@ . . . . A
0 0 0 0 0 0
16. Prove that if and are idempotent elements of A(V ), then and
are similar if and only if they have the same rank. (The rank of is
dim(Im( )).)
17. Let R be the …eld of real numbers. Prove that the n n matrix of all 1’s
is similar to the n n matrix (aij ) with a11 = n and all other entries 0.
Lemma 5.2.3 Let f (x) be in F [x], and let deg(f (x)) = n. Then as a vector
space over F , F [x]=(f (x)) is of dimension n. In fact, f1 + (f (x)), x + (f (x)),
: : : , xn 1 + (f (x))g is a basis.
where ff1 (x), : : : , fk (x)g are the invariant factors of V . Since f1 (x) is a multi-
ple of each fi (x), it follows that f1 (x)(F [x]=(fi (x))) = 0. Therefore, f1 (x)V =
0. What does this mean? It means that for each v in V , f1 (x) v = 0. But
f1 (x) v = f1 ( )(v) = 0. Therefore, f1 ( ) = 0. The polynomials g(x) in F [x]
such that g( ) = 0 form an ideal. This ideal is non-zero. Any non-zero ideal of
F [x] is generated by a unique monic polynomial. Clearly, no monic polynomial
g(x) of degree smaller than deg(f1 (x)) can satisfy g( ) = 0. If it did, then
g(x)V = 0 = g(x)(F [x]=(f1 (x))), which is impossible. Therefore, f1 (x) is the
monic generator of the ideal of polynomials g(x) such that g( ) = 0. In other
words, f1 (x) is the monic polynomial of least degree such that f1 ( ) = 0. But
what is the degree of f1 (x)? By 5.2.3, or 5.2.5, deg(f1 (x)) dim(V ). Using our
isomorphism A(V ) ! Fn , we have
Note that 5.2.6 just says that these minimum polynomials are of degree
dim(V ).
Theorem 5.2.9 Let V be a …nite dimensional vector space over a …eld F , and
let and be in A(V ). Then the F [x]-modules V and V are isomorphic if
and only if and are similar.
180 CHAPTER 5. LINEAR TRANSFORMATIONS
De…nition 5.2.10 Let V be a …nite dimensional vector space over the …eld F ,
and let be in A(V ). The invariant factors, and the elementary divisors, of the
module V are called the invariant factors, and the elementary divisors,
respectively, of the linear transformation .
Corollary 5.2.11 Let and be in A(V ). Then and are similar if and
only if they have the same invariant factors, or the same elementary divisors.
De…nition 5.2.12 Let (aij ) be in Fn , and let be any element of A(V ) whose
matrix relative to some basis of V is (aij ). The invariant factors, and ele-
mentary divisors of (aij ) are the invariant factors, and elementary divisors,
respectively, of .
Corollary 5.2.13 Two matrices in Fn are similar if and only if they have the
same invariant factors, or the same elementary divisors.
Corollary 5.2.14 Let be in A(V ), with dim(V ) = n. Then and (aij ) have
the same invariant factors, or the same elementary divisors, if and only if (aij )
is the matrix of relative to some basis of V .
How does one tell whether or not two matrices in Fn have the same invariant
factors? This is the same as asking how one tells whether or not two matrices in
Fn are similar. It turns out that there is a matrix in each similarity class that
5.2. THE RATIONAL CANONICAL FORM FOR MATRICES 181
puts on display its invariant factors. This particular canonical form (that is, this
special matrix in each similarity class) is called the rational canonical form.
We will get at it through V . Now V is a direct sum of cyclic modules. Let’s
examine the cyclic case …rst. So suppose that V is cyclic. This is expressed
by saying that is cyclic. So a linear transformation on a vector space V is
cyclic if the associated module V is a cyclic module. Suppose that is cyclic.
Then V F [x]=(m (x)), where m (x) is the minimum polynomial of . Let
v generate V . By 5.2.4,
n 1
B = fv, (v), : : : , (v)g
This matrix is, of course, uniquely determined by the polynomial m (x), and
one can read o¤ m (x) from it.
Corollary 5.2.17 The minimum polynomial of the companion matrix C(f (x))
of the monic polynomial f (x) is f (x). That is, mC(f (x)) (x) = f (x).
Theorem 5.2.18 Let V be a vector space of dimension n over the …eld F . Let
be a linear transformation on V , and let ff1 (x), f2 (x), : : : , fk (x)g be the
invariant factors of . Then V has a basis such that the matrix of relative to
it is 0 1
C(f1 (x))
B C(f2 (x)) 0 C
B C
R( ) = B .. C
@ 0 . A
C(fk (x))
5.2. THE RATIONAL CANONICAL FORM FOR MATRICES 183
as a real matrix has invariant factors just the polynomial x2 +1. The polynomial
x2 + 1 is also, of course, its only invariant factor as a complex matrix. But its
elementary divisors as a complex matrix are x + i and x i, while it has only
the single elementary divisor x2 + 1 as a real matrix. Its rational canonical form
both as a real and as a complex matrix is the matrix itself. In the next section,
we will get a canonical form that will display the elementary divisors. That
canonical form of a matrix might change if the …eld is enlarged, because the
elementary divisors might change.
PROBLEMS
1. Find the minimum polynomial of an idempotent linear transformation.
2. Find the minimum polynomial of a nilpotent linear transformation.
3. Find the minimum polynomial of the 3 3 matrix whose ij-entry is i ij .
m
9. Let be in A(V ), and suppose that is nilpotent. That is, = 0 for
some m > 0. Prove that has a matrix of the form
0 1
N1
B N2 0 C
B C
B N3 C
B C,
B .. C
@ 0 . A
Nk
and n1 n2 nk .
10. Prove that the set of similarity classes of nilpotent matrices in Fn is in
one-to-one correspondence with the set of partitions n1 + n2 + + nk = n
of n into the sum of positive integers with n1 n2 nk .
11. Prove that if the matrices Mi are non-singular in Fni , then the matrix
0 1
M1
B M2 0 C
B C
B M C
B 3 C
B .. C
@ 0 . A
Mk
Proof. The degree of m (x) is at most dim(V ), and hence m (x) has at
most dim(V ) distinct roots.
fx a1 , : : : , x a1 , x a2 , : : : , x a2 , : : : , x ak g
fvij : j = 1, 2, : : : , ni g:
De…nition 5.3.7 Let V be a vector space over F , and let be in A(V ). Suppose
that a is an eigenvalue of . Then E(a) = Ker( a) is the eigenspace of a.
The number dim(E(a)) is the geometric multiplicity of a.
De…nition 5.3.8 Let V be a vector space over F , and let be in A(V ). The
product of the elementary divisors (or equivalently, the product of the invariant
factors) of is the characteristic polynomial of , denoted c (x).
Note that deg(c (x)) = dim(V ). Also, similar linear transformations have
the same characteristic polynomial. Thus we de…ne the characteristic poly-
nomial of a matrix in Fn to be the characteristic polynomial of any in V
corresponding to it. (In 5.4, we will see how to compute the characteristic poly-
nomial of a matrix directly from that matrix.) Finally, note that if is in A(V )
and a is in F , then a is an eigenvalue of if and only if a is a root of c (x).
The multiplicity of a as a root of c (x) is called its algebraic multiplicity.
How do the two multiplicities of an eigenvalue compare?
Theorem 5.3.10 Let V be a vector space over F and let be in A(V ). Then
has a diagonal matrix if and only if the minimum polynomial of is a product
of distinct linear factors in F [x]:
Proof. Suppose that has a diagonal matrix. That is, suppose that V has
a basis fv1 , v2 , : : : , vn g consisting of eigenvectors of . Let f (x) = (x aj ),
where aj ranges over the distinct eigenvalues of . Then f ( )(vi ) = 0 for all
basis vectors vi , and so f ( )(v) = 0 for all v in V . Thus m (x) divides f (x), so
m (x) is the product of distinct linear factors in F [x]. (Actually, m (x) = f (x).)
Now suppose that the minimum polynomial of is a product of distinct
linear factor in F [x]. The module V is a direct sum V1 V2 Vk of cyclic
modules, and if each Vi has a basis such that the matrix of the restriction of to
Vi is diagonal, then union of these bases will be a basis of V such that the matrix
of with respect to it is diagonal. Therefore, we may assume that V is cyclic,
so that V F [x]=(m (x)). Let m (x) be the product (x ai ) of distinct
linear factors, and let fi (x) = m (x)=(x ai ). Then the fi (x) are relatively
prime and there are polynomials gi (x) in F [x] such that gi (x)fi (x) = 1.
Therefore, F [x]fi (x) = F [x]. Observe that F [x]fi (x)=(m (x)) is a submodule
of F [x]=(m (x)), is cyclic, is annihilated by x ai , and hence is isomorphic to
F [x]=(x ai ). These are primary modules for distinct primes x ai . It follows
that V is a direct sum of one dimensional subspaces, and that V has a basis
of eigenvectors.
Recalling that the minimum polynomial of a linear transformation restricted
to an invariant subspace is a factor of the original minimum polynomial, we get
the following result.
Corollary 5.3.11 Suppose that has a diagonal matrix. Let W be any sub-
module of V . Then W is a direct sum of submodules of dimension 1.
Corollary 5.3.12 Let be in A(V ). Then has a diagonal matrix if and only
if all the elementary divisors of are linear.
0 1
1 2
of (x 1)2 .
If does have a diagonal matrix diag(ai ), then V = V1 V2 Vn
with Vi F [x]=(x ai ). Thus the elementary divisors of are the members
of the family fx a1 , x a2 , : : : , x an g. The point is that if does have
a diagonal matrix, that matrix displays the elementary divisors of . Now for
any in A(V ), we are going to get a canonical matrix of that displays the
elementary divisors of and which is diagonal if has a diagonal matrix.
Let be in A(V ). Then V = V1 V2 Vk , where Vi is the pi (x)-
component of V . Each
with Vij cyclic. We want a special basis of Vij . To simplify notation, suppose
that V is cyclic primary. Thus V has a single elementary divisor p(x)q ,
where p(x) is a prime polynomial in F [x]. Of course, p(x)q is also the minimal
polynomial as well as the invariant factor of in this case. First we will do the
special case when p(x) is linear. This is an important special case. For example,
if the …eld F is the …eld of complex numbers, then every prime polynomial is
linear. Also, working through this case may illuminate the general case. So
suppose that p(x) = x — a. Let v be a generator of the module V . Consider
the family of vectors
v1 = ( a)q 1
(v), v2 = ( a)q 2
(v), : : : , vq = ( a)q q
(v).
basis.
(v1 ) = ( a)q 1
(v) = ( ( a))( a)q 1
(v)
q 1
= a( a) (v) = av1 ;
q 2
(v2 ) = ( a) (v) = ( ( a))( a)q 2
(v) +
q 1 q 2 q 1
( a) (v) = a( a) (v) + ( a) (v) = v1 + av2 ; : : : ;
(vq 1) = ( a)(v) = ( ( a))( a)(v) + ( a)2 (v)
= a( a)(v) + ( a)2 (v) = vq 2 + avq 1;
(vq ) = (v) = ( ( a))(v) + ( a)(v)
= av + ( a)(v) = vq 1 + avq .
Theorem 5.3.14 Let V be a vector space over the …eld F , and let be in A(V ).
Suppose that the minimum polynomial factors into linear factors in F [x]. Then
V has a basis such that the matrix of relative to it has the form
0 1
J1
B J2 C
B C
B . .. C
@ A
Jk
Note that it is not really proper to say the Jordan canonical form for .
There is no particular way in which to order the Ji . However, apart from
5.3. EIGENVECTORS AND EIGENVALUES; THE JORDAN CANONICAL FORM[EIGENVECTORS AND EIGE
where C(p(x)) is the companion matrix of p(x) and D is the r r matrix with
1 in the upper right hand corner and 0 elsewhere. We call J(p(x)q ) the Jordan
matrix of the prime power polynomial p(x)q :
In general, let be in A(V ), and write
V = V1 V2 Vk ,
where Vi are cyclic primary modules with elementary divisor qi (x). Pick a basis
of Vi as above. Putting these bases of the Vi together, we get a matrix of of
a certain form.
be the elementary divisors of . Then V has a basis such that the matrix of
relative to it has the form
0 1
J(q1 (x))
B J(q2 (x)) C
B C
J( ) = B .. C
@ . A
J(qk (x)) ;
Again, there is no particular order into which to put the J(qi (x)), and they
may of course not all be distinct. But up to such orderings, J( ) is unique.
Note that the matrix in 5.3.17 is the matrix in 5.3.14 in case all the ele-
mentary divisors are powers of a linear polynomial. Further, if the elementary
divisors are linear, then J( ) is indeed diagonal. Thus J( ) is diagonal if has
a diagonal matrix at all.
A matrix (aij ) is upper triangular if aij = 0 whenever j < i. Lower
triangular is de…ned analogously. When we say simply triangular, we mean
upper triangular. In case m (x) factors into linear factors in F [x], then J( )
is triangular. Therefore every matrix in Fn whose minimum polynomial is a
product of linear factors in F [x] is similar in Fn to a triangular matrix. This
special fact is easy to prove directly. It does not require information about the
modules V , for example. Furthermore, the converse holds. That is, a matrix in
Fn is similar in Fn to a triangular matrix if and only if its minimal polynomial
is a product of linear factors in F [x]. The pertinent theorem follows.
5.3. EIGENVECTORS AND EIGENVALUES; THE JORDAN CANONICAL FORM[EIGENVECTORS AND EIGE
fw2 , w3 , : : : , wn g
( akk )(F v1 + F v2 + + F vk ) F v1 + F v2 + + F vk 1;
n
so that i=1 ( aii ) annihilates V . Thus the minimum polynomial of divides
n
i=1 (x aii ), so factors into linear factors.
PROBLEMS
1. Suppose that dim(V ) = n and that 2 A(V ) has n distinct eigenvalues.
Prove that is cyclic.
2. Prove that if a 2 F is an eigenvalue of , then for f (x) 2 F [x], f (a) is an
eigenvalue for f ( ).
3. Suppose that a is an eigenvalue of the cyclic linear transformation . Let
m (x) = (x a)k f (x), where x a does not divide f (x). Prove that the
eigenvectors of belonging to a are precisely the non-zero elements of
( a)k 1 f ( )(V ).
4. Prove that if is cyclic and a is an eigenvalue of , then dim(Ker( a)) =
1.
5. Let a be an eigenvalue of , and let
6. Let V be a vector space over F and let be in A(V ). Prove that has a
diagonal matrix if and only if the characteristic polynomial c (x) of is
a product of linear factors in F [x] and the geometric multiplicity of each
eigenvalue a of is equal to its algebraic multiplicity.
7. Suppose that V = V1 V2 and that has a diagonal matrix relative to
some basis. Prove that V has a basis
fv1 , v2 , : : : , vr , w1 , w2 , : : : , ws g
20. Suppose that , , and are in A(V ) and all have triangulable matrices.
Suppose that = , = , and = . Prove that there is a
basis of V relative to which the matrices of , , and are triangular.
21. Let 2 A(V ). De…ne f on A(V ) by f ( ) = . Prove that f
is a linear transformation on the vector space A(V ) and that f has a
diagonal matrix if and only if does.
5.4 Determinants
In this section, we present an introduction to the theory of determinants. This
theory evolved in the eighteenth century, actually preceding the theory of ma-
trices. The reader is likely to be somewhat familiar with determinants, for
example, with their use in solving systems of linear equations (Cramer’s Rule,
5.4.9). Determinants also provide a means to compute the characteristic poly-
nomial of a matrix and hence of a linear transformation on a vector space (the
Cayley-Hamilton Theorem, 5.4.14).
198 CHAPTER 5. LINEAR TRANSFORMATIONS
There are essentially two ways to de…ne determinant— the classical way, and
via alternating forms. We choose the …rst, but do characterize them in terms of
alternating forms.
We shall consider determinants of square matrices with entries from an ar-
bitrary commutative ring with identity. There is a reason for this generality. It
will be necessary in any case to consider determinants of matrices with entries
from F [x], where F is a …eld. Once this is done, we may as well let the entries
come from any commutative ring with identity.
Let Sn denote the symmetric group on f1, 2, : : : , ng. Recall that an element
in Sn is even if it is the product of an even number of transpositions, and
is odd otherwise. Let ( 1) be 1 if is even and 1 if is odd. Note that
1
( 1) ( 1) = ( 1) and ( 1) = ( 1) if = . Here is the de…nition of
determinant.
De…nition 5.4.1 Let R be a commutative ring with identity, and let (aij ) be
an n n matrix with entries from R. Then the determinant of (aij ), denoted
det(aij ), is the element
X
( 1) a1 (1) a2 (2) an (n)
Each permutes f1, 2, : : : ; ng. Thus as i ranges through this set, so does (i).
1
Let = . Rearranging the terms of the product i ai (i) = i a ( (i)) (i) in
order of increasing (i) gives i ai (i) = i a (i)i . Therefore
X
det(aij ) = ( 1) a (1)1 a (2)2 a (n)n .
Lemma 5.4.3 Interchanging two rows or two columns of a matrix changes the
sign of the determinant of that matrix.
Lemma 5.4.4 If two rows or two columns of a matrix are equal, then the de-
terminant of that matrix is zero.
Proof. Note that 5.4.4 is not implied by 5.4.3. That lemma just says that
if two rows of (aij ) are equal. then det(aij ) = det(aij ), which does not imply
that det(aij ) = 0. For example, in the …eld of integers modulo 2, 1 = 1 6= 0.
200 CHAPTER 5. LINEAR TRANSFORMATIONS
Suppose that the qth and rth columns of (aij ) are the same. Let = (qr).
Then for each i, ai (i) = ai (i) , and
X Y X Y
det(aij ) = ( 1) ai (i) + ( 1) ai (i)
even i o dd i
X Y X Y
= ( 1) ai (i) ( 1) ai (i)
even i o dd i
X Y X Y
= ( 1) ai (i) ( 1) ai (i) =0
even i even i
Proof.
! !
X X Y X
det((aij )(bij )) = det aik bkj = ( 1) aik bk (i) ,
k i k
and
! !
X Y X Y
det(aij ) det(bij ) = ( 1) ai (i) ( 1) bi (i) .
i i
where f ranges over all the functions from the set f1, 2, ..., ng to itself, not just
the permutations. Given one of the functions f which is not a permutation,
there must be a pair of integers 1 s < t n with f (s) = f (t). Let be the
transposition (st). Then (i) = (i) if i 6= s, t and (s) = (t), (t) = (s).
Moreover, ( 1) = ( 1) . Thus
Y Y
( 1) aif (i) bf (i) (i) + ( 1) aif (i) bf (i) (i) = 0.
i i
5.4. DETERMINANTS 201
To see this, note that all the factors in each product are the same except possibly
when i = s, t. But
[asf (s) bf (s) (s) ][atf (t) bf (t) (t) ] = [asf (s) bf (s) (s) ][atf (t) bf (t) (t) ],
since f (s) = f (t), (s) = (t), and (t) = (s). Thus the terms involving
functions f that are not permutations cancel out in pairs, so that
! !
X X Y X X Y
( 1) aif (i) bf (i) (i) = ( 1) ai (i) b (i) (i) ,
f i i
1
where ranges over all permutations of f1, 2, : : : , ng. Let = . Then
i ai (i) b (i) (i) = i ai (i) bi (i) , just noting that the bjk in i ai (i) b (i) (i) are
exactly those such that (j) = k. Thus
X Y
det((aij )(bij )) = ( 1) ai (i) bi (i) :
; i
Thus conjugates have the same determinant. Therefore we could de…ne det( )
for any linear transformation on a …nite dimensional vector space V by letting
it be the determinant of any matrix of . We will return to determinants of
linear transformations later. Finally, note that
Proof.
X Y X X Y
det(aij ) = ( 1) ai (i) = ( 1) ai (i)
i j (1)=j i
X X Y
= a1j ( 1) ai (i)
j (1)=j i>1
For the general result, change (aij ) by moving its ith row to its …rst row, getting
a matrix (bij ). Then
n
X
det(aij ) = ( 1)i 1
det(bij ) = ( 1)i 1
( 1)j+1 b1j det B1j
j=1
n
X n
X
i+j
= ( 1) aij det Aij = aij ij .
j=1 j=1
Note that 5.4.6 and 5.4.7 can be combined. They assert that jk det(aij ) =
n n
i=1 aij ik = i=1 aji ki for all j and k, where ik is the Kronecker :
If 5.4.6 is looked at properly, it not only gives a method for calculating
det(aij ), but it also gives a method for calculating the inverse of the matrix
(aij ), if it exists. Furthermore, it tells us exactly when (aij ) does have an
inverse.
5.4. DETERMINANTS 203
Let adj(aij ), called the adjoint of (aij ), be the matrix whose (i; j) entry is
ji .Then 5.4.6 yields
det(aij )In = (aij ) adj(aij ) = (adj(aij ))(aij ).
Thus, if det(aij ) is a unit in R, we get
In = (aij )[adj(aij )= det(aij )] = [adj(aij )= det(aij )](aij ).
In other words, adj(aij )= det(aij ) is the inverse of (aij ) in case det(aij ) is a unit
in R. We noted earlier that det(aij ) is a unit if (aij ) is a unit. Therefore we
have
Corollary 5.4.8 Let (aij ) be in Rn . Then (aij ) is a unit in Rn if and only if
det(aij ) is a unit in R. Furthermore, if det(aij ) is a unit in R, then (aij ) 1 =
adj(aij )= det(aij ).
This corollary tells us a little more. Suppose that a matrix (aij ) in Rn has
a right inverse. Then there is a matrix (bij ) in Rn with (aij )(bij ) = In . Hence
det((aij )(bij )) = 1 = det(aij ) det(bij ). Since R is commutative, det(aij ) is a
unit in R. Hence (aij ) is a unit in Rn . Similarly, if (aij ) has a left inverse, it is
a unit.
From 5.4.6 and 5.4.7, we can also get Cramer’s Rule. Suppose that
a11 x1 + a12 x2 + + a1n xn = b1
a21 x1 + a22 x2 + + a2n xn = b2
..
.
an1 x1 + an2 x2 + + ann xn = bn
is a system of n equations in the n unknowns x1 , x2 , : : : , xn , with aij and bi in
R. The matrix (aij ) is called the matrix of coe¢ cients, and det(aij ) is called
the determinant of the system. Cramer’s Rule gives a su¢ cient condition for
such a system to have a unique solution.
Theorem 5.4.9 (Cramer’s Rule) Let
a11 x1 + a12 x2 + + a1n xn = b1
a21 x1 + a22 x2 + + a2n xn = b2
..
.
an1 x1 + an2 x2 + + ann xn = bn
be a system of linear equations with coe¢ cients aij and bi in the commutative
ring R with identity. If the matrix of coe¢ cients (or equivalently, the deter-
minant) of the system is a unit, then the system has a unique solution given
by
Xn
xk = bi ik = det(aij ),
i=1
where ij is the cofactor of aij .
204 CHAPTER 5. LINEAR TRANSFORMATIONS
n
X n X
X n n
X
bi ik = aij ik xj = aik ik xk = det(aij )xk .
i=1 j=1 i=1 i=1
n
Thus if det(aij ) is a unit in R, we get xk = i=1 bi ik =det(aij ). Thus there is
at most one solution. But
n
X n
X n
X n
X
ajk bi ik = bi ajk ik = bj det(aij ),
k=1 i=1 i=1 k=1
Proof. Let
A1
(aij ) =
0 A2
It su¢ ces to show that det(aij ) = det(A1 ) det(A2 ). Let A1 be r r, A2 be s s,
and S be the symmetric group Sr+s . Then
X r+s
Y
det(aij ) = ( 1) ai (i) .
2S i=1
If for some k > r, (k) r, then r+s i=1 ai (i) = 0. Thus only those
r+s
i=1 ai (i)
for which permutes fr + 1, r + 2, : : : , r + sg, and hence also permutes f1, 2,
: : : , rg, are non-zero. Every such permutation is uniquely a product , where
5.4. DETERMINANTS 205
X r
Y r+s
Y
det(aij ) = ( 1) ai (i) ai (i)
; i=1 i=r+1
X r
Y X r+s
Y
= ( 1) ai (i) ( 1) ai (i)
i=1 i=r+1
= det(A1 ) det(A2 )
where runs through the permutations with (i) = i for i > r, and runs
through those with (i) = i for i r.
Let f (x) = a0 + a1 x + + an 1 xn 1
+ xn be in R[x]. As in the case for
…elds, the companion matrix C(f (x)) is the matrix
0 1
0 0 0 0 a0
B 1 0 0 0 a1 C
B C
B 0 1 0 0 a2 C
B C
B .. .. .. C
B . . . C
B C
@ 0 0 0 0 an 2 A
0 0 0 1 an 1 :
Theorem 5.4.13 Let (aij ) be an n n matrix over a …eld F . Then the char-
acteristic polynomial of (aij ) is det(xIn (aij )).
where the Ci are the companion matrices of the invariant factors fi (x) of (aij ).
Thus for suitable (bij ), (cij ) = (bij ) 1 (aij )(bij ). We have
1
det(xIn (aij )) = det((bij ) ) det(xIn (aij )) det((bij )
1 1
= det((bij ) xIn (bij ) (bij ) (aij )(bij ))
= det(xIn (cij )) = fi (x),
c(x) = det(xIn A) = xn + an 1x
n 1
+ + a1 x + a0 ,
and Q(x) = adj(xIn A). Then we have c(x)In = (xIn A)Q(x), and c(x)In
c(A) = ni=1 ai (xi In Ai ), where an = 1. Write xi In Ai = (xIn A)Qi (x).
Then c(x)In c(A) = (xIn A) ni=0 ai Qi (x) = (xIn A)P (x) = (xIn A)Q(x)
c(A). Hence (xIn A)(Q(x) P (x)) = c(A). If Q(x) P (x) 6= 0, then the
matrix xIn (Q(x) P (x)) has an entry which is a polynomial in x of higher
degree than any entry in A(Q(x) P (x)). It follows that c(A) = 0. We have
proved the following theorem.
picking a basis of V and then taking the determinant of the matrix relative
to that basis. We will present this “coordinate free” de…nition of det( ). That
is the …rst important point. The second is this. We want to characterize det as
a function
det : Rn ! R
A number of its properties have been derived, and we will show that det is the
only function Rn ! R satisfying certain conditions. In fact, we take up this
point …rst.
For our commutative ring R, let Rn be the module of n-tuples of elements
of R, but view an element of Rn as a column
0 1
r1
B r2 C
B C
B .. C
@ . A
rn
det : Rn Rn Rn ! R
and we want to look at how det behaves with respect to the individual columns.
We know already that if two columns of (aij ) are equal then det(aij ) = 0, and
that if any column of (aij ) is 0, then det(aij ) = 0. We need two more facts that
follow from the de…nition. We can add elements of Rn , and we can multiply
them by scalars. This is done coordinate-wise, by de…nition. Let (aij ) be in Rn ,
and let 0 1
a1i
B a2i C
B C
ci = B . C
@ .. A
ani :
Further, if di is in Rn , then
det(c1 , c2 , : : : , ci 1 , ci + di , ci+1 , : : : , cn )
= det(c1 , : : : , ci , : : : , cn ) + det(c1 , : : : , di , : : : , cn ).
These two properties are expressed by saying that det is linear in each variable
ci . Here is a more general de…nition.
208 CHAPTER 5. LINEAR TRANSFORMATIONS
f : C1 C2 Cn ! D
f (c1 , : : : , ci 1 , ci + di , ci+1 , : : : , cn )
= f (c1 , : : : , ci 1 , ci , ci+1 , : : : , cn ) + f (c1 , : : : , ci 1, di , ci+1 , : : : , cn );
and
C = C1 = C2 = = Cn ,
f (c1 , c2 , : : : , ci , : : : , cj , : : : , cn )
= f (c1 , : : : , ci 1 , cj , ci+1 , : : : , cj 1, ci , cj+1 , : : : , cn ).
Proof. There is at least one such form. Let detr = r det. It is completely
routine to check that detr is an n-multilinear alternating form with detr (In ) = r.
Now let dr be any such form. Let ui be the ith column of In . Then (u1 , u2 , : : :
, un ) = In , so dr (u1 , u2 , : : : , un ) = r. Now we compute
dr (c1 , c2 , : : : , cn ):
There is a nice way to view 5.4.16. It is routine to check that the alternating
forms on Rn Rn form an R-module under the operation (f + g)(X) =
f (X) + g(X), and (rf )(X) = r(f (X)), where X is in Rn Rn and r is in
R. Our theorem then says that the module of n-multilinear alternating forms
on Rn Rn = Rn is free of rank 1, and in fact det is a generator. To
generalize a bit, the R-module Rn is free of rank n. Let M be any free R-module
of rank n. Then certainly the R-module of n-multilinear alternating forms on
M n , the Cartesian product of M with itself n times, is free of rank 1. The proof
is the same as above. Just choose fu1 , u2 , : : : , un g to be a basis of M . Then for
any n-multilinear alternating form f , f ( ai1 ui , : : : , ain ui ) = det(aij ) f (u1 ,
: : : , un ). Hence f is determined by f (u1 , : : : , un ). De…ning
!
X X
d ai1 ui , : : : , ain ui = det(aij )
i i
But
PROBLEMS
1. How many additions and multiplications are necessary in order to calculate
the determinant of a 5 5 matrix? a 10 10?
2. Calculate the determinant of the following matrices.
0 1 0 1
6 5 4 8 3 7 2 4
B 1 1 3 2 C B 3 6 0 5 C
B C B C
@ 7 9 4 5 A @ 0 5 4 3 A
2 1 2 1 3 1 0 8
0 1 0 1
75 63 18 39 1 4 9 16
B 61 45=2 119 67 C B 4 9 16 25 C
B C B C
@ 76 83 109=7 31 A @ 9 16 25 36 A
25 52 6 13 16 25 36 49
3. Are the vectors (1; 0; 3; 2; 1), (1; 1; 4; 3; 2), (2; 0; 0; 1; 1), (3; 4; 5; 0; 1),
(0; 1; 1; 1; 1) independent?
4. Prove that the determinant of
0 1
1 x1 x21 xn1 1
B 1 x2 x22 xn2 1 C
B C
A=B . .. .. C
@ .. . . A
1 xn x2n xnn 1
7. Prove that the eigenvalues of a triangular matrix are the diagonal entries
of that matrix.
x y+z =2
2x + y = 5
3x + 7y 2z = 6
10. Solve the same system as in Problem 9 where the coe¢ cients are taken
modulo 11.
11. Let V be a …nite dimensional vector space over a …eld F , and let A(V )
be the vector space of all linear transformations on V . Let be in A(V ),
and let f : A(V ) ! A(V ) be de…ned by f ( ) = . Prove that f
is linear and that det(f ) = 0.
12. Let A(V ) be the vector space of all linear transformations on the …nite
dimensional vector space V . Let be in A(V ). Let f ( ) = for in
A(V ). Prove that f is a linear transformation on the vector space A(V ),
and …nd det(f ) in terms of det( ).
13. Let S be a commutative subring of the ring of all n n matrices over the
commutative ring R. Let A be an m m matrix over S. View A as an
(nm) (nm) matrix over R. As such a matrix, its determinant detR (A)
is an element of R. As a matrix over S, its determinant detS (A) is an
element of S. Prove that detR (A) = det(detS (A)):
14. Prove that any 2 2 real matrix whose determinant is negative is similar
over the real numbers to a diagonal matrix.
15. Let n be any positive integer. Prove that over the complex numbers the
n n matrix 0 1
0 0 0 0 1
B 1 0 0 0 0 C
B C
B 0 1 0 0 0 C
B C
B .. .. C
B . . C
B C
B 0 0 0 0 0 C
B C
@ 0 0 0 0 0 A
0 0 0 1 0
is similar to a diagonal matrix.
16. Let (aij ) be an n n matrix over the …eld F . Prove that there are at most
n elements a in F such that det(aIn (aij )) = 0.
212 CHAPTER 5. LINEAR TRANSFORMATIONS
17. Let V be a vector space of dimension n over F , and let and be linear
transformations on V . Suppose that is non-singular. Prove that there
are at most n scalars a in F such that a + is singular.
1
18. Let A be an invertible matrix. Find the characteristic polynomial of A
in terms of that of A.
24. Prove that the set of all real matrices of the form
0 1
a b c d
B b a d c C
B C
@ c d a b A
d c b a
is a division ring. Prove that as a vector space over the real numbers, it
is of dimension 4 and has a basis f1, i, j, kg such that i2 = j 2 = k 2 = 1,
ij = ji = k, jk = kj = i, and ki = ki = j.
25. If R is any commutative ring, the group of units of Rn is denoted GLn (R).
Let S be any subgroup of the group of units R of R. Prove that N =
fA 2 GLn (R) : det(A) 2 Sg is normal in GLn (R), and that GLn (R)=N
R =S.
26. Prove that if the n n matrix A over the …eld F has rank m, then A
has an m m non-singular “submatrix” and has no larger non-singular
submatrix.
Therefore
(dij )(bij ) = (aij )(cij ).
Note that (cij ) and (dij ) are invertible. Indeed, if xi = k eki vk , then vj =
i cij xi = i cij k eki vk = k ( i cij eki )vk . Thus i eki cij = ik . In other words,
(eij )(cij ) = In . Similarly, (cij )(eij ) = In , and so (cij ) 1 = (eij ). Similarly, (dij )
is also invertible. We have
with (cij ) and (dij ) invertible matrices. There is terminology for matrices so
related.
since B = P AQ. Therefore A and B are both matrices of , and the proof is
complete.
Note that if m = n, X = Y , and xi = yi , then Q = P 1 , so that B =
P AP 1 . Hence, as in the vector space case, two square matrices are similar if
and only if they are matrices of the same endomorphism : X ! X relative to
appropriate (single) bases of X:
There are two fundamental ways to associate a module with a matrix. There
is the one for square matrices over a …eld in 5.2. That procedure can be gen-
eralized. Let Rn be the free R-module of rank n, that is, the direct sum of n
copies of the module R. As in 5.4, think of Rn as columns (ri ) of elements of
R. If (aij ) is in Rn , then Rn ! Rn : (ri ) ! ( j aij rj ) is an endomorphism of
Rn . It is just the endomorphism whose matrix with respect to the canonical
basis fe1 , e2 , : : : , en g, where ei is the column with 1 in the ith position and
zeroes elsewhere, is (aij ). The action of (aij ) on (ri ) is just matrix multiplica-
tion. Thus with a matrix A in Rn , we have an endomorphism of Rn , which we
also denote by A. This enables us to make Rn into an R[x]-module by de…ning
f (x)v = f (A)(v) for v in Rn . Denote this R[x]-module by M1 (A). Thus with
an n n matrix A over R we have associated an R[x]-module M1 (A). For …elds,
this is the construction carried out in 5.2. (M1 (A) corresponds to V ). If A
and B are in Rn , then M1 (A) M1 (B) if and only if A is similar to B. The
5.5. EQUIVALENCE OF MATRICES 215
proof is exactly the same as for the vector space case. Although M1 (A) is a
…nitely generated R[x]-module, R[x] is not necessarily a principal ideal domain,
and so we have no nice complete set of invariants for M1 (A) as we had in the
case when R was a …eld.
The other module associated with A is Rn =A(Rn ). However, we want to
be a little more general here. Let A be an m n matrix over R. Then the
homomorphism Rn ! Rm whose matrix with respect to the canonical bases of
Rn and Rm is A is given by matrix multiplication, and is also denoted by A.
Thus 0 1
Xn
(ri ) ! @ aij rj A = A(ri ),
j=1
Rm =A(Rn ) Rm =B(Rn ).
1 2
? ?
M2 - N2
2
216 CHAPTER 5. LINEAR TRANSFORMATIONS
Q 1
p
? ?
Rm - Rn
B
where A = P BQ with P and Q appropriate non-singular matrices.
Such pictures as above are called diagrams, and the condition 2 1 = 2 1
is expressed by saying that the …rst diagram commutes.
The converse of 5.5.3 holds when R is a PID. That is, if R is a PID, if A
and B are m n matrices over R, and if M2 (A) M2 (B), then A and B are
equivalent. In order to prove this, we need a theorem about submodules of free
modules of …nite rank. This theorem implies that …nitely generated modules
over a PID are direct sums of cyclic modules. We could actually derive it from
our results in 4.7, but we prefer to give an alternate (direct) proof. This theorem
will also give a nice canonical form for equivalence classes of equivalent matrices
over PID’s. Here is the theorem.
Theorem 5.5.4 Let M be a free module of …nite rank n over a PID R. Let N
be a submodule of M . Then there are an integer m n, a basis fe1 , e2 , : : : ,
en g of M , and elements r1 , r2 , : : : , rn of R such that fr1 e1 , r2 e2 , : : : , rm em g
is a basis of N , r1 jr2 , r2 jr3 , : : : , rm 1 jrm , and rn+1 , : : : , rn = 0. The ri are
unique up to associates.
dy1 + a3 e3 + + an en = a1 e1 + a2 e2 + + an en
We need only that a1 jr2 . But we already showed that if fe1 , e2 , : : : , en g is any
basis and a1 e1 + a2 e2 + +an en is in N , then a1 ja2 . Now a1 x1 + r2 x2 is in
N , whence a1 jr2 . Setting a1 = r1 gets the required ri ’s.
For the uniqueness of the ri ’s, we appeal to 4.7.15; those not zero or a unit
are just the invariant factors of M=N:
An immediate consequence is the fact that a …nitely generated module X
over a PID is a direct sum of cyclic modules. Let fx1 , x2 , : : : , xn g be a set of
generators of X. Let M be free of rank n with basis fe1 , e2 , : : : , en g. Sending
ei to xi induces an epimorphism : M ! X. Now there are a basis fy1 , y2 , : : :
, yn g and elements r1 , r2 , : : : , rm (m n) of R such that fr1 y1 , : : : , rm ym g
is a basis of Ker( ). Hence X M= Ker( ) Ry1 =Rr1 y1 Ryn =Rrn yn
(where rm+1 = = rn = 0), which is a direct sum of cyclic modules.
S - A(Rn )
? ?
T - B(Rn )
Rn - Rm
? ?
Rn - Rm
d1 , d2 , : : : , dr , 0 , : : : , 0
1; 1; :::; 1; 0; 0; :::; 0
d1 , d2 , : : : , dr , 0 , 0, : : : , 0
Proof. M1 (A) F [x]=F [x]f1 (x) F [x]=F [x]fr (x). If fi (x) = 1, then
F [x]=F [x]fi (x) = 0. Thus those not 1 are the invariant factors of M1 (A), and
the corollary follows.
Actually, a new proof that a square matrix satis…es its characteristic poly-
nomial is at hand. We need one additional fact.
with ad bc = 1.
A secondary matrix is of course invertible since its determinant is 1. Multi-
plying A on the left by S is a secondary row operation on A, and multiplying
A on the right by S is a secondary column operation on A.
Recall that a matrix (aij ) is in diagonal form if aij = 0 whenever i is not
j. The important fact is that any m n matrix over a PID can be put into a
canonical diagonal form with diagonal entries d1 , d2 , : : : , dr , 0, 0, : : : , 0 with di
dividing di+1 by applying a succession of elementary and secondary operations
to it. This gives an e¤ective way to compute invariant factors. Here is how it is
done. Let A = (aij ) be an m n matrix over a PID. We may interchange two
rows or two columns by multiplying A by suitable elementary matrices. Hence
we can get a11 6= 0 if A 6= 0. (If A = 0, there is nothing to do.) Let d be
a greatest common divisor of a11 and a12 . Write a11 = a1 d, a12 = a2 d, and
sa11 + ta12 = d. Then sa1 + ta2 = 1. Now
0 1 0 1
s a2 0 0 0 d 0 a13 a1n
B t a1 0 0 0 C B 0 a022 a23 a2n C
B C B a21 C
B 0 0 1 0 0 C B a31 a032 a33
0
a3n C
(aij ) B C=B C
B .. .. C B .. .. C
@ . . A @ . . A
0 0
0 0 0 1 am1 am2 am3 amn
Interchanging the second and third columns and repeating the process above
yields an equivalent matrix with the (1,2) and (1,3) entries zero. Proceeding
in this manner, we get a matrix with …rst row of the form (d, 0, 0, : : : , 0),
and, in fact, with d a greatest common divisor of the original a11 , a12 , : : : , a1n .
Now proceed in an analogous way to get a matrix with …rst column having only
the (1; 1) entry non-zero. This process, however, disturbs the …rst row, and the
matrix has the form 0 1
e f 2 f3 fn
B 0 C
B C
B .. C.
@ . A
0
If e divides each f , then subtracting suitable multiples of the …rst column from
the other columns brings the matrix into the form
0 1
e 0 0 0
B 0 C
B C
B=B . C.
@ .. A
0
If e does not divide some fi , then proceed as in the very …rst step. This yields
a matrix of the form 0 1
e1 0 0 0
B g2 C
B C
B .. C
@ . A
gm
222 CHAPTER 5. LINEAR TRANSFORMATIONS
Now we can proceed as before without disturbing the …rst row or column. The
rest should be clear. By induction on n, our procedure brings the matrix into
the diagonal form
0 1
d1
B d2 C
B C
B . .. C
B C
B C
B C,
B dr C
B 0 C
B C
B 0 C
@ A
..
.
with di jdi+1 .
Observe that the second elementary operation, namely, multiplying a row
or column by a unit, has not been used. If our PID were Z, or F [x], we could
make all the di positive, or monic, respectively, by so doing. In these cases, the
di are unique. However, if the PID is Z or F [x], we do not need the secondary
operations in our algorithm. For instance, suppose our ring is Z. By elementary
operations, we can make a11 > 0. Write a1i = qi a11 + ri , with 0 a11 .
Elementary operations then yield a matrix with …rst row (a11 , r2 , r3 , : : : , rn ).
Start all over. Either all the ri are 0 or there is an ri with 0 < ri < a11 . Put
ri in the (1; 1) position and repeat the process. This eventually brings us to
5.5. EQUIVALENCE OF MATRICES 223
a matrix with …rst row of the form (e1 , 0, 0, : : : , 0). Now work on the …rst
column of this new matrix. We get a matrix with …rst column having only its
(1; 1) entry non-zero, but we know nothing about the …rst row any more. So
start all over. The ei are getting smaller, and are positive. Thus we eventually
bring our matrix to the form with the only non-zero entry in the …rst row or
…rst column the (1; 1) entry. The rest should be clear. It is analogous to what
we did in the general PID case. The division algorithm for polynomials (4.5.7)
enables us to follow a similar procedure in the case our ring is F [x], where F is
a …eld.
PROBLEMS
1. Prove that the matrix
0 1
x 1 0
@ 0 x 1 A
6 5 x+2
is equivalent over Q[x] to the diagonal matrix with diagonal entries (1; 1; x3 +
2x2 + 5x + 6).
2. Reduce the following matrices of integers to canonical form.
0 1
11 8 6 3
B 6 7 5 1 C
2 8 6 B C
24 12 15 ; ; B B 3 8 1 4 C
C
4 2 10 @ 1 A
2 4 12
4 3 2 10
4. Reduce the following matrices over Q[x] to the canonical diagonal form
with diagonal entries
(f1 (x), f2 (x), : : : , fr (x), 0, : : : , 0),
with fi (x) monic and with fi (x) dividing fi+1 (x).
1 x x 1 + x2
x2 x x2
x x x2
x4 + 1 x7 4x3 + 1 x4 4x3 + 4x 5
2x4 + 3 2x7
2x + 4x3 2
4
3x + 4x3 + x2 + 10x 14
4
5. Prove that if the matrix A 6= 0 over a PID has canonical form with diagonal
entries (d1 , d2 , : : : , dr , 0, : : : , 0), then d1 is the greatest common divisor
of all the entries of A.
10. Let M be a module over the commutative ring R, and let be an endo-
morphism of M . For m in M and f (x) in R[x], let f (x) m = f ( )(m).
Prove that this makes M into an R[x]-module. Denote this module by
M . Prove that M M if and only if there is an automorphism of
1
the R-module M such that = :
11. Let M be a module over the commutative ring R. Let M [x] be the set
of all “polynomials” m0 + m1 x + +mk xk in x with coe¢ cients in
M . Prove that M [x] is, in a natural way, an R[x]-module. Let be an
endomorphism of the R-module M , and let
X X X
(x ) mi xi = mi xi+1 (mi )xi :
section, we will study vector spaces over R that come equipped with an added
structure that will enable us to de…ne such geometric notions as length and
angle. A parallel development for C will be indicated in the exercises.
Various special properties of the real numbers will be used. For example, R
is ordered. That is, in R there is a notion of an element a being less than
an element b, written a < b, or equivalently, b > a, and this relation satis…es
certain properties. We write a b if a < b or a = b. Numbers a such that 0 < a
are positive, and positive numbers have unique square roots. If a b, and if
c is positive, then ac bc. Such familiar properties of the real numbers will be
used throughout our discussion, and we assume that the reader is acquainted
with them.
How does one de…ne length, and so on, in R2 ? First, if x = (x1 , y1 ) and
y = (x2 , y2 ) are in R2 , then the inner product, or dot product, of x and y
is (x, y) = x1 x2 + y1 y2 . The product (x, y) is also denoted by x y. The other
geometric notions are de…ned in terms of this product. For example, the length
of x is (x, x)1=2 , and the angle between x and y is given by cos( ) = (x, y)=((x,
x)(y, y))1=2 . Analogous de…nitions are made for Rn . The inner product above
is a map R2 R2 ! R. What we will consider will be vector spaces V over
R that come equipped with a map V V ! R in terms of which the requisite
de…nitions can be made.
There are two classical examples. For the vector space Rn of n-tuples of
real numbers, de…ne Rn Rn ! R by (v, w) ! ni=1 vi wi , where v = (v1 , v2 ,
: : : , vn ) and w = (w1 , w2 , : : : , wn ). This makes Rn into an n-dimensional
Euclidean space. When we speak of the Euclidean space Rn , we mean Rn
with this particular inner product. The second example is the vector space
V of all continuous real functions on the closed interval [0; 1] with an inner
R1
product given by (f (x), g(x)) = 0 f (x)g(x)dx. This makes V into an in…nite
dimensional Euclidean space. Our concern will be almost wholly with …nite
dimensional spaces.
Theorem 5.6.3 Let V be a Euclidean space. Then for all r in R and v and w
in V
a. krvk = jrjkvk.
d. kv + wk kvk + kwk.
Proof. The equalities krvk = (rv, rv)1=2 = (r2 (v, v))1=2 = jrjkvk establish
(a).
(b) follows immediately from 5.6.1 (b).
Since
whence
kv + wk kvk + kwk,
and we have (d).
Part (c) is Schwarz’s inequality, and part (d) is the triangle inequality.
From 5.6.2, it follows easily that our notion of distance satis…es the usual
properties of ordinary distance, namely that
3. the distance from u to v plus the distance from v to w is not less than the
distance from u to w.
(v; w)
cos( ) = .
kvk kwk
5.6. EUCLIDEAN SPACES 227
f(1; 0; :::; 0); (0; 1; 0; :::; 0); :::; (0; 0; :::; 0; 1)g
If
Pnv1 , v2 , : : : ,vn is P
an orthonormal basis of a Euclidean space V , and if v =
n
a v
i=1 i i and w = i=1 bi vi , then
!
X X X X
(v; w) = ai vi ; bi v i = ai bj (vi ; vj ) = a i bi
i i i;j i
is orthonormal.
An immediate consequence is the following.
It follows that (v; w) = ( (v); (w)), and thus that preserves inner products.
If preserves inner products, then ( (v); (v)) = (v; v), and so k (v)k = kvk
and is orthogonal.
In particular, if is orthogonal then preserves orthogonality, or (v, w) = 0
if and only if ( (v), (w)) = 0. Also it is useful to observe that if fv1 , v2 , : : : ,
vn g is a basis, then is orthogonal if (vi , vj ) = ( (vi ), (vj )) for all i and j:
We now examine the matrix of an orthogonal transformation relative to an
orthonormal basis. Let be such a linear transformation on V and fv1 , v2 , : : :
, vn g such a basis of V . If (aij ) is the matrix of relative to this basis, then
(vj ) = i aij vi , and
!
X X X
(vi , vj ) = ( (vi ), (vj )) = ij = aki vk , akj vk = aki akj :
k k k
This means that (aij )0 (aij ) = In , the identity n n real matrix, where (aij )0
denotes the transpose of the matrix (aij ). Thus (aij ) 1 = (aij )0 . Another
way to look at it is that the rows of (aij ) form an orthonormal basis of the
Euclidean space Rn . On the other hand, let (aij ) be an n n matrix such that
(aij )0 = (aij ) 1 , and let fv1 , v2 , : : : , vn g be an orthonormal basis of V . Then
the linear transformation given by (vj ) = i aij vi is orthogonal. Indeed,
( (vi ), (vj )) = ( k aki vk , k akj vk ) = k aki akj = ij = (vi , vj ).
this polynomial has real roots. Thus f (x) is a product of linear factors.
The name of the following theorem derives from its role in determining the
principal axis of an ellipse. In plane analytic geometry, the standard form of
the equation of an ellipse is x2 =a2 + y 2 =b2 = 1. In this form, the axes of the
ellipse are parallel to the coordinate axes. For certain values of a, b, c, and d,
ax2 + bxy + cy 2 = d is the equation of an ellipse, but unless b = 0, the axes
are not parallel to the coordinate axes. The Principal Axis Theorem implies
that by a rotation of the axes, that is, by an orthogonal transformation, the
equation can be put in standard form, from which the lengths of the axes and
other information about the ellipse can be read o¤. It can also be used to deduce
results about the higher dimensional analogues of conic sections.
Corollary 5.6.19 Two real symmetric matrices are similar if and only if they
have the same characteristic polynomial.
Now we will get a canonical form for orthogonal matrices. The theorem is
this.
232 CHAPTER 5. LINEAR TRANSFORMATIONS
1. cf de = 1;
2. c2 + d2 = 1;
3. e2 + f 2 = 1, and
5.6. EUCLIDEAN SPACES 233
4. ce + df = 0:
cos( ) sin( )
sin( ) cos( )
PROBLEMS
1. Let V be a …nite dimensional real vector space over R, and let fv1 , v2 ,
: : : , vn g be a basis of V . For v = i ai vi and w = i bi vi , de…ne (v,
w) = i ai bi . Prove that this makes V into a Euclidean space.
2. If V is a Euclidean space with inner product (v, w), then for any positive
real number r, r(v, w) is an inner product on V .
3. Prove that R2 is a Euclidean space if inner products are given by ((a; b); (c; d)) =
(a b)(c d) + bd.
4. Prove that a Euclidean space V is a metric space via V V ! R : (v,
w) ! kv wk.
5. Prove that a linear transformation on a Euclidean space is orthogonal
if for some basis fv1 , v2 , : : : , vn g, (vi , vj ) = ( (vi ), (vj )).
6. Let fv1 , v2 , : : : , vn g be an orthonormal basis of the Euclidean space V ,
and let v be in V .
13. Let V be the vector space of n-tuples of complex numbers. For v = (c1 ,
c2 , : : : , cn ) and w = (d1 , d2 , : : : , dn ) in V , de…ne (v, w) = i ci di , where
c denotes the conjugate of the complex number c. Prove that for all u, v,
and w in V , and a and b in C,
(Cv)? = fw 2 V : (v, w) = 0g
is invariant under .
Fields
e. fa + bi : a; b 2 Qg;
f. fa + b(2)1=2 : a; b 2 Qg;
h. the quotient …eld F (x) of the integral domain F [x] of all polynomials in x
with coe¢ cients from a …eld F ;
i. Q[x]=(x2 + x + 1);
j. (Z=2Z)[x]=(x3 + x + 1);
237
238 CHAPTER 6. FIELDS
Any …eld F has a unique smallest sub…eld P , called the prime sub…eld of
F . It may be obtained by taking the intersection of all the sub…elds of F , or
alternately by taking the sub…eld generated by 1. If the integral multiples n 1
of 1 are distinct in F , then the map
Z!Z 1:n!n 1
is an isomorphism, and thus F contains a copy of the ring Z of integers and
hence of the …eld Q of rationals. It should be clear that in this case the prime
sub…eld of F is isomorphic to Q. Furthermore, if n a = 0 for any a 2 F , then
0 = n a = (n 1)a, so either a or n is zero. Therefore, the non-zero elements
of F all have in…nite additive order. This is expressed by saying that F has
characteristic zero. If m 1 = n 1 with m 6= n, then (m n) 1 = 0, and
thus there is a smallest positive integer p such that p 1 = 0. If p = qr with
1 q < p, then 0 = p 1 = (qr) 1 = (q 1)(r 1), whence r 1 = 0, and r = p.
It follows that p is a prime. In this case, the map
Z!Z 1:n!n 1
has kernel pZ. This is a ring homomorphism and is clearly onto. Thus Z 1
Z=pZ, which is a …eld. Therefore Z 1 is the smallest sub…eld of F . For a non-zero
element a of F , p a = (p 1)a = 0. Therefore the additive order of any non-zero
element of F is p. This is expressed by saying that F has characteristic p.
Theorem 6.1.2 The prime sub…eld of a …eld of characteristic 0 is isomorphic
to the …eld Q of rational numbers. The prime sub…eld of a …eld of prime char-
acteristic p is isomorphic to the …eld Z=pZ of integers modulo p.
If F is a sub…eld of a …eld K, then K is called an extension of F . In
particular, every …eld is an extension of its prime sub…eld. If K is an extension
of F , then K is, in particular, a vector space over F .
De…nition 6.1.3 Let K be an extension of F . The dimension of K as a vector
space over F is denoted by [K : F ] and called the degree of the extension.
The vector space K over F may not have a …nite basis, in which case we
have not really de…ned its dimension, and will just say that it is in…nite. Our
real concern with degrees will be with …nite ones. An extension of …nite degree
is called a …nite extension.
The …eld of complex numbers C is an extension of degree 2 of the …eld of
real numbers R. The set f1, ig is a basis of C as a vector space over R. For any
…eld F , the quotient …eld F (x) of F [x] is an extension of in…nite degree of F .
The set f1, x, x2 , x3 , : : : g is linearly independent over F .
Let K be an extension of F . If a 2 K, let F (a) denote the smallest sub…eld of
K containing F and a, and let F [a] denote the smallest subring of K containing
F and a. Note that F [a] is just the set of all polynomials c0 + c1 a + + cn an
in a with coe¢ cients in F , and that F (a) is the …eld of quotients of F [a]. We
distinguish two principal kinds of elements a in the extension K of F — those
with [F (a) : F ] …nite, and those with [F (a) : F ] not …nite.
6.1. SUBFIELDS AND EXTENSION FIELDS 239
Therefore the elements a of K that are algebraic over F are those that
generate …nite extensions F (a) of F . In the proof of 6.1.4, we saw that if
[F (a) : F ] = n, then f1; a; : : : ; an 1 g is a basis of F (a) over F , n is the degree
of a over F , and F [a] = F (a).
Suppose that two elements a and b in K are algebraic over F . The elements
a + b, ab, and a 1 if a 6= 0 are also algebraic over F , but this is not exactly
obvious. We will show that the set of all elements of K which are algebraic over
F is a sub…eld of K. We need some important preliminaries.
fai bj : i = 1, 2, , m; j = 1, 2, : : : , ng
240 CHAPTER 6. FIELDS
n
is a basis of L over F . Let c 2 L. Then c = i=1 ei bi with ei 2 K, and
ei = mj=1 fij aj with fij 2 F . Thus
0 1
n
X Xm X
c= @ fij aj A bi = fij aj bi .
i=1 j=1 i;j
Both [F (a) : F ] and [F (a)(b) : F (a)] are …nite, the latter because b is algebraic
over F and thus certainly over F (a). Thus [F (a)(b) : F ] is …nite. Therefore
every element in F (a)(b) is algebraic over F . The elements a + b, ab, and 1=b
are in F (a)(b).
It should be clear that F (a)(b) = F (b)(a). This …eld is denoted F (a; b).
More generally, if fa1 , a2 , : : : , an g is any …nite set of elements in K, where K is
an extension of F , then F (a1 , a2 , : : : , an ) denotes the sub…eld of K generated
by F and the elements a1 , a2 , : : : , an .
Proof. Let a and b be constructible real numbers. Then there are …elds F
and K with a 2 F , b 2 K, and appropriate …elds Fi between F and Q and Ki
between K and Q. Let L be the sub…eld generated by F and K. Let
Fi = Q(a1 , a2 , : : : , ai )
Q F1 F2 Fn = F Fn+1 Fn+m = L
is one of successive extensions each of degree one or two. Since a b and a=b (if
b 6= 0) are in L, the theorem follows.
Note that a constructible number has degree a power of 2 over Q.
Constructible numbers are numbers that can be constructed with ruler and
compass. What does this mean? Ruler and compass constructions are generally
done in elementary geometry courses. These are constructions done in a …nite
number of steps, starting with only the unit segment (unless speci…cally stated
otherwise). To use the ruler means to regard as drawn the line determined
by two given points. To use the compass means to regard as drawn the circle
having a given point as center and the length of a given line segment as radius.
Ruler and compass constructions are carried out in the plane. For example, such
constructions as those of a perpendicular bisector of a given line segment, and of
a line through a given point parallel to a given line are no doubt familiar. Other
possible ruler (more appropriately, straight-edge) and compass constructions
are those of points partitioning a given segment into n equal segments, of line
segments of length equal to the sum, di¤erence, product, or quotient of the
lengths of two given line segments. To say that a real number x is constructible
means that a line segment of length jxj is constructible. In particular, all rational
numbers are constructible. One can construct the square root of a given positive
real number. Thus square roots of positive rational numbers are constructible
with ruler and compass. Therefore, starting with the unit segment, we can
construct Q, and for any positive rational number x, we can construct x1=2 .
Hence we can construct Q(x1=2 ). Similarly, we can construct Q(x1=2 ; y 1=2 ) for
any other positive number y 2 Q(x1=2 ). Thus we see that a real number r is
constructible (with ruler and compass) if r is in a …eld of real numbers obtainable
from Q by successive extensions of degree 2.
Suppose that we are given a …eld F of real numbers. What new numbers
are constructible from it with ruler and compass? We can draw all circles whose
centers have coordinates in F and whose radii are in F , and we can draw all
242 CHAPTER 6. FIELDS
PROBLEMS
1. Let K be an extension of F , and let a 2 K. Prove that F [a] = F (a) if
and only if a is algebraic over F .
2. Let K be an extension of F , and let a and b be in K. Prove that F (a)(b) =
F (b)(a).
3. Let K be a …nite extension of F , and let a and b be in K. Prove that
[F (a; b) : F ] [F (b) : F ][F (a) : F ].
4. Prove that if f (x) is irreducible over F [x] and [K : F ] is relatively prime
to deg(f (x)), then f (x) is irreducible over K[x].
5. Let F be a sub…eld of an integral domain K. Prove that if K has …nite
dimension as a vector space over F , then K is a …eld.
6. Prove that the usual formula for the roots of a quadratic equation holds
for any …eld of characteristic not 2.
6.2. SPLITTING FIELDS 243
p
7. Suppose that [K : Q] = 2. Prove that K = Q( n), where n is a square-
free integer.
16. Prove that the angles of 72o and of 60o are constructible, while the angle
of (360=7)o is not constructible.
17. Prove that a regular pentagon and a regular hexagon are constructible,
while a regular septagon is not constructible.
Theorem 6.2.1 Let F be a …eld, and let f (x) be in F [x] and have degree at
least one. Then there is a …nite extension K of F containing a root of f (x).
244 CHAPTER 6. FIELDS
Proof. Let p(x) be a prime factor of f (x). Then F [x]=(p(x)) is a …eld, and
it contains F via the embedding
F ! F [x]=(p(x)) : a ! a + (p(x)).
a0 + a1 k + + an k n = 0:
a0 + a1 (x + (p(x))) + + an (x + (p(x)))n
= a0 + a1 x + + an xn + (p(x)) = 0.
f1 + (p(x)), x + (p(x)), , xn 1
+ (p(x))g,
Splitting …elds are also called root …elds. If K is a root …eld of f (x) 2 F [x],
then f (x) = a(x a1 )(x a2 ) (x an ) in K[x], and K = F (a1 , a2 , : : : , an ).
We know that every polynomial f (x) in F [x] has a splitting …eld. We want to
show that any two splitting …elds of f (x) are essentially the same.
Let : F ! G be an isomorphism between the …elds F and G. Then
induces an isomorphism : F [x] ! G[x] by
Corollary 6.2.4 Let p(x) be prime in F [x]. If a and b are roots of p(x) in an
extension of F , then there is an isomorphism : F (a) ! F (b) such that is
the identity on F and such that (a) = b.
Proof. We will induct on the degree of f (x). If deg(f (x)) = 1, then take
= . Let p(x) be a prime factor of f (x). Let a be a root of p(x) in K, and let
b be a root of (p(x)) in L. By 6.2.3, there is an isomorphism : F (a) ! G(b)
such that = on F , and such that (a) = b. Write f (x) = (x a)g(x) in
F (a)[x]. Now K is a splitting …eld for g(x) over F (a), and L is a splitting …eld
for (g(x)) over G(b). Since deg(g(x)) < deg(f (x)), there is an isomorphism
: K ! L such that agrees with on F (a). This completes the proof.
The following corollary is the result we are really after. In order to proceed
by induction, we were forced to prove the stronger result 6.2.5.
Corollary 6.2.6 Let f (x) 2 F [x], and let K and L be splitting …elds for f (x).
Then there is an isomorphism : K ! L such that is the identity on F .
Let F be a …eld, and let f (x) be an element of F [x] of degree > 1. There
is an algebraic extension K of F such that in K[x], f (x) factors into a product
of linear factors. It is natural to wonder if there is an algebraic extension K
of F such that every such f (x) 2 F [x] factors completely in K[x]. This is
indeed the case. In fact, there exists an algebraic extension K of F such that
every f (x) in K[x] of degree 1 factors into linear factors in K[x], and any
two such extensions of F are F -isomorphic, that is, if K and L are two such
extensions of F , then there is an isomorphism ' : K ! L such that '(a) = a
for all a 2 F . A …eld K such that every f (x) 2 K[x] of degree 1 factors
into a product of linear factors is said to be algebraically closed. This is
equivalent to every polynomial in K[x] of degree 1 having a root in K, and is
equivalent to K having no algebraic extensions. If K is an algebraic extension
246 CHAPTER 6. FIELDS
…elds. We need some information about them at this point. If K and L are
splitting …elds of f (x), then from 6.2.6 it follows that f (x) factors into distinct
linear factors over K if and only if it so factors over L.
The usual rules for di¤erentiation hold, and are easily veri…ed. Here is the test
for separability.
Theorem 6.2.9 A polynomial f (x) over the …eld F is separable if and only if
f (x) and f 0 (x) are relatively prime.
Proof. Since f (x) and f 0 (x) are both in F [x], their greatest common divisor
in K[x] for any extension K of F is the same as their greatest common divisor
in F [x]. Let K be the root …eld of f (x), and write f (x) = c ni=1 (x ai )ki in
K[x], with the ai distinct. We get
0 1
n
X Y
f 0 (x) = c @ki (x ai )ki 1 (x aj )kj A .
i=1 i6=j
so that f (x) and f 0 (x) have no root, and hence no factor, in common.
Proof. If f 0 (x) = 0, then f (x) and f 0 (x) are not relatively prime, whence
f (x) is inseparable. If f (x) is inseparable, then f (x) and f 0 (x) have a factor of
degree at least one in common. Since f (x) is prime, f (x) divides f 0 (x). But
deg(f 0 (x)) < deg(f (x)). Thus f 0 (x) = 0. If F has characteristic 0 and f (x)
is prime, then f (x) has degree at least one, and hence f 0 (x) 6= 0. If F has
248 CHAPTER 6. FIELDS
characteristic p and f (x) is prime, then it is apparent that f 0 (x) = 0 if and only
if f (x) = g(xp ).
It may not be clear that there exist prime inseparable polynomials. Here is
an example. Let P = Z=pZ be the …eld with p elements. Form the quotient
…eld K = P (y) of the polynomial ring P [y]. Let F be the sub…eld P (y p ) of
K. The polynomial xp y p 2 F [x] is inseparable since its derivative is 0. In
K[x], xp y p = (x y)p . If xp y p = f (x)g(x) is a non-trivial factorization
in F [x], then f (x) = (x y)n for 0 < n < p. Therefore y n 2 F . Now
F = P (y p ) consists of quotients of polynomials in y p with coe¢ cients in P .
Hence y n = h(y p )=k(y p ) with h(y p ) and k(y p ) in P (y p ), and so k(y p )y n = h(y p )
is in P [y p ]. But 0 < n < p clearly makes this impossible. Therefore xp y p is
prime in F [x], is separable, and has y as a root of multiplicity p in K.
We are in a position to use splitting …elds to describe completely all …nite
…elds. We begin by noticing a few elementary properties that a …nite …eld F
must have. The characteristic of F is a prime p. Hence F is a vector space of
…nite dimension n over its prime sub…eld P of p elements. Therefore F has pn
elements. Let q = pn . Every element of F satis…es the polynomial xq x in
P [x]. To see this, note that the multiplicative group F of non-zero elements
of F has q 1 elements. Thus for a 2 F , aq 1 = 1, whence aq = a. Now 0
certainly satis…es xq x. The polynomial xq x has q roots in F , hence splits
completely in F . Therefore F is the root …eld of xq x over the …eld P of p
elements. Any two such root …elds are isomorphic by 6.2.5. We have shown
that there is at most one …eld of order pn (up to isomorphism).
To show that there is a …nite …eld with pn = q elements, the natural thing
to do is to take the splitting …eld of xq x over the …eld P of p elements. This
gets a …eld F , but conceivably it is too small. If a and b are roots of xq x in
F , then
(a + b)q (a + b) = aq + bq a b = 0,
and (ab)q ab = aq bq ab = 0, using the facts that the characteristic of P is
p, so that py = 0 for any y 2 F , that aq = a, and that bq = b. Similarly, if
a 6= 0, then 1=a is a root of xq x. Therefore the root …eld F of xq x consists
entirely of roots of xq x. If this polynomial has no multiple roots, then F has
q = pn elements. By 6.2.8, this is the case. Thus we have the following theorem
about …nite …elds.
G. Since q o(G) and xq 1 has no more than q roots, it follows that q = o(G),
and that G is cyclic.
If F is any …eld of characteristic p, then
F ! F : a ! ap
K = F (a1 , a2 , : : : , an )
PROBLEMS
1. Find the degrees of the splitting …elds of the following polynomials over
Q.
a. x2 1 b. x2 + 1 c. x2 2
d. x3 2 e. x3 3 f. x4 5
g. x4 4 h. x4 + x2 + 1 i. xp 1, p a prime
2. Let K be an algebraic extension of F , and let f (x) 2 K[x]. Prove that
there is a …nite extension L of K such that in L[x], f (x) divides some
non-zero element of F [x].
3. Find the group of automorphisms of the …elds Q(21=3 ), Q(21=2 ), and Q(i).
4. Let C be the algebraic closure of a …eld F , and let K be a …eld between
F and C. Prove that C is the algebraic closure of K.
5. Prove that if F is countable, then so is F [x].
6. Prove that the algebraic closure of a countable …eld is countable.
7. Prove that the usual rules for di¤erentiation of polynomials hold. Thus
show that
12. Prove that the derivative of a polynomial f (x) over a …nite …eld F of
characteristic p is zero if and only if f (x) = (g(x))p for some g(x) 2 F [x].
13. Let F be a …eld of characteristic p 6= 0. Prove that if F ! F : a ! ap is
not onto, then F has a …nite inseparable extension.
14. Let K be an extension of the …eld F of characteristic p 6= 0. Let L = fa 2
K : aq 2 F for q some power of pg. Prove that L is a sub…eld of K.
Prove that [L : F ] is a power of p, if …nite.
15. Prove that if K is a separable extension of F , and L is any …eld between
K and F , then K is a separable extension of L.
16. Let K be a …nite extension of F , and let
fb1 , b2 , : : : , bn g
be a basis of K over F . For a 2 K, let abi = nj=1 aij bj . Prove that the
polynomial det(xIn (aij )) is a multiple of the minimum polynomial of a
over F . Prove that it is a power of the minimum polynomial of a over F:
fx1 , x2 , : : : , xn g
1;
a1 + a2 + a3 + a4 ;
6.3. GALOIS THEORY 253
a1 a2 + a1 a3 + a1 a4 + a2 a3 + a2 a4 + a3 a4 ;
a1 a2 a3 + a1 a2 a4 + a1 a3 a4 + a2 a3 a4 ;
and
a1 a2 a3 a4 .
For our purposes at the moment, the pertinent fact is this. If is any automor-
phism of K which induces a permutation of fa1 , a2 , : : : , an g, then …xes each
elementary symmetric polynomial in the ai ’s. This should be obvious.
c. G(K=F ) = G.
Theorem 6.3.8 Let K be a …nite extension of F . Then the following are equiv-
alent.
L ! G(K=L)
Lemma 6.3.12 Let f (x) be a monic polynomial with integer coe¢ cients. Let
f (x) = g(x)h(x) in Q[x] with g(x) and h(x) monic. Then g(x) and h(x) are in
Z[x].
Proof. Let a and b be the smallest positive integers such that ag(x) = m(x)
and bh(x) = n(x) are in Z[x]. We have abf (x) = m(x)n(x). Let m(x) =
r i s i
i=0 ai x , and let n(x) = i=0 bi x . Note that gcdfa0 , a1 , : : : , ar g = gcdfb0 ,
b1 , : : : , bs g = 1. If ab = 1, there is nothing to do. Otherwise, let p be a prime
dividing ab. Let i and j be the smallest indices such that p does not divide ai
and bj , respectively. Now p must divide the coe¢ cient : : : ai 2 bj+2 + ai 1 bj+1 +
ai bj + ai+1 bj 1 + : : : of xi+j , and it divides every term except ai bj . Therefore
ab = 1, and the lemma is proved.
This lemma is a classic. We will treat factorizations of polynomials over
certain integral domains in Chapter 8.
Theorem 6.3.13 The nth cyclotomic polynomial 'n (x) is in Z[x] and is irre-
ducible over Q[x].
Proof. Let ! be a primitive nth root of 1, and let f (x) be its minimum
polynomial over Q. Let p be a prime not dividing n, and let g(x) be the minimum
polynomial over Q of the primitive nth root ! p . By the lemma above, f (x) and
g(x) are in Z[x]. If f (x) 6= g(x), then they are relatively prime in Q[x], and
hence have no roots in common. They both divide xn 1 = f (x)g(x)h(x),
with h(x) 2 Z[x] also. The polynomial g(xp ) has ! as a root, whence g(xp ) =
f (x)k(x), again with k(x) 2 Z[x]. Now view the equalities
Theorem 6.3.14 The Galois group over Q of the nth cyclotomic polynomial
'n (x) is isomorphic to the group of units in the ring Z=nZ. In particular, it is
Abelian and has order '(n).
6.3. GALOIS THEORY 259
Proof. Let K be a root …eld of 'n (x) over Q, and let ! be a primitive
nth root of 1. Then K = Q(!), and if 2 G(K=Q), then (!) = ! i , where
0 < i < n and (i, n) = 1. Since 'n (x) is the minimum polynomial of ! over Q,
there is a 2 G(K=Q) for each such i. It is now completely routine to check
that ! i gives the desired isomorphism.
The proof above works with Q replaced by any …eld F of characteristic 0,
except that 'n (x) may not be irreducible over F and hence the map ! i may
not be onto. Thus we have
Theorem 6.3.16 Let F be a …eld of characteristic 0 which contains all the nth
roots of 1, and let a 2 F . Then the Galois group of xn a over F is isomorphic
to a subgroup of the additive group of Z=nZ. In particular, it is cyclic, and if
xn a is irreducible, it is cyclic of order n.
PROBLEMS
1. Prove that the root …eld of a separable polynomial is the root …eld of an
irreducible separable polynomial.
260 CHAPTER 6. FIELDS
2. Prove that the Galois group of a polynomial of degree n has order dividing
n!.
3. Prove that K is a …nite normal extension of F if and only if for any
extension L of K, every element of G(L=F ) induces an automorphism of
K.
4. Prove that if K is a …nite normal extension of F , if F L K, and if is
an F -isomorphism of L into K, then is induced by an F -automorphism
of K.
5. Prove that if K is a normal extension of F and if is an F -isomorphism
of K into itself, then is an F -automorphism of K.
6. Let K be a …nite extension of F . Prove that there is a (…nite) normal
extension N of F containing K such that if M is any normal extension
of F containing K, then there is a K-isomorphism of N into M . The
extension N of F is called a normal closure of the extension K of F .
Prove that any two normal closures of K over F are K-isomorphic.
7. Let K be a Galois extension of F , let 2 G(K=F ), and let L and
M be …elds between K and F . Prove that (L) = M if and only if
G(K=L) 1 = G(K=M ). (That is, prove that L and M are “conjugate”
sub…elds if and only if G(K=L) and G(K=M ) are conjugate subgroups.)
8. Prove that the Galois group G(K=F ) of a separable polynomial f (x) 2
F [x] is transitive on the roots of f (x) if and only if f (x) is irreducible in
F [x].
9. Prove that if K is a …nite separable extension of F , then the number of
…elds between K and F is …nite.
10. Prove that K(51=3 ) is the root …eld of x3 5 over the …eld K = Q(i31=2 ).
11. Find the Galois group of the polynomials
over the …eld Q. In each case, make explicit the correspondence between
subgroups and intermediate extensions, and determine which are normal.
12. Prove that xn 1 = djn 'd (x). Prove from this fact that for all n,
'n (x) 2 Z[x].
13. Calculate 'n (x) for n 12.
14. Let F be a …eld of prime characteristic p, and let n be a positive integer
prime to p. Prove that there are primitive nth roots of 1 in some extension
of F .
6.4. SOLVABILITY BY RADICALS 261
15. Prove that if p is prime, then the Galois group of xp 1 over Q is cyclic of
order p 1. Prove that the Galois group of xp a over Q is not necessarily
Abelian.
18. Let x be an indeterminate, and let y 2 F (x). Prove that F (x) = F (y) if
and only if there exist elements a, b, c, d 2 F such that ad bc 6= 0 and
such that y = (ax + b)=(cx + d). Find G(F (x)=F ).
De…nition 6.4.1 Let f (x) 2 F [x]. Then f (x) is solvable by radicals over
F if there is a …nite sequence
F = F0 F1 Fn
The condition on the Galois group of f (x) that is decisive for the solvability
of f (x) by radicals follows.
De…nition 6.4.2 A group G is solvable if it has a sequence of subgroups 1 =
G0 G1 G2 Gn = G such that each Gi is normal in Gi+1 and such
that Gi+1 =Gi is Abelian.
In particular, every Abelian group is solvable. A deep theorem only recently
proved (Feit and Thompson, 1963) asserts that every group of odd order is
solvable. However, the alternating groups An with n 5 are simple, and since
such An are non-Abelian, they are not solvable.
One should note that if G is …nite and solvable, then there is a sequence
of subgroups 1 = G0 G1 Gn = G such that Gi is normal in Gi+1 ,
and such that Gi+1 =Gi is cyclic of prime order. Indeed, if Gi+1 =Gi is Abelian
and has more than one element, then it has a subgroup H=Gi of prime order.
The subgroup H is normal in Gi+1 , and by induction, there is a …nite chain
H = H0 H1 Hr = Gi+1 with Hi normal in Hi+1 and with Hi+1 =Hi
cyclic of prime order. The rest should be clear. Therefore, in the case G is
…nite, the condition in 6.4.1 that Gi+1 =Gi be Abelian could be replaced by
the condition that Gi+1 =Gi be cyclic of prime order (or simply cyclic, if one
preferred).
Our goal is to show that f (x) 2 F [x] is solvable by radicals over F if and
only if the Galois group of f (x) over F is solvable. Keep in mind that our
…elds all have characteristic 0. First we record some basic properties of solvable
groups.
Theorem 6.4.3 Subgroups and quotient groups of solvable groups are solvable.
If N is a normal subgroup of G and if N and G=N are solvable, then G is
solvable.
Proof. Let S be a subgroup of the solvable group G. There is a chain 1 =
G0 G1 Gn = G of subgroups Gi of G such that Gi is normal in Gi+1
and Gi+1 =Gi is Abelian. Let Si = S \ Gi . Then 1 = S0 S1 Sn = S is
such a chain for S. Indeed, if x 2 Si+1 and y 2 Si , then xyx 1 2 S since x and
y are both in S, and xyx 1 2 Gi since x 2 Gi+1 , y 2 Gi , and Gi is normal in
Gi+1 . Hence Si is normal in Si+1 . The map Si+1 =Si ! Gi+1 =Gi : xSi ! xGi
is a monomorphism, whence Si+1 =Si is Abelian and S is solvable.
Now let G=N be a factor group of the solvable group G. In a similar vein,
one can verify that G0 N=N G1 N=N Gn N=N satis…es the conditions
necessary to make G=N solvable.
Suppose that N is a normal subgroup of G and that N and G=N are solvable.
Then there are the usual chains of subgroups 1 = N0 N1 Nr = N ,
and
1 = N=N G1 =N Gs =N = G=N
of N and G=N . It is completely routine to check that the chain of subgroups
1 = N0 N1 Nr G1 G2 Gs = G has the requisite
properties.
6.4. SOLVABILITY BY RADICALS 263
Lemma 6.4.4 Let f (x) 2 F [x] be solvable by radicals over F . Then there is
a …nite sequence F = F0 F1 Fn of …elds with each Fi a radical
extension of Fi 1 and a normal extension of F , and with Fn containing the root
…eld of f (x) over F .
Theorem 6.4.5 Let F be a …eld of characteristic 0, and let f (x) 2 F [x]. Then
f (x) is solvable by radicals over F if and only if the Galois group of f (x) over
F is solvable.
Each subgroup is normal in the next since each extension Ki+1 of Ki and K0 of F
is normal. For i 0, G(Kn =Ki )=G(Kn =Ki+1 ) is isomorphic to G(Ki+1 =Ki ) by
part of the Fundamental Theorem of Galois Theory (6.3.11), and G(Ki+1 =Ki ) is
Abelian by 6.3.17. Similarly, G(Kn =F )=G(Kn =K0 ) is isomorphic to G(K0 =F ).
By 6.3.15, G(K0 =F ) is Abelian. It follows from 6.4.3 that G(Kn =F ) is solvable.
The root …eld K of f (x) over F lies between F and Kn , and is normal over F .
By the Fundamental Theorem (6.3.11), the Galois group G(K=F ) of f (x) over
F is isomorphic to G(Kn =F )=G(Kn =K), and by 6.4.3, this quotient is solvable.
Hence the Galois group of f (x) over F is solvable.
The proof of the converse is a little complicated. In order not to lose sight
of it, we put the essence of the matter in the following lemma.
ai = a + ! i (a) + (! 2 )i 2
(a) + + (! p 1 i p 1
) (a).
…eld of s(x) over S. This implies that [K : S] n!. Since jG(K=S)j n!, it
follows from 6.3.8 that [K : S] = n! and that G(K=S) Sn . At this point,
we have of course proved 6.4.7. However, a couple of more observations about
symmetric polynomials are in order. Let E be the sub…eld of S generated
by F and the elementary symmetric polynomials in x1 , x2 , : : : xn . Then
F E X K. Since the coe¢ cients of s(x) (neglecting sign) are precisely
the elementary symmetric polynomials in x1 , x2 , : : : , xn , then s(x) 2 E[x]
and K is the root …eld of s(x) over E. Hence [K : E] n!. It follows that
S = E. In other words, the symmetric rational functions in x1 , x2 , : : : , xn are
rational functions of the elementary symmetric polynomials in x1 , x2 , : : : , xn .
Actually something stronger is true, but it has nothing to do with …eld theory.
A symmetric polynomial in x1 , x2 , : : : , xn is a polynomial in the elementary
symmetric polynomials. For example,
and
x21 x2 + x1 x22 + x1 x23 + x22 x3 + x2 x23 + x21 x3
= (x1 x2 + x2 x3 + x2 x3 )(x1 + x2 + x3 ) 3x1 x2 x3 :
We omit the proof in general, but it is only slightly di¢ cult.
We close this section by exhibiting a speci…c polynomial in Q[x] of degree 5
whose Galois group is S5 and hence is not solvable. Our Theorem 6.4.7 is not
applicable. However, its proof does enable one to get such a polynomial over a
sub…eld F of the …eld of real numbers. One only needs real numbers r1 , r2 , r3 ,
r4 , r5 such that if x1 , x2 , x3 , x4 , x5 are indeterminates, then
Q(x1 , x2 , x3 , x4 , x5 ) Q(r1 , r2 , r3 , r4 , r5 )
via an isomorphism such that (xi ) = ri for all i. Then the proof of 6.4.7
shows that the Galois group of f (x) = 5i=1 (x ri ) over the …eld generated by
Q and the coe¢ cients of f (x) is S5 . The existence of such a set of ri is most
easily shown by appealing to cardinality arguments. Pick a real number r1
which is not algebraic over Q. The …eld Q(r1 ) is countable, and only countably
many real numbers are algebraic over Q(r1 ). Pick a real number r2 which is not
algebraic over Q(r1 ). Proceeding in this way, we get real numbers r1 , r2 , r3 , r4 ,
r5 which …ll the bill.
A frequently used example of a polynomial in Q[x] which is not solvable by
radicals is x5 4x + 2. We will indicate that its Galois group over Q is S5 . First,
we need it to be irreducible over Q. This can be shown using the Eisenstein
criterion (8.2.9). The polynomial has exactly three real roots. Just graph it and
observe that it crosses the x axis exactly three times. It has no double roots
since it is irreducible over Q and hence separable. Hence it has exactly two
complex roots. Let K be the root …eld of x5 4x + 2 over Q in the …eld C of
complex numbers. If a and b are roots of x5 4x + 2, then there is an element
2 G(K=F ) such that (a) = b. This follows from the irreducibility of the
polynomial. The automorphism of C which takes a complex number a + bi to
a bi induces an automorphism of K which …xes the real roots of x5 4x + 2
266 CHAPTER 6. FIELDS
PROBLEMS
1. Prove that if F K and f (x) 2 F [x] is solvable over F , then f (x) is
solvable over K.
2. Prove that if G(K=F ) is …nite and solvable, and if f (x) 2 F [x] is solvable
over K, then f (x) is solvable over F .
6. Let f (x) be a separable polynomial in F [x], and let K be its root …eld
over F . Prove that f (x) is irreducible in F [x] if and only if G(K=F ) is
transitive as a group of permutations of the roots of f (x).
(a) x5 8x + 3
(b) x + x5 + x4 + x3 + x2 + x + 1
6
(c) x18 + x9 + 1
(d) x6 + x3 + 1
(e) x5 5x3 20x + 5.
F (a1 , : : : , ai 1, ai+1 , : : : , an ),
7.1 Introduction
Group theory is a vast subject which pervades almost every mathematical dis-
cipline. This chapter gives a brief introduction to some of the classical topics
in non-Abelian group theory. In Chapter 2, we presented the basic notions of
homomorphisms, normal subgroup, quotient group, direct sum, and so on. The
Fundamental Theorem of Finite Abelian Groups (2.7.6), and the fact that for
n 5, the alternating group An is simple (2.5.7) are the deepest results in that
chapter. In section 6.4, we introduced solvable groups (6.4.2) and derived some
of their properties. We will return to that topic in this chapter.
Since our discussion here is with groups that are not necessarily Abelian, we
will use multiplicative notation.
G = G0 G1 G2 Gn = feg
such that Gi+1 is normal in Gi , and Gi =Gi+1 is simple. Now suppose that
G = H0 H1 H2 Hm = feg
269
270 CHAPTER 7. TOPICS FROM GROUP THEORY
Proof. If a group has a composition series of length one, then the group is
simple and any two composition series are certainly equivalent. Now suppose
that a group G has a composition series
of length n > 1, and that if a group has a composition series of length less than
n, then any two composition series of that group are equivalent. Let
and
G = H0 H1 H1 \ G1 H2 \ G1 ::: Hm \ G1 = feg. (4)
Since Gi+1 \ H1 is a normal subgroup of Gi \ H1 and Gi Gi+1 , the Third
Isomorphism Theorem (2.3.12) yields
and Gi+1 (Gi \H1 ) is a normal subgroup of Gi since it is a product of two normal
subgroups. Since Gi =Gi+1 is a simple group, (Gi+1 (Gi \ H1 ))=Gi+1 is either
Gi =Gi+1 or Gi+1 =Gi+1 . That is, Gi+1 (Gi \ H1 ) is either Gi+1 or Gi . Therefore,
if we remove repetitions from
G1 G2 :. . Gn = feg,
7.2. THE JORDAN-HÖLDER THEOREM 271
and hence (1) and (3) (with repetitions removed) are equivalent. If G1 = H1 ,
then (1) and (2) are certainly equivalent. If G1 6= H1 , then G1 H1 is a normal
subgroup of G properly containing G1 , so G1 H1 = G. Thus G1 =(G1 \ H1 )
(G1 H1 )=H1 , and H1 =(H1 \ G1 ) (G1 H1 )=G1 = G=G1 . Therefore (3) (with
repetitions removed) and (4) (with repetitions removed) are equivalent, and the
theorem is proved.
PROBLEMS
1. Prove that if G has a composition series, then any normal subgroup of G
has a composition series.
2. Prove that if N is normal in G, and if G has a composition series, then so
does G=N .
3. Suppose that G has a composition series and that N is normal in G. Prove
that G has a composition series of which N is a member.
4. Suppose that G has a composition series. Prove that any normal series of
G has a re…nement that is a composition series.
5. (Zassenhaus’s Lemma) Let A and B be subgroups of a group G, and let
M and N be normal in A and B, respectively. Prove that
where the sum ranges over a set of representatives of the equivalence classes. If
an element a is in the center
of G, then Cl(a) = fag. The equation in the next theorem is called the class
equation, and it follows from the equations above.
Theorem 7.3.2 o(G) = o(Z(G)) + o(G)=o(N (a)), where the sum is over a
set of representatives a such that a 2
= Z(G).
PROBLEMS
1. Find all the Sylow 2-subgroups and all the Sylow 3-subgroups of S3 , S4 ,
and S5 .
2. Let H be a proper subgroup of the …nite group G. Prove that G is not
the set union of the conjugates of H.
3. Let G be a …nite group, let N be a normal subgroup of G, and let S be
a Sylow p-subgroup of G. Prove that S \ N is a Sylow p-subgroup of N ,
and that SN=N is a Sylow p-subgroup of G=N .
4. Prove that a group of order 28 has a normal subgroup of order 7. Prove
that a group of order 28 that does not have exactly 7 Sylow 2-subgroups
has a normal subgroup of order 4, and is Abelian.
5. If o(G) = pq with p and q primes, and p < q. Let S be a subgroup of G
of order p, and let T be a subgroup of G of order q.
6. Prove that any group of order less than 60 is either of prime order or has
a proper normal subgroup.
7. Prove that a group of order 108 has a normal subgroup of order 9 or 27.
8. Let G be …nite, and let S be a Sylow p-subgroup of G. If S is normal in
N , and if N is normal in G, prove that S is normal in G.
9. Prove that any simple group of order 60 is isomorphic to A5 .
10. If o(G) = pn , p a prime, and if H is a proper subgroup of G, then N (H) 6=
H.
G = G0 G1 G2 Gn = feg
such that Gi =Gi+1 is Abelian. Such a series will be called a solvable series.
We showed that subgroups and quotient groups of solvable groups are solvable,
276 CHAPTER 7. TOPICS FROM GROUP THEORY
and that if N is normal in G, then N and G=N solvable imply that G is solvable.
We noted further that a solvable group has a normal chain whose factors are
cyclic of prime order. Since a non-trivial p-group G has non-trivial center Z(G),
it follows by induction on o(G) that G is solvable. Indeed, Z(G) is a solvable
normal subgroup of G and G=Z(G) is solvable since it is a p-group smaller that
G. It follows that G is solvable. This fact is important enough to exhibit as a
theorem.
Actually, …nite p-groups satisfy a much stronger condition than that of solv-
ability. They are nilpotent, a concept we will come to shortly.
Recall that the commutator subgroup G0 of a group G is the subgroup
generated by all its commutators a 1 b 1 ab, where a and b are in G. The crucial
facts about G0 are that G0 is a fully invariant subgroup of G, G=G0 is Abelian,
and G0 is contained in every normal subgroup N such that G=N is Abelian. We
now de…ne the higher commutator subgroups of a group G inductively by
This chain is called the derived series of G. In the derived series, either each
inclusion is proper, or for some i, G(i) = G(i+n) for all n. Either situation may
occur. If G is …nite, the latter must occur, of course.
Theorem 7.4.2 A group G is solvable if and only if G(n) = feg for some n.
G = G0 G1 G2
7.4. SOLVABLE AND NILPOTENT GROUPS 277
where Gi+1 = [Gi ; G]. Recall that a subgroup H of a group G is fully invariant
if it is taken into itself by every endomorphism of G. It is an easy exercise to
show that every member of the lower central series of G is a fully invariant
subgroup of G. In particular, each member of the lower central series is a
normal subgroup of G. The factors of the series are certainly Abelian, since
Gi+1 = [Gi ; G] [Gi ; Gi ] = (Gi )0 .
Let Z 0 (G) = feg, Z(G) be the center of G, and in general Z i+1 (G) be given
by Z(G=Z i (G)) = Z i+1 (G)=Z i (G). The chain
Z(G=Z m i (G)) = Z m (i 1)
(G)=Z m i (G).
Theorem 7.4.9 A …nite group is nilpotent if and only if it is the direct product
of its Sylow subgroups.
PROBLEMS
1. Prove that (G H)0 = G0 H 0.
2. Q
Prove that if G1 , G2 , , Gn are solvable, then so is the direct product
Gi .
7.4. SOLVABLE AND NILPOTENT GROUPS 279
3. Prove that if m and n are integers such that m n, then (G=G(n) )(m) =
G(m) =G(n) .
4. Prove that any group of order pq is solvable, where p and q are primes.
5. Prove that any group of order p2 q is solvable, where p and q are primes.
6. Prove that a solvable group G 6= feg has a fully invariant Abelian subgroup
6= feg.
7. Prove that a …nite solvable group G 6= feg has a fully invariant Abelian
p-subgroup 6= feg.
9. Let A and B be two …nite solvable groups of the same order. Prove that
there is a one-to-one correspondence between the factors of a composition
series of A and the factors of a composition series of B such that the
corresponding factors are isomorphic.
11. Prove that every member of the lower central series of a group G is a fully
invariant subgroup of G.
12. Prove that every member of the upper central series of a group G is a
characteristic subgroup of G.
15. Prove that if A and B are subgroups of G, then [A; B] B if and only if
A N (B):
16. Prove directly from the de…nition of nilpotent groups that subgroups and
quotient groups of nilpotent groups are nilpotent.
19. Q
Let H be a subgroup of a …nite nilpotent group G. Prove that H =
(H \ Sp ), where Sp ranges over the Sylow subgroups of G.
280 CHAPTER 7. TOPICS FROM GROUP THEORY
where the product ranges over the primitive nth roots of 1, is in Z[x] (6.3.13).
Proof. Let D be a …nite division ring. Its center Z(D) is a …eld F , and
since D is a …nite dimensional vector space over F , D has q n elements, where
o(F ) = q and n is the dimension of D over F . Let a 2 D. Then it is easy to
check that
N (a) = fd 2 D : da = adg
is a subdivision ring of D. Since N (a) F , it is a vector space over F , so that
o(N (a)) = q n(a) for some integer n(a). Now consider the class equation for the
group D of non-zero elements of D. It is
qn 1=1 1+ a (q
n
1)=(q n(a) 1)
where the sum is over representatives a of each conjugate class for a not in the
center Z(D ) of the group D . Since q n(a) 1 divides q n 1, n(a) must divide
n. In fact, if n = n(a)m + r with 0 r < n(a), then
and since q n mn(a) 1 < q n(a) 1 and q n(a) 1 divides q n 1, it follows that
n mn(a) = 0. Therefore, in each term (q n 1)=(q n(a) 1) in the class equation,
n(a) divides n. Now consider the polynomials xn 1 = (x i ) and
Y
xn(a) 1 = (x j ),
where the i range over the nth roots of 1 and the j range over the n(a)th
roots of 1. Since no j is a primitive nth root of 1, n (x) = (x i ), where i
ranges over the primitive nth roots of 1, is relatively prime to xn(a) 1. Since all
these polynomials are in Z[x], 'n (x) divides (xn 1)=(xn(a) 1) in Z[x]. Hence
the integer 'n (q) divides (q n 1)=(q n(a) 1) in Z. From the class equation,
7.6. THE FUNDAMENTAL THEOREM OF ALGEBRA 281
this forces the integer 'n (q) to divide q 1. But 'n (q) = (q i ), where i
ranges over the primitive nth roots of 1. Each q i has absolute value jq ij
strictly greater than q 1 if n > 1. Indeed, if i = a + bi, then since ni = 1, i
has absolute value 1, so a2 + b2 = 1. Thus
and
jq 1j2 = q 2 2q + 1.
If n > 1, then a < 1. Since q is positive, it follows that
q2 2qa + 1 > q 2 2q + 1,
PROBLEMS
1. Assume that every positive real number has a square root. Prove that
every complex number has a square root.
2. Prove that every polynomial in R[x] factors in R[x] into a product of linear
and quadratic factors.
3. Let A be the set of all elements in C that are algebraic over Q. Use 7.6.1
to prove that A is algebraically closed. Generalize.
Chapter 8
8.1 Introduction
Ring theory is broadly divided into two areas— commutative ring theory and
non-commutative ring theory. Commutative ring theory has developed from
algebraic geometry and algebraic number theory. The prototype of the rings
studied in algebraic geometry is the ring F [x1 , x2 , : : : , xn ] of polynomials in
n indeterminates x1 , x2 , : : : , xn with coe¢ cients from a …eld F . In algebraic
number theory, the prototype is the ring Z of integers. Both of these areas are
vast and important mathematical topics, and are part of what is generally called
commutative algebra. Commutative algebra is the study of commutative rings.
The …rst topics we take up are from commutative ring theory. Section 8.2
takes up unique factorization domains, and the goal there is to show that if
R is a unique factorization domain, then so is R[x] (8.2.10). This is one of
several important theorems in ring theory that asserts that if R satis…es a certain
property, then so does R[x]. Section 8.3 is in that same spirit. There we prove
one of the most famous of these theorems –the Hilbert Basis Theorem (8.3.2).
That theorem says that if R is a commutative Noetherian ring, then so is R[x].
Section 8.4 presents a classical topic from commutative Noetherian rings, the
Noether-Lasker decomposition theory. These three topics from commutative
ring theory were chosen because they require minimum background, and yet
impart some of the ‡avor of commutative ring theory.
The prototype of the rings studied in non-commutative theory is the ring Fn
of all n n matrices over a …eld F , or more generally the ring Dn of all such ma-
trices over a division ring D. In fact, our only topic (8.5) from non-commutative
ring theory deals essentially with just these rings. We begin that topic with an
examination of the rings Dn . One of the best theorems in non-commutative
ring theory is the Wedderburn-Artin Theorem (8.5.9) which gives elegant nec-
essary and su¢ cient conditions for a ring to be isomorphic to some Dn . We will
prove this theorem via the more general Jacobson Density Theorem (8.5.8). It
is hoped that this topic will impart some of the ‡avor of non-commutative ring
283
284 CHAPTER 8. TOPICS IN RING THEORY
theory.
then s = k, and after renumbering, ai and bi are associates and mi = ni for all
i. If a and b are two non-zero elements of R, then we can write
a = uam 1 m2
1 a2 am
r ,
r
and
b = van1 1 an2 2 anr r ;
with u and v units, ai irreducible, and mi and ni 0. If this is done, then a
divides b if and only if mi ni for all i. Finally, note that an element a in R is
irreducible if and only if it is prime (Problem 3).
Lemma 8.2.2 In a UFD, any two elements have a greatest common divisor
(gcd).
Proof. Let a and b be in a UFD R. The only non-trivial case is the one
where neither a nor b is zero or a unit. Write a and b as above, and let ki be
the minimum of mi and ni . We claim that
De…nition 8.2.3 Let R be a UFD, and let f (x) be in R[x]. Then f (x) is
primitive if a gcd of the coe¢ cients of f (x) is 1. A gcd of the coe¢ cients of
f (x) is called the content of f (x), and denoted c(f (x)).
The content c(f (x)) is unique only up to multiplication by a unit. Note that
if f (x) is a non-zero element of R[x] where R is a UFD, then f (x) = c(f (x))f1 (x)
with f1 (x) primitive. Also note that if f (x) = c g(x) with g(x) primitive, then
c is a content of f (x). The following classical lemma is crucial.
(If i + j > m, take ai+j = 0, and similarly for bi+j :) Since p divides a0 , a1 , : : : ,
ai 1 and b0 , b1 , : : : , bj 1 , then p divides every term except possibly ai bj . But
p divides ci+j , whence p divides ai bj . Thus p divides either ai or bj , which is
impossible. Thus f (x)g(x) is primitive.
There are a number of corollaries of interest.
Corollary 8.2.5 Let R be a UFD, and let f (x) and g(x) be in R[x]. Then
c(f (x)g(x)) = c(f (x))c(g(x)), up to units.
Proof. Write f (x) = c(f (x))f1 (x), and g(x) = c(g(x))g1 (x) with f1 (x) and
g1 (x) primitive. Then
Corollary 8.2.6 Let R be a UFD, and let F be its quotient …eld. If f (x) is in
R[x] and if f (x) = g(x)h(x) in F [x], then f (x) = g1 (x)h1 (x) with g1 (x) and
h1 (x) in R[x] and with deg(g1 (x)) = deg(g(x)).
286 CHAPTER 8. TOPICS IN RING THEORY
Proof. Let r and s be non-zero elements of R such that rg(x) and sh(x)
are in R[x]. Then rsf (x) = (rg(x))(sh(x)), rg(x) = c1 g1 (x), and sh(x) =
c2 h1 (x) with ci in R and g1 (x) and h1 (x) primitive elements of R[x]. Thus
rsf (x) = c1 c2 g1 (x)h1 (x). Since g1 (x) and h1 (x) are primitive, rs divides c1 c2 .
Thus f (x) = (cg1 (x))h1 (x). Since the degree of g1 (x) is the degree of g(x), the
corollary follows.
The following two corollaries are immediate consequences of 8.2.6.
Corollary 8.2.7 Let R be a UFD, and let F be its quotient …eld. A primitive
polynomial f (x) in R[x] is irreducible in F [x] if and only if it is irreducible in
R[x].
Corollary 8.2.8 Let R be a UFD, and let F be its quotient …eld. If f (x) is
monic, f (x) is in R[x], and f (x) = g(x)h(x) in F [x], then f (x) = g1 (x)h1 (x)
in R[x] with g1 (x) and h1 (x) monic and having the same degrees as g(x) and
h(x), respectively.
Corollaries 8.2.6, 8.2.7, and 8.2.8 are of interest even for the case R = Z.
Here is the main theorem.
with each factor irreducible in R[x], with the ai and bi in R, and the fi (x)
and gi (x) polynomials of positive degree. Each fi (x) and gi (x) is primitive and
hence irreducible in F [x] by 8.2.7. In F [x], the ai and bi are units. Therefore,
viewing the equality in the PID F [x], m = n, and after rearrangement, fi (x)
8.2. UNIQUE FACTORIZATION DOMAINS 287
and gi (x) are associates in F [x]. Hence (ri =si )fi (x) = gi (x) with ri and si in
R, and so ri fi (x) = si gi (x). Since fi (x) and gi (x) are primitive, ri and si are
associates in R. Thus we have a1 a2 aj = b1 b2 bk v with v a unit. Since R
is a UFD, j = k, and after renumbering, ai and bi are associates in R. This
completes the proof.
Theorem 8.2.13 (The Eisenstein Criterion) Let R be a UFD, and let f (x) =
a0 + a1 x + : : : +an xn be in R[x]. If p is a prime in R such that p does not divide
an , p divides each of a0 , a1 , : : : , an 1 , and p2 does not divide a0 , then f (x) is
irreducible over F [x], where F is the quotient …eld of R.
Proof. Write f (x) = c(f (x))f1 (x). Then f (x) satis…es the hypothesis on
its coe¢ cients if and only if f1 (x) does, and f (x) is irreducible in F [x] if and
only if f1 (x) is irreducible in F [x]. Thus we may assume that f (x) is primitive.
Suppose that f (x) factors properly in F [x]. Then by 8.2.7,
with the bi and ci in R, r and s positive, and br and cs not zero. Since p
divides a0 , p divides b0 c0 = a0 . Since p2 does not divide a0 , p divides exactly
one of b0 and c0 , say b0 . If p divides all the bi , then p divides an . Thus p
does not divide some bk , and let k be the smallest such integer. Now ak =
bk c0 + bk 1 c1 + + b0 ck , p divides b0 , b1 , , bk 1 , and since k < n, p divides
ak . Thus p divides bk c0 . But p divides neither bk nor c0 . This contradiction
establishes the theorem.
Eisenstein’s Criterion is of special interest for the case R = Z.
An easy example of an integral domain that is not a UFD is this. Let F
be any …eld, and let R = F [x2 ; x3 ] be the subring of F [x] generated by F , x2 ,
and x3 . Then in R, x6 = x2 x2 x2 = x3 x3 , and both x2 and x3 are irreducible
elements. Neither element is prime. For example, x2 divides x3 x3 , but does
not divide x3 .
A less apparent example is the integral domain R = fm + n( 5)1=2 : m,
n 2 Zg. To show that R is not a UFD, it is convenient to introduce the map
N : R ! Z given by
One can readily check that N is multiplicative, that is, that if r and s are
in R, then N (rs) = N (r)N (s). Therefore, if u is a unit in R, then N (uu 1 ) =
288 CHAPTER 8. TOPICS IN RING THEORY
The factors are all irreducible. For example, N (3) = 9, so for any factor r of
3, N (r) = 1 or N (r) = 3. But N (r) = 1 implies that r = 1, and N (m +
n( 5)1=2 ) = m2 + 5n2 = 3 is impossible. Similarly, one can show that the other
factors are irreducible. Therefore R is not a UFD.
PROBLEMS
1. Prove that an element in a UFD is irreducible if and only if it is prime.
and let
b = van1 1 an2 2 : : : anr r ,
with the ai irreducible, mi and ni positive, and u and v units. Prove that
a divides b if and only if mi ni for all i.
5. Let R be a UFD, and let f (x) = ag(x) = bh(x) with a and b in R, and g(x)
and h(x) primitive elements of R[x]. Prove that a and b are associates in
R, and that g(x) and h(x) are associates in R[x].
x4 + 2y 2 x3 + 3y 3 x2 + 4 yx + y + 8y 2
7. Let R be a UFD. Prove that if f (x) is in R[x] and f (x) is reducible, then
so is f (x + a) for each a in R.
10. Prove that if R is a UFD and if f (x) is a monic polynomial with a root
in the quotient …eld of R, then that root is in R.
8.3. THE HILBERT BASIS THEOREM 289
I1 I2 I3
a. R is left Noetherian.
b. R satis…es the maximum condition on left ideals.
c. Every left ideal of R is …nitely generated.
Proof. Assume (a), and let S be a non-empty set of left ideals of R. Let
I1 be in S. If I1 is not a maximal element of S, then there is an element I2 of
S such that I1 is properly contained in I2 . If I2 is not maximal, then there is
an I3 in S such that I2 is properly contained in I3 , and so on. Since R is left
Noetherian, we eventually reach an In that is maximal in S. Thus (a) implies
(b).
Assume (b), and let I be a left ideal of R. Let r1 be in I, and let I1 be the
left ideal generated by r1 , that is, I1 = Rr1 . If I1 is not I, then there is an
element r2 in I and not in I1 . Let I2 be the ideal Rr1 + Rr2 generated by r1
and r2 . Then I2 properly contains I1 . If I2 is not I, then there is an element r3
in I and not in I2 . This process must stop. That we must come to an In such
that In = I. Otherwise we get a non-empty set fI1 , I2 , I3 , : : : g of left ideals
which has no maximal element. Thus I = In for some n, whence I is generated
by the …nite set fr1 , r2 , : : : , rn g. Thus (b) implies (c).
Assume (c), and let I1 I2 I3 be a chain of left ideals of R. By
(c), the left ideal I = [j Ij is generated by a …nite subset fr1 , r2 , : : : , rn g of I.
290 CHAPTER 8. TOPICS IN RING THEORY
Each ri is in some member of the chain of left ideals, and hence there exists an
n such that all ri are in In . Thus In = I, and so
In = In+1 = In+2 = .
g=f s1 xd n
fn1 srn xd n
fnrn
Lemma 8.4.1 Let R be a Noetherian ring. Then every ideal of R is the inter-
section of …nitely many irreducible ideals.
Proof. Let be the set of all ideals of R which are not intersections of
…nitely many irreducible ideals. We need to show that is empty. If is not
empty, then it has a maximal element I because R is Noetherian. Since I is not
irreducible, I = J \ K, where J and K are ideals properly containing I. Since
J and K properly contain I and I is a maximal element of , neither J nor K
is in . Hence J and K are …nite intersections of irreducible ideals, whence I
is a …nite intersection of irreducible ideals. This is a contradiction. Therefore
is empty, and the lemma is proved.
The next step is to show that in a Noetherian ring, irreducible ideals are
primary. Let I be an ideal, and let S be a subset of R. Then I : S is de…ned to
be the set fr 2 R : rS Ig. That is, I : S is the set of all r in R such that r s
is in I for all s in S. It is easy to check that I : S is an ideal of R:
I:b I : b2 I : b3 :. .
bx = bu + rab = bv + sbn+1
PROBLEMS
p
1. Let I1 , I2 , : : : , In and J be ideals of R, and let Ji = Ii . Prove that
3. Prove that if Q is primary for P , and if I and J are ideals such that
IJ Q and I is not contained in Q, then J P .
p
4. Let P and Q be ideals of R. Prove that Q is primary and Q = P if and
only if
(a) Q P,
(b) if a is in P , then an is in Q for some n, and
(c) if a b is in Q and a is not, then b is in P .
(a(cij )) = a((cij ) ).
Hence for any (cij ) in Cj , (cij ) = (cij )d. Now it is easy to see that the
endomorphism ring of Cj is isomorphic to D. Just associate each with its
corresponding d.
Now consider Cj as a right module over E(Cj ), or equivalently, over D. It
is then a right vector space of dimension n over D, and its endomorphism ring,
that is, the ring of linear transformations on Cj , is just the ring Dn .
We sum up our discussion so far in the following theorem.
Theorem 8.5.1 Let D be a division ring, and let Dn be the ring of all n n
matrices over D. Then Dn is a simple ring. As a left module over itself, Dn
is semisimple. In fact, Dn is the direct sum of n mutually isomorphic simple
modules. The endomorphism ring E of each of these simple Dn -modules S is
isomorphic to D, and the endomorphism ring of S considered as a module over
E is isomorphic to Dn . Analogous statements hold for Dn considered as a right
module over itself.
There are still some things about Dn that we would like to know. For
example, suppose that S is a simple left Dn -module. Is S isomorphic to the
8.5. SEMISIMPLE RINGS 297
simple modules Cj ? That is, does Dn have, up to isomorphism, just one simple
left module? The answer is yes, and to see that, let s be any non-zero element
of S. Then the map Dn ! S : r ! rs is an epimorphism of left Dn -modules.
Its kernel M is a left ideal, and since Dn =M S and S is simple, M is maximal
in Dn . That is, there is no left ideal strictly between M and Dn . Since Dn =
C1 C2 Cn and M 6= Dn , M does not contain some Cj . But since Cj
is simple, M \ Cj = 0. Since M is maximal in Dn , Dn = M Cj . Hence
Dn =M Cj . But Dn =M S. Therefore S Cj . Similarly, Dn has only one
simple right module (up to isomorphism).
The left module Dn is the direct sum of n simple left modules. It cannot
be the direct sum of m simple left modules with m 6= n since any simple left
Dn -module has dimension n as a vector space over D. This also follows from
the following more general theorem, which we will need later.
Theorem 8.5.2 Let M be a left module over any ring R. Suppose that M is
the direct sum S1 S2 Sn of simple modules Si . If M = T1 T2 Tm
with each Ti simple, then m = n, and after renumbering, Si Ti for all i.
S M=Tm T1 T2 Tm 1 S1 S2 Sn 1:
Corollary 8.5.3 Suppose that R is a ring, and as a left module over itself, R
is the direct sum of n minimal left ideals. If R is the direct sum of m minimal
left ideals, then m = n.
Theorem 8.5.5 Suppose that D and E are division rings, and that the rings
Dn and Em are isomorphic. Then m = n and D E.
298 CHAPTER 8. TOPICS IN RING THEORY
0 : S = fr 2 R : rS = 0g = 0:
fs1 , s2 , : : : , sn g
is a linearly independent subset of the right vector space S over the division ring
D, and if x1 , x2 , : : : , xn are elements of S, then there is an element r in R
such that rsi = xi for all i.
Proof. Before we start the proof of the theorem, let us see what it says in
a special case. If R is a simple ring, then certainly 0 : S = 0. If it turns out
that S is …nite dimensional over D, then taking fs1 , s2 , : : : , sn g to be a basis
of S, we see that every linear transformation of the vector space S over D is
just multiplication by an element of r. Since no element r of R annihilates all
of S, R is, in e¤ect, the ring of all linear transformations on the n-dimensional
vector space S. Thus R Dn . We will draw additional corollaries.
Now to the proof of the theorem. First we will prove the preliminary fact
that if T is a subspace of S of …nite dimension, and if s is in S and not in T ,
then there is an element r in R such that rT = 0 and rs 6= 0. We induct on
dim(T ). If dim(T ) = 0, let r = 1. Suppose that dim(T ) = n > 0, and that the
assertion is true for all subspaces of dimension less than n. Now T = V xD
with Dim(V ) = n 1. Let
I = fr 2 R : rV = 0g:
i(xd s) = ixd i s = 0,
such that r1 asn = xn rsn . We have (r + r1 a)si = rsi + r1 asi = rsi = xi for
i < n, and
(r + r1 a)sn = rsn + r1 asn = xn .
This concludes the proof of the theorem.
Suppose that R is a simple ring and that S is a simple left R-module. Then
the hypotheses of 8.5.8 are satis…ed, and as we pointed out, if S is of dimension
n over the division ring D = ER (S), then R Dn . How can we tell from R itself
whether or not S is …nite dimensional? Suppose that S is in…nite dimensional.
Then there is a subset fs1 , s2 , : : : g of S such that for all n, fs1 , s2 , : : : , sn g
is linearly independent. Let In be the left ideal fr 2 R : rsi = 0 for all i ng.
We have the chain I1 I2 I3 of left ideals, and by the proof of 8.5.8,
all the inclusions are proper ones. That is, R does not satisfy the descending
chain condition on left ideals. Thus, if R satis…es the descending chain condition
on left ideals, S would have to be …nite dimensional. A ring that satis…es the
dcc on left ideals is called left Artinian, and right Artinian rings are similarly
de…ned. Also, a left Artinian ring R has a minimal left ideal S, and S is then a
simple left R-module. Thus we have the following theorem.
A ring R which has a simple left module S such that 0 : S = 0 is called left
primitive.
Proof. A ring which is left semisimple is left Artinian (Problem 7). Then
8.5.11 follows from 8.5.9.
Ri = Si1 Sini :
ri = 1 ri = ri 1 = (e1 + e2 + + em )ri =
ri (e1 + e2 + + em ) = ei ri = ri ei ;
so that ei is the identity of the ring Ri . Now it should be clear that each Ri is
a left semisimple ring, Ri = Si1 Sini , and the Sij , j = 1, 2, : : : , ni are
mutually isomorphic. By 8.5.7, Ri is simple, and so is a ring of matrices over a
division ring.
Thus R1 \ Aj 6= 0 for some i. Since Ri and Aj are both minimal two sided
ideals, Ri = Aj . Thus each Aj equals some Ri , and the uniqueness of the Ri
follows.
PROBLEMS
1. Let be an isomorphism from the ring R to the ring S. Prove that if I is
a left ideal of R, then (I) is a left ideal of S. Prove that if I is minimal,
then so is (I). Prove that if the left R-module R is semisimple, then so
is the left S-module S. Prove that if R is left Artinian, then so is S.
2. Let R be a ring, and let I be a summand of the left R-module R. Prove
that I = Re with e2 = e.
302 CHAPTER 8. TOPICS IN RING THEORY
5. Let R be a ring such that every left ideal of R is a summand of the left
R-module R. Prove that if R is left Noetherian, then R is semisimple.
9. Let e1 and e2 be idempotents in the ring R. Prove that the left modules
Re1 and Re2 are isomorphic if and only if the right modules e1 R and e2 R
are isomorphic.
a b
0 c
11. Let D be a division ring. Let R be the ring of all upper triangular 2 2
matrices with entries from D. Is R simple? Is R semisimple?
14. Let G be the group with two elements. Prove that the group ring Q(G)
is semisimple. Do the same thing for the group G with three elements.
(Actually, Q(G) is semisimple for every …nite group G.)
16. Prove that a ring is semisimple if and only if it is left Artinian, and for all
minimal left ideals I, I 2 6= 0.
17. Prove that a simple ring with a minimal left ideal is semisimple.
8.5. SEMISIMPLE RINGS 303
18. Prove that the intersection J of all maximal left ideals of a ring R is a two
sided ideal. Prove that J is the intersection of all maximal right ideals of
R.
19. Let I be a left ideal of the Noetherian ring R. Prove that 1 + a has a left
inverse for all a in I if and only if I is contained in all the maximal left
ideals of R. (This is true without the assumption that R is Noetherian.
See the Appendix, Problem 9.) Prove that if 1 + a has a left inverse for
all a in I, then 1 + a has a right inverse for all a in I.
304 CHAPTER 8. TOPICS IN RING THEORY
Appendix: Zorn’s Lemma
While we have not hesitated to use the Axiom of Choice in its believable intuitive
form, and have done so without explicit mention, we have refrained from using
the more technical arguments involving Zorn’s Lemma. Such arguments are
used extensively in mathematics, especially in algebra. We will illustrate the
use of this important tool in this appendix.
Let I be a non-empty set, and let fSi gi2I be a family of non-empty sets
Si . Is there a function f : I ! [i Si such that f (i) is in Si for all i in I? The
assumption that there is such a function is the Axiom of Choice. The function
f “chooses” an element from each set Si . The Axiom of Choice can be neither
proved nor disproved from the usual axioms of set theory. However, it is an
extremely useful axiom, and used by most mathematicians. There are several
mathematical statements logically equivalent to the Axiom of Choice, and on
the intuitive level, some of them are not nearly as believable as the Axiom of
Choice. Zorn’s Lemma is one of these. We need some preliminary de…nitions
before stating it.
1. s s for all s in S;
2. if s t and if t u, then s u;
3. if s t and t s, then s = t.
305
306 APPENDIX: ZORN’S LEMMA
Zorn’s Lemma is logically equivalent to the Axiom of Choice, but this is not
obvious.
Note that S cannot be empty. If S were empty, then S is a chain in S, and
thus must have an upper bound in S. That is, S must have an element s such
that t s for all t in S:
Here is how Zorn’s Lemma will often be applied. Let A be a set. Let S be
a set of subsets of A. Then S is partially ordered by setting A1 A2 if A1 is a
subset of A2 . To conclude that S has a maximal element, one must show that
every chain C in S has an upper bound in S. This is usually done by showing
that the union of the elements of C is in S. Of course, S has to be a very special
set of subsets of A for this to be so.
Here are some applications of Zorn’s Lemma.
Theorem .0.18 Theorem. Let W be a subspace of the vector space V over the
…eld F . Then W is a summand of V .
Theorem .0.19 Let R be a ring with identity. Then R has a maximal left ideal.
Proof. Recall that a maximal left ideal of R is an ideal M of R such that
M 6= R and such that there is no left ideal strictly between M and R.
Let S be the set of all left ideals I of R such that I 6= R. The set S is
partially ordered by inclusion. For any chain C in R, [I2C I = J is a left ideal
of R and is not R since 1 is not in J. Also J is an upper bound of C. Thus S
has a maximal element.
Theorem .0.20 Let R be a semi-simple ring. Then any left R-module M is a
direct sum of simple modules.
Proof. Let M be a left R-module. The module M is a direct sum of simple
submodules if there is a set fSi gi2I of simple submodules of M such that every
element m in M can be written uniquely in the form m = i2I si with si in Si .
Of course, all but …nitely many si must be 0. Let S be the set of all sets X of
simple submodules of M such that the submodule generated by the submodules
in X is the direct sum of the submodules in X. The set S is partially ordered
by inclusion, and it should be clear that the union of a chain in S is in S, so
that chains in S have upper bounds in S. Thus S has a maximal element Y . We
need to show that M is the direct sum of the elements of Y . The only problem
is that the submodule N generated by the elements of Y may not be all of M .
Suppose that there is an element m in M such that m is not in N . Then since
R is semi-simple, Rm is the direct sum A1 Ak of simple modules. Some
Ai is not in Y , and it follows readily that Y [ fAi g is an element of S. That is,
Y is not maximal in S. This completes the proof.
Theorem .0.21 Any …eld F has an algebraic closure, and any two algebraic
closures of F are F -isomorphic.
Proof. Recall that an algebraic closure of F is an algebraic extension K
of F such that every polynomial in K[x] of degree at least one factors into
linear factors in K[x]. Let X be a set in one-to-one correspondence with the
set of all polynomials in F [x] of degree at least one. We denote by x(f ) the
element in X corresponding to f (x). Now form the ring F [X]. This is the ring
of all polynomials in indeterminates from X with coe¢ cients from F . Thus any
element in F [X] is an element of
F [x(f1 ), , x(fn )]
for some …nite number x(f1 ), ; x(fn ) of indeterminates. Let I be the ideal
of F [X] generated by all the elements f (x(f )). If I were F [X], then 1 =
n
i=1 gi fi (x(fi )) for some gi in F [X] and generators fi (x(fi )). The gi involve
only …nitely many x(f )’s, so each gi is a polynomial gi (x(f1 ), , x(fm )), with
m n. Let F1 be a …nite extension of F in which each fi x(fi )), i = 1, , n
has a root ri , and for i > n, set ri = 0. Substituting ri for x(fi ) in our relation
n
X
1= gi (x(f1 ); : : : ; x(fm ))fi (x(fi )),
i=1
308 APPENDIX: ZORN’S LEMMA
F ! K1 : a ! a + M ,
PROBLEMS
1. Let L be a linearly independent subset of a vector space V . Prove that V
has a basis containing L.
2. Let W be a subspace of a vector space V , and let X be a subspace of
V such that X \ W = 0. Prove that V = A W for some subspace A
containing X.
3. Let I be a left ideal of a ring R, with I 6= R. Prove that R has a maximal
left ideal M such that I M .
4. Prove that every ring with identity has a maximal two sided ideal.
309
1. Birkho¤, G., and MacLane, S., A Survey of Modern Algebra, 3rd ed., New
York: Macmillan, 1963.
2. Fraleigh, J. B., A First Course in Abstract Algebra, 2nd ed., Reading:
Addison-Wesley, 1976.
3. Gallian, J. A., Contempory Abstract Algebra, Lexington: D. C. Heath,
1986.
4. Herstein, I. N., Abstract Algebra, New York: Macmillan, 1986.
5. Herstein, I. N., Topics in Algebra, 2nd ed., New York: Wiley, 1975.
6. Hungerford, T. W., Algebra, New York: Holt, Rinehart, and Winston,
1974.
7. Jacobson, N., Basic Algebra I, San Francisco: Freeman, 1974.
8. Lang, S., Algebra, 2nd ed., Menlo Park: Addison-Wesley, 1984.
311
Index
312
INDEX 313
Vector, 82
Vector space, 79, 82
Vector space over a division ring, 295