7FFB? :H H Ii?ed7d7boi?i: Dehc7dh$:h7f H 7hhoic?j
7FFB? :H H Ii?ed7d7boi?i: Dehc7dh$:h7f H 7hhoic?j
7FFB? :H H Ii?ed7d7boi?i: Dehc7dh$:h7f H 7hhoic?j
I
J^_hZ;Z_j_ed
FTP SITE
NOW
AVAILABLE
DEHC7DH$:H7F;H
>7HHOIC?J>
M_b[oI[h_[i_dFheXWX_b_joWdZIjWj_ij_Yi
Norman R. Draper
Harry Smith
A Wiley-Interscience Publication
JOHN WILEY & SONS, INC.
New York . Chichester . Weinheim
. Brisbane
Singapore
. Toronto
18
Contents
Preface
About the Software
xiii
xvii
15
1.0
1.1
1.2
1.3
47
2.1
2.2
2.3
2.4
2.5
vi
CONTENTS
79
3.0
3.1
3.2
3.3
115
135
149
CONTENTS
vii
179
205
8.1 The Hat Matrix H and the Various Types of Residuals, 205
8.2 Added Variable Plot and Partial Residuals, 209
8.3 Detection of Influential Observations: Cook's Statistics, 210
8.4 Other Statistics Measuring Influence, 214
8.5 Reference Books for Analysis of Residuals, 214
Exercises for Chapter 8, 215
9 Multiple Regression: Special Topics
217
9.1
9.2
9.3
9.4
9.5
9.6
9.7
235
243
viii
CON TEN T S
251
=
3 and
277
299
327
369
CONTENTS
ix
387
17.1
17.2
17.3
17.4
Introduction, 387
Basic Form of Ridge Regression, 387
Ridge Regression of the RaId Data, 389
In What Circumstances Is Ridge Regression Absolutely the
Correct Way to Proceed?, 391
17.5 The Phoney Data Viewpoint, 394
17.6 Concluding Remarks, 395
Appendix 17 A Ridge Estimates in Terms of Least Squares Estimates,
396
Appendix 17B Mean Square Error Argument, 396
Appendix 17C Canonical Form of Ridge Regression, 397
Exercises for Chapter 17, 400
18 Generalized Linear Models (GUM)
401
409
20.1
20.2
20.3
20.4
427
CONTENTS
447
21.1
21.2
21.3
21.4
21.5
2l.6
461
23.1
23.2
23.3
Introduction, 473
The One-Way Classification: Standard Analysis and an Example,
474
Regression Treatment of the One-Way Classification Example,
477
23.4
473
CONTENTS
xi
505
567
585
593
True/False Questions
605
Answers to Exercises
609
xii
CON TEN T S
Tables
684
695
Index
697
The second edition had 10 chapters; this edition has 26. On the whole (but not
entirely) we have chosen to use smaller chapters, and so distinguish more between
different types of material. The tabulation below shows the major relationships between second edition and third edition sections and chapters.
Material dropped consists mainly of second edition Sections 6.8 to 6.13 and 6.15,
Sections 7.1 to 7.6, and Chapter 8. New to this edition are Chapters 16 on multicollinearity, 18 on generalized linear models, 19 on mixture ingredients, 20 and 21 on the
geometry of least squares, 25 on robust regression, and 26 on resampling procedures.
Small revisions have been made even in sections where the text is basically unchanged.
Less prominence has been given to printouts, which nowadays can easily be generated
due to the excellent software available, and to references and bibliography, which are
now freely available (either in book or computer form) via the annual updates in
Current Index to Statistics. References are mostly given in brief either in situ or close
by, at the end of a section or chapter. Full references are in a bibliography but some
references are also given in full in sections or within the text or in exercises, whenever
this was felt to be the appropriate thing to do. There is no precise rule for doing this,
merely the authors' predilection. Exercises have been grouped as seemed appropriate.
They are intended as an expansion to the text and so most exercises have full or
partial solutions; there are a very few exceptions. One hundred and one true/false
questions have also been provided; all of these are in "true" form to prevent readers
remembering erroneous material. Instructors can reword them to create "false" questions easily enough. Sections 24.5 and 24.6 have some duplication with work in Chapter
20, but we decided not to eliminate this because the sections contain some differences
and have different emphases. Other smaller duplications occur; in general, we feel
that duplication is a good feature, and so we do not avoid it.
Our viewpoint in putting this book together is that it is desirable for students of
regression to work through the straight line fit case using a pocket calculator and then
to proceed quickly to analyzing larger models on the computer. We are aware that
many instructors like to get on to the computer right away. Our personal experience
is that this can be unwise and, over the years, we have met many students who enrolled
for our courses saying "I know how to put a regression on the computer but I don't
understand what I am doing." We have tried to keep such participants constantly
in mind.
xiii